I have an AWS DynamoDb table which is write intensive. I've configured it in the provisioned capacity mode with 10,000
WCU and 1000
RCU.
I'm using AWS Datapipeline to export DynamoDb contents to S3. The pipeline is configured with the read throughput ratio 75%
.
It takes around ~2 hours
to export ~150GB
of data in this setting. When I increased the RCU to 10,000
the export completed in less than 20 minutes.
Is there any way in DataPipeline to increase the provisioined RCU only when my pipeline is running? As this pipeline is configured to run only once in a day.
You can’t control the DynamoDB capacity from within the data pipeline job.
However, you can use AWS Step Functions to orchestrate ETL jobs with other arbitrary steps. So, your solution could be a scheduled CloudWatch event that starts a Step Function to:
Additional Resources