---
title: "Populate Data"
chapter: true
weight: 153
---
## Stream Car Transactional Data to DynamoDB
### Populate Data In DynamoDB with the Cloud9 IDE
1. Navigate to [Cloud9](https://us-east-1.console.aws.amazon.com/cloud9/home?region=us-east-1#) by clicking the link, or you can also search **Cloud9** on the AWS search bar and click on the icon:
1. Once you’re in the Cloud9 IDE, you’ll need to set up the Python virtual environment. Run the following 2 commands separately:
`python3 -m virtualenv myenvpy3`
`source myenvpy3/bin/activate`
1. We’ll need to install the boto3 library. Run the following command: `pip install boto3`
1. In Cloud9, use the copy command to copy the script from S3 into the Cloud9 file system. Run the following command:
`aws s3 cp s3://rocksetawsworkshop/scriptReadS3.py .`
1. Open up **scriptReadS3.py** by double-clicking on the file. Paste your Access and Secret key you got from CloudShell:
1. In the Cloud9 terminal, export your keys like this – **please don’t add quotation marks around any of the values below**. Make sure to replace the values with the AWS credentials we got earlier:
`export AWS_ACCESS_KEY_ID=yourkey`
`export AWS_SECRET_ACCESS_KEY=yoursecret`
`export AWS_DEFAULT_REGION=us-east-1`
1. Now you can run the script by executing:
`python scriptReadS3.py`
1. Navigate back to [DynamoDB services](https://us-east-1.console.aws.amazon.com/dynamodbv2/home?region=us-east-1#tables). Click on **Tables → Update settings** via the left nav. Click on the table **rockset_carpurchases**. If you click on **explore table items**, you’ll see the records that were added from the script:
1. After you’ve got 10 or so records in **rockset_carpurchases**, kill the script:
```
Linux/MAC/Windows:
- Press "control" and "c" together
- Press "ctrl" and "c" together
```
**NOTE**: We’ll re-run the script during the Grafana portion of the workshop.
#### RECAP
We are able to simulate a real-time stream by running a script in Cloud9 that sends transactional data to DynamoDB.