Upload data from an Apify dataset to an Amazon S3 bucket. Providing various filters and transformation options, this Actor allows precise control over data structure, formatting, and upload settings to ensure seamless integration into your data pipeline.
This integration-ready Apify Actor uploads the content of an Apify dataset to an Amazon S3 bucket. You can use it to store data extracted by other Actors as either an integration or a standalone Actor.
To use this Actor, you will need an AWS IAM user with the necessary permissions. If you do not have one already, you can create a new IAM user by following the official AWS guide.
Parameter | Type | Required | Description |
---|---|---|---|
accessKeyId | string | ✅ | Your AWS access key ID used for authorization of the upload. |
secretAccessKey | string | ✅ | Your AWS secret access key used for authorization of the upload. |
region | string | ✅ | The AWS region where the target S3 bucket is located. |
bucket | string | ✅ | The name of the target S3 bucket. |
key | string | ✅ | The object key, which serves as an identifier for the uploaded data in the S3 bucket. It can include an optional prefix. If an object with the same key already exists, it will be overwritten with the uploaded data. |
datasetId | string | ✅ | The Apify dataset ID from which data will be retrieved for the upload. |
format | string | ❌ | The format of the uploaded data. Options: json , jsonl , html , csv , xml , xlsx , rss . Default: json . |
fields | array | ❌ | Fields to include in the output. If not specified, all fields will be included. |
omit | array | ❌ | Fields to exclude from the output. |
unwind | array | ❌ | Fields to unwind. If the field is an array, every element will become a separate record and merged with the parent object. If the unwound field is an object, it is merged with the parent object. If the unwound field is missing or its value is neither an array nor an object, it cannot be merged with a parent object, and the item gets preserved as is. If you specify multiple fields, they are unwound in the order you specify. |
flatten | array | ❌ | Fields to transform from nested objects into a flat structure. |
offset | integer | ❌ | Number of items to skip from the beginning of the dataset. Minimum: 0 . |
limit | integer | ❌ | Maximum number of items to upload. Minimum: 1 . |
clean | boolean | ❌ | If enabled, only clean dataset items and their non-hidden fields will be uploaded. See the documentation for details. Default: true . |
If the Actor encounters an issue, it will log an error and fail. Possible issues include:
The S3 Uploader is actively maintained. If you have any feedback or feature ideas, feel free to submit an issue.
Yes, if you're scraping publicly available data for personal or internal use. Always review Websute's Terms of Service before large-scale use or redistribution.
No. This is a no-code tool — just enter a job title, location, and run the scraper directly from your dashboard or Apify actor page.
It extracts job titles, companies, salaries (if available), descriptions, locations, and post dates. You can export all of it to Excel or JSON.
Yes, you can scrape multiple pages and refine by job title, location, keyword, or more depending on the input settings you use.
You can use the Try Now button on this page to go to the scraper. You’ll be guided to input a search term and get structured results. No setup needed!