Apify is the place to find, develop, buy and run cloud programs called actors. Use actors to turn any website into an API.
Write custom Node.js code and use any of the 400k+ npm packages available. Refer to the Pipedream Node docs to learn more.
Performs an execution of a selected actor in Apify. See the documentation
Executes a scraper on a specific website and returns its content as text. This action is perfect for extracting content from a single page.
Create or update a record in the key-value store of Apify. See the documentation
The Apify API unleashes the power to automate web scraping, process data, and orchestrate web automation workflows. By utilizing Apify on Pipedream, you can create dynamic serverless workflows to manage tasks like extracting data from websites, running browser automation, and scheduling these jobs to run autonomously. It integrates smoothly with Pipedream's capabilities to trigger actions on various other apps, store the results, and manage complex data flow with minimal setup.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
apify: {
type: "app",
app: "apify",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.apify.com/v2/users/me`,
headers: {
Authorization: `Bearer ${this.apify.$auth.api_token}`,
},
})
},
})
Develop, run and deploy your Node.js code in Pipedream workflows, using it between no-code steps, with connected accounts, or integrate Data Stores and File Stores.
This includes installing NPM packages, within your code without having to manage a package.json
file or running npm install
.
Below is an example of installing the axios
package in a Pipedream Node.js code step. Pipedream imports the axios
package, performs the API request, and shares the response with subsequent workflow steps:
// To use previous step data, pass the `steps` object to the run() function
export default defineComponent({
async run({ steps, $ }) {
// Return data to use it in future steps
return steps.trigger.event
},
})