Apify is the place to find, develop, buy and run cloud programs called actors. Use actors to turn any website into an API.
Write Python and use any of the 350k+ PyPi packages available. Refer to the Pipedream Python docs to learn more.
Performs an execution of a selected actor in Apify. See the documentation
Executes a scraper on a specific website and returns its content as text. This action is perfect for extracting content from a single page.
Create or update a record in the key-value store of Apify. See the documentation
The Apify API unleashes the power to automate web scraping, process data, and orchestrate web automation workflows. By utilizing Apify on Pipedream, you can create dynamic serverless workflows to manage tasks like extracting data from websites, running browser automation, and scheduling these jobs to run autonomously. It integrates smoothly with Pipedream's capabilities to trigger actions on various other apps, store the results, and manage complex data flow with minimal setup.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
apify: {
type: "app",
app: "apify",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.apify.com/v2/users/me`,
headers: {
Authorization: `Bearer ${this.apify.$auth.api_token}`,
},
})
},
})
Develop, run and deploy your Python code in Pipedream workflows. Integrate seamlessly between no-code steps, with connected accounts, or integrate Data Stores and manipulate files within a workflow.
This includes installing PyPI packages, within your code without having to manage a requirements.txt
file or running pip
.
Below is an example of using Python to access data from the trigger of the workflow, and sharing it with subsequent workflow steps:
def handler(pd: "pipedream"):
# Reference data from previous steps
print(pd.steps["trigger"]["context"]["id"])
# Return data for use in future steps
return {"foo": {"test":True}}