Rotating Proxies. Random for every request & Sticky ports. Unlimited bandwidth.
Write Python and use any of the 350k+ PyPi packages available. Refer to the Pipedream Python docs to learn more.
The Proxy Spider API lets you scrape and gather data from the web without the usual hassle of IP blocks or CAPTCHAs. By leveraging Pipedream's integration capabilities, you can automate the extraction of web data and manage proxy pools seamlessly. This means you can focus on what to do with the data you gather, rather than worrying about the technicalities of acquiring it. Within Pipedream's serverless platform, you could set up workflows that trigger based on a variety of events and use the Proxy Spider API to fetch data as needed.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
proxy_spider: {
type: "app",
app: "proxy_spider",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://proxy-spider.com/api/proxies.json`,
params: {
api_key: `${this.proxy_spider.$auth.api_key}`,
},
})
},
})
Develop, run and deploy your Python code in Pipedream workflows. Integrate seamlessly between no-code steps, with connected accounts, or integrate Data Stores and manipulate files within a workflow.
This includes installing PyPI packages, within your code without having to manage a requirements.txt
file or running pip
.
Below is an example of using Python to access data from the trigger of the workflow, and sharing it with subsequent workflow steps:
def handler(pd: "pipedream"):
# Reference data from previous steps
print(pd.steps["trigger"]["context"]["id"])
# Return data for use in future steps
return {"foo": {"test":True}}