Rotating Proxies. Random for every request & Sticky ports. Unlimited bandwidth.
Write custom Node.js code and use any of the 400k+ npm packages available. Refer to the Pipedream Node docs to learn more.
The Proxy Spider API lets you scrape and gather data from the web without the usual hassle of IP blocks or CAPTCHAs. By leveraging Pipedream's integration capabilities, you can automate the extraction of web data and manage proxy pools seamlessly. This means you can focus on what to do with the data you gather, rather than worrying about the technicalities of acquiring it. Within Pipedream's serverless platform, you could set up workflows that trigger based on a variety of events and use the Proxy Spider API to fetch data as needed.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
proxy_spider: {
type: "app",
app: "proxy_spider",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://proxy-spider.com/api/proxies.json`,
params: {
api_key: `${this.proxy_spider.$auth.api_key}`,
},
})
},
})
Develop, run and deploy your Node.js code in Pipedream workflows, using it between no-code steps, with connected accounts, or integrate Data Stores and File Stores.
This includes installing NPM packages, within your code without having to manage a package.json
file or running npm install
.
Below is an example of installing the axios
package in a Pipedream Node.js code step. Pipedream imports the axios
package, performs the API request, and shares the response with subsequent workflow steps:
// To use previous step data, pass the `steps` object to the run() function
export default defineComponent({
async run({ steps, $ }) {
// Return data to use it in future steps
return steps.trigger.event
},
})