Rotating Proxies. Random for every request & Sticky ports. Unlimited bandwidth.
Emit new event when you add or modify a new row in a table. See the docs here
Emit new event when new rows are returned from a custom query. See the docs here
Emit new event when a new table is added to a database. See the docs here
The Proxy Spider API lets you scrape and gather data from the web without the usual hassle of IP blocks or CAPTCHAs. By leveraging Pipedream's integration capabilities, you can automate the extraction of web data and manage proxy pools seamlessly. This means you can focus on what to do with the data you gather, rather than worrying about the technicalities of acquiring it. Within Pipedream's serverless platform, you could set up workflows that trigger based on a variety of events and use the Proxy Spider API to fetch data as needed.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
proxy_spider: {
type: "app",
app: "proxy_spider",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://proxy-spider.com/api/proxies.json`,
params: {
api_key: `${this.proxy_spider.$auth.api_key}`,
},
})
},
})
The MySQL application on Pipedream enables direct interaction with your MySQL databases, allowing you to perform CRUD operations—create, read, update, delete—on your data with ease. You can leverage these capabilities to automate data synchronization, report generation, and event-based triggers that kick off workflows in other apps. With Pipedream's serverless platform, you can connect MySQL to hundreds of other services without managing infrastructure, crafting complex code, or handling authentication.
import mysql from '@pipedream/mysql';
export default defineComponent({
props: {
mysql,
},
async run({steps, $}) {
// Component source code:
// https://github.com/PipedreamHQ/pipedream/tree/master/components/mysql
const queryObj = {
sql: "SELECT NOW()",
values: [], // Ignored since query does not contain placeholders
};
return await this.mysql.executeQuery(queryObj);
},
});