with Apify and Database?
Performs an execution of a selected actor in Apify. See the documentation
Run a specific task and return its dataset items. See the documentation
Executes a scraper on a specific website and returns its content as text. This action is perfect for extracting content from a single page.
The Apify API unleashes the power to automate web scraping, process data, and orchestrate web automation workflows. By utilizing Apify on Pipedream, you can create dynamic serverless workflows to manage tasks like extracting data from websites, running browser automation, and scheduling these jobs to run autonomously. It integrates smoothly with Pipedream's capabilities to trigger actions on various other apps, store the results, and manage complex data flow with minimal setup.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
apify: {
type: "app",
app: "apify",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.apify.com/v2/users/me`,
headers: {
Authorization: `Bearer ${this.apify.$auth.api_token}`,
},
})
},
})
The Database API on Pipedream allows users to execute SQL commands directly within workflows, enabling rich and dynamic data manipulation and storage. It supports PostgreSQL, MySQL, and SQLite, making it a versatile option for managing data across various database systems. With this API, users can perform tasks such as data insertion, querying, updates, and deletions, directly within their automations, facilitating real-time data processing and integration across multiple platforms.
import postgresql from "@pipedream/postgresql";
export default {
props: {
postgresql,
sql: {
type: "sql",
auth: {
app: "postgresql",
},
label: "PostreSQL Query",
},
},
async run({ $ }) {
const args = this.postgresql.executeQueryAdapter(this.sql);
const data = await this.postgresql.executeQuery(args);
return data;
},
};