The easiest way to extract structured data from any website with no coding.
Emits an event when a Browse AI task is completed. See the documentation
Emit new event when a task finishes with an error. See the documentation
Runs a robot on-demand with custom input parameters. See the documentation
Write custom Node.js code and use any of the 400k+ npm packages available. Refer to the Pipedream Node docs to learn more.
The Browse AI API enables the automation of data extraction from websites, turning web pages into organized data. It's built for non-coders and coders alike, allowing for custom web scraping and monitoring tasks. Within Pipedream, you can harness Browse AI to create intricate workflows that trigger actions in other apps based on the data you extract. Think of automated competitive analysis, price tracking, or content changes detection, all streamlined through Pipedream's serverless platform.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
browse_ai: {
type: "app",
app: "browse_ai",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.browse.ai/v2/robots`,
headers: {
Authorization: `Bearer ${this.browse_ai.$auth.api_key}`,
},
})
},
})
Develop, run and deploy your Node.js code in Pipedream workflows, using it between no-code steps, with connected accounts, or integrate Data Stores and File Stores.
This includes installing NPM packages, within your code without having to manage a package.json
file or running npm install
.
Below is an example of installing the axios
package in a Pipedream Node.js code step. Pipedream imports the axios
package, performs the API request, and shares the response with subsequent workflow steps:
// To use previous step data, pass the `steps` object to the run() function
export default defineComponent({
async run({ steps, $ }) {
// Return data to use it in future steps
return steps.trigger.event
},
})