Easy to use no-code web scraping and data extraction software.
Emit new event when an automation run has finished running. See the documentation
Write custom Node.js code and use any of the 400k+ npm packages available. Refer to the Pipedream Node docs to learn more.
Triggers a pre-built automation by providing the scraper ID. See the documentation
The Browserhub API offers automation and control over browser sessions, enabling users to create, manipulate, and extract data from web pages programmatically. Integrating Browserhub with Pipedream opens a world of possibilities for automating web interaction workflows, monitoring website changes, scraping data, and testing web applications. Pipedream's serverless platform allows for easy orchestration of API calls and data handling, making it a powerful tool for developers to harness the capabilities of Browserhub in a scalable and efficient manner.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
browserhub: {
type: "app",
app: "browserhub",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.browserhub.io/v1/status`,
headers: {
Authorization: `Bearer ${this.browserhub.$auth.api_key}`,
},
})
},
})
Develop, run and deploy your Node.js code in Pipedream workflows, using it between no-code steps, with connected accounts, or integrate Data Stores and File Stores.
This includes installing NPM packages, within your code without having to manage a package.json
file or running npm install
.
Below is an example of installing the axios
package in a Pipedream Node.js code step. Pipedream imports the axios
package, performs the API request, and shares the response with subsequent workflow steps:
// To use previous step data, pass the `steps` object to the run() function
export default defineComponent({
async run({ steps, $ }) {
// Return data to use it in future steps
return steps.trigger.event
},
})