with Scrapeless and Node?
Crawl any website at scale and say goodbye to blocks. See the documentation
Write custom Node.js code and use any of the 400k+ npm packages available. Refer to the Pipedream Node docs to learn more.
Retrieve the result of a completed scraping job. See the documentation
Endpoints for fresh, structured data from 100+ popular sites. See the documentation
Submit a new web scraping job with specified target URL and extraction rules. See the documentation
Scrapeless – your go-to platform for powerful, compliant web data extraction. With tools like Universal Scraping API, Scrapeless makes it easy to access and gather data from complex sites. Focus on insights while we handle the technical hurdles. Scrapeless – data extraction made simple.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
scrapeless: {
type: "app",
app: "scrapeless",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.scrapeless.com/api/v1/me`,
headers: {
"x-api-token": `${this.scrapeless.$auth.api_key}`,
},
})
},
})
Develop, run and deploy your Node.js code in Pipedream workflows, using it between no-code steps, with connected accounts, or integrate Data Stores and File Stores
This includes installing NPM packages, within your code without having to manage a package.json
file or running npm install
.
Below is an example of installing the axios
package in a Pipedream Node.js code step. Pipedream imports the axios
package, performs the API request, and shares the response with subsequent workflow steps:
// To use previous step data, pass the `steps` object to the run() function
export default defineComponent({
async run({ steps, $ }) {
// Return data to use it in future steps
return steps.trigger.event
},
})