with WebScraper.IO and PDF.co?
Emit new event when a page scraping job has completed. See the docs here
Creates a scraping job (scrapes a sitemap). See the docs here
Creates a sitemap for the selected website. See the docs here
Convert CSV, XLS, XLSX, DOC, DOCX, RTF, TXT, XPS, JPG, PNG, TIFF, URL, EMAIL to PDF. See docs here
Retrieves a list of scraping jobs for a sitemap. See the docs here
The WebScraper.IO API allows you to programmatically perform web scraping tasks, extracting structured data from websites. With the API, you can automate the gathering of web content for analysis, monitoring, and integration with other data sources. In Pipedream, you can leverage this API to build workflows that process, analyze, and act on the data you scrape without writing code for backend infrastructure.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
webscraper_io: {
type: "app",
app: "webscraper_io",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.webscraper.io/api/v1/sitemaps`,
params: {
api_token: `${this.webscraper_io.$auth.api_key}`,
},
})
},
})
PDF.co API on Pipedream opens up opportunities for automating document handling tasks. You can create PDFs from scratch, merge multiple documents, extract text or data, convert PDFs to different formats, and even perform complex operations like filling out forms programmatically. With Pipedream's serverless platform, these capabilities can be integrated into workflows that respond to events, schedule tasks, or trigger on specific conditions, streamlining processes that involve PDF manipulation.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
pdf_co: {
type: "app",
app: "pdf_co",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.pdf.co/v1/account/credit/balance`,
headers: {
"x-api-key": `${this.pdf_co.$auth.api_key}`,
},
})
},
})