Simple and powerful web scraping API with automated proxies rotation and Chrome JS rendering
Gets an answer to a question about a given webpage. See the documentation
Write Python and use any of the 350k+ PyPi packages available. Refer to the Pipedream Python docs to learn more.
Returns the full HTML content of a webpage specified by the URL. See the documentation:
Returns the visible text content of a webpage specified by the URL. See the documentation
WebScraping.AI API provides powerful tools for extracting data from websites, enabling users to retrieve structured information without the hassle of setting up a custom scraper. It handles proxy rotation, browsers, and CAPTCHAs, allowing you to focus on data collection. With Pipedream, you can harness this capability to create automated workflows that trigger on various events, process web content, and connect with countless other apps to feed data pipelines, monitor changes, or populate databases.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
webscraping_ai: {
type: "app",
app: "webscraping_ai",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.webscraping.ai/account`,
params: {
api_key: `${this.webscraping_ai.$auth.api_key}`,
},
})
},
})
Develop, run and deploy your Python code in Pipedream workflows. Integrate seamlessly between no-code steps, with connected accounts, or integrate Data Stores and manipulate files within a workflow
This includes installing PyPI packages, within your code without having to manage a requirements.txt
file or running pip
.
Below is an example of using Python to access data from the trigger of the workflow, and sharing it with subsequent workflow steps:
def handler(pd: "pipedream"):
# Reference data from previous steps
print(pd.steps["trigger"]["context"]["id"])
# Return data for use in future steps
return {"foo": {"test":True}}