Real Simple Syndication
Initiate the scraping process for a specific endpoint. See the documentation here.
Retrieve multiple RSS feeds and return a merged array of items sorted by date See documentation
The RSS app allows users to automatically fetch and parse updates from web feeds. This functionality is pivotal for staying abreast of content changes or updates from websites, blogs, and news outlets that offer RSS feeds. With Pipedream, you can harness the RSS API to trigger workflows that enable a broad range of automations, like content aggregation, monitoring for specific keywords, notifications, and data synchronization across platforms.
module.exports = defineComponent({
props: {
rss: {
type: "app",
app: "rss",
}
},
async run({steps, $}) {
// Retrieve items from a sample feed
const Parser = require('rss-parser');
const parser = new Parser();
const stories = []
// Replace with your feed URL
const url = "https://pipedream.com/community/latest.rss"
const feed = await parser.parseURL(url);
const { title, items } = feed
this.title = title
if (!items.length) {
$end("No new stories")
}
this.items = items
},
})
The Scrape-It.Cloud API allows you to automate the extraction of data from websites. It can parse, scrape, and retrieve content without the need for manual intervention. With this API on Pipedream, you can build workflows that trigger on various events and use the scraped data for numerous applications like data analysis, lead generation, and content aggregation.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
scrape_it_cloud: {
type: "app",
app: "scrape_it_cloud",
}
},
async run({steps, $}) {
const data = {
"url": `https://pipedream.com`,
}
return await axios($, {
method: "post",
url: `https://api.scrape-it.cloud/scrape`,
headers: {
"Content-Type": `application/json`,
"x-api-key": `${this.scrape_it_cloud.$auth.api_key}`,
},
data,
})
},
})