The easiest way to extract structured data from any website with no coding.
Emits an event when a Browse AI task is completed. See the documentation
Emit new event when a task finishes with an error. See the documentation
Runs a robot on-demand with custom input parameters. See the documentation
Run any Go code and use any Go package available with a simple import. Refer to the Pipedream Go docs to learn more.
The Browse AI API enables the automation of data extraction from websites, turning web pages into organized data. It's built for non-coders and coders alike, allowing for custom web scraping and monitoring tasks. Within Pipedream, you can harness Browse AI to create intricate workflows that trigger actions in other apps based on the data you extract. Think of automated competitive analysis, price tracking, or content changes detection, all streamlined through Pipedream's serverless platform.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
browse_ai: {
type: "app",
app: "browse_ai",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.browse.ai/v2/robots`,
headers: {
Authorization: `Bearer ${this.browse_ai.$auth.api_key}`,
},
})
},
})
You can execute custom Go scripts on-demand or in response to various triggers and integrate with thousands of apps supported by Pipedream. Writing with Go on Pipedream enables backend operations like data processing, automation, or invoking other APIs, all within the Pipedream ecosystem. By leveraging Go's performance and efficiency, you can design powerful and fast workflows to streamline complex tasks.
package main
import (
"fmt"
pd "github.com/PipedreamHQ/pipedream-go"
)
func main() {
// Access previous step data using pd.Steps
fmt.Println(pd.Steps)
// Export data using pd.Export
data := make(map[string]interface{})
data["name"] = "Luke"
pd.Export("data", data)
}