with Databricks and Heroku?
Emit new event on each webhook event. See the documentation
Retrieve the output and metadata of a single task run. See the documentation
Run a job now and return the id of the triggered run. See the documentation
The Databricks API allows you to interact programmatically with Databricks services, enabling you to manage clusters, jobs, notebooks, and other resources within Databricks environments. Through Pipedream, you can leverage these APIs to create powerful automations and integrate with other apps for enhanced data processing, transformation, and analytics workflows. This unlocks possibilities like automating cluster management, dynamically running jobs based on external triggers, and orchestrating complex data pipelines with ease.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
databricks: {
type: "app",
app: "databricks",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://${this.databricks.$auth.domain}.cloud.databricks.com/api/2.0/preview/scim/v2/Me`,
headers: {
Authorization: `Bearer ${this.databricks.$auth.access_token}`,
},
})
},
})
The Heroku API offers programmatic access to Heroku's cloud platform, enabling developers to automate, extend, and integrate their app's lifecycle events with other services. With Pipedream, you can harness the Heroku API for powerful automation, such as managing apps, dynos, add-ons, and configuring scaling operations. Pipedream's ability to connect with multiple services allows for creating complex workflows, such as syncing your development pipeline with external project management tools or triggering alerts based on app metrics.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
heroku: {
type: "app",
app: "heroku",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.heroku.com/account`,
headers: {
Authorization: `Bearer ${this.heroku.$auth.oauth_access_token}`,
"accept": `application/vnd.heroku+json; version=3`,
},
})
},
})