What do you want to automate

with Google Cloud and Databricks?

Prompt, edit and deploy AI agents that connect to Google Cloud, Databricks and 3,000+ other apps in seconds.

Trusted by 1,000,000+ developers from startups to Fortune 500 companies

Adyen logo
Appcues logo
Bandwidth logo
Checkr logo
ChartMogul logo
Dataminr logo
Gopuff logo
Gorgias logo
LinkedIn logo
Logitech logo
Replicated logo
Rudderstack logo
SAS logo
Scale AI logo
Webflow logo
Warner Bros. logo
Adyen logo
Appcues logo
Bandwidth logo
Checkr logo
ChartMogul logo
Dataminr logo
Gopuff logo
Gorgias logo
LinkedIn logo
Logitech logo
Replicated logo
Rudderstack logo
SAS logo
Scale AI logo
Webflow logo
Warner Bros. logo
Cancel All Runs with Databricks API on New Pub/Sub Messages from Google Cloud API
Google Cloud + Databricks
 
Try it
Cancel Run with Databricks API on New Pub/Sub Messages from Google Cloud API
Google Cloud + Databricks
 
Try it
Create Endpoint with Databricks API on New Pub/Sub Messages from Google Cloud API
Google Cloud + Databricks
 
Try it
Create Job with Databricks API on New Pub/Sub Messages from Google Cloud API
Google Cloud + Databricks
 
Try it
Create SQL Warehouse with Databricks API on New Pub/Sub Messages from Google Cloud API
Google Cloud + Databricks
 
Try it
New Pub/Sub Messages from the Google Cloud API

Emit new Pub/Sub topic in your GCP account. Messages published to this topic are emitted from the Pipedream source.

 
Try it
BigQuery - New Row from the Google Cloud API

Emit new events when a new row is added to a table

 
Try it
BigQuery - Query Results from the Google Cloud API

Emit new events with the results of an arbitrary query

 
Try it
Cancel All Runs with the Databricks API

Cancel all active runs for a job. The runs are canceled asynchronously, so it doesn't prevent new runs from being started. See the documentation

 
Try it
Cancel Run with the Databricks API

Cancel a job run. The run is canceled asynchronously, so it may still be running when this request completes. See the documentation

 
Try it
Create Endpoint with the Databricks API

Create a new vector search endpoint. See the documentation

 
Try it
Bigquery Insert Rows with the Google Cloud API

Inserts rows into a BigQuery table. See the docs and for an example here

 
Try it
Create Job with the Databricks API

Create a job. See the documentation

 
Try it
Integrate the Google Cloud API with the Databricks API
Setup the Google Cloud API trigger to run a workflow which integrates with the Databricks API. Pipedream's integration platform allows you to integrate Google Cloud and Databricks remarkably fast. Free for developers.

Overview of Google Cloud

The Google Cloud API opens a world of possibilities for enhancing cloud operations and automating tasks. It empowers you to manage, scale, and fine-tune various services within the Google Cloud Platform (GCP) programmatically. With Pipedream, you can harness this power to create intricate workflows, trigger cloud functions based on events from other apps, manage resources, and analyze data, all in a serverless environment. The ability to interconnect GCP services with numerous other apps enriches automation, making it easier to synchronize data, streamline development workflows, and deploy applications efficiently.

Connect Google Cloud

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
module.exports = defineComponent({
  props: {
    google_cloud: {
      type: "app",
      app: "google_cloud",
    }
  },
  async run({steps, $}) {
    // Required workaround to get the @google-cloud/storage package
    // working correctly on Pipedream
    require("@dylburger/umask")()
    
    const { Storage } = require('@google-cloud/storage')
    
    const key = JSON.parse(this.google_cloud.$auth.key_json)
     
    const storage = new Storage({
      projectId: key.project_id,
      credentials: {
        client_email: key.client_email,
        private_key: key.private_key,
      }
    })
    
    await storage.authClient.getCredentials()
    return {
      status: "success",
      authenticated: true,
      projectId: key.project_id,
      serviceAccount: key.client_email
    }
  },
})

Overview of Databricks

The Databricks API allows you to interact programmatically with Databricks services, enabling you to manage clusters, jobs, notebooks, and other resources within Databricks environments. Through Pipedream, you can leverage these APIs to create powerful automations and integrate with other apps for enhanced data processing, transformation, and analytics workflows. This unlocks possibilities like automating cluster management, dynamically running jobs based on external triggers, and orchestrating complex data pipelines with ease.

Connect Databricks

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
import { axios } from "@pipedream/platform"
export default defineComponent({
  props: {
    databricks: {
      type: "app",
      app: "databricks",
    }
  },
  async run({steps, $}) {
    return await axios($, {
      url: `https://${this.databricks.$auth.domain}.cloud.databricks.com/api/2.0/preview/scim/v2/Me`,
      headers: {
        Authorization: `Bearer ${this.databricks.$auth.access_token}`,
      },
    })
  },
})

Trusted by 1,000,000+ developers from startups to Fortune 500 companies

Adyen logo
Appcues logo
Bandwidth logo
Checkr logo
ChartMogul logo
Dataminr logo
Gopuff logo
Gorgias logo
LinkedIn logo
Logitech logo
Replicated logo
Rudderstack logo
SAS logo
Scale AI logo
Webflow logo
Warner Bros. logo
Adyen logo
Appcues logo
Bandwidth logo
Checkr logo
ChartMogul logo
Dataminr logo
Gopuff logo
Gorgias logo
LinkedIn logo
Logitech logo
Replicated logo
Rudderstack logo
SAS logo
Scale AI logo
Webflow logo
Warner Bros. logo