← Schedule + Databricks integrations

List Runs with Databricks API on Weekly schedule from Schedule API

Pipedream makes it easy to connect APIs for Databricks, Schedule and 2,400+ other apps remarkably fast.

Trigger workflow on
Weekly schedule from the Schedule API
Next, do this
List Runs with the Databricks API
No credit card required
Intro to Pipedream
Watch us build a workflow
Watch us build a workflow
8 min
Watch now ➜

Trusted by 1,000,000+ developers from startups to Fortune 500 companies

Adyen logo
Appcues logo
Bandwidth logo
Checkr logo
ChartMogul logo
Dataminr logo
Gopuff logo
Gorgias logo
LinkedIn logo
Logitech logo
Replicated logo
Rudderstack logo
SAS logo
Scale AI logo
Webflow logo
Warner Bros. logo
Adyen logo
Appcues logo
Bandwidth logo
Checkr logo
ChartMogul logo
Dataminr logo
Gopuff logo
Gorgias logo
LinkedIn logo
Logitech logo
Replicated logo
Rudderstack logo
SAS logo
Scale AI logo
Webflow logo
Warner Bros. logo

Developers Pipedream

Getting Started

This integration creates a workflow with a Schedule trigger and Databricks action. When you configure and deploy the workflow, it will run on Pipedream's servers 24x7 for free.

  1. Select this integration
  2. Configure the Weekly schedule trigger
    1. Connect your Schedule account
    2. Configure Cron Schedule
  3. Configure the List Runs action
    1. Connect your Databricks account
    2. Optional- Select a Job
    3. Optional- Configure Active Only?
    4. Optional- Configure Max Results
  4. Deploy the workflow
  5. Send a test event to validate your setup
  6. Turn on the trigger

Details

This integration uses pre-built, source-available components from Pipedream's GitHub repo. These components are developed by Pipedream and the community, and verified and maintained by Pipedream.

To contribute an update to an existing component or create a new component, create a PR on GitHub. If you're new to Pipedream component development, you can start with quickstarts for trigger span and action development, and then review the component API reference.

Trigger

Description:Trigger your workflow on one or more days each week at a specific time (with timezone support).
Version:0.0.1
Key:schedule-weekly

Schedule Overview

The Schedule app in Pipedream is a powerful tool that allows you to trigger workflows at regular intervals, ranging from every minute to once a year. This enables the automation of repetitive tasks and the scheduling of actions to occur without manual intervention. By leveraging this API, you can execute code, run integrations, and process data on a reliable schedule, all within Pipedream's serverless environment.

Trigger Code

import schedule from "../../schedule.app.mjs";

export default {
  name: "Weekly schedule",
  version: "0.0.1",
  key: "schedule-weekly",
  type: "source",
  description: "Trigger your workflow on one or more days each week at a specific time (with timezone support).",
  props: {
    schedule,
    cron: {
      propDefinition: [
        schedule,
        "cron",
      ],
      default: {
        cron: "0 10 * * 1",
      },
    },
  },
};

Trigger Configuration

This component may be configured based on the props defined in the component code. Pipedream automatically prompts for input values in the UI and CLI.
LabelPropTypeDescription
SchedulescheduleappThis component uses the Schedule app.
Cron Schedulecron$.interface.timer

Enter a cron expression

Trigger Authentication

The Schedule API does not require authentication.

About Schedule

Trigger workflows on an interval or cron schedule.

Action

Description:Lists all runs available to the user. [See the documentation](https://docs.databricks.com/en/workflows/jobs/jobs-2.0-api.html#runs-list)
Version:0.0.1
Key:databricks-list-runs

Databricks Overview

The Databricks API allows you to interact programmatically with Databricks services, enabling you to manage clusters, jobs, notebooks, and other resources within Databricks environments. Through Pipedream, you can leverage these APIs to create powerful automations and integrate with other apps for enhanced data processing, transformation, and analytics workflows. This unlocks possibilities like automating cluster management, dynamically running jobs based on external triggers, and orchestrating complex data pipelines with ease.

Action Code

import databricks from "../../databricks.app.mjs";

export default {
  key: "databricks-list-runs",
  name: "List Runs",
  description: "Lists all runs available to the user. [See the documentation](https://docs.databricks.com/en/workflows/jobs/jobs-2.0-api.html#runs-list)",
  version: "0.0.1",
  type: "action",
  props: {
    databricks,
    jobId: {
      propDefinition: [
        databricks,
        "jobId",
      ],
      optional: true,
    },
    activeOnly: {
      type: "boolean",
      label: "Active Only?",
      description: "Set to true to return only active runs",
      optional: true,
    },
    maxResults: {
      type: "integer",
      label: "Max Results",
      description: "Maximum number of runs to return",
      default: 100,
      optional: true,
    },
  },
  async run({ $ }) {
    const params = {
      job_id: this.jobId,
      active_only: this.activeOnly,
      limit: 100,
      offset: 0,
    };
    const allRuns = [];
    let total = 0;

    do {
      const { runs } = await this.databricks.listRuns({
        params,
        $,
      });
      if (!runs?.length) {
        break;
      }
      allRuns.push(...runs);
      params.offset += params.limit;
      total = runs?.length;
    } while (total === params.limit && allRuns < this.maxResults);

    if (allRuns?.length > this.maxResults) {
      allRuns.length = this.maxResults;
    }

    $.export("$summary", `Successfully retrieved ${allRuns.length} run${allRuns.length === 1
      ? ""
      : "s"}.`);

    return allRuns;
  },
};

Action Configuration

This component may be configured based on the props defined in the component code. Pipedream automatically prompts for input values in the UI.

LabelPropTypeDescription
DatabricksdatabricksappThis component uses the Databricks app.
JobjobIdstringSelect a value from the drop down menu.
Active Only?activeOnlyboolean

Set to true to return only active runs

Max ResultsmaxResultsinteger

Maximum number of runs to return

Action Authentication

Databricks uses API keys for authentication. When you connect your Databricks account, Pipedream securely stores the keys so you can easily authenticate to Databricks APIs in both code and no-code steps.

To retrieve your Personal Access Token:

  • Sign in to your Databricks account, and open your Databricks workspace.
  • From your workspace, click on your username in the top right corner,
  • Go to “User Settings” > "Developer" > “Access tokens”

Your domain is 1234 if your Databricks instance URL is https://1234.cloud.databricks.com/, except for the Accounts API whose domain is account regardless of instance.

About Databricks

Databricks is the lakehouse company, helping data teams solve the world’s toughest problems.

More Ways to Connect Databricks + Schedule

Get Run Output with Databricks API on Daily schedule from Schedule API
Schedule + Databricks
 
Try it
Get Run Output with Databricks API on Weekly schedule from Schedule API
Schedule + Databricks
 
Try it
Get Run Output with Databricks API on Custom Interval from Schedule API
Schedule + Databricks
 
Try it
Get Run Output with Databricks API on Monthly Schedule from Schedule API
Schedule + Databricks
 
Try it
List Runs with Databricks API on Daily schedule from Schedule API
Schedule + Databricks
 
Try it
Custom Interval from the Schedule API

Trigger your workflow every N hours, minutes or seconds.

 
Try it
Daily schedule from the Schedule API

Trigger your workflow every day.

 
Try it
Monthly Schedule from the Schedule API

Trigger your workflow on one or more days each month at a specific time (with timezone support).

 
Try it
Weekly schedule from the Schedule API

Trigger your workflow on one or more days each week at a specific time (with timezone support).

 
Try it
Get Run Output with the Databricks API

Retrieve the output and metadata of a single task run. See the documentation

 
Try it
List Runs with the Databricks API

Lists all runs available to the user. See the documentation

 
Try it
Run Job Now with the Databricks API

Run a job now and return the id of the triggered run. See the documentation

 
Try it

Explore Other Apps

1
-
24
of
2,400+
apps by most popular

HTTP / Webhook
HTTP / Webhook
Get a unique URL where you can send HTTP or webhook requests
Node
Node
Anything you can do with Node.js, you can do in a Pipedream workflow. This includes using most of npm's 400,000+ packages.
Python
Python
Anything you can do in Python can be done in a Pipedream Workflow. This includes using any of the 350,000+ PyPi packages available in your Python powered workflows.
OpenAI (ChatGPT)
OpenAI (ChatGPT)
OpenAI is an AI research and deployment company with the mission to ensure that artificial general intelligence benefits all of humanity. They are the makers of popular models like ChatGPT, DALL-E, and Whisper.
Premium
Salesforce
Salesforce
Web services API for interacting with Salesforce
Premium
HubSpot
HubSpot
HubSpot's CRM platform contains the marketing, sales, service, operations, and website-building software you need to grow your business.
Premium
Zoho CRM
Zoho CRM
Zoho CRM is an online Sales CRM software that manages your sales, marketing, and support in one CRM platform.
Premium
Stripe
Stripe
Stripe powers online and in-person payment processing and financial solutions for businesses of all sizes.
Shopify
Shopify
Shopify is a complete commerce platform that lets anyone start, manage, and grow a business. You can use Shopify to build an online store, manage sales, market to customers, and accept payments in digital and physical locations.
Premium
WooCommerce
WooCommerce
WooCommerce is the open-source ecommerce platform for WordPress.
Premium
Snowflake
Snowflake
A data warehouse built for the cloud
Premium
MongoDB
MongoDB
MongoDB is an open source NoSQL database management program.
Supabase
Supabase
Supabase is an open source Firebase alternative.
MySQL
MySQL
MySQL is an open-source relational database management system.
PostgreSQL
PostgreSQL
PostgreSQL is a free and open-source relational database management system emphasizing extensibility and SQL compliance.
Premium
AWS
AWS
Amazon Web Services (AWS) offers reliable, scalable, and inexpensive cloud computing services.
Premium
Twilio SendGrid
Twilio SendGrid
Send marketing and transactional email through the Twilio SendGrid platform with the Email API, proprietary mail transfer agent, and infrastructure for scalable delivery.
Amazon SES
Amazon SES
Amazon SES is a cloud-based email service provider that can integrate into any application for high volume email automation
Premium
Klaviyo
Klaviyo
Email Marketing and SMS Marketing Platform
Premium
Zendesk
Zendesk
Zendesk is award-winning customer service software trusted by 200K+ customers. Make customers happy via text, mobile, phone, email, live chat, social media.
Notion
Notion
Notion is a new tool that blends your everyday work apps into one. It's the all-in-one workspace for you and your team.
Slack
Slack
Slack is a channel-based messaging platform. With Slack, people can work together more effectively, connect all their software tools and services, and find the information they need to do their best work — all within a secure, enterprise-grade environment.
Microsoft Teams
Microsoft Teams
Microsoft Teams has communities, events, chats, channels, meetings, storage, tasks, and calendars in one place.
Schedule
Schedule
Trigger workflows on an interval or cron schedule.