A data warehouse built for the cloud
Snowflake offers a cloud database and related tools to help developers create robust, secure, and scalable data warehouses. See Snowflake's Key Concepts & Architecture.
Snowflake recommends you create a new user, role, and warehouse when you integrate a third-party tool like Pipedream. This way, you can control permissions via the user / role, and separate Pipedream compute and costs with the warehouse. You can do this directly in the Snowflake UI.
We recommend you create a read-only account if you only need to query Snowflake. If you need to insert data into Snowflake, add permissions on the appropriate objects after you create your user.
Visit https://pipedream.com/accounts. Click the button to Connect an App. Enter the required Snowflake account data.
You'll only need to connect your account once in Pipedream. You can connect this account to multiple workflows to run queries against Snowflake, insert data, and more.
Visit https://pipedream.com/new to build your first workflow. Pipedream workflows let you connect Snowflake with 1,000+ other apps. You can trigger workflows on Snowflake queries, sending results to Slack, Google Sheets, or any app that exposes an API. Or you can accept data from another app, transform it with Python, Node.js, Go or Bash code, and insert it into Snowflake.
Learn more at Pipedream University.
import snowflake from '@pipedream/snowflake';
export default defineComponent({
props: {
snowflake,
},
async run({ $ }) {
// Component source code:
// https://github.com/PipedreamHQ/pipedream/tree/master/components/snowflake
return this.snowflake.executeQuery({
sqlText: `SELECT CURRENT_TIMESTAMP()`,
binds: [],
});
},
});
The Datadog API, accessible through Pipedream, empowers you to programmatically interact with Datadog's monitoring and analytics platform. This enables developers to automate the retrieval of monitoring data, manage alert configurations, and synchronize service health information across systems. With Pipedream's serverless execution model, you can create intricate workflows that react to Datadog events or metrics, manipulate the data, and pass it on to other services or even Datadog itself for a cohesive operational ecosystem.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
datadog: {
type: "app",
app: "datadog",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.datadoghq.com/api/v1/user`,
headers: {
"DD-API-KEY": `${this.datadog.$auth.api_key}`,
"DD-APPLICATION-KEY": `${this.datadog.$auth.application_key}`,
},
})
},
})