with Splunk HTTP Event Collector and dbt Cloud?
Checks the health status of the Splunk HTTP Event Collector to ensure it is available and ready to receive events. See the documentation
Retrieve information about an environment. See the documentation
Sends multiple events in a single request to the Splunk HTTP Event Collector. See the documentation
Sends an event to Splunk HTTP Event Collector. See the documentation
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
splunk_http_event_collector: {
type: "app",
app: "splunk_http_event_collector",
}
},
async run({steps, $}) {
const data = {
"event": `Hello world!`,
"sourcetype": `manual`,
}
return await axios($, {
method: "post",
url: `${this.splunk_http_event_collector.$auth.api_url}:${this.splunk_http_event_collector.$auth.port}/services/collector`,
headers: {
"authorization": `Splunk ${this.splunk_http_event_collector.$auth.api_token}`,
},
params: {
channel: `2AC79941-CB26-421C-8826-F57AE23E9702`,
},
data,
})
},
})
The dbt Cloud API allows users to initiate jobs, check on their status, and interact with dbt Cloud programmatically. On Pipedream, you can harness this functionality to automate workflows, such as triggering dbt runs, monitoring your data transformation jobs, and integrating dbt Cloud with other data services. By leveraging Pipedream's serverless platform, you can create custom workflows that act on dbt Cloud events or use the dbt Cloud API to manage your data transformation processes seamlessly.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
dbt: {
type: "app",
app: "dbt",
}
},
async run({steps, $}) {
const baseUrl = this.dbt.$auth.access_url || `https://${this.dbt.$auth.region}.com/`
return await axios($, {
url: `${baseUrl}api/v3/accounts/`,
headers: {
"Authorization": `Token ${this.dbt.$auth.api_key}`,
"Accept": `application/json`,
},
})
},
})