with Mistral AI and Groq Cloud?
Emit new event when a new batch job is completed. See the Documentation
Emit new event when a new batch job fails. See the Documentation
Emit new event when a new AI model is registered or becomes available. See the Documentation
Create a new batch job, it will be queued for processing. See the Documentation
Creates a model response for the given chat conversation. See the documentation
Download a batch job results file to the /tmp directory. See the Documentation
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
mistral_ai: {
type: "app",
app: "mistral_ai",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://api.mistral.ai/v1/models`,
headers: {
Authorization: `Bearer ${this.mistral_ai.$auth.api_key}`,
"content-type": `application/json`,
},
})
},
})
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
groqcloud: {
type: "app",
app: "groqcloud",
}
},
async run({steps, $}) {
const data = {"messages": [{"role": "user", "content": "What is Pipedream?"}], "model": "llama3-8b-8192"}
return await axios($, {
method: "post",
url: `https://api.groq.com/openai/v1/chat/completions`,
headers: {
Authorization: `Bearer ${this.groqcloud.$auth.api_key}`,
},
data,
})
},
})