Hyperfast LLM running on custom built GPU and the fastest interface in the world.
Creates a model response for the given chat conversation. See the documentation
Write custom Node.js code and use any of the 400k+ npm packages available. Refer to the Pipedream Node docs to learn more.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
groqcloud: {
type: "app",
app: "groqcloud",
}
},
async run({steps, $}) {
const data = {"messages": [{"role": "user", "content": "What is Pipedream?"}], "model": "llama3-8b-8192"}
return await axios($, {
method: "post",
url: `https://api.groq.com/openai/v1/chat/completions`,
headers: {
Authorization: `Bearer ${this.groqcloud.$auth.api_key}`,
},
data,
})
},
})
Develop, run and deploy your Node.js code in Pipedream workflows, using it between no-code steps, with connected accounts, or integrate Data Stores and File Stores.
This includes installing NPM packages, within your code without having to manage a package.json
file or running npm install
.
Below is an example of installing the axios
package in a Pipedream Node.js code step. Pipedream imports the axios
package, performs the API request, and shares the response with subsequent workflow steps:
// To use previous step data, pass the `steps` object to the run() function
export default defineComponent({
async run({ steps, $ }) {
// Return data to use it in future steps
return steps.trigger.event
},
})