with l3mbda and fal.ai?
Adds a request to the queue for asynchronous processing, including specifying a webhook URL for receiving updates. See the documentation
Cancels a request in the queue. This allows you to stop a long-running task if it's no longer needed. See the documentation
Gets the response of a completed request in the queue. This retrieves the results of your asynchronous task. See the documentation
Gets the status of a request in the queue. This allows you to monitor the progress of your asynchronous tasks. See the documentation
The l3mbda API provides a platform to run JavaScript functions in the cloud, allowing you to execute code without setting up servers. On Pipedream, you can leverage the l3mbda API to create dynamic, serverless workflows. This enables you to run custom JavaScript functions as part of an automated process, integrate with other APIs, manipulate data, and respond to webhooks.
import { axios } from "@pipedream/platform"
export default defineComponent({
props: {
l3mbda: {
type: "app",
app: "l3mbda",
}
},
async run({steps, $}) {
return await axios($, {
url: `https://clerk.l3mbda.com/oauth/userinfo`,
headers: {
Authorization: `Bearer ${this.l3mbda.$auth.oauth_access_token}`,
},
})
},
})
import { fal } from "@fal-ai/client"
export default defineComponent({
props: {
fal_ai: {
type: "app",
app: "fal_ai",
}
},
async run({ steps, $ }) {
fal.config({
credentials: `${this.fal_ai.$auth.api_key}`,
});
const result = await fal.subscribe("fal-ai/lora", {
input: {
model_name: "stabilityai/stable-diffusion-xl-base-1.0",
prompt:
"Photo of a rhino dressed suit and tie sitting at a table in a bar with a bar stools, award winning photography, Elke vogelsang",
},
logs: true,
});
return result;
},
})