← Google PaLM

Chat with Google PaLM API

Pipedream makes it easy to connect APIs for Google PaLM and 2,000+ other apps remarkably fast.

Trigger workflow on
HTTP requests, schedules and app events
Next, do this
Chat with the Google PaLM API
No credit card required
Intro to Pipedream
Watch us build a workflow
Watch us build a workflow
8 min
Watch now ➜

Trusted by 800,000+ developers from startups to Fortune 500 companies

Adyen logo
Appcues logo
Bandwidth logo
Checkr logo
ChartMogul logo
Dataminr logo
Gopuff logo
Gorgias logo
LinkedIn logo
Logitech logo
Replicated logo
Rudderstack logo
SAS logo
Scale AI logo
Webflow logo
Warner Bros. logo
Adyen logo
Appcues logo
Bandwidth logo
Checkr logo
ChartMogul logo
Dataminr logo
Gopuff logo
Gorgias logo
LinkedIn logo
Logitech logo
Replicated logo
Rudderstack logo
SAS logo
Scale AI logo
Webflow logo
Warner Bros. logo

Developers Pipedream

Getting Started

Create a workflow to Chat with the Google PaLM API. When you configure and deploy the workflow, it will run on Pipedream's servers 24x7 for free.

  1. Configure the Chat action
    1. Connect your Google PaLM account
    2. Configure Prompt Text
    3. Optional- Configure Previous Messages
    4. Optional- Configure Temperature
    5. Optional- Configure Context
    6. Optional- Configure Candidate Count
    7. Optional- Configure Top K
    8. Optional- Configure Top P
    9. Optional- Configure Max Output Tokens
    10. Optional- Configure Stop Sequences
    11. Optional- Select one or more Harm Categories
  2. Select a trigger to run your workflow on HTTP requests, schedules or app events
  3. Deploy the workflow
  4. Send a test event to validate your setup
  5. Turn on the trigger

Integrations

Chat with Google PaLM API on New Requests from HTTP / Webhook API
HTTP / Webhook + Google PaLM
 
Try it
Chat with Google PaLM API on New Submission from Typeform API
Typeform + Google PaLM
 
Try it
Chat with Google PaLM API on New Submission (Instant) from Jotform API
Jotform + Google PaLM
 
Try it
Chat with Google PaLM API on New Custom Webhook Events from Stripe API
Stripe + Google PaLM
 
Try it
Chat with Google PaLM API on New Scheduled Tasks from Pipedream API
Pipedream + Google PaLM
 
Try it

Details

This is a pre-built, source-available component from Pipedream's GitHub repo. The component is developed by Pipedream and the community, and verified and maintained by Pipedream.

To contribute an update to an existing component or create a new component, create a PR on GitHub. If you're new to Pipedream component development, you can start with quickstarts for trigger span and action development, and then review the component API reference.

Chat on Google PaLM
Description:Chat using Google PaLM. [See the docs here](https://developers.generativeai.google/api/python/google/generativeai/chat)
Version:0.0.2
Key:google_palm_api-chat

Code

import app from "../../google_palm_api.app.mjs";
import constants from "../../common/constants.mjs";

export default {
  key: "google_palm_api-chat",
  name: "Chat",
  description: "Chat using Google PaLM. [See the docs here](https://developers.generativeai.google/api/python/google/generativeai/chat)",
  version: "0.0.2",
  type: "action",
  props: {
    app,
    promptText: {
      type: "string",
      label: "Prompt Text",
      description: "The text to be used as a prompt for the chat",
    },
    previousMessages: {
      type: "string[]",
      label: "Previous Messages",
      description: "The previous messages in the chat. If provided, will override the chat history",
      optional: true,
    },
    temperature: {
      type: "string",
      label: "Temperature",
      description: `The temperature to use for the chat. Values can range from [0.0,1.0], inclusive.
        A value closer to 1.0 will produce responses that are more varied and creative, while a value closer to 0.0 will typically result in more straightforward responses from the model.
        Defaults to \`0.5\``,
      optional: true,
    },
    context: {
      type: "string",
      label: "Context",
      description: "Text that should be provided to the model first, to ground the response",
      optional: true,
    },
    candidateCount: {
      type: "integer",
      label: "Candidate Count",
      description: "The maximum number of generated response messages to return. This value must be between [1, 8], inclusive. If unset, this will default to 1. Note: Only unique candidates are returned. Higher temperatures are more likely to produce unique candidates. Setting temperature=0.0 will always return 1 candidate regardless of the candidate_count.",
      optional: true,
      default: 1,
      min: 1,
      max: 8,
    },
    topK: {
      type: "string",
      label: "Top K",
      description: "The API uses combined nucleus and top-k sampling. top_k sets the maximum number of tokens to sample from on each step.",
      optional: true,
    },
    topP: {
      type: "string",
      label: "Top P",
      description: "  The API uses combined nucleus and top-k sampling. top_p configures the nucleus sampling. It sets the maximum cumulative probability of tokens to sample from. For example, if the sorted probabilities are [0.5, 0.2, 0.1, 0.1, 0.05, 0.05] a top_p of 0.8 will sample as [0.625, 0.25, 0.125, 0, 0, 0]. Typical values are in the [0.9, 1.0] range.",
      optional: true,
    },
    maxOutputTokens: {
      type: "integer",
      label: "Max Output Tokens",
      description: "Maximum number of tokens to include in a candidate. Must be greater than zero. If unset, will default to 64.",
      optional: true,
    },
    stopSequences: {
      type: "string",
      label: "Stop Sequences",
      description: "A set of up to 5 character sequences that will stop output generation. If specified, the API will stop at the first appearance of a stop sequence. The stop sequence will not be included as part of the response.",
      optional: true,
    },
    harmCategories: {
      type: "string[]",
      label: "Harm Categories",
      description: "To set safety settings, select the harm categories to set a threshold for",
      optional: true,
      options() {
        return constants.HARM_CATEGORIES.map(({
          value, label,
        }) => ({
          value,
          label,
        }));
      },
      reloadProps: true,
    },
  },
  async additionalProps() {
    const props = {};
    if (!this.harmCategories?.length) {
      return props;
    }
    for (const category of this.harmCategories) {
      props[`${category}_threshold`] = {
        type: "string",
        label: `${category} - Harm Block Threshold`,
        description: `Select the harm block threshold to set for the category ${category}`,
        options: this.getThresholdOptions(),
      };
    }
    return props;
  },
  methods: {
    getThresholdOptions() {
      return constants.HARM_BLOCK_THRESHOLD.map(({
        value, label,
      }) => ({
        value,
        label,
      }));
    },
    async chat({
      promptText,
      previousMessages,
      temperature,
      context,
      candidateCount,
      topK,
      topP,
      maxOutputTokens,
      stopSequences,
      safetySettings,
    }) {
      return this.app.chat({
        temperature,
        prompt: {
          context,
          messages: [
            ...previousMessages.map((message) => ({
              content: message,
            })),
            {
              content: promptText,
            },
          ],
        },
        candidate_count: candidateCount,
        top_k: topK
          ? +topK
          : undefined,
        top_p: topP
          ? +topP
          : undefined,
        max_output_tokens: maxOutputTokens,
        stop_sequences: stopSequences,
        safety_settings: safetySettings,
      });
    },
  },
  async run({ $ }) {
    const safetySettings = [];
    if (this.harmCategories?.length) {
      for (const category of this.harmCategories) {
        const threshold = constants.HARM_BLOCK_THRESHOLD.find(({ value }) => value === this[`${category}_threshold`]);
        safetySettings.push({
          category: (constants.HARM_CATEGORIES.find(({ value }) => value === category)).numValue,
          threshold: threshold?.numValue,
        });
      }
    }

    const response = await this.chat({
      promptText: this.promptText,
      previousMessages: this.previousMessages || [],
      temperature: parseFloat(this.temperature || "0.5"),
      context: this.context,
      candidteaCount: this.candidateCount,
      topK: this.topK,
      topP: this.topP,
      maxOutputTokens: this.maxOutputTokens,
      stopSequences: this.stopSequences,
      safetySettings,
    });
    $.export("$summary", "Successfully received response from Google PaLM");
    return response;
  },
};

Configuration

This component may be configured based on the props defined in the component code. Pipedream automatically prompts for input values in the UI and CLI.
LabelPropTypeDescription
Google PaLMappappThis component uses the Google PaLM app.
Prompt TextpromptTextstring

The text to be used as a prompt for the chat

Previous MessagespreviousMessagesstring[]

The previous messages in the chat. If provided, will override the chat history

Temperaturetemperaturestring

The temperature to use for the chat. Values can range from [0.0,1.0], inclusive.
A value closer to 1.0 will produce responses that are more varied and creative, while a value closer to 0.0 will typically result in more straightforward responses from the model.
Defaults to 0.5

Contextcontextstring

Text that should be provided to the model first, to ground the response

Candidate CountcandidateCountinteger

The maximum number of generated response messages to return. This value must be between [1, 8], inclusive. If unset, this will default to 1. Note: Only unique candidates are returned. Higher temperatures are more likely to produce unique candidates. Setting temperature=0.0 will always return 1 candidate regardless of the candidate_count.

Top KtopKstring

The API uses combined nucleus and top-k sampling. top_k sets the maximum number of tokens to sample from on each step.

Top PtopPstring

The API uses combined nucleus and top-k sampling. top_p configures the nucleus sampling. It sets the maximum cumulative probability of tokens to sample from. For example, if the sorted probabilities are [0.5, 0.2, 0.1, 0.1, 0.05, 0.05] a top_p of 0.8 will sample as [0.625, 0.25, 0.125, 0, 0, 0]. Typical values are in the [0.9, 1.0] range.

Max Output TokensmaxOutputTokensinteger

Maximum number of tokens to include in a candidate. Must be greater than zero. If unset, will default to 64.

Stop SequencesstopSequencesstring

A set of up to 5 character sequences that will stop output generation. If specified, the API will stop at the first appearance of a stop sequence. The stop sequence will not be included as part of the response.

Harm CategoriesharmCategoriesstring[]Select a value from the drop down menu.

Authentication

Google PaLM uses API keys for authentication. When you connect your Google PaLM account, Pipedream securely stores the keys so you can easily authenticate to Google PaLM APIs in both code and no-code steps.

Generate an API key or join the PaLM waitlist here, then enter your API key below.

About Google PaLM

Build generative AI applications with Google's PaLM 2 model.

More Ways to Use Google PaLM

Actions

Generate Embeddings with the Google PaLM API

Generate embeddings using Google PaLM. See the docs here

 
Try it
Generate Text with the Google PaLM API

Generate text using Google PaLM. See the docs here

 
Try it

Explore Other Apps

1
-
24
of
2,000+
apps by most popular

HTTP / Webhook
HTTP / Webhook
Get a unique URL where you can send HTTP or webhook requests
Node
Node
Anything you can do with Node.js, you can do in a Pipedream workflow. This includes using most of npm's 400,000+ packages.
Python
Python
Anything you can do in Python can be done in a Pipedream Workflow. This includes using any of the 350,000+ PyPi packages available in your Python powered workflows.
OpenAI (ChatGPT)
OpenAI (ChatGPT)
OpenAI is an AI research and deployment company with the mission to ensure that artificial general intelligence benefits all of humanity. They are the makers of popular models like ChatGPT, DALL-E, and Whisper.
Premium
Salesforce (REST API)
Salesforce (REST API)
Web services API for interacting with Salesforce
Premium
HubSpot
HubSpot
HubSpot's CRM platform contains the marketing, sales, service, operations, and website-building software you need to grow your business.
Premium
Zoho CRM
Zoho CRM
Zoho CRM is an online Sales CRM software that manages your sales, marketing, and support in one CRM platform.
Premium
Stripe
Stripe
Stripe powers online and in-person payment processing and financial solutions for businesses of all sizes.
Shopify Developer App
Shopify Developer App
Shopify is a user-friendly e-commerce platform that helps small businesses build an online store and sell online through one streamlined dashboard.
Premium
WooCommerce
WooCommerce
WooCommerce is the open-source ecommerce platform for WordPress.
Premium
Snowflake
Snowflake
A data warehouse built for the cloud
Premium
MongoDB
MongoDB
MongoDB is an open source NoSQL database management program.
Supabase
Supabase
Supabase is an open source Firebase alternative.
MySQL
MySQL
MySQL is an open-source relational database management system.
PostgreSQL
PostgreSQL
PostgreSQL is a free and open-source relational database management system emphasizing extensibility and SQL compliance.
Premium
AWS
AWS
Amazon Web Services (AWS) offers reliable, scalable, and inexpensive cloud computing services.
Premium
Twilio SendGrid
Twilio SendGrid
Send marketing and transactional email through the Twilio SendGrid platform with the Email API, proprietary mail transfer agent, and infrastructure for scalable delivery.
Amazon SES
Amazon SES
Amazon SES is a cloud-based email service provider that can integrate into any application for high volume email automation
Premium
Klaviyo
Klaviyo
Email Marketing and SMS Marketing Platform
Premium
Zendesk
Zendesk
Zendesk is award-winning customer service software trusted by 200K+ customers. Make customers happy via text, mobile, phone, email, live chat, social media.
Premium
ServiceNow
ServiceNow
The smarter way to workflow
Notion
Notion
Notion is a new tool that blends your everyday work apps into one. It's the all-in-one workspace for you and your team.
Slack
Slack
Slack is a channel-based messaging platform. With Slack, people can work together more effectively, connect all their software tools and services, and find the information they need to do their best work — all within a secure, enterprise-grade environment.
Microsoft Teams
Microsoft Teams
Microsoft Teams has communities, events, chats, channels, meetings, storage, tasks, and calendars in one place.