← HTTP / Webhook + Rev.ai integrations

Submit Transcription Job with Rev.ai API on New Requests (Payload Only) from HTTP / Webhook API

Pipedream makes it easy to connect APIs for Rev.ai, HTTP / Webhook and 1000+ other apps remarkably fast.

Trigger workflow on
New Requests (Payload Only) from the HTTP / Webhook API
Next, do this
Submit Transcription Job with the Rev.ai API
No credit card required
Into to Pipedream
Watch us build a workflow
Watch us build a workflow
7 min
Watch now ➜
Trusted by 250,000+ developers from startups to Fortune 500 companies:
Trusted by 250,000+ developers from startups to Fortune 500 companies

Developers Pipedream

Getting Started

This integration creates a workflow with a HTTP / Webhook trigger and Rev.ai action. When you configure and deploy the workflow, it will run on Pipedream's servers 24x7 for free.

  1. Select this integration
  2. Configure the New Requests (Payload Only) trigger
    1. Connect your HTTP / Webhook account
  3. Configure the Submit Transcription Job action
    1. Connect your Rev.ai account
    2. Configure media_url
    3. Optional- Configure skip_diarization
    4. Optional- Configure skip_punctuation
    5. Optional- Configure remove_disfluencies
    6. Optional- Configure filter_profanity
    7. Optional- Configure speaker_channels_count
    8. Optional- Configure delete_after_seconds
    9. Optional- Configure metadata
    10. Optional- Configure callback_url
    11. Optional- Configure phrases
  4. Deploy the workflow
  5. Send a test event to validate your setup
  6. Turn on the trigger

Details

This integration uses pre-built, open source components from Pipedream's GitHub repo. These components are developed by Pipedream and the community, and verified and maintained by Pipedream.

To contribute an update to an existing component or create a new component, create a PR on GitHub. If you're new to Pipedream component development, you can start with quickstarts for trigger span and action development, and then review the component API reference.

Trigger

Description:Get a URL and emit the HTTP body as an event on every request
Version:0.1.1
Key:http-new-requests-payload-only

Trigger Code

import http from "../../http.app.mjs";

// Core HTTP component
// Returns a 200 OK response, emits the HTTP payload as an event
export default {
  key: "http-new-requests-payload-only",
  name: "New Requests (Payload Only)",
  // eslint-disable-next-line
  description: "Get a URL and emit the HTTP body as an event on every request",
  version: "0.1.1",
  type: "source",
  props: {
    // eslint-disable-next-line
    httpInterface: {
      type: "$.interface.http",
      customResponse: true,
    },
    http,
  },
  async run(event) {
    const { body } = event;
    this.httpInterface.respond({
      status: 200,
      body,
    });
    // Emit the HTTP payload
    this.$emit({
      body,
    });
  },
};

Trigger Configuration

This component may be configured based on the props defined in the component code. Pipedream automatically prompts for input values in the UI and CLI.
LabelPropTypeDescription
N/AhttpInterface$.interface.httpThis component uses $.interface.http to generate a unique URL when the component is first instantiated. Each request to the URL will trigger the run() method of the component.
HTTP / WebhookhttpappThis component uses the HTTP / Webhook app.

Trigger Authentication

The HTTP / Webhook API does not require authentication.

About HTTP / Webhook

Get a unique URL where you can send HTTP or webhook requests

Action

Description:Starts an asynchronous job to transcribe speech-to-text for a media file. Add an optional callback URL to invoke when processing is complete.
Version:0.1.1
Key:rev_ai-submit-transcription-job

Action Code

// legacy_hash_id: a_NqiqvY
import { axios } from "@pipedream/platform";

export default {
  key: "rev_ai-submit-transcription-job",
  name: "Submit Transcription Job",
  description: "Starts an asynchronous job to transcribe speech-to-text for a media file. Add an optional callback URL to invoke when processing is complete.",
  version: "0.1.1",
  type: "action",
  props: {
    rev_ai: {
      type: "app",
      app: "rev_ai",
    },
    media_url: {
      type: "string",
      description: "Direct download media url. Ignored if submitting job from file",
    },
    skip_diarization: {
      type: "boolean",
      description: "Specify if speaker diarization will be skipped by the speech engine",
      optional: true,
    },
    skip_punctuation: {
      type: "boolean",
      description: "Specify if \"punct\" type elements will be skipped by the speech engine. For JSON outputs, this includes removing spaces. For text outputs, words will still be delimited by a space",
      optional: true,
    },
    remove_disfluencies: {
      type: "boolean",
      description: "Currently we only define disfluencies as 'ums' and 'uhs'. When set to true, disfluencies will be not appear in the transcript.",
      optional: true,
    },
    filter_profanity: {
      type: "boolean",
      description: "Enabling this option will filter for approx. 600 profanities, which cover most use cases. If a transcribed word matches a word on this list, then all the characters of that word will be replaced by asterisks except for the first and last character.",
      optional: true,
    },
    speaker_channels_count: {
      type: "integer",
      description: "Use to specify the total number of unique speaker channels in the audio.\n\nGiven the number of audio channels provided, each channel will be transcribed separately and the channel id assigned to the speaker label. The final output will be a combination of all individual channel outputs. Overlapping monologues will have ordering broken by the order in which the first spoken element of each monologue occurs. If speaker_channels_count is greater than the actual channels in the audio, the job will fail with invalid_media.\n\nNote:\n\nThe amount charged will be the duration of the file multiplied by the number of channels specified.\nWhen using speaker_channels_count each channel will be diarized as one speaker, and the value of skip_diarization will be ignored if provided",
      optional: true,
    },
    delete_after_seconds: {
      type: "integer",
      description: "Specify the number of seconds after job completion when job is auto-deleted. It may take up to 2 minutes after the scheduled time for the job to be deleted. The number of seconds provided must range from `0` seconds to `2592000` seconds (30 days).",
      optional: true,
    },
    metadata: {
      type: "string",
      description: "Optional metadata that was provided during submission",
      optional: true,
    },
    callback_url: {
      type: "string",
      description: "Optional callback url to invoke when processing is complete",
      optional: true,
    },
    phrases: {
      type: "any",
      description: "Array of phrases not found in normal dictionary. Add technical jargon, proper nouns and uncommon phrases as strings in this array to add them to the lexicon for this job.\n\nA phrase must contain at least 1 alpha character but may contain any non-numeric character from the Basic Latin set. A phrase can contain up to 12 words. Each word can contain up to 34 characters.",
      optional: true,
    },
  },
  async run({ $ }) {
    const data = {
      "media_url": this.media_url,
      "skip_diarization": this.skip_diarization,
      "skip_punctuation": this.skip_punctuation,
      "remove_disfluencies": this.remove_disfluencies,
      "filter_profanity": this.filter_profanity,
      "speaker_channels_count": this.speaker_channels_count,
      "delete_after_seconds": this.delete_after_seconds,
      "metadata": this.metadata,
      "callback_url": this.callback_url,
    };

    if (this.phrases) {
      data.phrases = {
        "custom_vocabularies": [
          {
            "phrases": this.phrases,
          },
        ],
      };
    }

    return await axios($, {
      url: "https://api.rev.ai/speechtotext/v1/jobs",
      method: "post",
      headers: {
        Authorization: `Bearer ${this.rev_ai.$auth.access_token}`,
      },
      data,
    });
  },
};

Action Configuration

This component may be configured based on the props defined in the component code. Pipedream automatically prompts for input values in the UI.

LabelPropTypeDescription
Rev.airev_aiappThis component uses the Rev.ai app.
media_urlmedia_urlstring

Direct download media url. Ignored if submitting job from file

skip_diarizationskip_diarizationboolean

Specify if speaker diarization will be skipped by the speech engine

skip_punctuationskip_punctuationboolean

Specify if "punct" type elements will be skipped by the speech engine. For JSON outputs, this includes removing spaces. For text outputs, words will still be delimited by a space

remove_disfluenciesremove_disfluenciesboolean

Currently we only define disfluencies as 'ums' and 'uhs'. When set to true, disfluencies will be not appear in the transcript.

filter_profanityfilter_profanityboolean

Enabling this option will filter for approx. 600 profanities, which cover most use cases. If a transcribed word matches a word on this list, then all the characters of that word will be replaced by asterisks except for the first and last character.

speaker_channels_countspeaker_channels_countinteger

Use to specify the total number of unique speaker channels in the audio.

Given the number of audio channels provided, each channel will be transcribed separately and the channel id assigned to the speaker label. The final output will be a combination of all individual channel outputs. Overlapping monologues will have ordering broken by the order in which the first spoken element of each monologue occurs. If speaker_channels_count is greater than the actual channels in the audio, the job will fail with invalid_media.

Note:

The amount charged will be the duration of the file multiplied by the number of channels specified.
When using speaker_channels_count each channel will be diarized as one speaker, and the value of skip_diarization will be ignored if provided

delete_after_secondsdelete_after_secondsinteger

Specify the number of seconds after job completion when job is auto-deleted. It may take up to 2 minutes after the scheduled time for the job to be deleted. The number of seconds provided must range from 0 seconds to 2592000 seconds (30 days).

metadatametadatastring

Optional metadata that was provided during submission

callback_urlcallback_urlstring

Optional callback url to invoke when processing is complete

phrasesphrasesany

Array of phrases not found in normal dictionary. Add technical jargon, proper nouns and uncommon phrases as strings in this array to add them to the lexicon for this job.

A phrase must contain at least 1 alpha character but may contain any non-numeric character from the Basic Latin set. A phrase can contain up to 12 words. Each word can contain up to 34 characters.

Action Authentication

Rev.ai uses API keys for authentication. When you connect your Rev.ai account, Pipedream securely stores the keys so you can easily authenticate to Rev.ai APIs in both code and no-code steps.

Get your access token at https://www.rev.ai/access_token

About Rev.ai

Accurate Speech-to-Text APIs for all of your speech recognition needs

More Ways to Connect Rev.ai + HTTP / Webhook

Get Transcript By Id with Rev.ai API on New Requests (Payload Only) from HTTP / Webhook API
HTTP / Webhook + Rev.ai
 
Try it
Get Transcript By Id with Rev.ai API on New Requests from HTTP / Webhook API
HTTP / Webhook + Rev.ai
 
Try it
Submit Transcription Job with Rev.ai API on New Requests from HTTP / Webhook API
HTTP / Webhook + Rev.ai
 
Try it
Get Transcript By Id with Rev.ai API on New event when the content of the URL changes. from HTTP / Webhook API
HTTP / Webhook + Rev.ai
 
Try it
Submit Transcription Job with Rev.ai API on New event when the content of the URL changes. from HTTP / Webhook API
HTTP / Webhook + Rev.ai
 
Try it
New Requests from the HTTP / Webhook API

Get a URL and emit the full HTTP event on every request (including headers and query parameters). You can also configure the HTTP response code, body, and more.

 
Try it
New Requests (Payload Only) from the HTTP / Webhook API

Get a URL and emit the HTTP body as an event on every request

 
Try it
New event when the content of the URL changes. from the HTTP / Webhook API

Emit new event when the content of the URL changes.

 
Try it
Send any HTTP Request with the HTTP / Webhook API

Send an HTTP request using any method and URL. Optionally configure query string parameters, headers, and basic auth.

 
Try it
Send GET Request with the HTTP / Webhook API

Send an HTTP GET request to any URL. Optionally configure query string parameters, headers and basic auth.

 
Try it
Send POST Request with the HTTP / Webhook API

Send an HTTP POST request to any URL. Optionally configure query string parameters, headers and basic auth.

 
Try it
Send PUT Request with the HTTP / Webhook API

Send an HTTP PUT request to any URL. Optionally configure query string parameters, headers and basic auth.

 
Try it
Return HTTP Response with the HTTP / Webhook API

Use with an HTTP Source that uses Return a custom response from your workflow as its HTTP Response

 
Try it