← SFTP (key-based auth) + AWS integrations

CloudWatch Logs - Put Log Event with AWS API on New Remote Directory Watcher from SFTP (key-based auth) API

Pipedream makes it easy to connect APIs for AWS, SFTP (key-based auth) and 2,400+ other apps remarkably fast.

Trigger workflow on
New Remote Directory Watcher from the SFTP (key-based auth) API
Next, do this
CloudWatch Logs - Put Log Event with the AWS API
No credit card required
Intro to Pipedream
Watch us build a workflow
Watch us build a workflow
8 min
Watch now ➜

Trusted by 1,000,000+ developers from startups to Fortune 500 companies

Adyen logo
Appcues logo
Bandwidth logo
Checkr logo
ChartMogul logo
Dataminr logo
Gopuff logo
Gorgias logo
LinkedIn logo
Logitech logo
Replicated logo
Rudderstack logo
SAS logo
Scale AI logo
Webflow logo
Warner Bros. logo
Adyen logo
Appcues logo
Bandwidth logo
Checkr logo
ChartMogul logo
Dataminr logo
Gopuff logo
Gorgias logo
LinkedIn logo
Logitech logo
Replicated logo
Rudderstack logo
SAS logo
Scale AI logo
Webflow logo
Warner Bros. logo

Developers Pipedream

Getting Started

This integration creates a workflow with a SFTP (key-based auth) trigger and AWS action. When you configure and deploy the workflow, it will run on Pipedream's servers 24x7 for free.

  1. Select this integration
  2. Configure the New Remote Directory Watcher trigger
    1. Connect your SFTP (key-based auth) account
    2. Configure timer
    3. Configure Root directory
    4. Configure Maximum watcher depth
  3. Configure the CloudWatch Logs - Put Log Event action
    1. Connect your AWS account
    2. Select a AWS Region
    3. Select a CloudWatch Log Groups
    4. Select a CloudWatch Log Streams
    5. Configure Message
    6. Configure Timestamp
    7. Optional- Configure Sequence token
  4. Deploy the workflow
  5. Send a test event to validate your setup
  6. Turn on the trigger

Details

This integration uses pre-built, source-available components from Pipedream's GitHub repo. These components are developed by Pipedream and the community, and verified and maintained by Pipedream.

To contribute an update to an existing component or create a new component, create a PR on GitHub. If you're new to Pipedream component development, you can start with quickstarts for trigger span and action development, and then review the component API reference.

Trigger

Description:Emit new events when files get created, changed or deleted from a remote directory. [See the docs](https://github.com/theophilusx/ssh2-sftp-client#orgfac43d1)
Version:0.1.1
Key:sftp-watch-remote-directory

SFTP (key-based auth) Overview

The SFTP (key-based auth) app on Pipedream allows you to securely transfer and manage files over a network. It uses SSH for data transfer and provides the same level of security as SSH, without requiring password authentication, making automation workflows more secure and less prone to human error. You can automate file uploads, downloads, synchronization tasks, and efficiently manage your remote files in a serverless environment. Integrating this with other Pipedream-supported apps enables you to craft powerful and secure data flow systems.

Trigger Code

import { createHash } from "crypto";
import app from "../../sftp.app.mjs";
import { DEFAULT_POLLING_SOURCE_TIMER_INTERVAL } from "@pipedream/platform";

export default {
  key: "sftp-watch-remote-directory",
  name: "New Remote Directory Watcher",
  description: "Emit new events when files get created, changed or deleted from a remote directory. [See the docs](https://github.com/theophilusx/ssh2-sftp-client#orgfac43d1)",
  version: "0.1.1",
  type: "source",
  dedupe: "unique",
  props: {
    app,
    timer: {
      type: "$.interface.timer",
      default: {
        intervalSeconds: DEFAULT_POLLING_SOURCE_TIMER_INTERVAL,
      },
    },
    db: "$.service.db",
    rootDirectory: {
      label: "Root directory",
      description: "Root directory to be watched. example: `/public`",
      type: "string",
      default: "/",
    },
    maxDepth: {
      label: "Maximum watcher depth",
      description: "Watch all subdirectories within of the root directory, considering the selected maximum depth.",
      type: "integer",
      min: 0,
    },
  },
  methods: {
    setFiles(files) {
      this.db.set("files", files);
    },
    getFiles() {
      return this.db.get("files") || [];
    },
    getChangesWithEvent(currentFiles, previousFiles) {
      const fileChanges = [];
      // deleted file detection;
      for (const prvFile of previousFiles) {
        const file = currentFiles.find((p) => p.path === prvFile.path);
        if (!file) {
          fileChanges.push({
            ...prvFile,
            event: "deleted",
          });
        }
      }
      // created, updated file detection;
      for (const file of currentFiles) {
        const prvFile = previousFiles.find((p) => p.path === file.path);
        if (prvFile) {
          // unchanged files (checking size and last update time)
          if (prvFile.size === file.size && prvFile.modifyTime === file.modifyTime) {
            continue;
          }
          // file has changed
          fileChanges.push({
            ...file,
            event: "updated",
          });
        } else {
          // new file detection
          fileChanges.push({
            ...file,
            event: "created",
          });
        }
      }
      return fileChanges;
    },
    async listDirectories(sftp, parent, currDepth) {
      const files = await sftp.list(parent);
      return files.filter((file) => file.type === "d")
        .map((directory) => ({
          ...directory,
          parent,
          path: `${parent}/${directory.name}`,
          depth: currDepth,
        }));
    },
    async listDirectoriesDeep(sftp, parent, maxDepth, currDepth) {
      if (currDepth > maxDepth) {
        return [];
      }
      const nextDepth = currDepth + 1;
      const rootDirectories = await this.listDirectories(sftp, parent, currDepth);
      const childrenDirectories = [];
      for (const item of rootDirectories) {
        const path = `${parent}/${item.name}`;
        const directories = await this.listDirectoriesDeep(sftp, path, maxDepth, nextDepth);
        childrenDirectories.push(...directories);
      }
      rootDirectories.push(...childrenDirectories);
      return rootDirectories;
    },
    async listAllFilesFromDirectories(sftp, directories) {
      const allFiles = [];
      for (const directory of directories) {
        const listingResult = await sftp.list(directory.path);
        const files = listingResult.filter((file) => file.type !== "d")
          .map((file) => ({
            ...file,
            directory: directory.path,
            path: `${directory.path}/${file.name}`,
            depth: directory.depth || 0,
          }));
        allFiles.push(...files);
      }
      return allFiles;
    },
    validateRootDirectory(rootDirectory) {
      if (!rootDirectory) {
        throw new Error("Must provide root directory");
      }
      if (!rootDirectory.startsWith("/")) {
        throw new Error("The root directory must to be the absolute path and start with a slash, such as: '/public'");
      }
    },
    getEventId(event) {
      return createHash("md5")
        .update(event.path + event.modifyTime)
        .digest("hex");
    },
    emitEvents(events) {
      // sort events by modification time in ascending order
      events.sort((a, b) => a.modifyTime - b.modifyTime);
      for (const fileChangeEvent of events) {
        this.$emit(fileChangeEvent, {
          id: this.getEventId(fileChangeEvent),
          summary: `${fileChangeEvent.event} ${fileChangeEvent.path}`,
          ts: new Date(fileChangeEvent.modifyTime),
        });
      }
    },
  },
  async run() {
    this.validateRootDirectory(this.rootDirectory);

    const sftp = await this.app.connect();

    let directories = [
      {
        path: this.rootDirectory,
      },
    ];
    directories = directories.concat(
      await this.listDirectoriesDeep(sftp, this.rootDirectory, this.maxDepth, 0),
    );

    const currentFiles = await this.listAllFilesFromDirectories(sftp, directories);
    const previousFiles = this.getFiles();
    const filesChangesWithEvent = this.getChangesWithEvent(currentFiles, previousFiles);
    this.emitEvents(filesChangesWithEvent);

    this.setFiles(currentFiles);
    await this.app.disconnect(sftp);
  },
};

Trigger Configuration

This component may be configured based on the props defined in the component code. Pipedream automatically prompts for input values in the UI and CLI.
LabelPropTypeDescription
SFTP (key-based auth)appappThis component uses the SFTP (key-based auth) app.
timer$.interface.timer
N/Adb$.service.dbThis component uses $.service.db to maintain state between executions.
Root directoryrootDirectorystring

Root directory to be watched. example: /public

Maximum watcher depthmaxDepthinteger

Watch all subdirectories within of the root directory, considering the selected maximum depth.

Trigger Authentication

SFTP (key-based auth) uses API keys for authentication. When you connect your SFTP (key-based auth) account, Pipedream securely stores the keys so you can easily authenticate to SFTP (key-based auth) APIs in both code and no-code steps.

This app holds details for an SFTP server you want to interact with in Pipedream.

Below, enter the hostname of the server you'd like to connect to, your username, and the private key you'd like to use to connect.

This app only supports key pair authentication, not authentication via password. If you need to use a username / password to connect to a host, please use the SFTP (password-based auth) app, instead.

About SFTP (key-based auth)

Interact with an SFTP (SSH File Transfer Protocol) Server using a private key for auth

Action

Description:Uploads a log event to the specified log stream. [See docs](https://docs.aws.amazon.com/AWSJavaScriptSDK/v3/latest/clients/client-cloudwatch-logs/classes/putlogeventscommand.html)
Version:0.1.1
Key:aws-cloudwatch-logs-put-log-event

AWS Overview

The AWS API unlocks endless possibilities for automation with Pipedream. With this powerful combo, you can manage your AWS services and resources, automate deployment workflows, process data, and react to events across your AWS infrastructure. Pipedream offers a serverless platform for creating workflows triggered by various events that can execute AWS SDK functions, making it an efficient tool to integrate, automate, and orchestrate tasks across AWS services and other apps.

Action Code

import common from "../../common/common-cloudwatch-logs.mjs";
import { toSingleLineString } from "../../common/utils.mjs";

export default {
  ...common,
  key: "aws-cloudwatch-logs-put-log-event",
  name: "CloudWatch Logs - Put Log Event",
  description: toSingleLineString(`
    Uploads a log event to the specified log stream.
    [See docs](https://docs.aws.amazon.com/AWSJavaScriptSDK/v3/latest/clients/client-cloudwatch-logs/classes/putlogeventscommand.html)
  `),
  version: "0.1.1",
  type: "action",
  props: {
    aws: common.props.aws,
    region: common.props.region,
    logGroupName: common.props.logGroupName,
    logStreamName: common.props.logStreamName,
    message: {
      type: "string",
      label: "Message",
      description: "The event message",
    },
    timestamp: {
      type: "integer",
      label: "Timestamp",
      description: "The Unix timestamp for when the event occurred, in milliseconds",
    },
    sequenceToken: {
      type: "string",
      label: "Sequence token",
      description: toSingleLineString(`
        The sequence token obtained from the response of the previous \`PutLogEvents\` call.
        An upload in a newly created log stream does not need a sequence token.
        **You can also get the sequence token using \`DescribeLogStreams\`**.
        If you call \`PutLogEvents\` twice within a narrow time period using the same value for sequenceToken,
        both calls might be successful or one might be rejected.
      `),
      optional: true,
    },
  },
  async run({ $ }) {
    const params = {
      logGroupName: this.logGroupName,
      logStreamName: this.logStreamName,
      logEvents: [
        {
          message: this.message,
          timestamp: this.timestamp,
        },
      ],
    };
    if (this.sequenceToken) {
      params.sequenceToken = this.sequenceToken;
    }
    const response = await this.putLogEvents(params);
    if (response.rejectedLogEventsInfo) {
      $.export("$summary", "Log event was rejected");
    } else {
      $.export("$summary", "Successfully stored log");
    }
    return response;
  },
};

Action Configuration

This component may be configured based on the props defined in the component code. Pipedream automatically prompts for input values in the UI.

LabelPropTypeDescription
AWSawsappThis component uses the AWS app.
AWS RegionregionstringSelect a value from the drop down menu.
CloudWatch Log GroupslogGroupNamestringSelect a value from the drop down menu.
CloudWatch Log StreamslogStreamNamestringSelect a value from the drop down menu.
Messagemessagestring

The event message

Timestamptimestampinteger

The Unix timestamp for when the event occurred, in milliseconds

Sequence tokensequenceTokenstring

The sequence token obtained from the response of the previous PutLogEvents call. An upload in a newly created log stream does not need a sequence token. You can also get the sequence token using DescribeLogStreams. If you call PutLogEvents twice within a narrow time period using the same value for sequenceToken, both calls might be successful or one might be rejected.

Action Authentication

AWS uses API keys for authentication. When you connect your AWS account, Pipedream securely stores the keys so you can easily authenticate to AWS APIs in both code and no-code steps.

Follow the AWS Instructions for creating an IAM user with an associated access and secret key.

As a best practice, attach the minimum set of IAM permissions necessary to perform the specific task in Pipedream. If your workflow only needs to perform a single API call, you should create a user and associate an IAM group / policy with permission to do only that task. You can create as many linked AWS accounts in Pipedream as you'd like.

Enter your access and secret key below.

About AWS

Amazon Web Services (AWS) offers reliable, scalable, and inexpensive cloud computing services.

More Ways to Connect AWS + SFTP (key-based auth)

Upload String as File with SFTP (key-based auth) API on New Emails sent to SES Catch-all Domain from AWS API
AWS + SFTP (key-based auth)
 
Try it
Upload String as File with SFTP (key-based auth) API on New Records Returned by CloudWatch Logs Insights Query from AWS API
AWS + SFTP (key-based auth)
 
Try it
Upload String as File with SFTP (key-based auth) API on New Scheduled Tasks from AWS API
AWS + SFTP (key-based auth)
 
Try it
Upload String as File with SFTP (key-based auth) API on New SNS Messages from AWS API
AWS + SFTP (key-based auth)
 
Try it
Upload String as File with SFTP (key-based auth) API on New Deleted S3 File from AWS API
AWS + SFTP (key-based auth)
 
Try it
New Remote Directory Watcher from the SFTP (key-based auth) API

Emit new events when files get created, changed or deleted from a remote directory. See the docs

 
Try it
New Scheduled Tasks from the AWS API

Creates a Step Function State Machine to publish a message to an SNS topic at a specific timestamp. The SNS topic delivers the message to this Pipedream source, and the source emits it as a new event.

 
Try it
New SNS Messages from the AWS API

Creates an SNS topic in your AWS account. Messages published to this topic are emitted from the Pipedream source.

 
Try it
New Inbound SES Emails from the AWS API

The source subscribes to all emails delivered to a specific domain configured in AWS SES. When an email is sent to any address at the domain, this event source emits that email as a formatted event. These events can trigger a Pipedream workflow and can be consumed via SSE or REST API.

 
Try it
New Deleted S3 File from the AWS API

Emit new event when a file is deleted from a S3 bucket

 
Try it
Upload File with the SFTP (key-based auth) API

Uploads a file or string in UTF-8 format to the SFTP server. See the documentation

 
Try it
CloudWatch Logs - Put Log Event with the AWS API

Uploads a log event to the specified log stream. See docs

 
Try it
DynamoDB - Create Table with the AWS API

Creates a new table to your account. See docs

 
Try it
DynamoDB - Execute Statement with the AWS API

This operation allows you to perform transactional reads or writes on data stored in DynamoDB, using PartiQL. See docs

 
Try it
DynamoDB - Get Item with the AWS API

The Get Item operation returns a set of attributes for the item with the given primary key. If there is no matching item, Get Item does not return any data and there will be no Item element in the response. See docs

 
Try it

Explore Other Apps

1
-
24
of
2,400+
apps by most popular

HTTP / Webhook
HTTP / Webhook
Get a unique URL where you can send HTTP or webhook requests
Node
Node
Anything you can do with Node.js, you can do in a Pipedream workflow. This includes using most of npm's 400,000+ packages.
Python
Python
Anything you can do in Python can be done in a Pipedream Workflow. This includes using any of the 350,000+ PyPi packages available in your Python powered workflows.
OpenAI (ChatGPT)
OpenAI (ChatGPT)
OpenAI is an AI research and deployment company with the mission to ensure that artificial general intelligence benefits all of humanity. They are the makers of popular models like ChatGPT, DALL-E, and Whisper.
Premium
Salesforce
Salesforce
Web services API for interacting with Salesforce
Premium
HubSpot
HubSpot
HubSpot's CRM platform contains the marketing, sales, service, operations, and website-building software you need to grow your business.
Premium
Zoho CRM
Zoho CRM
Zoho CRM is an online Sales CRM software that manages your sales, marketing, and support in one CRM platform.
Premium
Stripe
Stripe
Stripe powers online and in-person payment processing and financial solutions for businesses of all sizes.
Shopify
Shopify
Shopify is a complete commerce platform that lets anyone start, manage, and grow a business. You can use Shopify to build an online store, manage sales, market to customers, and accept payments in digital and physical locations.
Premium
WooCommerce
WooCommerce
WooCommerce is the open-source ecommerce platform for WordPress.
Premium
Snowflake
Snowflake
A data warehouse built for the cloud
Premium
MongoDB
MongoDB
MongoDB is an open source NoSQL database management program.
Supabase
Supabase
Supabase is an open source Firebase alternative.
MySQL
MySQL
MySQL is an open-source relational database management system.
PostgreSQL
PostgreSQL
PostgreSQL is a free and open-source relational database management system emphasizing extensibility and SQL compliance.
Premium
AWS
AWS
Amazon Web Services (AWS) offers reliable, scalable, and inexpensive cloud computing services.
Premium
Twilio SendGrid
Twilio SendGrid
Send marketing and transactional email through the Twilio SendGrid platform with the Email API, proprietary mail transfer agent, and infrastructure for scalable delivery.
Amazon SES
Amazon SES
Amazon SES is a cloud-based email service provider that can integrate into any application for high volume email automation
Premium
Klaviyo
Klaviyo
Email Marketing and SMS Marketing Platform
Premium
Zendesk
Zendesk
Zendesk is award-winning customer service software trusted by 200K+ customers. Make customers happy via text, mobile, phone, email, live chat, social media.
Notion
Notion
Notion is a new tool that blends your everyday work apps into one. It's the all-in-one workspace for you and your team.
Slack
Slack
Slack is a channel-based messaging platform. With Slack, people can work together more effectively, connect all their software tools and services, and find the information they need to do their best work — all within a secure, enterprise-grade environment.
Microsoft Teams
Microsoft Teams
Microsoft Teams has communities, events, chats, channels, meetings, storage, tasks, and calendars in one place.
Schedule
Schedule
Trigger workflows on an interval or cron schedule.