/

AWS S3 Storage

&

Supabase

Integrate AWS S3 Storage and Supabase

Connect AWS S3 Storage and Supabase nodes in your workflow. Integrate with any tool or database and ship powerful backend logic and APIs instantly - No code required!

Node stack

Supported AWS S3 Storage and Supabase Nodes

Add any other tools or your preferred database nodes. If an integration is not available generate your own using AI

Base64 File Loader

Fetch file from AWS S3 bucket in Base64 format.

Create File

Create a file with the given content in AWS S3 bucket.

Generate Public URL

Make a file in S3 bucket publically accessible and return its public URL

Generate Signed URL

Generate a public signed URL for a specific file in an AWS S3 bucket

Upload Base64 File

Uploads a Base64 file to AWS S3 bucket.

Upload File from External URL

Uploads a file from an external URL to AWS S3.

Zip Folder

Zips an existing folder in S3 bucket folder and returns a download url.

Supabase Trigger

Creates a database webhook on Supabase. In other words, this workflow will be executed when there is an event on the provided table. [Full Documentation](https://docs.buildship.com/trigger-nodes/supabase-trigger)

Create Row

Add new row(s) to your Supabase table.

Create Row with Token

Insert a JSON object to an existing table on Supabase via a POST request

Delete Row

Delete an existing table row using Supabase Realtime API endpoint.

Delete Row with Token

Delete an existing table row using Supabase Realtime API endpoint.

Get Columns

Gets all the values for the specified columns of a Supabase table.

Get Columns with Token

Read the specified columns of a Supabase table.

Get Row

Fetch a row from a Supabase table using the API URL, Key, table name, filter, and Authorization token.

Get Row with Token

Fetch a row from a Supabase table using the API URL, Key, table name, filter, and Authorization token.

List Supabase Projects

Returns a list of all projects you've previously created.

Postgres RPC

Initializes a Supabase client and calls a Postgres function as Remote Procedure Call.

Update Row

Update an existing table row using Supabase Realtime API endpoint

Update Row with Token

Update an existing table row using Supabase Realtime API endpoint

Upload File

Upload a file to a public Supabase storage bucket and return the download URL.

Quick start

How to integrate AWS S3 Storage and Supabase

Step 1 — Add the nodes to your workflow

Create a new workflow in BuildShip, click “Add node”, and select the AWS S3 Storage and Supabase actions you want to use.

Step 2 — Configure each node

Go to each node to authenticate (or add your API key) and fill in the required parameters.

Step 3 — Connect the nodes

Each node in BuildShip can connect to others by using their output variables. When you reference a variable from one node in another, BuildShip automatically links them in the workflow.

Step 4 — Test your workflow

Define your starting data in the Inputs node and choose what to do with the result in the Flow Output node. Finally, run a test to see your workflow in action.

blog posts & tutorials

Recommended Reads

Below are recommneded blogs that will help in your journey