Integrate Perplexity AI and Hugging Face to automate workflows with scalable backend

Connect Perplexity AI and Hugging Face nodes to in your workflow. Integrate with any tool or database and ship powerful backend logic and APIs instantly - No code required!

Getting Started

How To Connect Perplexity AI and Hugging Face

01

Step 1: Add Trigger Node

It is the node that will start your workflow execution.

02

Step 2: Connect Perplexity AI and Hugging Face

It is the node that will start your workflow execution.

03

Step 3: Integrate with any other tool or service

Add any other tools or your preferred database nodes. If an integration is not available generate your own using AI

04

Step 4: Execute Workflow

Once done, test and click ship to run as an API or scheduled job.

Node stack

Supported Triggers & Actions

Perplexity AI NODES

Perplexity AI Chat

Sends a question to Perplexity's AI models and generates a response for the given chat conversation.

Hugging Face NODES

Caption Image

Generate caption for the image using Hugging Face's [Salesforce/blip-image-captioning-large](https://huggingface.co/Salesforce/blip-image-captioning-large) model for image captioning pretrained on COCO dataset - base architecture (with ViT large backbone).

script

Image Classification

Get classification labels for your image using Hugging Face's [google/vit-base-patch16-224](https://huggingface.co/google/vit-base-patch16-224) model which is a transformer encoder model (BERT-like) pretrained on a large collection of images in a supervised fashion, namely ImageNet-21k, at a resolution of 224x224 pixels. Next, the model was fine-tuned on ImageNet (also referred to as ILSVRC2012), a dataset comprising 1 million images and 1,000 classes, also at resolution 224x224.

script

Text Summarization

Summarize long text using Hugging Face's [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) model which is a transformer encoder-encoder (seq2seq) model with a bidirectional (BERT-like) encoder and an autoregressive (GPT-like) decoder. BART is pre-trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text.

script

Text-To-Image

Generate image from text, using Hugging Face's [openskyml/dalle-3-xl](https://huggingface.co/openskyml/dalle-3-xl) test model very similar to Dall•E 3.

script

Text-To-Music

Generate music from text using Hugging Face's [facebook/musicgen-small](https://huggingface.co/facebook/musicgen-small) model capable of generating high-quality music samples conditioned on text descriptions or audio prompts.

script

Support

Need Help?

Here are some helpful resources to get you "unstuck"

Start building your
BIGGEST ideas
in the *simplest* of ways.