Back to Catalog

Automated Hugging Face paper summary fetching & categorization workflow

quentinquentin
789 views
2/3/2026
Official Page

How the Automated Workflow Works

Scheduled Fetching from Hugging Face โฐ

The workflow triggers every weekday at 8 AM, automatically fetching the latest papers from Hugging Face for easy access.

Duplication Check to Avoid Redundant Entries ๐Ÿ”

It ensures the paper's summary is not already stored in your Notion workspace, preventing duplicate records and keeping your database organized.

Content Analysis with OpenAI ๐Ÿง 

Using OpenAI's powerful capabilities, the workflow analyzes the fetched paper summary, extracts key insights, and categorizes the content for easier understanding.

Data Storage and Notification Integration ๐Ÿ“ฅ๐Ÿ””

Once the summary is processed, it's automatically stored in your Notion workspace, and a notification containing the paper details is sent to your designated Slack channel for quick reference.


Set Up Your Automated Workflow

Create Your n8n Account ๐Ÿ“

Start by registering for an n8n account and logging into the n8n cloud service.

Connect OpenAI, Notion, and Slack ๐Ÿ”—

Link your OpenAI, Notion, and Slack accounts by entering the appropriate tokens. This step will take approximately 10โ€“15 minutes to complete.

Import the Workflow Template ๐Ÿ“ฅ

Import the provided workflow template into your n8n instance to streamline the setup process.

Activate the Workflow for Daily Summaries ๐Ÿš€

After importing, simply enable the workflow, and youโ€™re all set to receive daily paper summaries automatically.

Setup Time โณ: Approximately 15โ€“20 minutes.


Why Use This Automated Workflow?

This automated workflow not only saves you time by fetching and categorizing the latest research papers but also helps streamline your Notion workspace and Slack notifications, allowing you to stay organized and efficient without manual intervention.


Results Presentation

slack.png

n8n Hugging Face Paper Summary Fetching & Categorization Workflow

This n8n workflow automates the process of fetching information about the latest papers from Hugging Face, summarizing them using OpenAI, and then categorizing and storing them in Notion. It also sends notifications to a Slack channel for new, uncategorized papers.

Description

This workflow streamlines the discovery and organization of new research papers published on Hugging Face. By regularly polling for new papers, generating concise summaries with AI, and intelligently categorizing them, it helps researchers and enthusiasts stay up-to-date without manual effort. Uncategorized papers are highlighted in Slack for review.

What it does

  1. Triggers on Schedule: The workflow runs on a predefined schedule (e.g., daily, hourly) to check for new papers.
  2. Fetches Hugging Face Papers: It makes an HTTP request to the Hugging Face API to retrieve a list of recent research papers.
  3. Parses HTML Content: Extracts relevant data (like paper titles, links, and abstracts) from the HTML response.
  4. Loops Over Items: Processes each fetched paper individually.
  5. Generates AI Summary: For each paper, it uses OpenAI to generate a concise summary of the paper's abstract.
  6. Categorizes Paper (Conditional Logic):
    • If the paper contains specific keywords (e.g., "AI", "Machine Learning", "NLP"), it categorizes it accordingly.
    • Else, it marks the paper as "Uncategorized".
  7. Stores in Notion: Adds the paper's details (title, link, summary, category) to a Notion database.
  8. Notifies on Slack (Uncategorized Papers): If a paper is marked as "Uncategorized", it sends a notification to a designated Slack channel, including the paper's title and link, for manual review and categorization.

Prerequisites/Requirements

  • n8n Instance: A running n8n instance (cloud or self-hosted).
  • Hugging Face: No specific API key is explicitly shown in the JSON, but access to Hugging Face's public paper listings is assumed.
  • OpenAI API Key: An OpenAI API key is required for the OpenAI node to generate summaries.
  • Notion Integration: A Notion integration and a Notion database set up to store the paper information. You'll need to grant the integration access to your database.
  • Slack Webhook/App: A Slack webhook URL or a Slack app credential configured to post messages to a specific channel.

Setup/Usage

  1. Import the Workflow: Download the provided JSON and import it into your n8n instance.
  2. Configure Credentials:
    • OpenAI: Set up your OpenAI API key credential in n8n.
    • Notion: Set up your Notion API key credential and ensure it has access to the target database.
    • Slack: Set up your Slack API credential or webhook URL.
  3. Customize Nodes:
    • Schedule Trigger (Node 839): Adjust the schedule as needed (e.g., every hour, daily).
    • HTTP Request (Node 19): Verify the Hugging Face URL if you want to target specific paper types or feeds.
    • HTML (Node 842): Review and adjust the CSS selectors if the structure of the Hugging Face page changes, to ensure correct data extraction.
    • OpenAI (Node 1250): Adjust the prompt for generating summaries if you need a different style or length.
    • If (Node 20): Customize the conditions for paper categorization based on keywords relevant to your interests.
    • Notion (Node 487): Map the incoming data fields to your Notion database properties (e.g., Title, URL, Summary, Category).
    • Slack (Node 40): Configure the Slack channel and message content for uncategorized paper notifications.
  4. Activate the Workflow: Once configured, activate the workflow to start automated fetching and processing.

Related Templates

Auto-create TikTok videos with VEED.io AI avatars, ElevenLabs & GPT-4

๐Ÿ’ฅ Viral TikTok Video Machine: Auto-Create Videos with Your AI Avatar --- ๐ŸŽฏ Who is this for? This workflow is for content creators, marketers, and agencies who want to use Veed.ioโ€™s AI avatar technology to produce short, engaging TikTok videos automatically. Itโ€™s ideal for creators who want to appear on camera without recording themselves, and for teams managing multiple brands who need to generate videos at scale. --- โš™๏ธ What problem this workflow solves Manually creating videos for TikTok can take hours โ€” finding trends, writing scripts, recording, and editing. By combining Veed.io, ElevenLabs, and GPT-4, this workflow transforms a simple Telegram input into a ready-to-post TikTok video featuring your AI avatar powered by Veed.io โ€” speaking naturally with your cloned voice. --- ๐Ÿš€ What this workflow does This automation links Veed.ioโ€™s video-generation API with multiple AI tools: Analyzes TikTok trends via Perplexity AI Writes a 10-second viral script using GPT-4 Generates your voiceover via ElevenLabs Uses Veed.io (Fabric 1.0 via FAL.ai) to animate your avatar and sync the lips to the voice Creates an engaging caption + hashtags for TikTok virality Publishes the video automatically via Blotato TikTok API Logs all results to Google Sheets for tracking --- ๐Ÿงฉ Setup Telegram Bot Create your bot via @BotFather Configure it as the trigger for sending your photo and theme Connect Veed.io Create an account on Veed.io Get your FAL.ai API key (Veed Fabric 1.0 model) Use HTTPS image/audio URLs compatible with Veed Fabric Other APIs Add Perplexity, ElevenLabs, and Blotato TikTok keys Connect your Google Sheet for logging results --- ๐Ÿ› ๏ธ How to customize this workflow Change your Avatar: Upload a new image through Telegram, and Veed.io will generate a new talking version automatically. Modify the Script Style: Adjust the GPT prompt for tone (educational, funny, storytelling). Adjust Voice Tone: Tweak ElevenLabs stability and similarity settings. Expand Platforms: Add Instagram, YouTube Shorts, or X (Twitter) posting nodes. Track Performance: Customize your Google Sheet to measure your most successful Veed.io-based videos. --- ๐Ÿง  Expected Outcome In just a few seconds after sending your photo and theme, this workflow โ€” powered by Veed.io โ€” creates a fully automated TikTok video featuring your AI avatar with natural lip-sync and voice. The result is a continuous stream of viral short videos, made without cameras, editing, or effort. --- โœ… Import the JSON file in n8n, add your API keys (including Veed.io via FAL.ai), and start generating viral TikTok videos starring your AI avatar today! ๐ŸŽฅ Watch This Tutorial --- ๐Ÿ“„ Documentation: Notion Guide Need help customizing? Contact me for consulting and support : Linkedin / Youtube

Dr. FirasBy Dr. Firas
39510

Automate invoice processing with OCR, GPT-4 & Salesforce opportunity creation

PDF Invoice Extractor (AI) End-to-end pipeline: Watch Drive โžœ Download PDF โžœ OCR text โžœ AI normalize to JSON โžœ Upsert Buyer (Account) โžœ Create Opportunity โžœ Map Products โžœ Create OLI via Composite API โžœ Archive to OneDrive. --- Node by node (what it does & key setup) 1) Google Drive Trigger Purpose: Fire when a new file appears in a specific Google Drive folder. Key settings: Event: fileCreated Folder ID: google drive folder id Polling: everyMinute Creds: googleDriveOAuth2Api Output: Metadata { id, name, ... } for the new file. --- 2) Download File From Google Purpose: Get the file binary for processing and archiving. Key settings: Operation: download File ID: ={{ $json.id }} Creds: googleDriveOAuth2Api Output: Binary (default key: data) and original metadata. --- 3) Extract from File Purpose: Extract text from PDF (OCR as needed) for AI parsing. Key settings: Operation: pdf OCR: enable for scanned PDFs (in options) Output: JSON with OCR text at {{ $json.text }}. --- 4) Message a model (AI JSON Extractor) Purpose: Convert OCR text into strict normalized JSON array (invoice schema). Key settings: Node: @n8n/n8n-nodes-langchain.openAi Model: gpt-4.1 (or gpt-4.1-mini) Message role: system (the strict prompt; references {{ $json.text }}) jsonOutput: true Creds: openAiApi Output (per item): $.message.content โ†’ the parsed JSON (ensure itโ€™s an array). --- 5) Create or update an account (Salesforce) Purpose: Upsert Buyer as Account using an external ID. Key settings: Resource: account Operation: upsert External Id Field: taxid_c External Id Value: ={{ $json.message.content.buyer.tax_id }} Name: ={{ $json.message.content.buyer.name }} Creds: salesforceOAuth2Api Output: Account record (captures Id) for downstream Opportunity. --- 6) Create an opportunity (Salesforce) Purpose: Create Opportunity linked to the Buyer (Account). Key settings: Resource: opportunity Name: ={{ $('Message a model').item.json.message.content.invoice.code }} Close Date: ={{ $('Message a model').item.json.message.content.invoice.issue_date }} Stage: Closed Won Amount: ={{ $('Message a model').item.json.message.content.summary.grand_total }} AccountId: ={{ $json.id }} (from Upsert Account output) Creds: salesforceOAuth2Api Output: Opportunity Id for OLI creation. --- 7) Build SOQL (Code / JS) Purpose: Collect unique product codes from AI JSON and build a SOQL query for PricebookEntry by Pricebook2Id. Key settings: pricebook2Id (hardcoded in script): e.g., 01sxxxxxxxxxxxxxxx Source lines: $('Message a model').first().json.message.content.products Output: { soql, codes } --- 8) Query PricebookEntries (Salesforce) Purpose: Fetch PricebookEntry.Id for each Product2.ProductCode. Key settings: Resource: search Query: ={{ $json.soql }} Creds: salesforceOAuth2Api Output: Items with Id, Product2.ProductCode (used for mapping). --- 9) Code in JavaScript (Build OLI payloads) Purpose: Join lines with PBE results and Opportunity Id โžœ build OpportunityLineItem payloads. Inputs: OpportunityId: ={{ $('Create an opportunity').first().json.id }} Lines: ={{ $('Message a model').first().json.message.content.products }} PBE rows: from previous node items Output: { body: { allOrNone:false, records:[{ OpportunityLineItem... }] } } Notes: Converts discount_total โžœ per-unit if needed (currently commented for standard pricing). Throws on missing PBE mapping or empty lines. --- 10) Create Opportunity Line Items (HTTP Request) Purpose: Bulk create OLIs via Salesforce Composite API. Key settings: Method: POST URL: https://<your-instance>.my.salesforce.com/services/data/v65.0/composite/sobjects Auth: salesforceOAuth2Api (predefined credential) Body (JSON): ={{ $json.body }} Output: Composite API results (per-record statuses). --- 11) Update File to One Drive Purpose: Archive the original PDF in OneDrive. Key settings: Operation: upload File Name: ={{ $json.name }} Parent Folder ID: onedrive folder id Binary Data: true (from the Download node) Creds: microsoftOneDriveOAuth2Api Output: Uploaded file metadata. --- Data flow (wiring) Google Drive Trigger โ†’ Download File From Google Download File From Google โ†’ Extract from File โ†’ Update File to One Drive Extract from File โ†’ Message a model Message a model โ†’ Create or update an account Create or update an account โ†’ Create an opportunity Create an opportunity โ†’ Build SOQL Build SOQL โ†’ Query PricebookEntries Query PricebookEntries โ†’ Code in JavaScript Code in JavaScript โ†’ Create Opportunity Line Items --- Quick setup checklist ๐Ÿ” Credentials: Connect Google Drive, OneDrive, Salesforce, OpenAI. ๐Ÿ“‚ IDs: Drive Folder ID (watch) OneDrive Parent Folder ID (archive) Salesforce Pricebook2Id (in the JS SOQL builder) ๐Ÿง  AI Prompt: Use the strict system prompt; jsonOutput = true. ๐Ÿงพ Field mappings: Buyer tax id/name โ†’ Account upsert fields Invoice code/date/amount โ†’ Opportunity fields Product name must equal your Product2.ProductCode in SF. โœ… Test: Drop a sample PDF โ†’ verify: AI returns array JSON only Account/Opportunity created OLI records created PDF archived to OneDrive --- Notes & best practices If PDFs are scans, enable OCR in Extract from File. If AI returns non-JSON, keep โ€œReturn only a JSON arrayโ€ as the last line of the prompt and keep jsonOutput enabled. Consider adding validation on parsing.warnings to gate Salesforce writes. For discounts/taxes in OLI: Standard OLI fields donโ€™t support per-line discount amounts directly; model them in UnitPrice or custom fields. Replace the Composite API URL with your orgโ€™s domain or use the Salesforce nodeโ€™s Bulk Upsert for simplicity.

Le NguyenBy Le Nguyen
942

Automate RSS to social media pipeline with AI, Airtable & GetLate for multiple platforms

Overview Automates your complete social media content pipeline: sources articles from Wallabag RSS, generates platform-specific posts with AI, creates contextual images, and publishes via GetLate API. Built with 63 nodes across two workflows to handle LinkedIn, Instagram, and Blueskyโ€”with easy expansion to more platforms. Ideal for: Content marketers, solo creators, agencies, and community managers maintaining a consistent multi-platform presence with minimal manual effort. How It Works Two-Workflow Architecture: Content Aggregation Workflow Monitors Wallabag RSS feeds for tagged articles (to-share-linkedin, to-share-instagram, etc.) Extracts and converts content from HTML to Markdown Stores structured data in Airtable with platform assignment AI Generation & Publishing Workflow Scheduled trigger queries Airtable for unpublished content Routes to platform-specific sub-workflows (LinkedIn, Instagram, Bluesky) LLM generates optimized post text and image prompts based on custom brand parameters Optionally generates AI images and hosts them on Imgbb CDN Publishes via GetLate API (immediate or draft mode) Updates Airtable with publication status and metadata Key Features: Tag-based content routing using Wallabag's native system Swappable AI providers (Groq, OpenAI, Anthropic) Platform-specific optimization (tone, length, hashtags, CTAs) Modular designโ€”duplicate sub-workflows to add new platforms in \~30 minutes Centralized Airtable tracking with 17 data points per post Set Up Steps Setup time: \~45-60 minutes for initial configuration Create accounts and get API keys (\~15 min) Wallabag (with RSS feeds enabled) GetLate (social media publishing) Airtable (create base with provided schemaโ€”see sticky notes) LLM provider (Groq, OpenAI, or Anthropic) Image service (Hugging Face, Fal.ai, or Stability AI) Imgbb (image hosting) Configure n8n credentials (\~10 min) Add all API keys in n8n's credential manager Detailed credential setup instructions in workflow sticky notes Set up Airtable database (\~10 min) Create "RSS Feed - Content Store" base Add 19 required fields (schema provided in workflow sticky notes) Get Airtable base ID and API key Customize brand prompts (\~15 min) Edit "Set Custom SMCG Prompt" node for each platform Define brand voice, tone, goals, audience, and image preferences Platform-specific examples provided in sticky notes Configure platform settings (\~10 min) Set GetLate account IDs for each platform Enable/disable image generation per platform Choose immediate publish vs. draft mode Adjust schedule trigger frequency Test and deploy Tag test articles in Wallabag Monitor the first few executions in draft mode Activate workflows when satisfied with the output Important: This is a proof-of-concept template. Test thoroughly with draft mode before production use. Detailed setup instructions, troubleshooting tips, and customization guidance are in the workflow's sticky notes. Technical Details 63 nodes: 9 Airtable operations, 8 HTTP requests, 7 code nodes, 3 LangChain LLM chains, 3 RSS triggers, 3 GetLate publishers Supports: Multiple LLM providers, multiple image generation services, unlimited platforms via modular architecture Tracking: 17 metadata fields per post, including publish status, applied parameters, character counts, hashtags, image URLs Prerequisites n8n instance (self-hosted or cloud) Accounts: Wallabag, GetLate, Airtable, LLM provider, image generation service, Imgbb Basic understanding of n8n workflows and credential configuration Time to customize prompts for your brand voice Detailed documentation, Airtable schema, prompt examples, and troubleshooting guides are in the workflow's sticky notes. Category Tags social-media-automation, ai-content-generation, rss-to-social, multi-platform-posting, getlate-api, airtable-database, langchain, workflow-automation, content-marketing

Mikal Hayden-GatesBy Mikal Hayden-Gates
188