15 templates found
Category:
Author:
Sort:

✨🔪 Advanced AI powered document parsing & text extraction with Llama Parse

Description This workflow automates document processing using LlamaParse to extract and analyze text from various file formats. It intelligently processes documents, extracts structured data, and delivers actionable insights through multiple channels. How It Works Document Ingestion & Processing 📄 Monitors Gmail for incoming attachments or accepts documents via webhook Validates file formats against supported LlamaParse extensions Uploads documents to LlamaParse for advanced text extraction Stores original documents in Google Drive for reference Intelligent Document Analysis 🧠 Automatically classifies document types (invoices, reports, etc.) Extracts structured data using customized AI prompts Generates comprehensive document summaries with key insights Converts unstructured text into organized JSON data Invoice Processing Automation 💼 Extracts critical invoice details (dates, amounts, line items) Organizes financial data into structured formats Calculates tax breakdowns, subtotals, and payment information Maintains detailed records for accounting purposes Multi-Channel Delivery 📱 Saves extracted data to Google Sheets for tracking and analysis Sends concise summaries via Telegram for immediate review Creates searchable document archives in Google Drive Updates spreadsheets with structured financial information Setup Steps Configure API Credentials 🔑 Set up LlamaParse API connection Configure Gmail OAuth for email monitoring Set up Google Drive and Sheets integrations Add Telegram bot credentials for notifications Customize AI Processing ⚙️ Adjust document classification parameters Modify extraction templates for specific document types Fine-tune summary generation prompts Customize invoice data extraction schema Test and Deploy 🚀 Test with sample documents of various formats Verify data extraction accuracy Confirm notification delivery Monitor processing pipeline performance

Joseph LePageBy Joseph LePage
19466

🎦🚀 YouTube video comment analysis agent

🎦🚀 YouTube Video Comment Analysis Agent This n8n workflow is designed to help YouTube creators analyze video details and comments to generate a comprehensive and actionable report. The workflow provides insights into video performance, audience engagement, and viewer feedback, helping creators identify trends, interests, and opportunities for future content creation. --- ✨ Key Features Video Performance Analysis: Extracts metrics like views, likes, and comments to evaluate the video's success. Comment Sentiment Analysis: Determines the tone of comments (positive, neutral, or negative) to understand audience sentiment. Recurring Themes Detection: Identifies common topics or questions in comments to highlight viewer interests. Engagement Drivers: Pinpoints what aspects of the video resonated most with viewers. Actionable Recommendations: Offers strategies for creating follow-up content or improving future videos. Keyword Suggestions: Extracts frequently mentioned terms for better discoverability on YouTube. Collaboration Opportunities: Suggests potential partnerships based on viewer feedback or related channels. --- 🛠️ How to Use Set Up Workflow Variables: Add your GOOGLEAPIKEY and the VIDEO_ID of the YouTube video you want to analyze in the "Workflow Variables" node. Ensure your Google API key has access to the YouTube Data API. Run the Workflow: Trigger the workflow manually or through another workflow using the "Execute Workflow Trigger" node. The workflow will fetch video details and comments using pagination to ensure all data is captured. Generate Insights: The workflow processes video details and comments to create a detailed report with actionable insights. Outputs include sentiment analysis, engagement drivers, content opportunities, and audience profiling. View or Share Results: The report is converted into Markdown and can be emailed via Gmail or saved to Google Drive as a document. --- 🌟 Value from This Workflow Gain a deeper understanding of your audience's preferences and feedback. Identify trends and engagement drivers to replicate success in future videos. Discover new content opportunities based on viewer questions and suggestions. Improve discoverability by leveraging keyword suggestions extracted from comments. Build stronger connections with your audience by addressing their needs effectively.

Joseph LePageBy Joseph LePage
4860

Gmail to Telegram: email summaries with OpenAI GPT-4o

🧑‍💼 Who is this for? This workflow is for anyone who receives too many emails and wants to stay informed without drowning in their inbox. If you're constantly checking your Gmail and wish you had someone summarizing messages and sending just the important parts to your phone, this is for you. Especially useful for solopreneurs, customer support, busy professionals, or newsletter addicts. 🧠 What problem is this workflow solving? Email is powerful, but also overwhelming. Important info gets buried in threads, and staying on top of things can mean hours wasted scanning messages. This workflow turns that chaos into clarity: as soon as a new email arrives, you get a concise AI-generated summary in Telegram — straight to your pocket. No more checking Gmail constantly. No more missing key updates. Just a clean, human-style summary, written in the language you choose. ⚙️ What this workflow does Watches your Gmail inbox for new messages Prepares the content, including sender, subject, and message body Sends it to OpenAI to generate a friendly, casual summary Delivers that summary to your Telegram chat All in seconds, completely automated. 🛠️ Setup Connect your accounts: Gmail, Telegram, and OpenAI credentials must be added to the respective nodes. Set your Telegram chat ID: Use a bot like @userinfobot to get it. Customize the language in the Set summary language node (default is English). Activate the workflow — and watch it go. 🧩 How to customize this workflow to your needs You can make this workflow your own in a few easy ways: Summarize only some emails: Add a Filter node after the Gmail trigger (e.g., only messages from certain senders). Change the tone or detail of summaries: Tweak the system prompt in the Summary generation agent. Use a different model: Swap OpenAI’s GPT-4o for another provider like Claude or DeepSeek. Translate to your preferred language: Just change "english" to "español", "français", etc.

Lucía Maio BriosoBy Lucía Maio Brioso
2497

Stock portfolio analysis with Perplexity AI, GPT-4, and Google Sheets

📊 Dynamic Portfolio Advisor – Daily Stock Market Intelligence with Google Sheets Description: This advanced AI-powered n8n workflow automatically delivers a daily market intelligence briefing tailored to your stock holdings portfolio stored in Google Sheets. It uses real-time data from Perplexity AI, combines it with your portfolio, and generates personalized insights, risk alerts, and trade suggestions — all delivered via Telegram or any messaging app of your choice. For step-by-step build of workflows like this, check out: https://www.youtube.com/@Automatewithmarc ⚙️ How It Works: 🕒 Daily Trigger  Starts every day at a scheduled time (default: 10 AM) to fetch the most recent market data. 📈 Holdings Fetch  Reads your current portfolio dynamically from Google Sheets — no hardcoding required. 🧠 AI Analysis Agent  Combines: Market headlines Company-specific developments Macroeconomic updates  And analyzes how they might affect your holdings. 🔍 Perplexity Web Research Tool  Finds and summarizes the most relevant stock market news from the past 24 hours. 💬 Telegram Delivery  Sends a customized summary of: Market highlights Asset-specific impacts Opportunities and risks Actionable trade ideas (buy/sell/hold) 🛠️ Tools & Integrations: Google Sheets (live holdings feed) Perplexity AI (real-time market research) OpenAI GPT (financial summarization) Telegram (output, customizable) 💡 Use Cases: Portfolio-aware market intelligence Automated investor briefing assistant Risk alert + opportunity scanner Daily trade idea generator Finance bloggers or equity analysts streamlining prep work 📍Note: You can easily replace Telegram with Slack, Email, Notion, or any output tool supported by n8n. This template is perfect for active investors, financial advisors, or automation-savvy traders who want to turn AI and data into actionable daily signals.

Automate With MarcBy Automate With Marc
1731

WhatsApp group chat with your vector database — no Facebook Business required

Enable smart, real-time answers in your WhatsApp groups using a custom webhook, Pinecone vector database, and no Facebook Business setup. > 🟡 Note: This template uses a custom WhatsApp webhook. It does not use the official WhatsApp Business API. --- 👥 Who is this for? This workflow is designed for individuals and teams who want to enable smart WhatsApp group automation — without going through Meta’s official WhatsApp Business API. Ideal for small businesses, internal teams, communities, and personal power users. --- ❓ What problem is this solving? Setting up WhatsApp bots with intelligent responses often requires approval from Meta and a verified business account. This workflow removes those barriers by using a self-hosted webhook to handle incoming messages and respond using a document-trained AI via Pinecone. --- ⚙️ What this workflow does Connects a regular WhatsApp number to a custom webhook Adds the bot to any group chat (it stays silent unless mentioned) Indexes documents from Google Drive into Pinecone Responds with intelligent, context-aware answers from your custom knowledge base Auto-updates its knowledge every minute as the document changes --- 🛠️ Setup Step 1: Connect Google Drive Set up your Google Drive credentials in n8n Step 2: Configure Pinecone Create an index in Pinecone Dimension: 1536 Select this index in both Pinecone nodes Click Test Workflow to ingest your document into Pinecone Step 3: Get Access to the WhatsApp Webhook Fill out this form to request access You’ll receive a WhatsApp confirmation for linking Step 4: Test WhatsApp Integration ✅ One-on-one test: Send a message from another number 👥 Group test: Add the bot to a group; it will only respond when tagged --- 🧩 How to customize this workflow Modify the system prompt inside the AI agent node to control tone and behavior Update the connected Google Doc to match your specific domain (e.g. FAQs, SOPs, product manuals) Adjust the Pinecone sync frequency if you want updates more or less often --- 📚 Use cases Customer Support: Instant, intelligent replies in WhatsApp without live agents Team Knowledge Bot: Tag the bot for quick access to SOPs and internal docs Community Groups: Automate common questions while keeping noise low Personal AI Assistant: A WhatsApp chatbot trained on your notes and files --- 📝 Sticky Note Suggestion 💬 What this template does: > Enables an AI bot in your WhatsApp group that answers questions based on a Google Doc you provide. It uses a custom webhook, Google Drive, and Pinecone. 🔧 Requirements: > Google Drive account > Pinecone account with an index (dimension 1536) > Access to the custom WhatsApp webhook (see setup steps)

Cecilia MukimaBy Cecilia Mukima
1606

Send tweets every minute to Mattermost

This workflow executes every minute and fetches the recent tweets from Twitter with the search query n8n_io. The workflow is built on the concept of polling. Cron node: The Cron node triggers the workflow every minute. Based on your use-case you can configure the time. You can even use the Interval node to trigger the workflow at a certain time interval. Twitter node: The Twitter node searches for the tweets that contain n8n_io and returns the most recent tweets. You can specify a different search query based on your use-case. Set node: The Set node sets the data that we pass on to the next nodes in the workflow. You can set only the values that you require in your workflow. Function node: All the magic happens in this node. The Twitter node returns all the recent tweets, including the ones which were returned earlier. The Function node, using the getWorkflowStaticData() method, only returns the tweets that are new, i.e., the tweets that were not returned in the previous workflow. Mattermost node: The Mattermost node sends the tweets from the Function node to the Twitter notifications channel. If you don't use Mattermost and want to share this data on a different platform, replace this node with the appropriate node.

Harshil AgrawalBy Harshil Agrawal
1092

Post unassigned Zendesk tickets to Slack

> This has been updated to support the Query feature added to the Zendesk node in 0.144.0 This workflow will post all New and Open tickets without an agent assigned to a Slack channel on a schedule. The function node is used in this example to merge multiple inputs into one output message which is then used as the Slack message. The output in Slack will be similar to the below message, The "TICKET_ID" will be a link to the ticket. > Unassigned Tickets TICKETID [STATUS] - TICKETSUBJECT Usage Update the Cron schedule, The default value is 16:30 daily. Update the Credentials in the Zendesk nodes Update the Credentials and Channel in the Slack Node Grab a coffee and enjoy! Zendesk Query In the Zendesk node we are using the query assignee:none status<pending this returns all New and Open tickets with no assignee allowing us to remove the extra nodes.

Jonathan BennettsBy Jonathan Bennetts
963

Jotform to KlickTipp integration - webinar registration

Community Node Disclaimer: This workflow uses KlickTipp community nodes. How It Works: Jotform Webinar Registry Integration: This workflow streamlines the process of handling webinar registrations submitted via JotForm. It ensures the data is correctly formatted and seamlessly integrates with KlickTipp. Data Transformation: Input data is validated and transformed to meet KlickTipp’s API requirements, including formatting phone numbers, converting dates, and validating URLs. Key Features JotForm Trigger: Captures new form submissions, including participant details and webinar preferences. Data Processing: Standardizes and validates input fields Converts phone numbers to numeric-only format with international prefixes. Transforms dates into UNIX timestamps. Validates LinkedIn URLs and applies fallback URLs if validation fails. Scales numerical fields, such as work experience, for specific use cases. Subscriber Management in KlickTipp: Adds or updates participants as subscribers in KlickTipp. Includes custom field mappings, such as: Personal information (name, email, phone number). Webinar details (chosen webinar, start date/time). Preferences (reminder intervals, questions for presenters). Tags contacts for segmentation: Adds fixed and dynamic tags to contacts. Error Handling: Validates critical fields like phone numbers, URLs, and dates to prevent incorrect data submissions. Setup Instructions Install and Configure Nodes: Set up the JotForm and KlickTipp nodes in your n8n instance. Authenticate your JotForm and KlickTipp accounts. Custom Field Preparation in KlickTipp: Create the necessary custom fields to match the data structure: | Field Name | Field Type | |-----------------------------------------|------------------| | Jotform \| URL Linkedin | URL | | Jotform \| Workexperience in Years | Decimal Number | | Jotform \| Webinar start timestamp | Date & Time | | Jotform \| Questions/Notes | Text | | Jotform \| Webinar | Text | | Jotform \| Reminder | Text | After creating fields, allow 10-15 minutes for them to sync. If fields don’t appear, reconnect your KlickTipp credentials. Field Mapping and Adjustments: Verify and customize field assignments in the workflow to align with your specific form and subscriber list setup. Workflow Logic Trigger via JotForm Submission: A new form submission from JotForm initiates the workflow Data Transformation: Processes raw form data to ensure compatibility with KlickTipp’s API. Add to KlickTipp Subscriber List: Adds participants to the designated KlickTipp list, including webinar-specific details. Get all tags from KlickTipp and create a list: Fetches all existing Tags and turns them into an array Define tags to dynamically set for contacts: Definiton of variables that are received from the form submission and should be converted into tags Merge tags of both lists: Checks whether the list of existing tags in KlickTipp contains the tags which should be dynamically set based on the form submission Tag creation and tagging contacts: Creates new tags if it previously did not exist and then tags the contact Benefits Efficient lead generation: Contacts from forms are automatically imported into KlickTipp and can be used immediately, saving time and increasing the conversion rate. Automated processes: Experts can start workflows directly, such as welcome emails or course admissions, reducing administrative effort. Error-free data management: The template ensures precise data mapping, avoids manual corrections and reinforces a professional appearance. Testing and Deployment: Test the workflow by filling the form on Jotform and verifying data updates in KlickTipp. Notes: Customization: Update field mappings within the KlickTipp nodes to align with your account setup. This ensures accurate data syncing. Resources: Jotform KlickTipp Knowledge Base help article Use KlickTipp Community Node in n8n Automate Workflows: KlickTipp Integration in n8n

KlickTippBy KlickTipp
814

Auto expense tracker from LINE messages with GPT-4 and Google Sheets

Overview Do you often forget to record expenses? 你是不是會常常忘記紀錄花費? Let Spending Tracker Bot help you! 讓 Spending Tracker Bot 來幫你! This AI image/text Spending Tracker LINE Bot Workflow allows you to quickly create a customized spending tracker robot without writing a line of code. At any time, you can speak or send a photo, and the AI will parse it and automatically log the expense to your cloud ledger. 這套 AI 圖片文字記帳 LINE Bot Workflow ,讓你不用寫一行程式碼,就能快速打造一個量身訂製的記帳機器人。無論何時,只需要口述或發送一張照片,AI 就會幫你整理好自動計入雲端帳本 <br /> Preparation ① Enable the Google Sheets API in GCP and complete the OAuth setup ② Create the Google Sheet and populate the field names (Feel free to modify based on your own needs) ③ Configure the Webhook URL in the LINE Developers Console ④ OpenAI API Key ① 在 GCP 啟用 Google Sheets API,並完成 OAuth ② 建立並填好 Google Sheet 欄名 (按照自己的需求做更動) ③ 於 LINE Developers 控制台設定 Webhook URL ④ OpenAI API Key <br /> Node Configurations Webhook Purpose: The URL is used to receive incoming requests from LINE. Configuration: Paste this URL into the Webhook URL field in your LINE Developers Console. 用途: 要接收 Line 的 URL 設定: 將 URL 放到 Line Webhook URL Switch based on Expense Type & Set/Https Purpose: To distinguish whether the incoming message is text or an image. Configuration: Use a Switch node to route the flow accordingly. 用途: 區分 text 或 image 設定: switch 分流 AI Agent Purpose: To extract and organize the required fields. Configuration: Chat Model & Structured Output Parser. 用途: 整理出需要的欄位 設定: Chat Model & Structured Output Parser Create a deduplication field Purpose: To prevent duplicate entries by creating a unique "for_deduplication" field. Configuration: Join multiple field names using hyphens (-) as separators. 用途: 確保不會重複寫入,先創建一個"去重使用"欄位 設定: 用 - 連接多個欄位 Aggregrate & Merge_all Purpose: To prevent duplicate entries in the data table. Configuration: Read the Google sheet, extract the existing "fordeduplication" column into a dedupeList, and compare it against the newly generated "fordeduplication" value from the previous step. 用途: 防止重複寫入資料表 設定:讀取雲端表,將原本的"去重使用欄位"整理成dedupeList,與前一步整理好的"去重使用"欄位做比對 Response Switch Purpose: To route data and send appropriate replies based on the content. Configuration: Use the replyToken to respond after branching logic. Depending on the result, either write to the data table or return a message: ✅ Expense recorded successfully: &lt;for_deduplication&gt; Irrelevant details or images will not be logged. ⚠️ This entry has already been logged and will not be duplicated. 用途: 資料分流,回應訊息 設定:使用 replyToken ,資料分流後,寫入資料表或回應訊息 ✅ 記帳成功 : &lt;去重使用欄位&gt; 不相關明細或圖片,不會計入 ⚠️ 此筆資料已記錄過,不會重複記帳 <br /> Step by step teaching notes 【Auto Expense Tracker from LINE Messages with GPT-4 and Google Sheets】 【AI 圖片文字記帳 Line Bot,自動記帳寫入 Google Sheet】

Catalina KuoBy Catalina Kuo
575

Automate company data enrichment with ZoomInfo, Serper & Oxylabs to Google Sheets

This workflow automatically enriches company domain lists with comprehensive business information scraped from ZoomInfo, organizing the data in Google Sheets for sales teams and researchers. Who's it for Sales teams building prospect databases with accurate company information Marketing professionals researching target companies for outreach campaigns Business development teams qualifying leads with revenue and employee data Researchers collecting structured company data for market analysis Lead generation specialists enriching domain lists with contact details How it works The workflow processes unprocessed domains from a Google Sheet, searches for their ZoomInfo profiles using Serper API, scrapes the company pages through Oxylabs proxy service, and extracts structured business data. Each domain is marked as processed to prevent duplicates, and the workflow includes proper rate limiting to respect API limits. What it does Loads unprocessed domains from your Google Sheets database Searches ZoomInfo using targeted queries via Serper API for each domain Validates search results and extracts relevant ZoomInfo profile URLs Scrapes company pages using Oxylabs to bypass anti-scraping protection Extracts structured data including company details, address, revenue, and employee count Updates Google Sheets with enriched company information Tracks processing status to prevent reprocessing the same domains Requirements Serper API account with search credits (Get API key) Oxylabs subscription for web scraping proxy service (Sign up here) Google Sheets API access with OAuth2 authentication Google Sheets template - Make a copy of this template sheet with pre-configured columns How to set up Make a copy of the Google Sheets template - Click here to copy the template to your Google Drive Configure API credentials in the respective HTTP Request nodes: Add Serper API key in the search node Set up Oxylabs username/password in the scraping node Set up Google Sheets authentication using OAuth2 Update the Google Sheets document ID in all Google Sheets nodes to point to your copied template Add your domain list to the sheet with 'processed' column empty or false Run the workflow using the manual trigger How to customize the workflow Search query modification: Update the search query in the Serper node for different geographic focus (currently set for Czech Republic) Data extraction fields: Modify the Google Sheets column mapping to include/exclude specific company data points Rate limiting: Adjust wait times between requests to match your API rate limits Batch processing: Configure the split batch size for processing domains in smaller groups Error handling: Customize the continue-on-error settings based on your data quality requirements Scheduling: Replace Manual Trigger with Schedule Trigger for automated daily/weekly runs Output data includes Complete company name and official address Phone numbers and contact information Revenue figures and employee headcount Industry classifications and business categories LinkedIn company profile URLs Geographic location details (city, state, country, postal code) Processing status tracking for workflow management Note: This workflow includes comprehensive error handling to ensure domains are always marked as processed, preventing infinite loops while maintaining data integrity. Rate limiting is built-in to respect API quotas and avoid service interruptions.

Naveen ChoudharyBy Naveen Choudhary
556

[eBay] negotiation API MCP server

Complete MCP server exposing 2 Negotiation API operations to AI agents. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Credentials Add Negotiation API credentials Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works This workflow converts the Negotiation API into an MCP-compatible interface for AI agents. • MCP Trigger: Serves as your server endpoint for AI agent requests • HTTP Request Nodes: Handle API calls to https://api.ebay.com{basePath} • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Returns responses directly to the AI agent 📋 Available Operations (2 total) 🔧 FindEligibleItems (1 endpoints) • GET /findeligibleitems: Find Eligible Listings 🔧 SendOfferToInterestedBuyers (1 endpoints) • POST /sendoffertointerestedbuyers: Send Discount Offer 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Path parameters and identifiers • Query parameters and filters • Request body data • Headers and authentication Response Format: Native Negotiation API responses with full data structure Error Handling: Built-in n8n HTTP request error management 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Cursor: Add MCP server SSE URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n HTTP request handling and logging • Extensible: Easily modify or add custom logic &gt; 🆓 Free for community use! Ready to deploy in under 2 minutes.

David AshbyBy David Ashby
452

Scrape Shopify store data with RapidAPI and save to Google Sheets

An automated workflow that scrapes Shopify store information and product data using the Shopify Scraper API from RapidAPI, triggered by a user submitting a website URL, then logs data into Google Sheets for easy access and analysis. --- Node-by-Node Explanation On form submission Triggers when a user submits a Shopify store website URL. Store Info Scrap Request Sends a POST request to shopify-scraper4.p.rapidapi.com/shopinfo.php to fetch store metadata (name, location, domain, etc.). Products Scarp Request Sends a POST request to shopify-scraper4.p.rapidapi.com/products.php to retrieve detailed product data (titles, prices, tags, etc.). Append Store Info Google Sheets Appends store metadata into the "Shop Info" sheet in Google Sheets. Append Products Data In Google Sheets Appends product data into the "Products" sheet in Google Sheets. --- Use Case Ideal for businesses or analysts who want to quickly gather Shopify store insights and product catalogs without manual data collection, enabling data-driven decision-making or competitive analysis. --- Benefits Automates Shopify data extraction with the powerful Shopify Scraper API on RapidAPI. Saves time by collecting and organizing data automatically into Google Sheets. Easily scalable and adaptable for multiple Shopify stores. --- 🔑 How to Get API Key from RapidAPI Shopify Scraper Follow these steps to get your API key and start using it in your workflow: Visit the API Page 👉 Click here to open Shopify Scraper API on RapidAPI Log in or Sign Up Use your Google, GitHub, or email account to sign in. If you're new, complete a quick sign-up. Subscribe to a Pricing Plan Go to the Pricing tab on the API page. Select a plan (free or paid, depending on your needs). Click Subscribe. Access Your API Key Navigate to the Endpoints tab. Look for the X-RapidAPI-Key under Request Headers. Copy the value shown — this is your API key. Use the Key in Your Workflow In your n8n workflow (HTTP Request node), replace: text "x-rapidapi-key": "your key" with: text "x-rapidapi-key": "YOURACTUALAPI_KEY"

Sk developer By Sk developer
307