Website content scraper & SEO keyword extractor with GPT-5-mini and Airtable
This workflow allows you to scrape website content, clean the HTML, extract structured information using GPT-5-mini, and store the results along with SEO keywords into Airtable. Ideal for building keyword lists and organizing web content for SEO research. --- Setup Instructions Prerequisites n8n Community or Cloud instance Airtable account with a base and table ready OpenAI API Key with access to GPT-5-mini --- Airtable Structure Ensure your Airtable table has the following fields: | Field Name | Type | Notes | | ------------ | ------- | ------------------------------- | | Website Name | String | Name or URL of the website | | Data | String | Cleaned website text | | Keyword | String | Extracted SEO keyword list | | Status | Options | Values: Todo, In progress, Done | --- Node Setup ✅ Form Trigger: Collects website URL from the user. ✅ HTTP Request: Fetches the website content. ✅ HTML Cleaner (Code Node): Strips out styles, tags, and whitespace to get clean text. ✅ Topic Extractor (AI Agent + GPT-5-mini): Extracts topic-wise information from the cleaned website content. ✅ Text Cleaner (Code Node): Removes unwanted symbols like and . ✅ Keyword Extractor (AI Agent + GPT-5-mini): Generates a list of 90 important SEO keywords. ✅ Airtable Upsert: Stores the cleaned data, keywords, and status in Airtable. --- Key Features ✅ Automatic website content scraping ✅ Clean HTML and extract plain text ✅ Use GPT-5-mini for topic-wise information extraction ✅ Generate 90-keyword SEO lists ✅ Store and manage data in Airtable --- Use Cases SEO Keyword Research Competitor Website Content Analysis Structured Website Data Collection --- Additional Workflow Recommendations ✅ Rename Nodes for Clarity | Current Name | Suggested Name | | ------------ | ------------------------------- | | Website Name | Website URL Input Form | | HTTP Request | Fetch Website Content | | Code | HTML to Plain Text Cleaner | | Split Out1 | Clean Text Splitter | | AI Agent1 | Topic Extractor (GPT-5-mini) | | Code1 | Text Cleanup Formatter | | Split Out2 | Final Text Splitter | | AI Agent | Keyword Extractor (GPT-5-mini) | | Airtable | Airtable Data Upsert | | Wait1 | Delay Before Merge | | Merge | Combine Data for Airtable | ---
Read an RSS feed
Companion workflow for RSS Feed Read node docs
Webpage change detection & alerts with Google Suite and hash tracking
This n8n template can monitor and detect changes to a webpage's contents and notify you only when a change occurs. Great to keep an eye on and track publicly available documents such as company TOS, government policy or competitor pages. How it works A scheduled trigger is used so we can run everyday to automate this process. A website page is then fetched with the HTTP request node and the contents we want to track are extracted using the HTML node. To detect changes, we generate a hash on the contents with the cryptography node and compare it with previously seen hashes using the "remove duplicates" node. If the hash was seen before, the workflow stops here. Finally, when new changes are detected a copy of the contents are uploaded to Google Drive and a logged into a Google sheet. A notification email can also be sent if action is required. How to use Update the URL you want to track in the node named "variables" and ensure the HTML node has updated selectors to get the content you want. Ensure the timezone is set correctly when using the Scheduled Trigger node. Requirements Google Sheets, Drive and Gmail for storing and notifying about changes. Webpages should ideally be publicly accessible. If not, you may need to switch the HTTP request node with a webscraping service. Customising this workflow Not using Google? Easier swap to other Service providers such as Miscrosoft365. Need more URLs? Try modifing the variables node to accept multiple URLs though the HTML node will need to be customised.
Add a song to your Spotify queue
No description available.
AI-powered viral trend analysis for TikTok and Instagram with GPT-4
Automatically discover, analyze, and report the most viral TikTok and Instagram videos in your niche every day. This workflow leverages AI and Apify to help you stay ahead of social media trends. --- What This Workflow Does Scrapes trending videos from TikTok and Instagram using Apify. Filters and analyzes content based on engagement, growth rate, and recency. Uses AI (OpenAI GPT-4 Vision & GPT-4) to provide visual and trend analysis for each video. Identifies “super viral” content and sends instant Slack alerts. Saves results to Google Sheets and sends a daily email report with the top trends. --- Use Cases Social Media Managers: Instantly spot viral trends to inform your content strategy. Content Creators: Get daily inspiration from the fastest-growing videos in your niche. Marketing Teams: Monitor competitor performance and adapt to new trends quickly. Agencies: Automate trend research and reporting for multiple clients. --- How to Configure Workflow Configuration: Add your API keys, set engagement thresholds, and adjust main workflow settings. Search Config: Enter your keywords, hashtags, language, and time window for content discovery. Scrape TikTok & Scrape Instagram: Set up your Apify API credentials and endpoints. Save to Google Sheets: Connect your Google account and select the destination spreadsheet. Send Daily Digest Email: Add the recipient’s email address and customize the message if needed. Send Super Viral Alert (optional): Configure your Slack webhook for instant notifications. > Tip: Test each integration and double-check your credentials before activating the workflow. --- Requirements Apify account and API token Google account for Sheets integration OpenAI API key (for AI analysis) Slack webhook URL (optional, for viral alerts) --- Stay ahead of the curve and never miss a viral trend again!
Seo blog content automation with GPT-4o-mini and human approval in Google Docs
Overview This n8n workflow automates the entire content creation process for SEO blog posts, from topic submission and AI drafting to human approval, revision, and final storage in Google Docs. It ensures high-quality, SEO-optimized content is generated efficiently while keeping a human in the loop for quality control. Prerequisites To use this workflow, you need the following accounts and credentials: Google Sheets Account: To manage the content tracker. The sheet must contain columns for Topic, Reference URL, Title, Status, and Link to document. OpenAI API Key (for GPT-4o-mini): To power the initial content drafting and subsequent revisions. Gmail Account: To send the content for human approval and wait for feedback. Google Docs Account: To create and store the final approved blog posts. How It Works The workflow operates in four main stages: Topic Submission, AI Content Creation, Human Approval & Revision Loop, and Final Publishing. Stage 1: Topic Submission and Tracking This stage captures a new blog topic and adds it to the content tracker. On form submission (Form Trigger): The workflow is initiated when a user submits a Topic and an optional Reference link through the form. Append row in sheet (Google Sheets): The new topic and reference link are added as a new row in your content tracker sheet. Get Topic from Google Sheets: The workflow fetches the newly added row, including the system-generated row_number for tracking updates. Stage 2: AI Content Creation (Initial Draft) The workflow uses an expert AI agent to generate the first draft of the blog post. Copywriter AI Agent: Acting as an expert SEO content strategist, the agent receives the topic and reference URL. The agent improves the provided topic into an SEO-optimized title. It writes a complete, conversational, SEO-friendly blog post (800–1200 words) using proper Markdown headings and lists. OpenAI Chat Model (GPT-4o-mini): Powers the agent's generation. Structured Output Parser: Ensures the output is in a JSON format with separate title and content keys. Set Data: The AI-generated title and content are mapped to workflow variables (Topic Title, Content) for easy use in subsequent nodes. Stage 3: Human Approval and Revision Loop This is the critical quality control step where a human reviews the draft and decides on the next action. Send Content for Approval (Gmail): The generated title and content are sent via email to the approver (<your email>). This uses a custom form that allows the user to choose Yes (Approve), No (Request Revision), or Cancel. The form also includes a Content Feedback textarea. Approval Result (Switch): The workflow pauses until the approver submits the form, and then directs the flow based on their choice. If "Yes" (Approved): Proceeds to the final publishing stage. The Update Topic Status on Google Sheets node is executed, setting the Status to Approved and logging the new Title. If "No" (Revision Requested): Proceeds to the revision loop. Copywriter Revision Agent: This agent takes the original content, the topic title, and the user's Content Feedback. It is instructed to incorporate the feedback, preserving the existing structure and tone, and outputs a revised blog post. The flow loops back through the Set Data node and then returns to the Send Content for Approval node for a new review. If "Cancel": Stops the workflow, and the Update Topic Status on Google Sheets node is executed. Stage 4: Final Publishing Once approved, the content is stored in Google Docs and the tracker is updated. Create Blog file (Google Docs): A new Google Doc is created using the approved Topic Title. Add blog content in file (Google Docs): The final Content (in Markdown format) is inserted into the newly created document. Update sheet with blog post link (Google Sheets): The final tracker update logs the Link to document and the Published date. Customization Tip Recommend using AI humanazier and AI detector tools before approval to make the SEO score better.
Automatically add travel time blockers to calendar using Google Directions API
Automatically add Travel time blockers before Appointments This bot automatically adds Travel time blockers to your calendar, so you never come late to an appointment again. How it works Trigger: The workflow is initiated daily at 7 AM by a "Schedule Trigger". AI Agent: An "AI Agent" node orchestrates the main logic. Fetch events: It uses the getcalendarevents tool to retrieve all events scheduled for the current day. Identify events with location: It then filters these events to identify those that have a specified location. Check for existing travel time Blockers: For each event with a location, it checks if a Travel time blocker already exists. Events that do not* have such a blocker are marked for processing. Calculate travel time: Using the Google Directions API it determines how lot it takes to get to the location of the event. The starting location is by default your Home Address, unless there is a previous event within 2 hours before the event, in which case it will use the location of that previous event. Create Travel time blocker: Finally, it uses the createcalendarevent tool to create the Travel time blocker with a duration equal to the calculated travel time + 10 minutes for buffer. Set up steps Set Variables Home address Blocker name Mode of Transportation Connect your LLM Provider Connect your Google Calendar Connect your Google Directions API
Generate SEO content with Claude AI & competitor analysis using Apify
SEO Content Generation Workflow (Basic Version) - n8n Template Instructions Who's it for This workflow is designed for SEO professionals, content marketers, digital agencies, and businesses who need to generate optimized meta tags, H1 headings, and content briefs at scale. Perfect for teams managing multiple clients or large keyword lists who want to automate competitor analysis and SEO content creation without the complexity of vector databases. How it works The workflow automates the entire SEO content creation process by analyzing your target keywords against top competitors, then generating optimized meta elements and comprehensive content briefs. It uses AI-powered analysis combined with real competitor data to create SEO-friendly content that's tailored to your specific business context. The system processes keywords in batches, performs Google searches, scrapes competitor content, analyzes heading structures, and generates personalized SEO content using your company information for maximum relevance. Requirements Required Services and Credentials Google Sheets API: For reading configuration and updating results Anthropic API: For AI content generation (Claude Sonnet 4) Apify API: For Google search results Firecrawl API: For competitor website scraping Template Spreadsheet Copy this template spreadsheet and configure it with your information: Template Link How to set up Step 1: Copy and Configure Template Make a copy of the template spreadsheet Fill in the Client Information sheet: Client name: Your company or client's name Client information: Brief business description URL: Website address Tone of voice: Content style preferences Restrictive instructions: Topics or approaches to avoid Complete the SEO sheet with your target pages: Page: Page you're optimizing (e.g., "Homepage", "Product Page") Keyword: Main search term to target Awareness level: User familiarity with your business Page type: Category (homepage, blog, product page, etc.) Step 2: Import Workflow Import the n8n workflow JSON file Configure all required API credentials in n8n: Google Sheets OAuth2 Anthropic API key Apify API key Firecrawl API key Step 3: Test Configuration Activate the workflow Send your Google Sheets URL to the chat trigger Verify that all sheets are readable and credentials work Test with a single keyword row first Workflow Process Overview Phase 0: Setup and Configuration Copy template spreadsheet Configure client information and SEO parameters Set up API credentials in n8n Phase 1: Data Input and Processing Chat trigger receives Google Sheets URL System reads client configuration and SEO data Filters valid keywords and empty H1 fields Initiates batch processing Phase 2: Competitor Research and Analysis Searches Google for top 10 results per keyword using Apify Scrapes first 5 competitor websites using Firecrawl Extracts heading structures (H1-H6) from competitor pages Analyzes competitor meta tags and content organization Processes markdown content to identify heading hierarchies Phase 3: Meta Tags and H1 Generation AI analyzes keyword context and competitor data using Claude Incorporates client information for personalization Generates optimized meta title (65 characters maximum) Creates compelling meta description (165 characters maximum) Produces user-focused H1 (70 characters maximum) Uses structured output parsing for consistent formatting Phase 4: Content Brief Creation Analyzes search intent percentages (informational, transactional, navigational) Develops content strategy based on competitor analysis Creates detailed MECE page structure with H2 and H3 sections Suggests rich media elements (images, videos, infographics, tables) Provides writing recommendations and detail level scoring (1-10 scale) Ensures SEO optimization while maintaining user relevance Phase 5: Data Integration and Updates Combines all generated content into unified structure Updates Google Sheets with new SEO elements Preserves existing data while adding new content Continues batch processing for remaining keywords Key Differences from Advanced Version This basic version focuses on core SEO functionality without additional complexity: No Vector Database: Removes Supabase integration for simpler setup Streamlined Architecture: Fewer dependencies and configuration steps Essential Features Only: Core competitor analysis and content generation Faster Setup: Reduced time to deployment Lower Costs: Fewer API services required How to customize the workflow Adjusting AI Models Replace Anthropic Claude with other LLM providers in the agent nodes Modify system prompts for different content styles or languages Adjust character limits for meta elements in the structured output parser Modifying Competitor Analysis Change number of competitors analyzed (currently 5) by adding/removing Scrape nodes Adjust scraping parameters in Firecrawl nodes for different content types Modify heading extraction logic in JavaScript Code nodes Customizing Output Format Update Google Sheets column mapping in the final Code node Modify structured output parser schema for different data structures Change batch processing size in Split in Batches node Adding Quality Controls Insert validation nodes between workflow phases Add error handling and retry logic to critical nodes Implement content quality scoring mechanisms Extending Functionality Add keyword research capabilities with additional APIs Include image optimization suggestions Integrate social media content generation Connect to CMS platforms for direct publishing Best Practices Setup and Testing Always test with small batches before processing large keyword lists Monitor API usage and costs across all services Regularly update system prompts based on output quality Maintain clean data in your Google Sheets template Content Quality Review generated content before publishing Customize system prompts to match your brand voice Use descriptive node names for easier workflow maintenance Keep competitor analysis current by running regularly Performance Optimization Process keywords in small batches to avoid timeouts Set appropriate retry policies for external API calls Monitor workflow execution times and optimize bottlenecks Troubleshooting Common Issues and Solutions API Errors Check credential configuration in n8n settings Verify API usage limits and billing status Ensure proper authentication for each service Scraping Failures Firecrawl nodes have error handling enabled to continue on failures Some websites may block scraping - this is normal behavior Check if competitor URLs are accessible and valid Empty Results Verify keyword formatting in Google Sheets Ensure competitor websites contain the expected content structure Check if meta tags are properly formatted in system prompts Sheet Update Errors Ensure proper column mapping in final Code node Verify Google Sheets permissions and sharing settings Check that target sheet names match exactly Processing Stops Review batch processing limits and timeout settings Check for errors in individual nodes using execution logs Verify all required fields are populated in input data Template Structure Required Sheets Client Information: Business details and configuration SEO: Target keywords and page information Results Sheet: Where generated content will be written Expected Columns Keywords: Target search terms Description: Brief page description Type de page: Page category Awareness level: User familiarity level title, meta-desc, h1, brief: Generated output columns This streamlined version provides all essential SEO content generation capabilities while being easier to set up and maintain than the advanced version with vector database integration.
Generate Interior Moodboards with Claude/Gemini Agents, Hugging Face Image Generation, and PDF Export
🎨 Template Overview This comprehensive n8n workflow automates the complete process of generating professional interior design moodboards from concept to client delivery. Users submit a design brief through a form, and the system automatically generates 12 AI-powered images, compiles them into a beautifully formatted two-page PDF moodboard, and emails the final deliverable. Key Features: Form-based design brief submission AI-powered image prompt generation (12 detailed prompts per project) Automated image generation via Hugging Face API Nextcloud cloud storage with public URL sharing Professional two-page HTML/PDF moodboard creation Automated email delivery with PDF attachment Technologies Used: Claude Sonnet 4 (OpenRouter), Google Gemini 2.5 Pro, Hugging Face FLUX.1-schnell, Nextcloud, Gotenberg PDF Service, Gmail --- ⚙️ Self-Hosted Requirements This template requires the following self-hosted or third-party services: Nextcloud Instance - For cloud file storage and public URL generation Gotenberg PDF Service - For HTML to PDF conversion (can be self-hosted via Docker) OpenRouter API Access - For Claude Sonnet 4 AI agent Google Gemini API Access - For secondary AI processing Hugging Face API Access - For FLUX.1-schnell image generation Gmail Account - For email delivery (or any SMTP service) --- 📋 Setup Instructions Step 1: Configure API Credentials OpenRouter (Claude Sonnet 4) Sign up at openrouter.ai Generate an API key Add credentials to the "OpenRouter Chat Model" node Google Gemini API Visit Google AI Studio Create an API key Add credentials to the "Google Gemini Chat Model1" node Hugging Face API Register at huggingface.co Generate an access token from Settings → Access Tokens Add credentials to the "Image Generator" node using HTTP Header Auth Header name: Authorization, Value: Bearer YOUR_TOKEN Nextcloud Set up a Nextcloud instance or use a hosted provider Generate an app password from Settings → Security Configure credentials in all Nextcloud nodes: "Create a folder" "Upload Image" "Share a file" Gotenberg PDF Service Self-host using Docker: docker run --rm -p 3000:3000 gotenberg/gotenberg:8 Or use a hosted instance Update the URL in the "PDF creator" node Configure HTTP Basic Auth credentials if required Gmail Enable 2-Factor Authentication on your Google account Generate an App Password from Google Account settings Add OAuth2 credentials to the "Send PDF" node Step 2: Customize Workflow Settings Email Extractor Node Review the stripPlus variable (default: true) This removes "+tag" portions from email addresses for folder naming Nextcloud Folder Structure Default path: /moodboard/{username}/ Modify in "Create a folder" node if needed Image Generator Settings Model: FLUX.1-schnell (fast generation, good quality) Adjust model in "Image Generator" node if needed Alternative models: FLUX.1-dev, Stable Diffusion XL PDF Generation Settings Default timeout: 360 seconds Page size: A4 (210mm × 297mm) Adjust in "PDF creator" node headers if needed Step 3: Test the Workflow Activate the Form Open the "Moodboard Form" node Copy the webhook URL Access the form in your browser Submit a Test Request Fill in the form fields: Title: Short, descriptive name (e.g., "Modern Minimalist Bedroom") Description: Detailed design brief with colors, materials, mood, lighting Email: Your test email address Submit and monitor workflow execution Verify Each Stage Check Nextcloud folder creation Monitor image generation progress (12 images) Review HTML moodboard generation Confirm PDF creation Check email delivery Step 4: Configure Form Embedding (Optional) Embed the form on your website: html <iframe src="YOURN8NFORMWEBHOOKURL" width="100%" height="800" frameborder="0"> </iframe> --- 🔄 Workflow Structure Form Input & Data Extraction Moodboard Form - Collects project title, description, and user email Email Extractor - Extracts username from email for folder organization Storage Setup Create a Folder - Creates personalized Nextcloud directory using email username AI Concept Generation Conceptualization Agent (Claude Sonnet 4) - Analyzes design brief and generates 12 detailed image prompts (300-500 words each) Images 1-11: Individual design elements (furniture, materials, details, styling) Image 12: Comprehensive 3D rendered view integrating all elements Image Processing Loop Concept Splitter - Separates 12 prompts into individual items Loop Over Items - Processes each prompt sequentially: Image Generator - Sends prompt to Hugging Face FLUX.1-schnell API Upload Image - Stores generated image in Nextcloud folder Share a File - Creates public shareable URL Set Image Title and URL - Formats data for aggregation URL Collection URL Aggregate - Combines all 12 public image URLs Clean URLs - Extracts and formats URLs into a structured list Moodboard Compilation Moodboard Generator Agent (Google Gemini 2.5 Pro) - Creates professional two-page HTML document: Page 1: Visual moodboard with all 12 images (Image 12 prominently featured 2-3x larger) Page 2: Administrative summary with design overview, color palette, materials, and project details PDF Generation & Delivery Binary Converter - Transforms HTML to base64-encoded binary format PDF Creator - Converts HTML to print-ready PDF via Gotenberg service Send PDF - Emails final moodboard PDF to user --- 🎯 Node Descriptions Moodboard Form Collects moodboard generation requests for any design topic. Users input a title, detailed description (colors, materials, patterns, textures, lighting), and email address for delivery. Email Extractor Extracts the username portion from email addresses, optionally stripping "+tags" for clean folder naming and user identification. Create a Folder Creates a dedicated Nextcloud folder using the extracted email username, organizing moodboard outputs by user. Conceptualization Agent AI agent that analyzes design briefs to generate 12 detailed image prompts (300-500 words each). Performs conceptual analysis of styles, colors, materials, and spatial requirements, outputting structured JSON. Concept Splitter Splits the 12 generated image prompts into individual items for parallel processing through the image generation pipeline. Loop Over Items Processes each prompt sequentially, generating images, uploading to Nextcloud, and creating public URLs. Image Generator Sends detailed prompts to Hugging Face FLUX.1-schnell API for AI-powered image generation, transforming written design concepts into high-quality visuals. Upload Image Uploads each generated moodboard image to the user's Nextcloud folder with appropriate naming conventions. Share a File Creates publicly shareable Nextcloud links for each uploaded image, enabling external viewing without authentication. Set Image Title and URL Formats image metadata (title and URL) for downstream aggregation. URL Aggregate Combines all 12 image URLs into a single consolidated output for moodboard compilation. Clean URLs Extracts and formats URLs from the aggregated data into a clean, structured list with count. Moodboard Generator Agent Transforms design concepts into professional two-page HTML moodboards. Analyzes project details and 12 image URLs, selecting appropriate visual styles. Creates artistic Page 1 with Image 12 as hero element, and comprehensive Page 2 with design documentation. Binary Converter Prepares HTML for PDF conversion by transforming it into binary format with proper encoding and filename ("index.html") for Gotenberg compatibility. PDF Creator Converts HTML moodboard into print-ready PDF with proper A4 dimensions, page breaks, and high-quality image resolution. Send PDF Emails the finalized PDF moodboard to the user's submitted email address with project details and PDF attachment. --- 🎨 Customization Options Design Styles The Moodboard Generator Agent automatically selects from 10 layout styles: Modern Sectional Grid Material Board Collage Editorial Magazine Clean Minimalist Split-Screen Dramatic Asymmetric Feature Centered Showcase Modular Block System Organic Flow Layered Depth Image Generation Models Replace FLUX.1-schnell with alternatives in the "Image Generator" node: black-forest-labs/FLUX.1-dev - Higher quality, slower stabilityai/stable-diffusion-xl-base-1.0 - Classic SD XL Email Templates Customize the email message in the "Send PDF" node to include: Brand messaging Next steps Support contact information Pricing information --- 🐛 Troubleshooting Images Not Generating Verify Hugging Face API token is valid Check API rate limits and quotas Increase timeout in "Image Generator" node (default: unlimited) PDF Generation Fails Ensure Gotenberg service is accessible Verify HTML output contains all 12 image URLs Check timeout settings (default: 360s) Review Gotenberg logs for specific errors Nextcloud Upload Errors Confirm folder creation succeeded Verify Nextcloud credentials and permissions Check available storage space Ensure WebDAV is enabled Email Not Received Verify Gmail OAuth2 credentials Check spam/junk folders Confirm email address is valid Review Gmail API quotas --- 📊 Performance Notes Average execution time: 5-8 minutes (depends on image generation) Image generation: ~20-30 seconds per image (12 images = 4-6 minutes) PDF generation: ~30-60 seconds Total data processed: ~15-25 MB per workflow execution --- 🔐 Security Considerations Store all API keys in n8n credentials (never hardcode) Use environment variables for sensitive configuration Implement rate limiting on the form webhook Consider adding CAPTCHA to prevent abuse Regularly rotate API keys and passwords Use HTTPS for all external communications --- 📝 License & Attribution This template is provided as-is for the n8n community. Feel free to modify and adapt to your needs. AI Models Used: Claude Sonnet 4 (Anthropic via OpenRouter) Google Gemini 2.5 Pro (Google) FLUX.1-schnell (Black Forest Labs via Hugging Face) --- 🤝 Support & Contributions For questions or improvements, please reach out through the n8n community forum or submit issues/PRs to enhance this template. Created by: Jameson Kanakulya Template Version: 1.0 Last Updated: November 2025
Article to threaded Bluesky posts with JinaAI and Gemini/GPT
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. Who is this for? Automation enthusiasts, content creators, or social media managers who post article-based threads to Bluesky and want to automate the process end-to-end. What problem is this solving? Manual content repackaging and posting can be repetitive and time-consuming. This workflow automates the process from capturing article URLs (via Telegram or RSS) to scraping content, transforming it into a styled thread, and posting on Bluesky platform. What this workflow does Listens on Telegram or fetches from RSS feeds (AI Trends, Machine Learning Mastery, Technology Review). Extracts content from URLs using JinaAI. Converts the article into a neat, scroll-stopping thread via LangChain + Gemini / OpenAI ChatGPT. Splits the thread into multiple posts. The first post is published with “Create a Post”, while subsequent posts are replies. Adds short delays between posting to avoid rate limits. Setup Add credentials for Telegram Bot API, JinaAI, Google Gemini, and Bluesky App Password. Add or customize RSS feeds if needed Test with a sample URL to validate posting sequence. How to customize Swap out RSS feeds or trigger sources. Modify prompt templates or thread formatting rules in the LangChain/Gemini node. Adjust wait times or content parsing logic. Replace Bluesky with another posting target if desired. Made by: Khaisa Studio Need customs workflows? Contact Me!
Retweet cleanup with scheduling for X/Twitter
Who’s it for Social media managers, creators, and brand accounts that rely on retweets for reach but want an automated, hands-off cleanup after campaigns to keep profiles tidy and on-brand. What it does / How it works On a schedule, the workflow resolves your handle to a user ID, fetches recent tweets, filters retweets only, and safely unretweets them using batching and delays to respect rate limits. A dedicated CONFIG (Set Fields) node centralizes variables (e.g., targetusername, maxresults, batchdelayminutes) so you can adjust behavior without touching logic. API endpoints used GET /2/users/by/username/{username} – resolve handle → user ID GET /2/users/{id}/tweets?tweet.fields=createdat,referencedtweets – fetch recent tweets (identify retweets via referenced_tweets.type === "retweeted") DELETE /2/users/{id}/retweets/{tweet_id} – unretweet Example response payloads GET /2/users/by/username/{username} json { "data": { "id": "2244994945", "name": "Twitter Dev", "username": "TwitterDev" } } GET /2/users/{id}/tweets (truncated) json { "data": [ { "id": "1760000000000000000", "text": "RT @someone: …", "referenced_tweets": [{ "type": "retweeted", "id": "1759999999999999999" }], "created_at": "2025-01-15T09:10:11.000Z" } ], "meta": { "result_count": 20 } } DELETE /2/users/{id}/retweets/{tweet_id} json { "data": { "retweeted": false } } Use cases Brand hygiene: Auto-unretweet promos after 48–72h. Campaign cadence: Remove event retweets once the event ends. Feed freshness: Clear low-priority retweets on a rolling basis. How to set up Open CONFIG (Set Fields) and replace placeholders: targetusername = "yourhandle" max_results = 100 (per fetch) batchdelayminutes = 2 (throttle between batches) Connect X/Twitter credentials in n8n (no keys hard-coded in HTTP nodes). Run once with small values, verify logs, then enable the schedule. > Optional enhancements: add a dead-letter path (Error Trigger → Set → Sheets/Email/Slack) and a notification node (e.g., Slack) for execution feedback.
Automated ClickUp time tracking reports & task performance analytics via Gmail
Description Automate your time tracking and task performance reporting with this comprehensive workflow that transforms ClickUp data into actionable insights. This automation pulls tasks directly from ClickUp, processes due dates, priorities, assignees, and time logs, then generates stakeholder-friendly HTML reports with performance metrics and accountability dashboards. Perfect for teams that need daily ClickUp reporting, time tracking, and automated performance insights without manual data compilation. Features 📅 Daily Task Logs - Complete task overview with due date status, sprint points, and time spent 👤 Performance Metrics - Assignee analysis highlighting top performers and improvement areas 📊 Time Analysis - Status breakdown comparing estimated vs. actual hours logged ⚠ Alert System - Overdue and due-today notifications for immediate action 📧 Professional Reports - Polished HTML emails delivered automatically via Gmail 🎯 Accountability Dashboard - Transparent communication for faster decision-making Requirements ClickUp Access: Workspace admin or member permissions Time tracking enabled on relevant tasks Tasks with assignees, due dates, and time estimates configured Technical Setup: n8n instance (v1.0+) ClickUp API token with read permissions for tasks and time tracking Gmail account with app-specific password or OAuth2 Time tracking data populated in ClickUp tasks Setup Instructions ClickUp Configuration: Enable time tracking in your ClickUp workspace and ensure tasks have time estimates and assignees API Access: Generate ClickUp API token from Settings > Apps > API Token and test permissions Gmail Setup: Create app-specific password (enable 2FA first) and configure SMTP credentials in n8n Import Template: Load workflow JSON and update ClickUp workspace/space/list IDs in the configuration nodes Time Zone: Adjust date calculations and reporting periods for your local timezone Recipients: Configure email distribution list in Gmail node for managers and stakeholders Test Execution: Run manual test to verify data retrieval, time log processing, and report delivery Schedule Setup: Configure daily automation (recommended: 9 AM after team check-ins) Perfect for teams that need daily ClickUp reporting, time tracking, and automated performance dashboards — without manual data entry. Keywords: n8n ClickUp automation, time logs, daily task report, ClickUp to Gmail workflow, automated team performance report, sprint analytics automation, due date alerts, assignee productivity tracking