XHS-Downloader
MCP ServerFree小红书(XiaoHongShu、RedNote)链接提取/作品采集工具:提取账号发布、收藏、点赞、专辑作品链接;提取搜索结果作品、用户链接;采集小红书作品信息;提取小红书作品下载地址;下载小红书作品文件
Capabilities15 decomposed
xiaohongshu work url parsing and metadata extraction
Medium confidenceParses XiaoHongShu (RedNote) work URLs to extract structured metadata including post ID, author information, caption text, image/video URLs, and engagement metrics. Uses HTTP request interception with cookie-based authentication to bypass platform anti-scraping measures and retrieve JSON API responses from XHS endpoints, then deserializes and normalizes the response into a standardized work object with media asset references.
Implements cookie-based session authentication with automatic refresh logic and XHS-specific JSON API endpoint targeting, rather than HTML parsing or Selenium-based browser automation, enabling 10-50x faster extraction with lower resource overhead
Faster and more reliable than browser automation tools (Selenium, Puppeteer) because it directly calls XHS JSON APIs after cookie authentication, avoiding DOM parsing and browser overhead
watermark-free media download with format conversion
Medium confidenceDownloads image and video files from XiaoHongShu work URLs and removes platform watermarks by fetching clean media assets directly from XHS CDN endpoints. Supports batch downloading with customizable file naming patterns (template-based: {work_id}_{index}_{timestamp}), automatic format conversion (MP4 video codec normalization, JPEG/PNG image optimization), and resumable downloads with partial file recovery using HTTP range requests.
Implements a dedicated Download Manager class with resumable HTTP range request support and FFmpeg-based codec normalization, rather than simple file.write() operations, enabling recovery from network interruptions and guaranteed output format compatibility
More robust than generic download tools because it handles XHS-specific CDN authentication, implements resumable downloads with partial file tracking, and automatically normalizes video codecs for cross-platform compatibility
sqlite-based download history and work metadata persistence
Medium confidenceStores all downloaded works, extracted links, and search results in a SQLite database with tables for works (work_id, title, author, media_urls, download_status), downloads (download_id, work_id, timestamp, file_paths), and searches (search_query, result_count, timestamp). Implements deduplication logic to prevent re-downloading the same work, tracks download status (pending, completed, failed), and enables querying download history by date range, author, or content type. Database schema includes indexes on frequently-queried columns (work_id, timestamp) for performance.
Implements SQLite schema with deduplication indexes and download status tracking, enabling efficient duplicate detection and resumable downloads, rather than simple file-based logging
More reliable than file-based logging because it provides structured querying, deduplication, and transactional consistency, enabling complex analysis and preventing accidental re-downloads
cookie-based authentication with automatic session refresh
Medium confidenceManages XiaoHongShu session authentication by storing and refreshing cookies in a persistent cookie jar. Reads cookies from browser storage (via browser extension or manual export) or accepts cookies as configuration input. Implements automatic cookie refresh logic that detects expired sessions (HTTP 401 responses) and attempts to refresh cookies using stored refresh tokens or re-authentication flow. Validates cookie freshness before each request and logs authentication failures for debugging.
Implements automatic cookie refresh detection (HTTP 401 response handling) with fallback re-authentication flow, rather than requiring manual cookie updates, enabling long-running processes without user intervention
More reliable than manual cookie management because it automatically detects and refreshes expired sessions, reducing authentication failures and enabling unattended operation
customizable file naming and folder organization with template variables
Medium confidenceSupports template-based file naming and folder organization using variable substitution. Naming templates can include variables like {work_id}, {author}, {title}, {timestamp}, {index} which are replaced with actual values from work metadata. Implements folder structure templates (e.g., {author}/{timestamp}/{work_id}) for organizing downloads into hierarchical directories. Validates template syntax and provides default templates for common use cases (flat structure, author-based organization, date-based organization).
Implements variable substitution with metadata-driven template expansion and automatic special character sanitization, rather than fixed naming schemes, enabling flexible organization without code changes
More flexible than tools with fixed naming schemes because it supports arbitrary folder hierarchies and file naming patterns, enabling users to organize downloads according to their own preferences
batch processing with rate limiting and error recovery
Medium confidenceSupports batch downloading of multiple XHS URLs with configurable rate limiting to avoid triggering XHS anti-scraping measures. Implements exponential backoff retry logic for failed downloads (retry up to 3 times with increasing delays), tracks download progress across the batch, and provides detailed error reports for failed items. Rate limiting is configurable (requests per second, delay between downloads) and can be adjusted based on observed XHS response patterns.
Implements exponential backoff retry logic with configurable rate limiting and detailed error tracking, rather than simple sequential processing, enabling robust batch operations that recover from transient failures
More reliable than simple batch scripts because it automatically retries failed downloads, implements rate limiting to avoid IP blocking, and provides detailed error reports for debugging
configuration management with settings.json persistence and validation
Medium confidenceManages all user-configurable parameters through a settings.json file with schema validation and default values. Supports configuration hierarchy: command-line arguments override settings.json, which overrides built-in defaults. Implements configuration validation (type checking, range validation for numeric fields, enum validation for choice fields) and provides clear error messages for invalid configurations. Automatically migrates settings.json schema when application version changes, preserving user settings while adding new fields.
Implements configuration hierarchy (CLI args > settings.json > defaults) with schema validation and automatic migration, rather than hard-coded defaults, enabling flexible configuration without code changes
More maintainable than tools with hard-coded configuration because it supports persistent settings, command-line overrides, and automatic schema migration, reducing user friction and supporting multiple deployment scenarios
user profile link extraction and work collection aggregation
Medium confidenceExtracts and aggregates work links from XiaoHongShu user profiles across multiple collection types: published works, bookmarked/saved posts, liked posts, and custom albums. Uses paginated API requests to the XHS user profile endpoint with cursor-based pagination, iterating through all available pages to build a complete inventory of work URLs. Stores extracted links in SQLite database with metadata (collection type, extraction timestamp, user ID) for deduplication and tracking.
Implements cursor-based pagination state management with SQLite deduplication tracking, rather than simple list accumulation, enabling recovery from interruptions and prevention of duplicate URL extraction across multiple runs
More complete than manual profile browsing because it automatically handles pagination across all work collections and stores results persistently, avoiding manual copy-paste and enabling batch processing of multiple profiles
search result link extraction and filtering
Medium confidenceExecutes XiaoHongShu search queries and extracts work and user links from paginated search results. Sends search requests to XHS search API endpoint with query parameters (keyword, filters, sort order), processes paginated JSON responses containing work and user cards, and extracts URLs with optional filtering by content type, engagement metrics, or publication date. Results are stored in SQLite with search metadata for reproducibility.
Implements search result pagination with XHS-specific ranking algorithm awareness and stores search metadata (query, timestamp, result count) in SQLite, enabling reproducible search result tracking and trend analysis over time
More systematic than manual search browsing because it automates pagination, stores results persistently, and enables filtering and analysis of search trends across multiple queries and time periods
multi-interface request routing and execution mode dispatch
Medium confidenceImplements a single entry point (main.py) that dispatches execution to five distinct user interfaces based on command-line arguments and configuration: Terminal UI (TUI) for interactive use, CLI for single-command automation, Browser UserScript for in-browser convenience, REST API Server for programmatic integration, and MCP Server for AI assistant integration. Each interface converges on the core XHS class, which coordinates content extraction, download, and storage operations through a shared processing pipeline.
Implements a unified architecture where five distinct interfaces (TUI, CLI, UserScript, REST API, MCP) all converge on a single XHS core class, rather than maintaining separate codebases, enabling consistent behavior and simplified maintenance across all deployment modes
More flexible than single-interface tools because it supports interactive, scripted, browser-based, programmatic, and AI-integrated workflows from the same codebase, reducing deployment complexity and enabling seamless switching between use cases
terminal user interface (tui) with clipboard monitoring and interactive settings
Medium confidenceProvides an interactive terminal-based UI built with the Textual framework that monitors system clipboard for XHS URLs, displays real-time download progress, and allows users to configure settings without editing JSON files. Implements a multi-panel layout with URL input field, download queue display, progress bars, and settings editor. Clipboard monitoring runs in a background thread that detects new XHS URLs and automatically queues them for download when enabled.
Implements background clipboard monitoring thread with Textual event loop integration, enabling real-time URL detection and automatic queue management without blocking the UI, rather than polling-based clipboard checks that freeze the interface
More user-friendly than CLI-only tools because it provides visual feedback, real-time progress tracking, and interactive settings management without requiring JSON file editing or command-line knowledge
command-line interface (cli) with selective media index specification
Medium confidenceProvides a command-line interface for single-command downloads with support for specifying which images/videos to download from a multi-media post. Accepts XHS URL as argument, optional image index range (e.g., --images 1-3 to download only images 1-3), and configuration overrides (--output-dir, --naming-template). Executes download synchronously and exits with status code indicating success/failure, suitable for shell scripts and CI/CD pipelines.
Implements selective media index parsing with range syntax support (1-3, 1,2,3) and validates indices against actual post media count before download, rather than blindly accepting any index specification
More scriptable than interactive tools because it supports non-blocking execution, exit status codes, and configuration overrides, enabling seamless integration into shell scripts and CI/CD pipelines
browser userscript integration with dual-mode operation
Medium confidenceProvides a Tampermonkey/Greasemonkey UserScript that runs in the browser and enables in-browser downloads or server-push mode. In standalone mode, the script uses browser APIs (Fetch API, Blob) to download files directly to the user's Downloads folder. In server-push mode (script_server=true in config), the script detects XHS URLs on the page, extracts work IDs, and sends download requests to a running XHS-Downloader instance via HTTP POST, offloading processing to the server while providing browser-side convenience.
Implements dual-mode operation where the same UserScript can function standalone (browser-based downloads) or server-push (task delegation), with configuration-driven mode selection, rather than requiring separate scripts for each mode
More convenient than CLI/TUI tools because it integrates directly into the browser workflow, enabling one-click downloads without switching windows or opening terminals
rest api server with json request/response protocol
Medium confidenceExposes XHS-Downloader functionality as a REST API server running on port 5556 with FastAPI framework. Provides endpoints for work detail retrieval (/xhs/detail), download submission (/xhs/download), and download status polling (/xhs/status). Accepts JSON request bodies with XHS URLs and configuration parameters, processes requests asynchronously using a task queue, and returns JSON responses with download status, file paths, and error messages. Supports CORS for cross-origin requests from web applications.
Implements FastAPI-based REST API with asynchronous task queue and CORS support, rather than simple HTTP server, enabling concurrent request handling and cross-origin web application integration
More scalable than CLI/TUI tools because it supports concurrent requests, enables programmatic integration with web applications, and can be deployed as a containerized microservice
model context protocol (mcp) server integration for ai assistants
Medium confidenceExposes XHS-Downloader as an MCP Server running on port 5556, enabling AI assistants (Claude, ChatGPT with plugins, etc.) to call XHS-Downloader functions as tools. Implements MCP protocol handlers for work extraction, download submission, and status checking. AI assistants can invoke these tools within their reasoning loops, enabling autonomous content extraction and download workflows orchestrated by the AI model.
Implements MCP Server protocol handlers that expose XHS-Downloader as callable tools for AI assistants, enabling autonomous content extraction within AI reasoning loops, rather than requiring manual user invocation
Enables AI-driven automation that CLI/TUI/API tools cannot achieve; AI assistants can autonomously decide when to extract content, analyze results, and adapt workflows based on outcomes
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with XHS-Downloader, ranked by overlap. Discovered automatically through the match graph.
XHS-Downloader
小红书(XiaoHongShu、RedNote)链接提取/作品采集工具:提取账号发布、收藏、点赞、专辑作品链接;提取搜索结果作品、用户链接;采集小红书作品信息;提取小红书作品下载地址;下载小红书作品文件
xiaohongshu-mcp
MCP for xiaohongshu.com
img2dataset
Easily turn a set of image urls to an image dataset
Agent-Reach
Give your AI agent eyes to see the entire internet. Read & search Twitter, Reddit, YouTube, GitHub, Bilibili, XiaoHongShu — one CLI, zero API fees.
local-deep-research
Local Deep Research achieves ~95% on SimpleQA benchmark (tested with GPT-4.1-mini). Supports local and cloud LLMs (Ollama, Google, Anthropic, ...). Searches 10+ sources - arXiv, PubMed, web, and your private documents. Everything Local & Encrypted.
mcp-smart-crawler
A command-line tool acting as an MCP (ModelContextProtocol) server, using Playwright to crawl web content for AI models.
Best For
- ✓Developers building content aggregation tools targeting XHS
- ✓Data analysts collecting social media metrics from Chinese platforms
- ✓Teams migrating content from XHS to other platforms
- ✓Content creators archiving their own XHS posts
- ✓Researchers collecting media datasets from XHS
- ✓Teams building content backup/migration pipelines
- ✓Content creators maintaining a personal archive of their own posts
- ✓Researchers analyzing download patterns and content trends
Known Limitations
- ⚠Requires valid XHS session cookies; authentication fails if cookies expire or are revoked
- ⚠Rate-limited by XHS servers; batch extraction of 100+ URLs may trigger temporary IP blocks
- ⚠Cannot extract private/deleted posts or content from suspended accounts
- ⚠Metadata structure may change with XHS platform updates, requiring code maintenance
- ⚠Video codec conversion adds 30-120 seconds per video depending on resolution and duration
- ⚠XHS CDN may serve region-locked content; downloads may fail from certain geographic locations
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
Repository Details
Last commit: Apr 21, 2026
About
小红书(XiaoHongShu、RedNote)链接提取/作品采集工具:提取账号发布、收藏、点赞、专辑作品链接;提取搜索结果作品、用户链接;采集小红书作品信息;提取小红书作品下载地址;下载小红书作品文件
Categories
Alternatives to XHS-Downloader
Are you the builder of XHS-Downloader?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →