os-level passive context capture with automatic enrichment
Runs a background daemon (PiecesOS) that monitors OS-level events across all applications in real-time, capturing code snippets, browser tabs, chat messages, documents, and highlights without user intervention. The Workstream Pattern Engine ingests millions of micro-events and routes them through on-device classification models (TF-IDF, SVMs, LSTMs, RNNs) to automatically detect code, extract metadata, flag sensitive data (PII/credentials), and associate context (source app, timestamp, related files/tabs). Captured data is stored locally in a proprietary database with optional cloud sync via Pieces Drive.
Unique: Uses OS-level daemon with Workstream Pattern Engine to passively capture millions of micro-events across all applications in real-time, automatically enriching with on-device ML models (TF-IDF, SVM, LSTM) rather than requiring manual tagging or bookmarking. Hardware-accelerated offline models enable real-time memory association without cloud transmission.
vs alternatives: Captures context automatically across all tools without user action, unlike GitHub Gist or Pastebin which require manual save, and unlike browser bookmarks which lack code-specific enrichment and sensitive data detection.
natural language search across 9-month memory with time-based filtering
Indexes all captured snippets, documents, and activity with vector embeddings, enabling semantic search via natural language queries. Users can search across 9 months of personal context and filter by time-based queries (e.g., 'code I wrote last Tuesday', 'snippets from the past week'). The search engine ranks results by relevance and associates results with the 'bigger picture' — implied relationship graph linking snippets to related chats, tabs, and documents. Queries are processed locally by default; optional cloud search available via Pieces Drive.
Unique: Combines vector-based semantic search with time-based filtering and implicit relationship graphs linking snippets to related activity (chats, tabs, documents), enabling 'bigger picture' context retrieval rather than isolated snippet matching. Local-first processing avoids cloud transmission of search queries.
vs alternatives: Searches personal context (not generic knowledge), supports time-based filtering, and associates results with related activity — unlike GitHub Gist search or IDE snippet managers which lack temporal filtering and activity correlation.
local-first data storage with optional cloud sync
All captured context and snippets are stored in a local, proprietary database on the user's machine by default. Cloud sync via Pieces Drive is optional and user-controlled — users can enable/disable sync at any time. No data is transmitted to cloud unless explicitly enabled. Local storage uses vector embeddings for semantic search and supports 9 months of retention with automatic deletion of older data.
Unique: Stores all data locally by default with optional cloud sync via Pieces Drive, giving users explicit control over cloud transmission. Uses proprietary database format with vector embeddings for local semantic search.
vs alternatives: Keeps data local by default (unlike cloud-first tools like GitHub Gist), enables offline access (unlike cloud-only solutions), and gives users control over sync (unlike automatic cloud backup).
hardware-accelerated on-device ml inference for real-time classification
Uses hardware acceleration (GPU, NPU, or CPU optimization — specific method undocumented) to run on-device ML models (TF-IDF, SVM, LSTM, RNN) in real-time as context is captured. Models classify code, detect language, associate context, and flag sensitive data without cloud transmission. Hardware acceleration enables low-latency inference on millions of micro-events per day.
Unique: Uses hardware acceleration (method undocumented) to run on-device ML models in real-time, enabling low-latency classification and context association without cloud transmission. Processes millions of micro-events per day.
vs alternatives: Runs inference locally without cloud latency (unlike cloud-based ML services), processes in real-time as code is captured (unlike batch processing), and avoids cloud transmission of sensitive code (unlike cloud ML APIs).
automatic language detection and code metadata extraction
On-device models automatically detect programming language, framework, and code type (function, class, snippet, etc.) from captured code. Extracted metadata is stored with the snippet and used for search, filtering, and context association. Detection runs in real-time without user input or cloud transmission.
Unique: Automatically detects language, framework, and code type from captured snippets using on-device models, enabling semantic filtering and search without user tagging. Detection is real-time and requires no cloud transmission.
vs alternatives: Detects language automatically (unlike manual tagging), runs locally (unlike cloud-based language detection), and enables semantic search (unlike keyword-only search).
pieces drive cloud sync with optional team collaboration
Optional cloud sync service (Pieces Drive) that synchronizes local memory to cloud storage for backup, multi-device access, and team collaboration. Users can enable/disable sync at any time. Sync mechanism (incremental, full, real-time) is undocumented. Team collaboration via Pieces Drive enables shared memory across team members with role-based access control.
Unique: Provides optional cloud sync (Pieces Drive) for backup and multi-device access, with team collaboration features (shared memory, role-based access). Sync is user-controlled and can be disabled at any time.
vs alternatives: Enables multi-device access (unlike local-only storage), provides backup (unlike unprotected local storage), and supports team collaboration (unlike personal-only tools).
context-aware copilot with multi-llm backend selection
Provides an AI copilot that accepts user queries and automatically injects personal context (saved snippets, activity history, related documents) before routing to a user-selected LLM backend. Supports Claude (4 Sonnet, Opus), Gemini 2.5, OpenAI models, and Ollama (local). The copilot 'knows what you know, not just what the LLM knows' — meaning it personalizes responses based on your saved code, patterns, and project context. Integrates via MCP (Model Context Protocol) server built into PiecesOS, enabling direct injection into Claude, GitHub Copilot, Cursor, and Goose.
Unique: Injects personal context (saved snippets, activity history) into user-selected LLM via MCP protocol, enabling copilot functionality that understands your specific codebase and patterns. Supports multiple LLM backends (Claude, OpenAI, Gemini, Ollama) with user-controlled switching, avoiding lock-in to a single provider.
vs alternatives: Personalizes LLM responses with your own code and context (unlike GitHub Copilot which uses generic training data), supports multiple LLM backends (unlike Copilot which is OpenAI-only), and integrates via MCP (unlike proprietary copilot APIs which are tool-specific).
code snippet transformation and language conversion
Accepts saved code snippets and applies transformations: change programming language, improve readability, optimize performance, or refactor for specific patterns. Transformations are executed by the selected LLM with personal context injected, enabling suggestions that align with your coding style and project patterns. Output can be previewed, edited, and re-saved to memory.
Unique: Transforms code with personal context injected, enabling suggestions that align with your coding style and project patterns rather than generic LLM defaults. Integrates with multi-LLM backend selection, allowing user to choose transformation engine.
vs alternatives: Personalizes transformations with your context (unlike generic LLM code conversion which ignores your patterns), integrates with your saved snippets (unlike standalone code converters), and supports multiple LLM backends.
+6 more capabilities