ChipOS Documentation
The complete guide to the AI Operating System for Hardware Design. Learn how to leverage world-leading RAG technology for semiconductor development.
What is ChipOS?
ChipOS is a complete AI operating system designed specifically for semiconductor engineers. Built on top of VS Code, it provides a standalone development environment with integrated knowledge management, AI-powered assistance, and seamless integration with modern AI tools.
For Hardware Engineers
- Store and search datasheets, RTL specs, and verification plans
- Generate SystemVerilog and UVM testbenches with AI
- Query your knowledge base using natural language
For AI/ML Engineers
- Deploy and benchmark models on Metis hardware
- Create inference pipelines with YAML configuration
- Access SDK documentation through RAG-powered search
Installation
ChipOS is a standalone application with its own data folder. It won't interfere with your existing VS Code installation.
Open the DMG and drag to Applications
The app is notarized by Apple — no security warnings.
Launch ChipOS
Find it in your Applications folder or Spotlight.
Quick Start
Get ChipOS connected to your AI assistant in under 5 minutes.
Open a Workspace
In the Controls panel, click "Open Workspace"
to select your project folder. ChipOS automatically installs 12 Voyager SDK
commands to .claude/commands/
Controls → Open Workspace → Select Folder
Connect AI Provider
Click "Connect AI Provider" and select your preferred tool:
Claude Code
Cursor
Windsurf
Gemini
Open Dashboard (Optional)
Click "Open Dashboard" to access the full ChipOS UI with Knowledge Base, AI Chatbot, and more.
Start Using Commands
In Claude Code terminal, try your first command:
/voyager-help How do I create a detection pipeline?
Claude will query the ChipOS knowledge base and provide contextual answers!
Configuration
ChipOS settings can be configured through VS Code's settings UI or settings.json.
| Setting | Default | Description |
|---|---|---|
| chipos.backend.mode | "hosted" | Backend mode: "hosted" (cloud) or "local" (Docker) |
| chipos.backend.hostedUrl | https://chipos-server-*.run.app | Cloud backend API URL |
| chipos.backend.apiUrl | http://localhost:8181 | Local backend API URL |
| chipos.showWelcomeOnStartup | true | Open dashboard on startup |
Knowledge Base
The Knowledge Base is your central repository for hardware documentation. Store datasheets, RTL specifications, verification plans, and SDK documentation — all searchable with AI.
Web Crawling
Automatically crawl documentation websites using sitemap discovery.
- Sitemap-first crawling
- Recursive link following
- HTML to markdown conversion
Repository Ingestion
Index entire GitHub repositories with AST-aware code parsing.
- 25+ programming languages
- Function/class extraction
- Dependency graph building
Document Upload
Upload and index PDF datasheets and documentation.
- PDF text extraction
- DOCX support
- Markdown/text files
Sidebar Browser
Browse all indexed sources directly in the sidebar.
- Tree view navigation
- Source filtering
- Quick search
AI Chatbot
The built-in AI chatbot provides hardware-aware conversations with full syntax highlighting for code.
Supported Languages
The chatbot automatically queries your knowledge base for hardware-specific questions, providing accurate, contextual answers.
MCP Server
The Model Context Protocol (MCP) server enables AI assistants to access your knowledge base. ChipOS runs a lightweight HTTP-based MCP server on port 8051.
Lightweight
HTTP-based protocol, no process spawning
Auto-Discovery
Tools automatically registered and documented
Rate-Limited
Built-in quota management for LLM APIs
Available MCP Tools
| Tool | Description |
|---|---|
| perform_rag_query | Search the knowledge base with semantic understanding |
| get_available_sources | List all indexed documentation sources |
| health_check | Check backend service health status |
| session_info | Get current session and user information |
Claude Commands
When you open a workspace in ChipOS, 12 Voyager SDK slash commands are automatically installed
to .claude/commands/. These
work on macOS, Linux, and Windows.
Get help from the ChipOS RAG knowledge base
Create a complete YAML pipeline configuration
Compile and deploy a model for Metis hardware
Run inference with various input sources
Benchmark model performance metrics
Create a Python app using InferenceStream
Debug pipeline, model, or inference issues
Run tests for SDK components
Build native operators and examples
Add a custom model to the model zoo
List and search available models
RAG Technology
ChipOS features a proprietary Retrieval-Augmented Generation architecture designed specifically for the complexity of semiconductor design.
Multi-Embedding Strategy
Simultaneously uses 3 different embedding models for comprehensive coverage:
cAST Chunking
Context-aware Abstract Syntax Tree chunking preserves code structure:
Embedding Models
ChipOS supports multiple embedding providers with automatic fallback and no vendor lock-in.
| Provider | Model | Dimensions | Use Case |
|---|---|---|---|
| Voyage AI | voyage-code-3 | 1024 | Default — Code-optimized |
| OpenAI | text-embedding-3-small | 1536 | General semantic search |
| OpenAI | text-embedding-3-large | 3072 | High-precision retrieval |
| text-embedding-004 | 768 | Alternative provider | |
| Ollama | nomic-embed-text | 1024 | Local/on-premise deployment |
Embeddings are generated in batches of 100 with automatic rate limiting and exponential backoff for API quota management.
Semantic Search
Unlike keyword matching, ChipOS understands engineering intent and connects concepts across documents.
How It Works
Query Embedding
Your question is converted to a vector using the same embedding model
Vector Similarity
HNSW index finds nearest neighbors using cosine similarity (sub-millisecond on millions of vectors)
Context Retrieval
Top-K relevant chunks are retrieved with source attribution
Hybrid Search
Combines vector similarity with BM25 keyword matching for comprehensive coverage.
Semantic Component
- Understands meaning and context
- Handles synonyms and paraphrasing
- Connects related concepts
Lexical Component
- Exact term matching (BM25)
- CamelCase/snake_case aware
- Handles technical identifiers
Auto-Code Extraction
ChipOS automatically identifies, extracts, and indexes code snippets from documentation — no manual tagging required.
When you ask "show me an example of a UVM sequence", ChipOS doesn't just find pages that mention sequences — it finds actual, runnable code examples.
Extracted Metadata
Imports
Definitions
Function Calls
Parent Scope
Node Types
Dependencies
Corrective RAG
ChipOS implements a 5-agent self-correcting RAG pipeline based on academic research.
Five-Agent Pipeline
Context Retrieval
Initial document search
Relevance Eval
Grade with confidence
Query Refine
Rewrite if poor results
External KB
Web search fallback
Synthesize
Generate answer
Iterative Refinement
Up to 2 query rewrites if initial results are poor
Confidence Scoring
LLM-based relevance grading with thresholds
External Fallback
Web search when internal KB is insufficient
Source Attribution
Full query history and citation tracking
Advanced RAG
State-of-the-art retrieval enhancements based on RAPTOR, GraphRAG, and production RAG research.
Unified Search
Unified Search resolves the issue where repository-ingested content wasn't searchable through the knowledge base API. Now all content sources are searched simultaneously.
ChipOS stores content in two tables: chipos_crawled_pages for
web documentation and
chipos_code_examples for repository code. Unified Search
queries both simultaneously with content-type awareness.
chipos_crawled_pages
- Web crawled documentation
- Uploaded documents (PDF, DOCX)
- Hierarchical summaries
chipos_code_examples
- Repository-ingested code
- Documentation from repos (README, docs/)
- Config files (YAML, JSON, TOML)
Content Type Classification
Every chunk is automatically tagged with a content type during ingestion:
Code
py, ts, js, go, rs, java, etc.
Documentation
md, txt, rst, markdown
Config
yaml, json, toml, ini, env
Query Classification
ChipOS analyzes query intent using keyword patterns to route queries to the optimal search strategy.
| Intent | Keywords | Strategy |
|---|---|---|
| Code Specific | "function", "class", "def", "implementation" | Prioritize code examples table |
| Conceptual | "how to", "what is", "explain", "guide" | Prioritize documentation + summaries |
| API Reference | "api", "endpoint", "parameter", "schema" | Prioritize API docs |
| Debugging | "error", "bug", "fix", "exception" | Search both code and docs |
Context Expansion
When retrieving code chunks, ChipOS automatically includes surrounding chunks from the same file to provide full context.
Parent-Child Relationships
Each code chunk stores metadata about its position in the original file:
file_hash
Groups chunks from same file
chunk_index
Position in file (0-indexed)
total_chunks
Total chunks in file
file_path
Full path for context
When a code chunk is retrieved, Context Expansion fetches adjacent chunks (before and after) to provide surrounding code, making it easier for AI assistants to understand the full context.
Cross-Encoder Reranking
Two-stage retrieval: first fetch candidates with vector search, then rerank with a CrossEncoder model for 15-20% better relevance.
Reranking Pipeline
Initial Search
Fetch 5× candidates
CrossEncoder
Score each candidate
Top Results
Return best N
| Setting | Default | Description |
|---|---|---|
USE_RERANKING |
true | Enable/disable reranking |
RERANKING_MODEL |
ms-marco-MiniLM-L-6-v2 | CrossEncoder model (fast + accurate) |
RERANK_CANDIDATE_MULTIPLIER |
5 | Fetch N× results for reranking |
RERANK_MAX_CANDIDATES |
100 | Maximum candidates to rerank |
RAG Health Monitoring
ChipOS automatically evaluates RAG quality after content is uploaded or crawled, detecting issues before they impact users.
Healthy
75-100% score
RAG working well
Degraded
50-74% score
Some issues detected
Unhealthy
<50% score
Critical problems
Health Score Algorithm
health_score = (
retrieval_success_rate * 0.40 + # Did queries return results?
avg_relevance_score * 0.40 + # How relevant were results?
system_health * 0.20 # No API errors?
)
Voyage AI Integration
Voyage AI's voyage-code-3 is now the
default embedding model, offering 13-17% better code retrieval than OpenAI with 33%
less memory usage.
| Metric | voyage-code-3 | text-embedding-3-small |
|---|---|---|
| Code Retrieval | 92.4% | 79.2% |
| Dimensions | 1024 | 1536 |
| Memory/vector | ~4KB | ~6KB |
| MRL Support | Yes | No |
Configuration
# Settings → RAG Settings → LLM Provider
LLM_PROVIDER=voyage
EMBEDDING_MODEL=voyage-code-3
# API Key (Settings → API Keys)
VOYAGE_API_KEY=pa-xxxx
System Architecture
Port 8181 • FastAPI + Socket.IO
Port 8051 • HTTP-based MCP
Port 8052 • PydanticAI
Technology Stack
| Component | Technology | Purpose |
|---|---|---|
| Desktop App | VS Code (Electron) | Cross-platform native app |
| UI Framework | React + TypeScript + Vite | Webview panels and sidebar |
| UI Components | Radix UI + TanStack Query | Accessible primitives + data fetching |
| Backend | Python + FastAPI + uv | High-performance API server |
| Vector DB | PostgreSQL + pgvector | HNSW indexes for similarity search |
| Code Parsing | tree-sitter | AST-aware chunking (25+ languages) |
| MCP Protocol | FastMCP (HTTP) | AI assistant integration |
| AI Agents | PydanticAI | Corrective RAG pipeline |
| Observability | Logfire | Distributed tracing and logging |
| Deployment | Docker + Google Cloud Run | Containerized microservices |
Database Schema
ChipOS uses Supabase (PostgreSQL 15+) with the pgvector extension for efficient vector similarity search.
chipos_crawled_pages
Stores crawled documentation with embeddings
- • embedding: vector(1536)
- • bm25_tokens: tsvector
- • url, title, content
- • source_id, metadata
chipos_code_examples
Indexed code with multi-dimensional embeddings
- • embedding: vector(1536)
- • code_embedding: vector(768)
- • late_embedding: vector(1024)
- • definitions, calls, imports[]
chipos_sources
Registered documentation sources
- • name, url, type
- • status, last_crawled
- • page_count, metadata
chipos_code_relationships
Code dependency graph
- • source_id, target_id
- • relationship_type
- • weight, metadata
AI Integrations
ChipOS connects seamlessly with your favorite AI coding assistants through the MCP server.
Claude Code
Full integration with auto-installed slash commands
Fully SupportedCursor
MCP server configuration for context injection
Fully SupportedWindsurf
Native MCP tool access for knowledge queries
Fully SupportedCommands Reference
Query the ChipOS knowledge base for hardware-related help.
/voyager-help How do I configure a detection pipeline with YOLOv8?
Create a complete YAML pipeline configuration for inference.
/new-pipeline yolov8n --input camera --output display
Compile and deploy a model to Metis hardware.
/deploy yolov8n --target m.2
Benchmark model performance metrics (latency, throughput, power).
/bench yolov8n --iterations 1000
API Reference
The ChipOS backend exposes REST and MCP APIs for integration.
REST Endpoints
| Method | Endpoint | Description |
|---|---|---|
| GET | /api/health | Health check |
| POST | /api/rag/query | RAG query with context |
| GET | /api/sources | List indexed sources |
| POST | /api/sources/crawl | Start web crawl |
| POST | /api/sources/ingest | Ingest GitHub repository |
Settings Overview
ChipOS provides extensive configuration options to customize RAG behavior, code extraction, and system performance.
Features
- Dark Mode - Toggle light/dark theme
- Projects - Enable task management
- Logfire - Structured logging
- Disconnect Screen - Server status display
API Keys
- OpenAI - RAG queries & embeddings
- Google - Alternative LLM provider
- GitHub - Private repo crawling
- Encrypted storage with show/hide toggle
RAG Settings
- LLM Provider - OpenAI, Gemini, Ollama
- Hybrid Search - Vector + keyword search
- Agentic RAG - Code-aware retrieval
- Reranking - Cross-encoder scoring
- Crawling - Batch size, concurrency
Code Extraction
- Block Length - Min/max characters
- Language Patterns - Per-language detection
- Prose Filtering - Remove non-code content
- Code Summaries - AI-generated descriptions
For detailed descriptions of all 50+ settings, recommended configurations, and troubleshooting tips.
Troubleshooting
MCP Server Not Connecting
If Claude/Cursor can't find the MCP server:
- Verify the backend is running (check Controls panel)
- Ensure port 8051 is not blocked by firewall
- Try restarting ChipOS with
Cmd+Shift+P → Restart Services
Slash Commands Not Working
If /voyager-help doesn't respond:
- Check that
.claude/commands/exists in your workspace - Verify commands were auto-installed (12 .md files should exist)
- Restart Claude Code terminal session
Empty RAG Results
If queries return no results:
- Check that sources are indexed in the Knowledge Base sidebar
- Verify embeddings were generated (check server logs)
- Try a broader query to test the connection