Docs

ContextuAI Solo User Guide

Installation, features, models, and troubleshooting

Installation

Download the installer for your platform and run it — no build tools, no terminal commands, no dependencies to install.

Step 1: Download

Go to the GitHub Releases page and download the latest version for your OS:

PlatformFileNotes
Windows.msi or .exeDouble-click to install. You may need to click "More info" → "Run anyway" on the SmartScreen prompt.
macOS.dmgOpen the DMG, drag ContextuAI Solo to Applications. On first launch, right-click → Open to bypass Gatekeeper.
Linux.AppImage or .debFor AppImage: chmod +x then run. For .deb: sudo dpkg -i contextuai-solo.deb

Step 2: Launch

Open ContextuAI Solo from your applications. On first launch, the Setup Wizard guides you through choosing an AI provider and configuring your profile. That's it — you're ready to go.

No API key? No problem. Select Ollama as your provider in the wizard — Solo will download a free local model and you can start chatting with zero cost, fully offline.

Building from Source (Advanced)

If you prefer to build from source (contributors, developers):

# Prerequisites: Node.js 18+, Python 3.11+, Rust (rustup.rs) git clone https://github.com/contextuai/contextuai-solo.git cd contextuai-solo # Frontend cd frontend && npm install # Backend cd ../backend && pip install -r requirements.txt # Run (two terminals) # Terminal 1: cd backend && CONTEXTUAI_MODE=desktop ENVIRONMENT=desktop python -m uvicorn desktop_app:app --host 127.0.0.1 --port 18741 # Terminal 2: cd frontend && npm run dev # Open http://localhost:1420

Setup Wizard

On first launch, a 3-step wizard walks you through configuration:

  1. Profile — Enter your name, business name, and select your industry (16 options: SaaS, E-commerce, Healthcare, Finance, Education, Marketing, Consulting, Real Estate, Legal, Manufacturing, Non-Profit, Freelancing, and more)
  2. AI Provider — Choose how you want to run AI:

    Built-in Local AI Free · No Setup

    Solo ships with a built-in Model Hub — 39 GGUF models you can download and run directly inside the app. No API keys, no external tools, no Ollama, no Python — just pick a model, click download, and start chatting. Everything runs on your CPU, fully offline.

    This is the fastest way to get started. The wizard lets you pick a recommended model based on your RAM:

    Your RAMRecommended ModelDownload Size
    4-8 GBQwen 3.5 0.8B or Gemma 3 1B~700 MB
    8-16 GBQwen 3 8B or DeepSeek R1 7B~4-5 GB
    16-32 GBGemma 4 12B or Qwen 3 14B~7-9 GB
    32+ GBGemma 4 27B or DeepSeek R1 32B~16-20 GB
    New: Gemma 4 — Google's latest model family (April 2025). Gemma 4 12B rivals models twice its size on reasoning and instruction-following. Gemma 4 27B is one of the strongest open models available. Both run fully on CPU inside Solo — our top recommendation for 16GB+ machines.

    Cloud Providers BYOK

    If you have API keys from cloud providers, you can add them for access to the latest frontier models:

    • Anthropic Claude — Sonnet 4, Opus, Haiku (Get API key)
    • OpenAI — GPT-4o, GPT-4o Mini, GPT-4 Turbo, o1-preview (Get API key)
    • Google Gemini — 2.0 Flash, 2.0 Pro, 1.5 Flash (Get API key)
    • AWS Bedrock — Claude and Titan models
    • Ollama — If you already run Ollama locally, Solo can connect to it as an additional provider

    You can use local + cloud together — switch between them anytime from the model dropdown in chat.

  3. Brand Voice — Define your target audience, content topics, and brand tone
Zero-cost path: Pick a built-in local model in the wizard and you're chatting in minutes — no accounts, no API keys, no subscriptions. Add cloud providers later from Settings if you want.

Your First Chat

After the wizard, you land on the Chat page:

  1. Type a message in the input box at the bottom
  2. Select an AI model from the dropdown (shows provider badges: Local/Cloud)
  3. Optionally pick a persona to give the AI specialized context
  4. Press Enter to send — the AI streams its response in real-time

Conversations are saved as sessions in the left sidebar, grouped by date with message counts.

Solo — Chat Dashboard

The main chat interface with model selector, persona picker, and session sidebar

AI Chat

The chat interface is the heart of Solo:

  • Multi-model support — Switch between providers and models mid-conversation using the dropdown
  • Real-time streaming — Responses appear token-by-token; click the stop button to abort
  • Session management — Create, rename, archive, delete, and search sessions from the sidebar
  • Persona selection — Attach a persona to give the AI access to your database, API, or custom instructions
  • Markdown rendering — Code blocks with syntax highlighting, tables, lists, and rich formatting
  • Thinking mode — Models that support reasoning (Qwen 3, DeepSeek R1) show a collapsible thinking section
  • Dark/Light mode — Toggle from the top bar; applies across all pages
Shortcut: Press Ctrl+N (or Cmd+N) to start a new chat instantly. Use Shift+Enter for a new line without sending.

Model Hub

Configure cloud AI providers and download local models from Settings → AI Providers.

Solo — Model Hub

Browse and configure AI models — cloud providers and local GGUF models

Cloud Providers

Enter your API key for any supported provider. Click Test Connection to verify it works. Toggle the active provider to set your default model.

Local Models

Download GGUF models with one click. Solo auto-detects your RAM and recommends compatible models. Progress is shown in real-time during download.

Solo — Installed Models

Manage installed local models — sync status, storage usage, ready to chat

AI Mode Toggle

Switch between Local and Cloud mode. In local mode, all inference runs on your CPU — no internet required after model download.

Personas

Personas are AI identities that connect to your real systems. Solo includes 12 persona types:

Persona TypeWhat It Does
Nexus AgentGeneral-purpose AI with custom system prompts
Web ResearcherSearch the web and scrape pages
PostgreSQLQuery PostgreSQL in natural language — AI writes the SQL
MySQLQuery MySQL databases with auto-generated SQL
MSSQLConnect to Microsoft SQL Server
SnowflakeQuery your Snowflake data warehouse
MongoDBQuery MongoDB document databases
GitHubBrowse repos, issues, and pull requests
GitLabAccess repos and CI/CD pipelines
API ConnectorCall any REST API endpoint
File OperationsRead, write, and parse local files
SlackSend and read messages in Slack channels
Solo — Personas Library

Browse and manage your AI personas

Creating a Persona

Solo uses a 2-step wizard:

  1. Choose Type — Select from the persona type dropdown (each type shows an icon and description)
  2. Configure Details — Enter name, category, description, connection credentials, and system prompt

For database types, use the Test Connection button to verify credentials before saving.

Solo — Create Persona Wizard

Wizard-based persona creation with type selection and credential configuration

Privacy: All credentials are stored locally in SQLite. Nothing leaves your machine.

Agent Library

Solo ships with 93 pre-built business agents across 12 departments. Each has a specialized system prompt, recommended model, and tool configurations.

Solo — Agent Library

Browse 93 agents across 12 departments — search, filter by role, view details

DepartmentExample Agents
C-SuiteCEO Strategic Advisor, CFO Financial Strategist, COO Operations Optimizer, CTO Technology Advisor
Marketing & SalesContent Strategist, SEO Specialist, Social Media Manager, Brand Voice Designer, Email Campaign Builder
Finance & OperationsFinancial Analyst, Budget Planner, Invoice Processor, Tax Advisor, Revenue Forecaster
Legal & ComplianceContract Reviewer, Compliance Checker, IP Advisor, Privacy Policy Drafter
HR & PeopleRecruiter Assistant, Job Description Writer, Performance Review Helper
Design & UXUI/UX Advisor, Brand Identity Designer, Presentation Builder, Color Palette Generator
Data & AnalyticsData Analyst, SQL Query Builder, Dashboard Designer, Statistical Modeler
IT & SecurityDevOps Assistant, Security Auditor, Infrastructure Planner, Incident Response Helper
Product ManagementProduct Manager, Feature Prioritizer, User Story Writer, Roadmap Planner
Startup & VenturePitch Deck Builder, Business Model Canvas Creator, Go-to-Market Strategist
OperationsProcess Optimizer, Supply Chain Analyst, Quality Assurance Planner
SpecializedIndustry-specific agents and custom roles

Using Agents

Browse by department or filter by role: Researcher, Writer, Analyst, Designer, Developer, Reviewer, Planner, or Custom. Click any agent card to view its full system prompt, tools, and model override.

Creating a Custom Agent

Click + Create Agent and configure: name, role, description, system prompt, tool access, model preference, category, and public/private visibility.

Solo — Create Custom Agent

Create a custom agent with role, system prompt, and tool configuration

Crews

Crews are multi-agent teams that collaborate on complex tasks. The crew system uses a 5-step creation wizard.

Solo — Crew Dashboard

Crew dashboard with stats (Total, Running, Completed, Failed) and crew list

5-Step Crew Wizard

  1. Crew Details — Name, description, optional blueprint template, and AI model selection
  2. Execution Mode — Choose how agents collaborate:
    ModeHow It WorksBest For
    SequentialAgents run one after another; output feeds into the nextResearch → Write → Edit
    ParallelAll agents run simultaneously on the same inputSEO + Social + Email
    PipelineStaged processing with checkpointsDraft → Review → Publish
    AutonomousA coordinator agent decides the flow dynamicallyOpen-ended tasks
  3. Agent Team — Add agents from the library or create custom ones inline. Reorder for sequential execution.
  4. Connections — Bind social channels (Telegram, Discord, LinkedIn) with an approval toggle
  5. Review — Confirm everything and click Create Crew

Running a Crew

Click Run on any crew. Enter your input text and watch real-time progress: status bar, duration, token count, cost, step timeline, and per-agent metrics. You can cancel a run mid-execution.

The Runs tab shows execution history with status, duration, and cost for every past run.

Blueprints

Solo ships with 15 pre-built templates to jumpstart your workflows.

Solo — Blueprints

Browse blueprint templates — search, filter by category, preview content

10 Workshop Blueprints

Multi-agent brainstorming sessions across 5 categories (Strategy, Content, Marketing, Product, Research):

  • C-Suite Strategy Session
  • Startup Pitch Review
  • Marketing Campaign Planning
  • Product Roadmap Workshop
  • Financial Analysis Session
  • Talent & Culture Review
  • Legal & Compliance Audit
  • Data Strategy Workshop
  • Technology Partner Evaluation
  • Custom Workshop

5 Team Templates

Pre-configured agent teams for engineering workflows:

  • Migration Squad — Code migration across stacks
  • MVP Builder — Rapid prototyping
  • Code Reviewer — Quality and standards
  • Full-Stack App — End-to-end application
  • API Designer — REST schemas and design

Creating Custom Blueprints

Click + Create Blueprint and define: name, description, category, tags, and markdown content. Your custom blueprints appear alongside the built-in ones.

Solo — Create Blueprint

Create a custom blueprint with markdown content and tags

Workspace

The Workspace is for project-based multi-agent collaboration with structured outputs and exportable artifacts.

Solo — Workspace

Project list with status badges — create, run, and review multi-agent projects

3-Step Project Wizard

  1. Project Details — Name, description, AI model, and optional blueprint template
  2. Agent Selection — Browse the 93-agent library, search, and check agents to add (count badge updates live)
  3. Review & Create — Confirm selections and launch

Viewing Results

Completed projects have two tabs:

  • Discussion Tab — Step-by-step agent contributions in a threaded conversation view
  • Compiled Output Tab — Final consolidated result with a copy button for easy export

Project statuses: Draft (not yet started), Running (agents working), Completed (results ready), Failed (error occurred).

Connections

Connect Solo to external messaging platforms for automated content publishing and AI-powered responses.

Solo — Connections

Platform connection cards with status badges and direction toggles

PlatformAuth MethodDirectionStatus
Telegram BotBot tokenInbound + OutboundLive
Discord BotBot token + Public Key + App IDInbound + OutboundLive
LinkedInOAuth 2.0 (Client ID/Secret)Outbound onlyLive
Twitter/XAPI Key + Secret + Access TokensOutboundComing Soon
InstagramOAuth (Business/Creator account)OutboundComing Soon
FacebookOAuth (Page required)OutboundComing Soon

Setting Up Telegram

  1. Open Telegram, search for @BotFather
  2. Send /newbot and follow the prompts to get your bot token
  3. In Solo, go to Connections → Telegram → paste the token
  4. Toggle Inbound/Outbound as needed and click Connect
  5. For inbound messages, set up a webhook with ngrok: ngrok http 18741

Setting Up Discord

  1. Go to the Discord Developer Portal
  2. Create a New Application → go to Bot → copy the token
  3. Enable Privileged Gateway Intents (Message Content, Server Members)
  4. In Solo, paste the Bot Token, Public Key, and Application ID

Setting Up LinkedIn

  1. Create an app at LinkedIn Developers
  2. Request "Share on LinkedIn" access for your app
  3. In Solo, enter Client ID and Client Secret
  4. Click Connect — an OAuth browser window will open for authorization

Connected platforms can be bound to crews in Step 4 of the crew wizard, enabling automated content distribution with human-in-the-loop approval.

Best models for connections: Qwen 3 8B or 14B gives the best results for auto-reply and content generation.

Approval Queue

When crews or agents auto-reply to incoming messages, every response lands in your approval queue first. Nothing gets sent without your explicit approval.

Solo — Approval Queue

Review AI-drafted replies before they are sent — approve, edit, or reject

  1. Set up triggers — Link a channel (Telegram, Discord) to a crew or agent with "require approval" enabled
  2. AI drafts a reply — Incoming messages are processed by your AI; the draft appears in the queue
  3. Review & send — Edit the response if needed, then approve to send or reject to discard

Triggers support cooldown periods to prevent spam, and you can link a crew instead of a single agent for more sophisticated multi-step responses.

BYOK — Bring Your Own Key

Solo supports 5 AI providers. Configure in Settings → AI Providers:

ProviderModelsCost
Anthropic ClaudeSonnet 4, Opus, HaikuBYOK
OpenAIGPT-4o, GPT-4o Mini, GPT-4 Turbo, o1-previewBYOK
Google Gemini2.0 Flash, 2.0 Pro, 1.5 FlashBYOK
AWS BedrockClaude, TitanBYOK
Ollama (Local)39 GGUF modelsFree

API keys are stored in your browser's localStorage. They are never sent to any server other than the respective AI provider.

Local Models

Solo includes a Model Hub with 39 GGUF models across 9 families. One-click download — no API key, no internet after download, no data leaving your machine.

RAM Requirements

Your RAMModels You Can Run
4 GBQwen 3.5 0.8B, Gemma 3 1B, Llama 3.2 1B
8 GBQwen 3 8B, DeepSeek R1 7B, Mistral 7B, Qwen 2.5 Coder 7B
16 GBQwen 3 14B, Gemma 4 12B, Phi-4 14B, DeepSeek R1 14B
32 GBQwen 3 32B, Gemma 4 27B, DeepSeek R1 32B, Gemma 3 27B
48+ GBLlama 3.1 70B, DeepSeek R1 70B

9 Model Families

  • Gemma 4 New · Recommended — Google's latest (April 2025). 12B and 27B variants. Best-in-class instruction following, reasoning, and multilingual support at its size. Our top pick for 16GB+ machines.
  • Qwen 3.5 / 3 / 2.5 — Alibaba's versatile family. General chat, reasoning, and coding (Qwen 2.5 Coder is excellent for IDE integration). Wide range from 0.8B to 32B.
  • DeepSeek R1 — Advanced reasoning with visible thinking mode. 7B to 70B. Great for analysis, math, and complex problem-solving.
  • Gemma 3 — Google's previous generation. Still solid for lightweight tasks (1B to 27B).
  • Llama 3 — Meta's open models. 1B to 70B. Strong general-purpose performance.
  • Mistral — Fast and lightweight (7B). Good for quick responses on lower-spec machines.
  • Phi-4 — Microsoft's 14B model. Punches above its weight on reasoning tasks.

Inference powered by llama-cpp-python — CPU-only, no GPU required. Models stored in ~/.contextuai-solo/models/.

Built-in Coding Server

Solo exposes an OpenAI-compatible API at localhost:18741/v1/chat/completions. Point your IDE at it for free, offline code completion.

IDE Setup

IDEHow to Connect
VS CodeContinue or Copilot extension → set base URL to http://localhost:18741/v1
CursorSettings → Models → Add custom model endpoint
WindsurfOpenAI-compatible provider → set base URL
Aideraider --openai-api-base http://localhost:18741/v1
Any toolAny tool that speaks the OpenAI API format — just change the base URL
Best coding models: Download Qwen 2.5 Coder (7B for 8GB RAM, 14B for 16GB, 32B for 32GB) or DeepSeek R1 for reasoning-heavy tasks.

Settings

Access settings from the sidebar. 5 tabs:

Solo — Settings

Settings page with AI Providers, Brand Voice, Appearance, Data & Export, and About tabs

  • AI Providers — Configure API keys, download local models, test connections, set active provider
  • Brand Voice — Business name, industry, brand prompt, target audience, content topics
  • Appearance — Theme (Light, Dark, System) and font size (Small, Medium, Large)
  • Data & Export — Export/import database as JSON, clear all data
  • About — App version, check for updates, technology stack links

Brand Voice

Configure your brand identity so AI-generated content matches your tone. Settings → Brand Voice:

  • Business Name — Used in agent prompts and content generation
  • Industry — 12 options (Tech, Marketing, Finance, Healthcare, etc.) to inform domain knowledge
  • Brand Voice Prompt — Free-form instructions: tone, style, personality, things to avoid
  • Target Audience — Who your content is aimed at
  • Content Topics — Key subjects the AI should focus on

A dynamic preview shows how your brand voice configuration looks before saving.

Data & Export

All your data lives locally. From Settings → Data & Export:

  • Export Data — Download a full JSON backup of all chats, personas, agents, crews, and settings
  • Import Data — Restore from a previous backup file
  • Clear All Data — Permanently delete everything (use with caution)
Important: Solo has no cloud sync. Back up regularly via export if your data matters to you. Exports go to your Downloads folder.

Keyboard Shortcuts

ShortcutAction
Ctrl+N / Cmd+NNew chat session
EnterSend message
Shift+EnterNew line in message
EscapeStop AI generation

Data Storage

Everything is stored locally on your machine:

WhatWhere
Database (chats, personas, agents, crews)~/.contextuai-solo/data/contextuai.db
Local AI models (GGUF files)~/.contextuai-solo/models/
API keys & provider configBrowser localStorage
Data exportsDownloads folder

No telemetry, no cloud calls (except to your chosen AI provider), no data collection.

Tech Stack

LayerTechnology
Desktop ShellTauri v2 (Rust) — lightweight, secure, cross-platform
FrontendReact 19 + Vite + TypeScript 5.9
StylingTailwind CSS + Framer Motion
BackendFastAPI (Python 3.11+)
DatabaseSQLite via async adapter
Local Inferencellama-cpp-python (CPU-only)
AI ProvidersAnthropic, OpenAI, Google, AWS Bedrock, Ollama
Agent FrameworkStrands Agents SDK
IconsLucide React

Troubleshooting

Backend won't start

  • Verify Python 3.11+: python --version
  • Install dependencies: cd backend && pip install -r requirements.txt
  • Check port 18741 is free: netstat -ano | findstr 18741 (Windows) or lsof -i :18741 (Mac/Linux)

Frontend won't start

  • Verify Node.js 18+: node --version
  • Clean install: rm -rf node_modules && npm install
  • Windows rollup error: npm install @rollup/rollup-win32-x64-msvc

Chat returns errors

  • Verify API key in Settings → AI Providers → Test Connection
  • If using Ollama, make sure it's running: ollama serve
  • Check the backend terminal for error details

Local model download fails

  • Check internet connection (download only — inference is fully offline)
  • Ensure enough disk space in ~/.contextuai-solo/models/
  • Try a smaller model first (Gemma 3 1B is ~700MB)

Telegram/Discord webhook not receiving messages

  • Ensure ngrok is running: ngrok http 18741
  • Set the webhook URL in the platform's settings to your ngrok HTTPS URL
  • Check Discord privileged intents are enabled (Message Content, Server Members)

Database reset

  • Export your data first from Settings → Data & Export
  • Delete ~/.contextuai-solo/data/contextuai.db
  • Restart the backend — it recreates the database with defaults