Product / Desktop · Early access beta

Valence

Cross-model AI, one interface.

Move between OpenAI, Anthropic, Google, Azure, xAI, OpenRouter, and Ollama without surrendering your keys, your context, or your control — with local memory recall, kids-safe profiles, and parent oversight in the same desktop surface.

See the working surface before you commit.

Actual captures from the current desktop beta. The point is evidence: how Valence handles conversations, local memory, parental controls, settings, and oversight inside one calm, inspectable surface.

Live conversation workspace

Actual light-theme conversation view from the current beta, showing the centered reading rail, calmer composer, and clearer message hierarchy.

These are real captures from the current desktop beta. Select a view, then open it full size for a closer read.

Designed for switching, continuity, and local control.

Valence keeps the working surface calm while giving power users the pieces that actually change outcomes: provider choice, tool access, persistent context, and inspectable system boundaries.

Switching

Cross-model continuity

Start with Claude, get a second opinion from GPT-4o, finish with Gemini, and keep the same working thread. @mentions reroute the conversation instead of forcing restarts.

  • Context preserved across provider switches
  • Fast handoffs with @claude, @gpt, and @gemini

Tooling

MCP integration

Connect stdio-based Model Context Protocol servers for filesystems, databases, APIs, and other task-specific tools when the model needs real leverage.

  • Tool access lives beside the chat surface
  • Managed from the desktop settings flow

Parallel work

Multi-tab conversations

Run separate threads in parallel, keep provider selection per tab, and cross-reference the important parts when one workflow branches into another.

  • Independent context per tab
  • Useful for research, coding, and comparison work

Profiles

User personas that travel with the task

Create workflow-specific system prompts and preferences for code review, writing, research, or any custom role without rebuilding the workspace every time.

  • Swap posture without opening a new app
  • Good for teams and repeatable solo workflows

Persistence

Local session history

Conversations save locally as JSON so you can reopen exact context later, maintain your own archive, and keep the raw history on your own machine.

  • Resume long-running threads without reconstruction
  • Export and back up the archive on your terms

Posture

No middleman

Valence does not proxy your traffic. API keys stay local and requests go directly to the providers you choose, including fully local flows through Ollama.

  • Your keys, your provider contracts, your boundaries
  • Ollama support for zero-cloud runs when needed

Surface

Modern WebView UI

The desktop surface uses WebView2 for readable markdown, syntax highlighting, and a dense but calm composition suited to long sessions. Legacy ImGui mode remains available for overlay-style usage.

  • Readable for code, prose, and tool output
  • Responsive layout tuned for desktop work

Memory

Local memory recall

Use #remember and #recall to keep important knowledge available across sessions without leaving the device. Memory becomes part of the working surface instead of a hidden side system.

  • On-device recall for long-lived context
  • Better continuity across projects and sessions

Evaluation

Compare models side by side

Send one prompt to 2–3 providers in parallel when you need a second opinion, a cost check, or a faster read on which model is right for the job.

  • Streaming comparison without copy-paste
  • Useful for quality, speed, and cost tradeoffs

Maintainability

Extensible architecture

SRP LEGO-style components, factory-driven provider adapters, and event-based coordination keep the system understandable as the product line expands.

  • Designed to add providers without rewriting the surface
  • Clear seams between UI, providers, storage, and tools

Seven major providers, no translation layer.

Use the best model for each job without leaving the product surface. Cloud, enterprise, and local options all live behind the same interaction model.

OpenAI

GPT-4o, GPT-4, GPT-3.5

Fast generalist coverage for writing, coding, and multimodal tasks.

Cloud

Anthropic

Claude 4, Claude 3.5 Sonnet, Haiku

Strong analysis, careful reasoning, and review-heavy workflows.

Cloud

Google

Gemini 2.0, Gemini Pro

Fast synthesis and broad model coverage for research and summarization.

Cloud

Azure OpenAI

Enterprise GPT deployments

Fits controlled enterprise environments where deployment boundaries matter.

Enterprise

xAI

Grok-3

Another model family in the mix when you want a different read on the same problem.

Cloud

OpenRouter

Multi-model gateway

Useful when you want access to a broader model menu without leaving the Valence surface.

Gateway

Ollama

Llama 3, Mistral, Phi-3, and other local models

The local-first path for private or air-gapped workflows that cannot depend on external calls.

Local

7
Provider families supported
Cross-model switches per thread
0
Helix middleman servers
100%
Local data storage posture

Layered for growth, boring where it matters.

The architecture favors explicit interfaces, service boundaries, and local persistence over hidden automation. That keeps the product easier to extend and easier to trust.

WebView layer

Readable surface

Modern HTML, CSS, and JavaScript deliver markdown, code rendering, and long-session usability through WebView2.

Orchestration

Coordinated flow

ChatWindow and the configuration flow coordinate lifecycle, navigation, and user-facing state without burying the logic.

Components

Single-responsibility pieces

Header, message area, and input surfaces stay intentionally scoped so the interface can evolve without cascading rewrites.

Provider system

ILLMProvider abstractions

Provider adapters sit behind a shared interface and factory pattern so new model families can be added with clear seams.

Abilities & MCP

Commands plus real tools

Built-in abilities, trigger systems, and MCP support give the model a practical action layer when the workflow calls for it.

Profiles

Task-specific posture

ProfileService manages custom personas, system prompts, and preferences so different workflows can keep their own operating discipline.

Persistence

Conversation durability

ChatStorageService keeps full message metadata in JSON so sessions can be reopened, exported, and audited later.

Settings

Explicit configuration

SettingsFactory, validation, API key management, and provider configuration stay inspectable instead of disappearing into the background.

Theme

Consistent visual language

ThemeColors.ModernHive keeps the UI coherent across components so density and readability hold up as the surface expands.

Built for people who compare, inspect, and keep their work.

Valence is most useful when one model is not enough, context matters, and the operator wants direct control over where the work goes next.

Multi-perspective analysis

Ask several models to work the same prompt and use the differences to sharpen the answer instead of picking one provider forever.

  • Side-by-side comparison for 2–3 models
  • Context preserved across provider switches
  • Useful for tradeoffs, risk, and quality checks

Development workflow

Use different model strengths for review, debugging, and documentation while keeping the same project thread and tool attachments visible.

  • Syntax-highlighted code blocks
  • MCP tools for filesystems and databases
  • Ollama for sensitive local codebases

Research and writing

Brainstorm with one provider, outline with another, refine with a third, then keep the history as part of the research trail.

  • Profiles for different writing postures
  • Full conversation export
  • Cross-model fact checking without tab chaos

Learning and exploration

When one explanation does not land, switch teaching styles instantly and keep the thread moving instead of starting over somewhere else.

  • Different explanation styles per model
  • Save and resume learning sessions
  • Good for deep, iterative understanding

Cost optimization

Use expensive models for hard work, cheaper models for routine tasks, and local models when experimentation or privacy matters more than raw frontier performance.

  • Switch by cost tier without leaving the app
  • No Valence subscription on top of provider costs
  • Free local inference through Ollama

Privacy-conscious AI

Keep keys local, keep conversations local, and decide explicitly when cloud providers enter the loop.

  • No proxy architecture
  • Fully local option with Ollama
  • Your data, your control, your audit trail

Start with the trial. Keep the system if it fits.

Valence is built for people who want model choice, local control, and a working surface they can actually inspect. Purchase the early-access license, download the current build, or go deeper through the docs and roadmap.