Your assistant's data lives wherever your assistant is hosted. For self-hosted installations, that means everything stays on your machine. If you're running in our managed cloud (coming soon) or in your own cloud infrastructure, your data stays within that environment.
Regardless of hosting option, your assistant thinks through an AI model in the cloud. Here's exactly what that means.
All of the following data stays within your assistant's environment and never leaves it — except where noted. For self-hosted installations, this means it stays on your machine.
~/.vellum/protected/~/.vellum/protected/trust.json~/.vellum/workspace/skills/~/.vellum/workspace/config.json~/.vellum/workspace/~/.vellum/workspace/data/~/.vellum/workspace/data/db/Your messages and context go to the AI model provider.
Every time you send a message, your assistant assembles a context bundle:
This bundle is sent to the AI model provider (Anthropic) to generate a response. That's how your assistant thinks — the AI model runs in the cloud.
What this means practically: If you tell your assistant “I'm working on a secret project called Nightfall,” that information may be saved as a local memory and included in future AI model calls when it's relevant. The AI provider processes it to generate a response but does not use it to train models (per their API terms). Still, it does leave your assistant's environment temporarily. We want you to know that.
API calls to connected services.
When your assistant checks your calendar, sends an email, or browses a web page, it makes API calls to those services. The data in those calls is whatever's needed for the action. These are standard API calls — the same ones any app would make.
Optional telemetry.
If you opt in (Settings > Privacy), Vellum collects:
Both are off by default. You can toggle them independently. If you leave them off, nothing is sent.
The credential system is designed with the assumption that the AI is compromised and acting maliciously. We literally don't want the assistant to be able to access your credentials — it can say what it wants to do with them, and you can approve that, but it never gets the actual values.
Credentials are stored in the macOS Keychain (or an AES-256-GCM encrypted file on Linux), inside a protected directory with restricted file permissions. When you need to enter a credential, a dedicated popup UI appears — you submit it there, not in the chat. This ensures the credential is captured and stored directly without ever passing through the assistant or the AI model.
All credential-bearing operations run through a Credential Execution Service (CES) — a separate process that handles authentication in isolation. The assistant says “make this API call with my Gmail credentials,” and the CES executes it deterministically, with no AI involved, without exposing the credential to the assistant's context or the AI model.
This is a hard process boundary, not just a software abstraction. The CES has its own grant system, audit log, and cleanup rules for temporary credential materializations.
If you accidentally type a credential into the chat instead of using the secure popup, Vellum detects it and blocks it from reaching the AI model. This is the second layer of protection — the popup ensures credentials are captured securely, and the scanner catches anything that slips through.
Beyond chat input, secrets can end up in unexpected places — returned in a tool result or embedded in a file. Vellum runs a secret scanner that:
When a secret is detected, the system can redact it, warn you, block the operation, or prompt for confirmation — depending on your configuration.
Not everyone who messages your assistant gets the same access.
Trusted contacts and unknown actors have heavily restricted access compared to you. Sensitive actions from trusted contacts require your explicit approval through guardian-in-the-loop notifications.
You can start a private conversation that gets its own isolated memory scope. Memories from a private conversation can't surface in other conversations. The private conversation can still access your shared memory pool, but nothing flows back out.
Use this for sensitive topics where you want the assistant to have context during the conversation but don't want those details persisting into your general memory.
When your assistant controls your screen, additional safety measures kick in:
You see an overlay showing what the assistant is doing at each step, and you can stop it at any time.
Every tool your assistant uses has a risk level:
When you approve an action, you choose the scope: one-time, 10 minutes, this conversation, always, or never. These rules accumulate in your trust configuration, so the assistant learns your boundaries over time without you having to re-approve the same actions.
Your messages are processed by Anthropic's Claude. Anthropic's API terms state that data sent through the API is not used for model training. Prompts and responses are processed to generate outputs and are subject to Anthropic's data retention policies.
Read Anthropic's Privacy Policy for full details. We chose Anthropic for their approach to safety and data handling, but we also believe you should verify their policies yourself.
If you have information you don't want sent to an AI model:
~/.vellum/protected/trust.json to see what you've auto-approved.Currently, Vellum runs as a self-hosted installation on your machine. Your assistant is a process on your computer, and all data lives locally.
Additional hosting options are coming soon:
Regardless of where your assistant is hosted, the same permission system, trust rules, credential isolation, and channel security apply. Your assistant's data stays within its hosting environment and is never aggregated centrally.