Platform
CDE runs on the data you provide for the discovery problem at hand. Choose cloud-hosted, self-hosted, or air-gapped deployment based on data locality, compute policy, and how close the engine needs to sit to the workflow.

Your Data
Unlike systems that require millions of samples, CDE discovers governing equations and causal structures from the datasets you already collect.
Your data stays yours. CDE trains per run on the data you provide. No pre-existing model or external dataset is required before discovery begins.
Why teams choose this deployment path
Upload episodes via POST /v1/data/upload with structured JSON payloads. Profile data, list datasets, and manage uploads programmatically.
client.upload_episodes(), client.profile_data(), client.generate_data() — typed methods with full IDE support.
Upload, profile, and generate data through Model Context Protocol tools. Agents can provision data without writing HTTP calls.
POST observations to /v1/streams/{id}/ingest as they arrive. Server-sent events for live status. Automatic trigger for discovery runs.
Data Model
All data enters CDE as episodes — structured observations over time. Only timestamps and observations are required. Spatial, relational, hierarchical, and causal modalities are automatically detected when present.
Core
Spatial
Relational
Hierarchical
Causal
Quality
Supported input formats: JSON episodes (native), CSV, Parquet, and HDF5 (via SDK conversion). Molecular dynamics adapter payloads are automatically normalized.
Compute
CDE estimates compute requirements from your data profile and discovery mode. Resource planning is automatic.
Vareon provisions and manages GPU compute. Automatic hardware selection based on data profile and discovery mode. Pay for gpu_seconds consumed. No infrastructure to manage.
Deploy CDE on your own infrastructure. License fee only — bring your own GPUs. Full control over data residency and network topology.
Fully isolated deployment with offline license activation. No outbound network required. Suitable for classified environments, defense, and regulated industries.
Hardware selection
CPU or GPU, automatic
Runtime budget
Estimated per-run, capped
Concurrency
Multiple runs in parallel
Billing
Compute consumed
Bring Your Own Agent
CDE's API surface is programmatic end to end. Bring your own research agent or orchestration layer and use CDE as the discovery engine behind it.
An external agent can formulate hypotheses, design experiments, launch runs, and review results through CDE's structured task system without replacing the discovery engine itself.
Typical agent capabilities
Integration endpoint
POST /v1/agents/sessions/byo/scientist
{
"project_id": "...",
"objective": "Discover governing equations...",
"dataset_id": "...",
"provider_id": "..."
}
Your agent receives structured tasks, claims its work items, executes discovery runs through the full API, and submits results. CDE handles governance, provenance, and negative controls automatically.
Model Provider
CDE's agent cognition layer supports any OpenAI-compatible model endpoint. Use your existing provider for planning, hypothesis generation, and result interpretation while CDE handles the actual discovery workflow.
Configure the provider once with a base URL, default model, and API key. CDE routes cognition calls to your endpoint using the standard chat completions protocol.
Configuration
How it works
Contact us to scope the right deployment path for your data, compute, and operational constraints.