/experiments
Most experiment tracking stops at artifacts and metrics. But the most valuable output of an experiment is the team’s interpretation: what worked, what didn’t, and what to do next. That context typically lives in someone’s head. When they leave, the next person starts from scratch.
Outcomes captures what your team tried, what happened, and what the team decided and why. Every decision logged is institutional knowledge that persists regardless of team changes.
Experiment Timeline
The main view shows all experiments for the currently selected project, sorted by date or impact.Experiment Cards
Each experiment appears as a card showing:- Name and creation date
- Source type icon — research resource, hypothesis, incident, or recommendation
- Status badge — Configure, In Progress, Complete, or Validated
- Target metric and observed delta (color-coded: green positive, red negative, gray pending)
- Tags for grouping and pattern detection
Metric Trend Chart
A collapsible line chart at the top tracks your target metric across completed experiments, showing improvement (or lack thereof) over time. When collapsed, key trend values remain visible in the summary bar.Filtering and Search
| Control | What it does |
|---|---|
| Search bar | Instant filtering by name, hypothesis, initiative, or tags |
| Status chips | All / Configure / In Progress / Complete / Validated (with counts) |
| Source filters | Toggle by source type (Papers / Hypotheses) |
| Sort | By date (newest first) or by impact (largest delta first) |
Creating an Experiment
Click + New Experiment in the top-right corner. The create form adapts based on source type:- From a Resource
- Custom Hypothesis
- From an Incident
Select the Search papers source mode. Search for a paper, repo, or model by title. Remyx autocompletes from its resource index.
| Field | Description |
|---|---|
| Name | Short descriptive name |
| Resource | Search and select from the index |
| Hypothesis | What you expect to happen |
| Target metric | Dropdown of metrics configured for this project |
| Project | Which initiative this belongs to |
| Tags | Comma-separated labels for grouping |
| Target repository | GitHub repo for the implementation (optional) |
| Tracker link | Link to Linear, Jira, or GitHub issue (optional) |
Experiment Detail
URL:/experiments/dashboard/<experiment_id>
A two-column layout showing the full lifecycle of a single experiment.
Origin Section
For research-sourced experiments, the Origin section shows the launch context — built automatically on first load (~2-4 seconds):| Field | Description | Editable? |
|---|---|---|
| Resource title | Link to the resource viewer | — |
| Abstract excerpt | One-sentence summary | Click to edit |
| Key methods | Technique badges extracted from the resource | Add/remove inline |
| Target repository | Repo where the implementation lands | Change triggers context rebuild |
| Implementation plan | AI-generated plan referencing actual file paths | Collapsible, editable, regeneratable |
| Docker image | Pre-built environment reference | Read-only |
Analysis Card
Combined Hypothesis and Decision in a single card:- Hypothesis — the team’s prediction, always visible at top
- Decision — logged after results are in; includes text, author, and timestamp; click to edit
Implement Section
A compact bar for Claude Code integration:- Copy-paste command to run Claude Code with the Remyx MCP connection
- Link to Connectors for setup
Activity Feed
Unified chronological feed combining:- Comments with @mention support, edit/delete
- System events from the knowledge graph (experiment created, status changed, decision logged, PR opened)
Sidebar
| Section | What it shows |
|---|---|
| Status | Dropdown: Configure → In Progress → Complete → Validated |
| Metric | Target metric, observed delta, confidence level |
| Resources | Linked artifacts: PR, ticket, repo, dataset, tracking run, custom links |
| Related Experiments | Bidirectional linking with cross-project search |
| Project | Initiative context from project settings |
| Tags | Editable tag list |
Logging a Decision
The most important step in the ExperimentOps workflow. After reviewing results:- Scroll to the Decision section in the Analysis card
- Write what the team decided and why
- The decision is timestamped and attributed to the author
“Ship to 100%. The re-ranker specifically helps with multi-topic tickets where the old retriever returned tangentially related articles. Three retrieval experiments now, all positive. This is our best direction.”
Related
Insights
See cross-experiment patterns and recommended next steps
Overview
Leadership portfolio across all initiatives
Connectors
Link GitHub, Linear, Jira for bidirectional sync
Projects & Settings
Configure metrics, repos, and integrations per project