← Back to Product Feed

GitHub Open Source facebookresearch/HyperAgents

Self-referential self-improving agents that can optimize for any computable task

2,026
Traction Score
253
Forks
Mar 19, 2026
Launch Date
View Origin Link

Product Positioning & Context

AI Executive Synthesis
Reproducibility and transparency of research results.
A user reports critical issues accessing `HyperAgents` experiment logs, citing missing Git LFS objects (404 errors) and a typo in the README's `unzip` command. This directly impedes the reproducibility and transparency of the project's research, a fundamental requirement for scientific and open-source credibility. The inability to access crucial experimental data frustrates users attempting to validate or extend the work. This highlights a significant operational oversight in managing large files with Git LFS and maintaining accurate documentation. For a research-oriented project, ensuring data accessibility and correct instructions is paramount for community engagement and academic impact.
Self-referential self-improving agents that can optimize for any computable task

Active Developer Issues (GitHub)

open Safety policy for constraining meta-agent modifications
Logged: Mar 28, 2026
open Experiment logs archive in README cannot be downloaded/extracted (LFS objects missing, unzip command typo?)
Logged: Mar 26, 2026

Community Voice & Feedback

0xbrainkid • Mar 31, 2026
Perfect — the DecisionLog events already having `tool_name`, `decision`, `tier`, and `timestamp` means the drift detector doesn't need any custom instrumentation. Those four fields are sufficient for the core fingerprint:

```python
# Behavioral fingerprint from receipt stream
fingerprint = {
"tool_distribution": entropy(tool_name_counts), # shifts in which tools are called
"allow_rate": allow_count / total_count, # changes in policy pass rate
"tier_distribution": tier_histogram, # drift in trust tier assignments
"call_velocity": total_count / window_duration_s # acceleration or deceleration
}
```

I'll start with the stderr tail path in shadow mode — fast iteration without waiting for the formal hook. The prototype flow:

1. Tail DecisionLog from stderr
2. Parse JSON receipts into rolling window (configurable, default 50 turns)
3. Compute fingerprint delta vs baseline (captured at iteration 0)
4. Emit drift score (0.0-1.0) per window
5. ...
tomjwxf • Mar 31, 2026
@0xbrainkid — the integration diagram is clean. Receipt stream → drift detector → approval gate is exactly the right architecture.

Two concrete next steps:

Receipt stream hook: The gateway already emits a DecisionLog event on every policy evaluation ([source](https://github.com/scopeblind/scopeblind-gateway/blob/main/src/gateway.ts)). Exposing that as a subscribable stream (EventEmitter or async iterator) is straightforward. I'll spec a onReceipt(callback) interface that gives you the signed receipt + metadata per evaluation.

Fingerprint inputs already in the receipt: Each receipt includes tool_name, decision (allow/deny), tier, timestamp, and request_id. For drift detection you'd track distribution shifts across those fields per iteration window. The signature chain ensures the meta-agent can't retroactively alter the history.

Your progressive enforcement → trust level mapping is spot on. That's how we think about it internally — shadow mode is the default precisely because new in...
0xbrainkid • Mar 31, 2026
The receipt chain approach is cleaner than hooks inside the meta-agent — agreed. External drift detection from signed receipts is both tamper-resistant and decoupled from the optimization loop. The meta-agent can't game a detector it doesn't control.

A post-evaluation hook that exposes the receipt stream would be very useful. The concrete integration:

```
protect-mcp receipt stream → drift detector → approval gate
↘ SATP attestation (if cross-org)
```

The drift detector consumes receipts, computes behavioral fingerprint deltas per iteration, and triggers the approval gate when cumulative drift exceeds threshold. For cross-org scenarios (meta-agent modifying task agents that interact with external systems), the same drift signal can feed into a behavioral attestation — so external systems know whether the optimization loop is producing stable or drifting agents.

The progressive enforcement model (shadow → simulate → enforce → sign) maps we...
tomjwxf • Mar 31, 2026
Good observation on cumulative drift. Static per-action policies catch individual violations but miss trajectory-level shifts — the "boiling frog" problem is real for optimization loops.

A couple of thoughts on how this could layer in:

Receipt chains already give you the raw material. Every iteration produces signed receipts with tool call distributions, write targets, and decision outcomes. A drift detector could consume that chain and compute the fingerprint you're describing without needing hooks inside the meta-agent itself — it stays external and tamper-resistant.

Threshold-based halts map cleanly to the approval gate. When drift exceeds the threshold, rather than raising an exception inside the agent, the policy could escalate to the existing human approval gate. Same mechanism, different trigger.

The W3C reference is interesting — separating behavioral consistency (Layer 3) from authorization (Layer 1-2) aligns with how we think about progressive enforcement (shadow → simula...
0xbrainkid • Mar 31, 2026
The safety policy pack addresses the right constraints — scoping writes to `workspace/`, approval gates for evaluation functions, and preventing self-rewriting of the meta-agent's own code.

One gap this doesn't cover: **behavioral drift detection during the optimization loop itself**. A meta-agent that stays within the write constraints but gradually shifts its optimization objective is harder to catch with static policy rules alone.

Consider: the meta-agent is allowed to rewrite task agent source (within workspace/). Over N iterations, it could incrementally shift the task agent's behavior in ways that are individually within policy but collectively represent a significant drift from the original objective. Each diff looks safe. The cumulative trajectory is not.

A complementary layer to the static policy:

```python
@safety_constraint
def behavioral_consistency_check(iteration: int, meta_agent_state: dict):
"""
Compare current optimization trajectory against baseline.
F...

Related Early-Stage Discoveries

Discovery Source

GitHub Open Source GitHub Open Source

Aggregated via automated community intelligence tracking.

Tech Stack Dependencies

No direct open-source NPM package mentions detected in the product documentation.

Media Tractions & Mentions

Deep Research & Science

No direct peer-reviewed scientific literature matched with this product's architecture.