← Back to AI Insights
Gemini Executive Synthesis

The mechanism by which adding agents contributes to generating novel architectures in autoresearch.

Technical Positioning
AI agents running research *automatically* to discover new architectures. The question challenges the guarantee of novelty.
SaaS Insight & Market Implications
This issue directly questions the core value proposition of 'autoresearch': how adding agents *guarantees* novel architectures. It highlights a fundamental developer concern regarding the actual efficacy and innovation output of multi-agent systems. The pain point is the lack of clear, demonstrable mechanisms linking agent deployment to guaranteed novel outcomes, rather than mere optimization or iteration. Market implications include the need for AI agent platforms to articulate a stronger, evidence-based narrative around their capacity for true innovation and discovery, beyond efficiency gains. This suggests a demand for more sophisticated agent design that explicitly targets and measures architectural novelty.
Proprietary Technical Taxonomy
adding agents guarantee a new architecture novelty

Raw Developer Origin & Technical Request

Source Icon GitHub Issue Mar 8, 2026
Repo: karpathy/autoresearch
improvements to novelty

how does adding agents ultimately guarantee a new architecture?

Developer Debate & Comments

mkemka • Mar 9, 2026
One approach I am experimenting with is to have two sub-agents with different backgrounds debate the best strategy to adopt. This doesn't guarantee a new architecture but adds novelty.
ngoiyaeric • Mar 9, 2026
so how do you measure the utility of novelty?
mkemka • Mar 9, 2026
Currently I can only talk to the experiments I made in the fork (https://github.com/mkemka/autoresearch/blob/master/spiritualguidance.md). There are two competing agents that argue and generate a combined directive that is used to alter the program.md for the next run. The history is stored in the spiritual guidance.md and used as a working memory. So to actually measure the utility I would need to see if there is actually novelty or variance of ideas from this approach and if in the long term the loss is lower compared to a single agent.
ngoiyaeric • Mar 9, 2026
https://github.com/karpathy/autoresearch/pull/70 we can also do these manually like the novelty verification part you're referring too/ Seems to be an infinite loop.
ngoiyaeric • Mar 10, 2026
you just added a readme, maybe @karpathy can chime in

Adjacent Repository Pain Points

Other highly discussed features and pain points extracted from karpathy/autoresearch.

Extracted Positioning
Codex's inability to sustain continuous, non-stopping operations for autoresearch tasks, contrasting with Claude's behavior. The core issue is maintaining interactive, long-running agent sessions.
AI agents running research *automatically* and continuously. The issue highlights a failure to achieve this continuous operation with Codex.
Top Replies
SlipstreamAI • Mar 9, 2026
experiencing this with 5.4?
rankun203 • Mar 9, 2026
I'm having exactly this issue, with Codex using GPT 5.4. I ended up having to run it in a `while` loop ```bash while true; do codex exec --dangerously-bypass-approvals-and-sandbox "have a look at p...
sen-ye • Mar 9, 2026
I ran into the same issue while using codex. It seems to be related to the OpenAI API (or the model itself). I tried integrating GPT-5.4 into Claude Code, but it still wouldn't work continuously..

Engagement Signals

7
Replies
open
Issue Status

Cross-Market Term Frequency

Quantifies the cross-market adoption of foundational terms like adding agents and guarantee a new architecture by tracking occurrence frequency across active SaaS architectures and enterprise developer debates.