The Inference Report

March 27, 2026

The dominant pattern across today's trending repos is agents that handle research and synthesis at scale. Mvanhorn's last30days-skill and virattt's dexter both crawl multiple information sources, Reddit, X, YouTube, financial data, then distill findings into structured summaries. Bytedance's deer-flow goes further, orchestrating long-running workflows that combine research, code generation, and artifact creation across sandboxed environments. These aren't just wrappers around LLM APIs. They implement memory systems, tool composition, and task decomposition to handle work that spans hours rather than seconds. The throughput difference matters: a financial research agent that can maintain context across dozens of sources and iterations solves a real coordination problem that copy-pasting into ChatGPT doesn't.

The infrastructure supporting this work is consolidating around two layers. The orchestration layer, agentscope, oh-my-claudecode, deer-flow, handles agent lifecycle, message routing, and subagent delegation. The foundation layer, llama_index for document retrieval, chandra for OCR on complex layouts, vllm-omni for multimodal inference, handles the data ingestion and model serving that agents depend on. Twentyhq's twenty and arc53's DocsGPT are building user-facing products on top of this stack, treating agents as the delivery mechanism rather than the novelty. What's notable is the absence of monolithic frameworks. Instead you're seeing point solutions that compose well: a team picks an orchestrator, plugs in their preferred models and retrieval systems, and adds domain-specific tools. The winning repos here solve a concrete step in that pipeline, not the entire pipeline itself.

Jack Ridley

Trending