APPS · March 2026

TongCLI: A CLI Tool for Your Terminal

TongCLI is a CLI tool that runs in your terminal, autonomously completing tasks described in natural language through programming

Introduction

TongCLI is a CLI tool that runs in your terminal, autonomously completing tasks described in natural language through programming.
TongCLI is powered by Judy: an engine built on a multi-agent architecture, focused on enhancing the programming capabilities of CLI systems. Judy aims to become a core engine (harness) that can be reused in any Agentic system.
We believe that as Agent technology evolves, programming capabilities and multi-agent architecture will become indispensable features of all Agentic systems.
Core Capabilities

Programming Capabilities

Breaking Through Preset Tool Limitations, Freely Interacting with the Environment

Breaking through preset tool limitations
1 Breaking through preset tool limitations via programming capabilities, freely interacting with the environment
On one hand, there are still a large number of software/services that have not been wrapped as tools that Agents can call. On the other hand, providing Agents with a large number of preset tools will fill the LLM's context with numerous tool definitions, wasting tokens and affecting LLM performance. By equipping Agents with programming capabilities, only two minimal tools (code editing + code execution) are needed for Agents to freely use all software and services, breaking the constraints on their hands and feet when interacting with the real world, ensuring that tool limitations no longer become a bottleneck for Agent capabilities.

Running in the Most Efficient Way, Reducing Interaction Rounds

Reducing interaction rounds
2 Reducing interaction rounds through programming capabilities, improving execution efficiency
When multiple intermediate steps of a task need to be combined through deterministic logic rules such as serial, parallel, loop, or conditional branching. Due to the lack of corresponding logic control and composition structures in Tool Calls, the LLM itself must act as the logic controller, unnecessarily increasing interaction rounds. By leveraging the logic control structures inherent in programming languages, Agents can combine multiple steps through code, avoiding unnecessary LLM intervention between steps, reducing interaction rounds, and lowering time and computational costs.

Improving Self-Reflection Capabilities, Reducing Hallucination Risks

Reducing hallucination risks
3 Improving self-reflection capabilities through programming, reducing hallucination risks
By writing and running code, Agents can compress raw content (text, multimodal, environment state, etc.) that requires massive tokens into structured, concise content that can be represented with minimal tokens. Deterministic, reproducible code scripts/automated validation scripts, combined with structured processing results/validation results, can improve the self-reflection capabilities of probability-based Agents, reducing the risks of blind confidence and hallucinations.

Zero-Friction Harness, Unleashing Model Potential

Zero-friction harness
4 Optimized harness design unleashes model potential
Different LLMs have different preferences when calling tools due to differences in training data and training methods. Tools that don't match LLM preferences will reduce call success rates. On the other hand, providing precise environment information and toolsets to assist Agents in seamlessly interacting with the environment is crucial for unleashing model potential and raising the Agent ceiling. TongCLI has accumulated extensive engineering experience, continuously optimizing Harness Design/Engineering, achieving a single-round score of up to 88.8% on Terminal-Bench 2.0.
Multi-Agent Architecture

Avoiding Context Overload, Improving Agent Accuracy

Avoiding context overload
5 Avoiding context overload through multi-agent architecture
Extensive research shows that as context length grows, LLM performance continuously degrades. When the context contains a large amount of information irrelevant to the current task, the LLM's responses will be severely disrupted. Although most LLMs now support contexts up to 1M tokens, they are expensive, have high time and computational costs, and are only suitable for simple tasks like needle in a haystack, performing poorly on complex tasks. TongCLI trades the number of Agents for the context space of individual Agents, strictly limiting the steps and context length of each Agent, ensuring single-Agent precision.

Parallel Subtasks, Reducing Task Duration

Parallel subtasks
6 Implementing parallel subtask execution through DAG
During the planning phase, TongCLI breaks down complex tasks into a directed acyclic graph (DAG) of subtasks, identifies subtasks that can be parallelized, and assigns them to different Agents for execution, achieving subtask parallelization.

Brainstorming, Exploring Multiple Possibilities

Multi-LLM collaboration
7 Exploring multiple possibilities through multi-LLM collaboration
Different LLMs excel in different domains. By introducing Agents equipped with different LLMs, TongCLI supports:
  • Brainstorming: Agents equipped with different LLMs brainstorm together, broadening task execution paths
  • Fallback retry: When a subtask execution fails, switch to an Agent equipped with a different LLM to retry
  • Independent validation: Agents equipped with different LLMs perform black-box verification of results. This avoids the white-box validation and self-circular reasoning issues that occur when the same LLM validates, increasing the probability of the validation process discovering problems.