🤖 AI Summary
To address core challenges in LLM-driven natural-language shells—including uncontrolled outputs, poor interpretability, and difficult recovery—this paper proposes NaSh, a user-centric protected command-line shell. Methodologically, NaSh introduces the first systematic protection paradigm for NL shells, integrating lightweight symbolic reasoning, structured instruction constraints, operational impact modeling, and an interactive feedback protocol. Crucially, it enables real-time intent clarification, operation preview, and execution rollback—without requiring LLM fine-tuning or reinforcement learning. This transforms the LLM from a “black-box oracle” into a “controllable collaborator,” markedly enhancing intervenability and auditability. Evaluation on realistic terminal tasks demonstrates a 72% reduction in erroneous command execution and an average user recovery time of only 1.8 seconds, validating both effectiveness and practical utility.
📝 Abstract
We explore how a shell that uses an LLM to accept natural language input might be designed differently from the shells of today. As LLMs may produce unintended or unexplainable outputs, we argue that a natural language shell should provide guardrails that empower users to recover from such errors. We concretize some ideas for doing so by designing a new shell called NaSh, identify remaining open problems in this space, and discuss research directions to address them.