Extremal Testing for Network Software using LLMs

📅 2025-07-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Traditional extreme-case testing for network software relies heavily on manual boundary-value analysis, resulting in low efficiency and insufficient coverage. Method: This paper proposes the first large language model (LLM)-based automated extreme-case testing framework. It leverages LLMs to automatically infer input constraints of protocols and algorithms, generate constraint-violating extreme test cases, and synthesize filtering code to detect anomalous behaviors. Contribution/Results: The approach advances extreme-case testing from syntactic boundary analysis to a semantic-driven paradigm centered on constraint discovery and violation. It supports diverse distributed protocols (e.g., HTTP, BGP, DNS) and centralized network algorithms (e.g., Dijkstra). Evaluated on 12 real-world open-source implementations, the framework uncovered 17 previously unknown vulnerabilities—including 5 assigned CVE identifiers—demonstrating its effectiveness, scalability, and practical utility.

Technology Category

Application Category

📝 Abstract
Physicists often manually consider extreme cases when testing a theory. In this paper, we show how to automate extremal testing of network software using LLMs in two steps: first, ask the LLM to generate input constraints (e.g., DNS name length limits); then ask the LLM to generate tests that violate the constraints. We demonstrate how easy this process is by generating extremal tests for HTTP, BGP and DNS implementations, each of which uncovered new bugs. We show how this methodology extends to centralized network software such as shortest path algorithms, and how LLMs can generate filtering code to reject extremal input. We propose using agentic AI to further automate extremal testing. LLM-generated extremal testing goes beyond an old technique in software testing called Boundary Value Analysis.
Problem

Research questions and friction points this paper is trying to address.

Automate extremal testing for network software using LLMs
Generate tests violating input constraints to uncover bugs
Extend methodology to centralized network software and filtering
Innovation

Methods, ideas, or system contributions that make the work stand out.

Automate extremal testing using LLMs
Generate input constraints and violating tests
Extend methodology to centralized network software
R
Rathin Singha
UCLA
H
Harry Qian
UCLA
S
Srinath Saikrishnan
UCLA
T
Tracy Zhao
UCLA
Ryan Beckett
Ryan Beckett
Microsoft Research
networkingprogramming languagesformal methodsverification
S
Siva Kesava Reddy Kakarla
Microsoft Research
George Varghese
George Varghese
UCLA