topics: [“control engineering”, “PID”, “AI”, “LLM”, “robotics”]
As AI and LLM-based control attract increasing attention,
many people—including myself—tend to think:
With that expectation, I designed and implemented an
AITL (Adaptive / Intelligent Triple-Layer) control architecture,
combining PID × FSM × LLM, and evaluated it against a standalone PID controller
using Python-based simulations.
The conclusion, stated plainly, was this:
Modern control theory—especially PID control—is far more complete and robust
than commonly assumed, and even with AITL, it is difficult to demonstrate
a clear advantage over a well-designed PID controller.
This article documents that evaluation process
and the sober conclusions drawn from it.
The idea behind AITL is straightforward:
The underlying expectations were:
In particular, I expected benefits in systems with:
A critical point in this evaluation was not to trivialize PID control.
The baseline PID controller was implemented with:
Under these conditions, a properly tuned PID controller showed:
In short, a well-designed PID controller is extremely strong.
Next, FSM and LLM layers were added:
Structurally, this is sound.
It simply separates what humans already do.
However, in practice, several issues emerged:
To be completely honest:
Under these conditions, no clear win over standalone PID was observed.
From both performance and operational perspectives:
The reasons are straightforward:
A properly designed PID controller is already close to optimal
—that was the strongest takeaway from this study.
This does not mean AITL is meaningless.
AITL may be valuable in contexts such as:
In other words:
AITL is not a technology to directly improve control performance,
but a framework for structuring design and decision-making.
AI- and LLM-based control is appealing.
But once implemented and compared honestly:
Modern control theory is already very strong.
This is not a failure.
It is the result of testing reality against expectation.
AITL should not be positioned as a replacement for PID,
but rather as a conceptual scaffold to revisit
when truly necessary.
“AI will dramatically improve everything”
—reality is not that simple.
But:
Trying it, and drawing a clear boundary,
is itself a meaningful engineering outcome.
I hope this serves as a useful reference
for others facing similar questions.