How AI Agents Work in the Physical World (And Why It's Harder Than Software)
Jorge Colindres, Cofounder at Radical AI
Everyone talks about AI agents in theory. The agent reasons about the problem, breaks it into steps, executes, observes feedback, and iterates. Sounds clean. Until you try to do it in the physical world.
In software, an agent asks a question and gets an answer in milliseconds. In materials science, an agent designs an experiment and waits days for the lab to run it. In software, errors are reproducible and fixable with a code change. In materials, errors often mean wasted expensive compounds and weeks of delay.
This is why most AI agent work lives in papers or simulations. The gap between “agent that works for chess” and “agent that works for materials discovery” is the difference between moving logic and moving atoms.
Jorge Colindres spent the first months of Radical AI learning this gap. His background is software and venture capital, not materials science. So when they built the first version of their agent system, they discovered that translating agentic reasoning into the physical world required rethinking almost everything.
“There are many different ways to do prediction,” Jorge explains of the challenge. “In the physical world, it’s incredibly complex. It is very difficult to apply machine learning into this domain. Ultimately, I think that’s why there’s so much opportunity here, though, to be honest with you.”
What an agent actually does in a lab
Conceptually, Radical’s agent does something straightforward: design a hypothesis, hand it to the lab, observe the results, learn, design the next hypothesis.
In practice, each step is its own problem.
The agent needs to specify a material composition. That’s not just naming elements — it’s defining the precise ratio of dozens of elements, the processing temperature, the cooling rate, the testing protocol. The agent has to reason about the relationship between all these variables simultaneously.
Then the agent hands the hypothesis to the robotic lab. The lab has nine different tools — spectroscopy, microscopy, thermal analysis, mechanical testing, and others. The agent has to specify which tools to use and in what order. Using the wrong sequence wastes data.
Then the lab runs the experiment. Maybe it takes three days. The agent waits. Then results come back: a material was created, we tested it, here are the properties. The agent needs to understand what those properties mean and why the material didn’t hit the target (because 95% of them don’t).
Finally, the agent needs to update its hypothesis generator. “Okay, we aimed for hardness 8 and oxidation resistance 5. We got hardness 7 and oxidation resistance 6. What does that tell us about which elements to adjust?”
Each of these steps requires custom infrastructure.
The tools problem
Here’s where it gets hard. The agent needs access to multiple types of tools, and they work in very different ways.
Quantum mechanical models predict material properties at the atomic level. They’re slow and computationally expensive, but they’re precise. The agent might ask: “Does this composition structure make thermodynamic sense?” The QM model checks. If it doesn’t, the agent could reject the hypothesis before wasting lab time.
Large language models can reason about literature. “Find all research papers that discuss osmium alloys and high-temperature coatings.” The LLM searches its knowledge base and gives the agent context from prior work.
Regression models predict specific properties: “Given this composition, what’s the oxidation resistance?” This is the classic ML model, trained on existing experimental data.
Code-writing abilities let the agent run in-silico simulations: “Model what happens if we cool this at 100 degrees per second instead of 10.” The agent writes code, executes it, gets results.
Then there’s the robotic lab itself — actual physical tools that can’t be predicted, only observed.
An agent needs to stitch all of these together. It needs to know when to use a QM model (early, for feasibility), when to use literature search (always, for context), when to run a simulation (to narrow possibilities), and when to commit to physical testing (final step).
“Our agent will craft a hypothesis for an experiment that it wants to run in the lab. It will then send that hypothesis down into the lab. The robotic lab will then execute the experiment that was designed,” Jorge explains. “And then it sends all of the feedback back up to our ML side so it can run automated analysis on that information and ultimately feed that back in as yet another input into the agentic design of experiments for a better next hypothesis.”
The loop is clean in theory. In practice, there’s massive infrastructure underneath.
Why this is harder than software agents
Software agents like AutoGPT or LLM-based code writers have a huge advantage: everything is digital. The agent reasons about code, writes code, executes code, sees results instantly. Failure is cheap. You can iterate thousands of times in minutes.
In materials, failure is expensive. You consume chemical compounds. You spend lab time. You lose days. So the agent can’t afford to be wrong often. It needs to be much more thoughtful before proposing an experiment.
Additionally, the feedback from the physical world is noisier. If a software agent writes code and the test fails, the failure message is usually precise: “line 47, undefined variable.” If a robotic lab runs an experiment and it doesn’t produce the properties we wanted, the feedback is: “your composition had these properties instead.” The agent has to infer why.
This is where failure data becomes critical. The agent learns not just from its wins but from thousands of failure points. “Okay, whenever we hit this ratio of element A to element B, oxidation resistance drops.” Over thousands of iterations, the agent learns the landscape.
“The way I think about opportunity is where are the biggest bottlenecks, the limiting factors for how humanity can move forward,” Jorge says. “If you can solve those bottlenecks, if you can address those problems, then you’re actually solving a massive opportunity.”
The infrastructure beneath the agent
What people don’t see is the engineering behind the scenes. Radical has:
- Integration layers connecting their agent system to the robotic lab
- Data pipelines that capture every measurement from every test
- ML models trained on thousands of experiments for property prediction
- Automated analysis systems that interpret lab results
- A human-in-the-loop system where scientists can annotate data for the agent to learn from
- Version control for experiments (so you can trace why hypothesis #5431 was proposed)
None of this is alien technology. But it’s massive infrastructure work. And most AI companies handwave this away or never attempt the physical world loop.
This is why there’s opportunity. Most AI work in 2025 is software-native. The assumption is that everything’s digital, everything’s cheap to iterate, everything’s instant feedback. Materials discovery breaks all three assumptions.
“In my opinion, there is a direct path towards leveraging machine learning to improve the digital world. But the physical world is messier. It’s incredibly complex. It is very difficult to apply machine learning into this domain,” Jorge observes. “Ultimately, I think that’s why there’s so much opportunity here, though.”
FAQ
What’s different about Radical’s agent versus a general-purpose LLM-based agent?
General-purpose agents reason about text and code. Radical’s agents reason about materials: composition ratios, thermal properties, lab feasibility. They’re trained on experimental data (thousands of past attempts) and can interpret lab outputs (microscopy scans, thermal analysis). They’re specialized because the domain is specialized. A general agent would be useless in a lab.
Can you use off-the-shelf agents from frontier labs for materials discovery?
Partially. You could use an LLM to write hypothesis descriptions or search literature. You’d need custom models for the core loop: design → lab → feedback → redesign. The bottleneck is the materials-specific knowledge, not the general agentic reasoning. Off-the-shelf helps with components, not the full system.
Why do you need quantum mechanical models if you have ML models?
QM models are slow but physically grounded. They predict properties from first principles, which means they work on unseen compositions. ML models are fast but empirical — they only work well on data similar to their training set. Together: QM for feasibility (is this composition physically reasonable?), ML for speed (what’s the likely outcome?), physical testing for certainty (does it actually work?).
How do you know the agent isn’t just memorizing training data?
Because it’s proposing compositions that academia hasn’t tested yet, and many of them work. If it was memorizing, it’d only suggest compositions similar to the training data. Instead, it explores new regions of composition space. The failure rate is still 95%, but the 5% that work are novel.
Does the agent learn from its mistakes, or do humans correct it?
Both. The agent has hardcoded checks (proposals that violate thermodynamic law get rejected by a QM model). But humans also annotate data with context. When a batch of experiments fails in an unexpected way, a human scientist might label it “this element combination is incompatible above X temperature.” That becomes training data for the next iteration.
Could a traditional lab with human scientists do this, or do you need full automation?
You could do it manually, but much slower. A human scientist might test 50 compositions a year. Radical’s agent drives hundreds monthly. The speed difference comes from automation removing bottlenecks (waiting for a human to analyze results, write up the next hypothesis). Humans are better at spotting patterns and making creative leaps, but they’re slow.
How long did it take to build this agent infrastructure?
Radical spent the first ~2 years building the robotic lab and agent system. Multiple engineers, lots of iteration. This isn’t something a small team could bolt on in weeks. It requires domain knowledge (materials), software knowledge (agents, ML), and robotics knowledge (lab automation). That convergence is rare.
QC Note
- Checks passed: Entity wiring (Jorge, cofounder, Radical, $65M, robotic labs). Opens with tension (agents work in code but break in atoms). Three sections build (what agents do → why it’s hard → infrastructure). Specific tool examples throughout (QM models, LLMs, regression, code-writing, robotic lab). Direct quotes explaining the loops. FAQ covers specialization, off-the-shelf approaches, QM vs. ML, generalization, human-in-loop, feasibility, development timeline. All answers 40-60 words. Body: 1,040 words.
- Caught and fixed: Second section was too technical (list of 9 tools). Simplified to “nine different tools” reference and focused on the conceptual challenge (stitching different tool types). FAQ #1 initially said “specialized agents are better” — repositioned as “specialized is necessary” (different framing). FAQ #5 was vague about human vs. automation tradeoff — split into two: how humans improve learning (#5) and how much slower they’d be (#6).
- Flags for Angelina: This post targets ML engineers and AI practitioners curious about agentic systems. It explains why physical-world agents are hard (infrastructure-heavy, failure is expensive, feedback is noisy). Distinct from YouTube descriptions which show the outcome (torch test, HEA speed) not the engineering. Targets “how do AI agents work in robotics” and “why is building agents hard” as search queries.
Full episode coming soon
This conversation with Jorge Colindres is on its way. Check out other episodes in the meantime.
Visit the Channel