9/23/2025 • LLM, RAG, AgenticAI, Learning, AIReality • Rohit Harsh

Invisible Labor of AI: The Truth About RAG, Agents, and 'Learning'

Last year companies spent around $30-$40 billion in enterprise investments in AI. Yet an MIT study found that only 5% were able to deliver ROI. That leaves $28.5–38 billion in wasted investment more than NASA’s entire annual budget (~$25B), larger than the GDP of Iceland or Cyprus, and greater than Spotify’s global revenue.

The statistics are shocking, but misleading if taken at face value. If 95% of the projects failed, the first instinct is to blame the technology. However, the same study shows that 5% of the projects succeeded, which means the technology does work, but many organizations cannot explain how to harness this new technology. They launch pilots without explaining what success looks like, try to replicate the demo into their own messy environment, they measure outputs without explaining their business impact, and they abandon projects without explaining what went wrong.

It isn’t AI that needs rescuing but the explanations around it.

The Study

MIT study revealed four patterns that define the GenAI Divide:

And the core barrier to scaling is not infrastructure, regulation, or talent. It is learning. Most GenAI systems do not retain feedback, adapt to context, or improve over time.

And this raises the fundamental question.

What is Learning

When we say “AI learns”, what exactly is learning? The model itself is frozen - GPT-5, Claude does not update from your prompts, one interaction does not make the next interaction smarter. Every interaction starts from the scratch. So where exactly is learning?

Three Types of “Learning” (that mostly don’t happen)

  1. Model Learning (doesn’t exist in production): LLMs are trained, then deployed and frozen. They didn’t learn from the interactions by themselves. The best interaction with the customer does not make LLM learn anything, and so does the worst interaction.

  2. System Learning (The promise of RAG/fine tuning): Companies try to make learning through:

    • RAG (Retrieval Augmented Generation) - but this just add the documents, and not the understanding or learning.
    • Fine-Tuning - Expensive, risky and frozen in time.
    • Prompt Engineering: the prompts drift and there is no guarantee which one is the perfect.
  3. Organizational Learning: If the AI doesn’t learn and the system barely adapts, who is supposed to be learning. The humans and organizations around the AI. This is where the real “learning” must happen - in processes, workflows, and institutional knowledge about what works and what doesn’t.

The Agentic AI Paradox

Now companies are betting on “Agentic AI”, the AI that can take actions, not only generate text. But if we can’t explain why the text output was wrong, then how we will explain where the agent went wrong and more importantly fix the next interaction? For example, an agent booking the wrong flight, will the next interaction will be of the right booking or we will just hope and depends on the probability of the Agentic AI.

Questions this raises:

Maybe the problem is more philosophical than technical. We’re trying to implement “learning machines” in organizations that themselves don’t have mechanisms to learn from implementation.

What would it mean to build learning around AI rather than expecting learning from AI?

The Human Work Behind “AI Learning”

When AI systems do work, here’s what’s actually happening behind the scenes:

A Typical Implementation Journey

“RAG” or “Agentic AI” are carefully designed, sophisticated solutions, configured and maintained by humans at every step.

It starts with curating knowledge base, PDFs, support tickets, screenshots, tribal knowledge and so on. Once all is collected, they are not plug and play for the LLM, they need cleaning, scaffolding, and labeling the content so that it can be embedded into a vector database. And every use case may require different kind of curation.

Next, comes the knowledge map, each domain concept is written in a way where the definitions, synonyms, and relationships are clearly defined. Without this, LLM can happily create new relationships which may not even exist and hallucinate.

Next, is the retrieval which needs to be tuned after understanding the user’s intent. Once the user asks the query, the knowledge base has to be retrieved based on the knowledge anchors identified in the user’s query. Based on the knowledge anchors, the detailed knowledge base is fetched. Once the documents are fetched, it needs to be re-ranked so recent documents outrank old ones, duplicates are removed, and shallow entries don’t drown out detailed ones. This tuning is manual, trial-and-error, and continuous.

On top of this sits the Prompt Engineering. The model exactly needs to be told how it should behave, how to identify the knowledge is not sufficient and when to trigger the tools, along with the definition of the tools and the input and how the output should be consumed. And when the prompts failed, they need to be re-written. Nothing about the model adapts on its own.

Agentic AI adds another complexity which makes the system capable to execute actions. Each “agent” is a set of scripts written to orchestrate queries, call APIs, or chain reasoning steps. If it fails, the next attempt starts again from scratch. The only improvement comes if humans redesign the agent logic.

The model remains frozen. The scaffolding around it evolves and only because people push it forward.

But this raises the question: If this much human work is required, what makes some attempts successful while others fail?

The Feedback Story

After all the above steps, the system is put into production and Sarah was assigned the responsibility to manage the system.

The system started serving real user’s request from Billing queries, to technical details etc. Some users really liked it, others had hesitations, but overall, people were actually using it.

As days passed, Sarah looked at the feedback. Users who got good answers rarely bothered to comment. Users who got wrong answers were vocal, but even then, many who got bad answers didn’t care enough to complain.

Sarah has the feedback but was clueless what to do with that. Users were telling whats wrong but how to trace it back, from where the LLM hallucinated or where the LLM was not able to provide the answer. More importantly, how to make sure it does not happen again.

She started building a dashboard from which queries are coming in and based on that, the system is fetching which knowledge documents, how the documents are ranked. And finally what are surfaced infront of the LLM for the final answer.

But within weeks, there were thousands of interactions logged and Sarah cannot handle all this alone. So she did what most people do, focus on the loudest complaints.

What about the other interactions which she never got the chance to look. The hallucinations, the system failing to provide any answers. was accepted by the users. This slowly degrading trust quietly moved the people away from the system.

Months later, usage started declining. Not because of dramatic failures, but because of a thousand small disappointments that no one reported and Sarah never had time to find.

The AI stayed frozen. The problems that mattered most were the ones humans never had the bandwidth to solve.

The truth is, we don’t know exactly what the successful 5% are doing differently. But we do know that whatever they’re doing, it’s not because their AI is learning.

The Philosophical Question

When organization build elaborate processes around frozen AI, where does the intelligence reside. In the AI or in the human systems that guides it?

The successful 5% do not have “learning AI” but “learning organizations” that have figured out the real manual work behind the frozen intelligence. And may be that’s enough, or may be its something entirely different what we call as “Learning” or “Intelligence” in general.

Conclusion

The $28 billion question isn’t whether AI can learn - it clearly can’t. It’s whether we are willing to acknowledge that the ‘intelligence’ in our AI systems is actually human intelligence, carefully maintained through invisible labor. The successful 5% haven’t built learning machines. They’ve built machines that require constant human learning to function.

Maybe that’s enough. Or maybe we are building something we have not yet identified, intelligence that exists not in the model, not in the system, but in the gap between human and machine.

Appreciate this article

Join the discussion

Sign in to comment