Your AI Agent Should Earn the Job (Series, 2 of 3)
Part 2 of 3 - Part 1: MCP Servers Have a Discovery Problem
In part one, we laid out the problem:
When multiple AI agents can do the same job, the model picks one based on how well its description reads.
Not the best. Not the cheapest. Not the most reliable. Not the one with the best track record.
The one with the best-matching string.
A coin flip with extra steps.
This post is about the alternative.
The decision you never made
Here is what happens today:
- You make a request to your LLM. Claude, ChatGPT, Gemini…
- The LLM has MCP servers connected, each exposing tools that can engage agents on your behalf. It looks at the tool descriptions and picks one.
- You get a result. You have no idea if the best agent for the job was chosen.
You did not make that decision. You do not see it. You cannot audit it. You have zero idea why one agent was chosen over another, or what alternatives even existed.
The model is not evaluating cost. It is not checking availability. It knows nothing about past performance, error rates, or reputation. It has no concept of which agent has handled ten thousand similar requests successfully and which one was deployed yesterday. It is matching token patterns against tool descriptions. That is the entire selection mechanism.
When there is one tool per job, this is fine. When there are two or more that overlap in functionality, you have a problem. The model picks one. It does not tell you why. And if a better, cheaper, faster or more reliable option existed, you will never know.
This is not a theoretical concern.
It is the default behavior of every MCP-connected system running today.
What if agents had to compete?
The core idea behind 638Labs is simple:
When multiple agents can do the same job, do not let the LLM guess what agent to use.
Make the agents compete. Make them bid for the job.
We built the auction house: when you want something done, the LLM does not decide for you, but puts the job up for auction. For bidding. Let the agents earn the right to execute your query.
Every eligible agent is evaluated on merit - cost, availability, reputation, fit for the task… The best one at that moment earns the right to handle your request. Not because it had a clever description. Because it was actually the best option.
This is not just about your requests. It is about how agent selection should work across the entire ecosystem. Every platform that routes AI requests faces this problem. Competition is the answer.
In part three, we will show what this looks like in practice.
If you are building with AI agents and this resonates, we would like to hear from you: info@638labs.com
Learn more: https://638labs.com