Artificial Analysis, the San Francisco-based platform that independently benchmarks over 100 AI models, has launched a Model Recommender tool. The service lets users specify their priorities across intelligence, speed, and cost, then spits out a ranked list of models that fit.
The pitch
Here's the problem the tool is trying to solve: there are too many models. Artificial Analysis tracks everything from Claude Opus 4.6 to Gemma 3n, across metrics like output speed, latency, pricing, context window size, and their own composite Intelligence Index. For a developer trying to pick between, say, a cheap and fast option for a customer service bot versus a pricier reasoning model for code generation, the sheer volume of data on the platform's model comparison page is genuinely overwhelming.
The Model Recommender collapses all that into a questionnaire. You set your constraints and the tool does the filtering.
According to the original Russian-language announcement (the source material was posted in Russian on X), the engine accounts for inference speed via API, multimodality support, license type, and lets users manually weight quality metrics for things like code generation, hallucination resistance, and agentic task performance. I couldn't independently verify every parameter the source claims is available, since the tool's page is heavily JavaScript-rendered and doesn't expose its full feature set to web scrapers. But the general pitch tracks with what Artificial Analysis already publishes.
Does it matter?
The honest answer: maybe. The concept isn't new. Plenty of developers already use Artificial Analysis as a de facto comparison shopping tool. Andrew Ng has publicly praised the platform for filling a gap that the LMSYS Chatbot Arena and Hugging Face leaderboards don't cover, particularly around API performance and pricing data.
But turning a leaderboard into a recommendation engine is a different thing. Leaderboards let you browse. Recommenders make choices for you. And the quality of those choices depends entirely on how well the underlying scoring works.
Artificial Analysis does have credibility here. The company, founded in 2024, runs its own Intelligence Index v4.0, which aggregates 10 evaluations including GDPval-AA, Terminal-Bench Hard, SciCode, Humanity's Last Exam, and their own AA-Omniscience hallucination benchmark. They run everything on dedicated hardware with standardized prompting. According to CB Insights, the founders maintain that nobody pays to appear on the public leaderboard, with revenue coming from enterprise subscriptions and private custom benchmarking for AI companies.
That firewall matters. If the recommender is just surfacing the same independent data in a more digestible format, fine. If model providers start paying for better placement... well, that's a different conversation entirely.
What's missing
The tool appears to focus on text-based LLMs accessed via API. Artificial Analysis does benchmark image, video, and speech models separately (they run popular arena-style comparisons for those), but it is unclear whether the recommender covers them. The source material mentions multimodality as a filter, but I couldn't confirm whether that means "models that accept images as input" or something broader.
There's also a question of freshness. The AI model landscape moves fast. When I checked the Artificial Analysis homepage, their Intelligence Index already listed models like GPT-5.2 and Claude Opus 4.6, but new models drop weekly. How quickly does the recommender update? The site doesn't say.
Who actually needs this?
Developers and teams making deployment decisions. Not researchers, not hobbyists comparing ChatGPT and Claude for personal use. The kind of person who cares about the difference between 288 tokens per second and 201 tokens per second on GPT-5 variants, or who needs to know that Gemma 3n E4B costs $0.03 per million tokens while Claude Opus 4.6 runs at $1.20.
For that audience, a recommendation engine that accounts for the full matrix of tradeoffs (intelligence scores, speed, latency, price, context window, license openness) could save real time. Whether it saves enough time over just reading the existing comparison charts is the question I can't answer yet.
The tool is live at artificialanalysis.ai/models/recommend.




