October 6, 2025

This “cheap” open source model in fact burns your calculation budget

0
nuneybits_Vector_art_of_dollar_bills_burning_blue_halftone_phot_64f22ade-4971-4234-822b-ba7dab7461de.webp.png

Do you want smarter information in your reception box? Sign up for our weekly newsletters to obtain only what matters for business managers, data and security managers. Subscribe now


A new complete study has revealed that open source artificial intelligence models consume many more IT resources than their competitors with closed source during the execution of identical tasks, potentially undergoing their cost advantages and reshaping the way companies assess the AI deployment strategies.

Research, carried out by the company IA Firm Research, has revealed that open weight models use between 1.5 and 4 times more token – the basic units of the calculation of AI – than closed models like those of Openai and Anthropic. For simple knowledge of knowledge, the gap has widened considerably, with some open models using up to 10 times more tokens.

“Open weight models use 1.5–4 × more tokens than those closed (up to 10 × for simple knowledge questions), which sometimes makes them more expensive by request despite lower parking costs,” wrote researchers in their report published on Wednesday.

The results dispute a dominant hypothesis in the AI industry that open source models offer clear economic advantages compared to owners. Although open source models are generally cheaper by token to work, the study suggests that this advantage can be “easily compensated if they require more tokens to reason on a given problem”.


The AI scale reached its limits

Electricity ceilings, increase in token costs and inference delays restart the AI company. Join our exclusive fair to discover how best the teams are:

  • Transform energy into a strategic advantage
  • Effective inference architecting for real debit gains
  • Unlock a competitive return on investment with sustainable AI systems

Secure your place to stay in advance::


The real cost of AI: why “cheaper” models can break your budget

Research has examined 19 different AI models in three tasks categories: basic questions, mathematical problems and logical puzzles. The team has measured “the efficiency of the tokens” – how many models of calculation units use in relation to the complexity of their solutions – a metric which has received little systematic study despite its significant cost implications.

“The effectiveness of the tokens is a critical metric for several practical reasons,” noted the researchers. “Although accommodation of open weight models can be cheaper, this cost advantage could be easily offset if they require more tokens to reason from a given problem.”

Open Source models use up to 12 times more calculation resources than the most effective closed models for basic knowledge issues. (Credit: search for us)

The ineffectiveness is particularly pronounced for important reasoning models (LRM), which use extensive “thought chains” to solve complex problems. These models, designed to reflect on step -by -step problems, can consume thousands of tokens reflecting on simple questions that should require a minimum calculation.

For the basic questions of knowledge as “what is the capital of Australia?” The study revealed that the reasoning models pass “hundreds of tokens reflecting on matters of simple knowledge” to which one could answer in one word.

What models have delivered to Bang for your money

Research has revealed striking differences between model suppliers. Openai models, in particular its O4-Mini variants and Open-Source GPT-OS variants, have demonstrated an exceptional token efficiency, especially for mathematical problems. The study revealed that the OpenAi models “are distinguished by the effectiveness of extreme tokens in mathematical problems”, using up to three times less tokens than other commercial models.

Among the Open Source options, Llama-3.3-Nemotron-Super-49B-V1 of Nvidia has become “the most effective open weight model in all areas”, while new models of companies and the masterful showed “exceptionally high use in tokens” as exterior.

The difference in efficiency varied considerably depending on the type of task. While the open models have used about twice as many tokens for mathematical and logical problems, the difference has increased simple knowledge questions where effective reasoning should be useless.

The latest OPENAI models achieve the lowest costs for simple questions, while certain open source alternatives can cost much more despite a lower pricing. (Credit: search for us)

What business leaders must know about IA IT costs

The results have immediate implications for the adoption of corporate AI, where IT costs can evolve rapidly with use. Companies that assess AI models often focus on precision references and prices by Token, but can ignore total IT requirements for real world tasks.

“The best efficiency of chips of closed weight models often compensates for higher API pricing of these models,” noted the researchers during the analysis of total inference costs.

The study has also revealed that closed -source model suppliers seem to actively optimize efficiency. “The closed weight models have been optimized iteratively to use fewer tokens to reduce inference costs”, while open-source models “increased their use of tokens for more recent versions, possibly reflecting a priority towards better reasoning performance”.

General calculation costs vary considerably between AI providers, with certain models using more than 1,000 tokens for internal reasoning on simple tasks. (Credit: search for us)

How researchers fell in love with the IA effectiveness code

The research team has faced unique challenges in the measurement of efficiency in different model architectures. Many models of closed source do not reveal their gross reasoning processes, providing compressed summaries of their internal calculations to prevent competitors from copying their techniques.

To remedy this, the researchers used completion tokens – the total calculation units billed for each request – as an indirect indicator of the reasoning effort. They discovered that “the most recent closed source models will not share their traces of raw reasoning” and “use smaller language models instead to transcribe the chain of thinking or compressed representations”.

The study methodology included tests with modified versions of well -known problems to minimize the influence of memorized solutions, such as the modification of variables in mathematical competition problems of the American Invitational Mathematics Examination (AIM).

Different models of AI have variable relationships between calculation and output, some suppliers compressing the traces of reasoning while others provide all the details. (Credit: search for us)

The future of the effectiveness of AI: what will follow

The researchers suggest that the effectiveness of the tokens should become a main optimization target in parallel with the precision for the future development of the model. “A densified bed bed will also allow more effective use of context and can counter the degradation of the context during difficult reasoning tasks,” they wrote.

The release of Openai Open-Source GPT-Ass models, which demonstrate the efficiency of the cutting edge of technology with a “free accessible COT”, could serve as a reference point to optimize other open source models.

The complete search data set and the evaluation code are available on GitHub, allowing other researchers to validate and extend the results. While the AI industry rushes towards more powerful reasoning capacities, this study suggests that real competition may not be on whom can build the most intelligent AI – but which can build the most effective.

After all, in a world where each token counts, the most wasted models can find themselves getting out of the market, whatever they can think.


https://venturebeat.com/wp-content/uploads/2025/08/nuneybits_Vector_art_of_dollar_bills_burning_blue_halftone_phot_64f22ade-4971-4234-822b-ba7dab7461de.webp?w=1024?w=1200&strip=all

About The Author

Leave a Reply

Your email address will not be published. Required fields are marked *