A new method for assessing uncertainty represents a major advance toward trust in the training of artificial intelligence models.

Publié le 24 June 2025 à 12h53
modifié le 24 June 2025 à 12h53

The rapid rise of artificial intelligence models across various sectors necessitates strong assurances of their reliability. A major challenge lies in accurately assessing the uncertainty of predictions. Current models struggle to define their knowledge boundaries, leaving users uncertain about their robustness. A new uncertainty analysis method proposes to enhance trust in the training of these models. The integration of such an approach revolutionizes the landscape of AI, making predictions more reliable and discoveries more immediate.

An innovative method to quantify the uncertainty of AI models

Researchers from the Pacific Northwest National Laboratory of the Department of Energy have developed a method to evaluate the uncertainty associated with artificial intelligence models, specifically those known as neural network potentials. This advancement focuses on assessing the responses provided by these models, thereby allowing for the detection of when a prediction exceeds the limits of their training.

Context and necessity of evaluation

Scientists wish to leverage the speed of predictions offered by artificial intelligence. The use of these technologies provides a competitive advantage in the field of materials and pharmaceuticals. Nevertheless, a dilemma persists between the speed of these forecasts and their accuracy. The need for a system that guarantees adequate trust in AI results is becoming urgent. The millions of dollars invested in this sector testify to this.

How the new method works

The method developed by the PNNL team quantifies the uncertainty of AI models while identifying areas requiring further training. Researchers have found that some uncertainty models exhibit overconfidence, even when prediction errors are significant. This overconfidence is common in deep neural networks, but the SNAP method helps mitigate this phenomenon.

Collaboration and availability of tools

The authors of the study, Jenna Bilbrey Pope and Sutanay Choudhury, have implemented a public availability process for their method via GitHub. This allows researchers worldwide to integrate these tools into their work, thereby promoting increased collaboration and innovation. This sharing of knowledge is essential for advancing the scientific community.

Impact on chemical and materials research

The results of the study show that the method can be used to evaluate how predictive models, such as MACE, are trained to estimate the energy of different families of materials. This type of calculation has significant implications for understanding the more time- and energy-intensive methods that are generally reserved for supercomputers. The researchers’ work paves the way for greater confidence in simulations performed using AI.

Benefits for autonomous laboratories

This advancement also aims to facilitate the incorporation of AI workflows into the daily operations of laboratories, transforming artificial intelligence into a reliable laboratory assistant. A model’s ability to clearly indicate its knowledge boundaries will allow scientists to approach predictions with a measurable level of confidence. The ambition of this methodology is to propel AI into new uses.

Statements from researchers

Team members emphasize the importance of reliable assessment of predictions. “AI must be able to detect its limits,” Choudhury explained. This will encourage cautious statements, such as: “This prediction provides 85% confidence that catalyst A is superior to catalyst B, according to your criteria.”

Reactions in the media and industry

The fallout from this research resonates well beyond the scientific community. The strong involvement of major technology companies, with expenditures exceeding $320 billion in artificial intelligence, reflects a challenge towards effectiveness that also rests on the reliability of models. The issues related to trust in AI are now at the heart of the debates.

For more information, it is possible to consult previous articles on the subject. For example, the reaction of writers to a computer joke is a hot topic in the field of AI: link here. In-depth thoughts on the spending imperatives of technology companies can also be found in this article.

Frequently Asked Questions about uncertainty evaluation in training artificial intelligence models

What is the importance of evaluating uncertainty in artificial intelligence models?
Evaluating uncertainty allows measuring the reliability of predictions provided by an artificial intelligence model, which is crucial for making informed decisions, especially in scientific and industrial contexts.

How does the method developed by PNNL improve the training of artificial intelligence models?
The method quantifies uncertainty in the predictions of models, helping researchers identify the limits of their models and determine areas needing additional training.

What does the term ‘active learning’ mean in this context?
‘Active learning’ refers to a process where the model learns by identifying difficult or uncertain examples, helping enhance its performance and accuracy over time.

Why is overconfidence in AI models a problem?
Overconfidence can lead to misinterpretations of results, increasing the risk of errors in decisions based on predictions, especially when models are poorly trained.

Can artificial intelligence models really predict material properties before experimentation?
Yes, with well-trained AI models, it is possible to predict material properties with good accuracy, reducing the time and costs associated with experimental trials.

What methodology did the PNNL researchers use to validate their uncertainty evaluation method?
They conducted benchmarks with advanced atomic-scale materials chemistry models, such as MACE, to assess the accuracy and reliability of predictions.

What role does the SNAP model play in uncertainty evaluation?
The SNAP model provides a framework for integrating neural network potentials while considering uncertainty, thereby enhancing the reliability of results provided by AI models.

How does the current methodology change scientists’ perceptions of AIs?
It allows scientists to view AIs as trusted tools capable of assisting in research and development, rather than as ‘black boxes’ whose decisions are difficult to interpret.

actu.iaNon classéA new method for assessing uncertainty represents a major advance toward trust...

the theory about Jony Ive’s AI hardware device is becoming increasingly credible

explorez la théorie captivante sur le dispositif matériel d'intelligence artificielle imaginé par jony ive, qui gagne en crédibilité. découvrez comment ses concepts innovants pourraient révolutionner notre interaction avec la technologie et redéfinir l'avenir des objets connectés.

how artificial intelligence has invested the world of perfumery

découvrez comment l'intelligence artificielle transforme l'industrie de la parfumerie, de la création de nouvelles fragrances à l'optimisation des procédés, en alliant innovation technologique et art de la senteur.

The influence of AI on our language: a study reveals that humans express themselves like ChatGPT

découvrez comment l'intelligence artificielle, à travers des outils comme chatgpt, façonne notre manière de communiquer. cette étude approfondie révèle des tendances fascinantes sur l'évolution de notre langage et les similitudes croissantes entre les expressions humaines et celles générées par l'ia.

Thomas Wolf from Hugging Face: the ambition to democratize robotics through open source

découvrez comment thomas wolf, co-fondateur de hugging face, vise à démocratiser la robotique grâce à l'open source. explorez ses idées innovantes et son engagement pour rendre la technologie accessible à tous.

the 20 most powerful AI models of June 2025: discover the detailed ranking

découvrez notre classement détaillé des 20 modèles d'intelligence artificielle les plus performants de juin 2025. explorez les innovations et les avancées qui façonnent l'avenir de la technologie.

Cédric O facing accusations of conflicts of interest, but receiving support from the HATVP

découvrez comment cédric o se retrouve au cœur de controverses concernant des accusations de conflit d'intérêts, tout en recevant le soutien inattendu de la haute autorité pour la transparence de la vie publique (hatvp).