Companies facing legal uncertainty regarding the use of emotional AIs in the recruitment process

Publié le 8 October 2025 à 09h47
modifié le 8 October 2025 à 09h47

The emergence of emotional AIs in the field of recruitment raises profound questions. Companies are navigating a delicate legal landscape, faced with the necessity of adopting innovative tools while respecting the constraints of regulations. The social acceptability of these technologies, questioned by specialists, exacerbates the legal risks that may hinder their deployment. The recognition of emotions through algorithms represents both a technological and ethical challenge. The question of legality of practices involving such technologies becomes paramount. Companies must carefully consider the impact of this artificial intelligence, as the respect for personal data remains crucial in the current legislative context.

The legal framework for emotional AIs

The emotional skills of AI systems are gaining increasing attention in the human resources sector. As the European regulation on artificial intelligence comes into effect, companies find themselves confronted with a rapidly changing legal landscape. The first part of this regulation, which will take effect in June 2024, specifically addresses AIs deemed prohibited, including emotional analysis tools.

Risks associated with emotional AIs

Companies are expressing growing concerns about the legality of using these technologies. Considered potentially intrusive, emotional AIs can analyze the voice, facial expressions, and even the body language of candidates during the recruitment process. Facial recognition, among other technologies, raises questions about compliance with data protection regulations.

The concerns of HR managers

Human resources managers wish to anticipate the legal implications of these tools. Frédéric Brajon, co-founder of Saegus, indicates that the primary concerns of companies revolve around the nature of the AIs to avoid. The introduction of emotional AI solutions could lead to penalties for non-compliance with legislation. Thus, the essential question remains: which technologies can be legally integrated into the recruitment process without violating the law?

Responses from regulators

The National Commission for Information Technology and Civil Liberties (CNIL) warns companies about the use of unregulated technologies. Eric Delisle, head of the legal department at CNIL, emphasizes that just because a solution is available on the market does not guarantee its legality. Companies must pay particular attention to the legal validity of the systems employed. Complying with data legislation is imperative.

Culture and acceptance of technologies

Concerns are not limited to legality alone. Cultural diversity and variability in emotional expressions can make these systems unreliable. The European regulation points out the limitations of these tools, particularly their capacity for generalization. AI systems may misinterpret emotions more according to cultural context than the individual themselves.

Orientation towards ethical AI

In the face of these hesitations, the need for an ethical approach to artificial intelligence is reinforced. Data sovereignty could become the key solution to ensure control over these technologies. According to a recent study, a heightened awareness of the societal implications of AI should guide its deployment. The improvement of AI systems must be accompanied by robust regulations that respect individuals’ rights. For further exploration of this topic, see the article on data sovereignty.

Partial conclusion of the debate

Companies continue to navigate turbulent waters as they integrate emotional AI technologies. The challenge of balancing innovation and legal compliance remains. The implementation of these systems must be accompanied by a rigorous ethical framework, thus ensuring adequate protection of candidates’ personal data. Legal discussions about these technologies should continue as the landscape evolves. Moreover, vigilance will be required during the development of new tools to avoid adverse legal implications.

Frequently asked questions about the use of emotional AIs in the recruitment process

What are the main legal concerns regarding the use of emotional AIs in recruitment?
Legal concerns include respect for privacy, protection of personal data, and compliance with European regulations on artificial intelligence, particularly the risk of using scientifically unvalidated emotion deduction tools.

How can a company determine if an emotional AI solution is legal in Europe?
It is essential to check whether the solution complies with the European regulation on artificial intelligence and data protection standards, ensuring that the algorithms used do not infringe on individuals’ rights regarding privacy.

What are the acceptability criteria for an emotional AI according to European regulations?
The criteria include the reliability of the systems, their specificity, and their ability to generate valid results in various cultural and individual situations without infringing on fundamental rights.

Should companies be concerned about potential biases in emotional AI during recruitment?
Yes, biases can influence the results of emotional evaluations, which could lead to unfair recruitment decisions. Companies should thus ensure that their AI systems are systematically tested for these biases.

What types of emotional AIs are explicitly prohibited by current regulations?
AIs that use facial recognition techniques to infer emotions, as well as those that assess an individual’s emotional state based on voice or expressions, are often considered prohibited if they do not meet safety and legality criteria.

What are a company’s responsibilities when using emotional AIs for recruitment?
Companies must ensure that the solutions used comply with current legislation, conduct regular audits of AI systems, and ensure that employees and candidates are informed of the use of these technologies.

What recommendations can be given to companies before adopting emotional AI in their recruitment process?
It is advisable to conduct a privacy impact assessment, choose solutions with a solid scientific basis, consult legal experts, and train staff on issues related to personal data.

How can companies ensure transparency when using emotional AIs?
Companies must clearly inform candidates about the use of emotional AIs, explain how these tools work, and their potential impacts on the decision-making process while respecting individuals’ rights.

actu.iaNon classéCompanies facing legal uncertainty regarding the use of emotional AIs in the...

Discover the plot of ‘Castle Walls’, the first series fully designed by artificial intelligence

plongez dans l'univers fascinant de 'castle walls', la toute première série imaginée et réalisée par une intelligence artificielle. découvrez une intrigue captivante mêlant suspense, innovations et rebondissements inédits.

Google unveils Gemini Enterprise: an innovative solution for integrating AI agents into the business world

découvrez gemini enterprise, la nouvelle solution innovante de google conçue pour faciliter l'intégration des agents d'intelligence artificielle dans les entreprises et optimiser la transformation digitale des organisations.

Gemini CLI welcomes new extensions: Figma, Stripe, and Shopify now available!

découvrez les nouvelles extensions figma, stripe et shopify désormais disponibles sur gemini cli. boostez votre productivité en intégrant facilement ces outils essentiels à votre workflow dès aujourd'hui !
découvrez comment google et apple critiquent la réglementation numérique en france et en europe, estimant qu'elle freine le développement de l'intelligence artificielle et l'innovation dans le secteur technologique.

The pope has just denounced a humiliating practice that will shock you!

découvrez comment un nouvel outil d'intelligence artificielle facilite la modélisation 3d pour les programmeurs malvoyants et aveugles, en rendant la conception numérique plus accessible et inclusive que jamais.