University of Luxembourg, Deutscher Bundestag, LIST

The experts (left to right): Tomer Libal, Patrick Glauner, Francesco Ferrero

Note from the editors: The article was updated on 14th March 2024, with the news of the EU parliament vote on the AI Act.

Yesterday, the majority of the EU parliament voted in favour of the EU AI Act, which defines the future rules for dealing with artificial intelligence (AI) in the EU. The new law can now enter into force. The EU Commission first published a draft in April 2021. Since then, the future rules have been heavily negotiated.

What do researchers think about the legal text? And what do they think is most important to drive forward the development of AI for the benefit of the general public? Here are the statements of three scientific experts based either in Luxembourg or with a connection to Luxembourg:

Prof. Patrick Glauner, Deggendorf Institute of Technology  

Patrick Glauner holds a PhD in AI from the SnT - Interdisciplinary Centre for Security, Reliability and Trust at the University of Luxembourg. He previously studied at Imperial College London. He has been Professor of Artificial Intelligence at the Deggendorf Institute of Technology since the age of 30. He is also the Ramon O'Callaghan Professor of Technology Management and Innovation at Woxsen University in India. Previously, he worked at the European Organisation for Nuclear Research (CERN), among others. As an expert, he has advised the German Bundestag and the French National Assembly, among others. He is listed by CDO Magazine as one of the world's leading professors in the field of data.

Photo © German Bundestag

First of all, it is to be welcomed that politicians are thinking about artificial intelligence. However, I personally take a very critical view of the EU AI Act and can only see very few positive aspects of it.  

The main positive aspect is the planned ban on social scoring by governments. This refers to AI systems that can be used by authorities to assess the trustworthiness of people. This would allow governments to monitor the behaviour and socio-economic status of citizens and collect data for this purpose.  

However, the EU also wants to ban social scoring by private companies. However, social scoring could certainly have useful applications in the private sector, for example in customer service. If a particularly upset customer can raise their concerns more quickly with a human dialogue partner instead of a chatbot thanks to social scoring AI, this would be an improved service.

In my opinion, the EU AI Act is a fear-driven regulation that ignores the opportunities of this technology. The regulations go very far and also cover aspects that, from a scientific point of view, do not even fall under AI - this is an attempt to regulate the computer world, so to speak.  

However, there is no major regulatory gap for AI systems. AI applications in cars or aeroplanes, which pose real risks to life and limb, have long been regulated for good reasons.  

The control of a technology for a single, specific application is called vertical regulation. The EU AI Act, on the other hand, wants to regulate all AI technology across the board for all conceivable applications, i.e. horizontally. This will inevitably lead to additional and conflict-laden regulations. Compared to China and the USA, the EU has a gap when it comes to the transfer of AI technology into the economy. We are not closing this gap through more regulation, but through increased innovation.

Universities are not directly affected by the EU AI Act. Research is exempt from the Act for the time being. However, universities are affected if they use AI in teaching or in their value chain. And if companies fund fewer third-party projects because the EU law makes it more difficult to commercialise research results, companies will no longer fund research at all or will fund it outside the EU - in the UK, for example. This could also slow down the universities in Europe.

I am already observing increasing uncertainty in SMEs and large corporations regarding the consequences of the Act. As with the Data Protection Regulation, many definitions in the AI Act are unclear and create a lot of room for interpretation. I would have liked a shorter, more precise law. A 140-page work helps neither science nor competitiveness.

My wish would be for the EU to massively promote innovation in the next legislative period. This is not only about significantly higher funding amounts, but also about the distribution procedures. It must become easier for scientists to apply for funding. Approval must also be faster. Nowadays, people sometimes wait years for funding for a project - the EU needs to rethink this completely. Computing time is very expensive; public funding for more data centres and investment in open source development would also be welcome so that everyone can use the technologies.  

The potential of AI in Europe is enormous, whether in engineering or manufacturing. AI could also alleviate the shortage of skilled labour that we are particularly experiencing in Europe, for example in the healthcare sector. AI can support doctors in making diagnoses and therapy decisions by comparing millions of cases with each other, leaving doctors more time to communicate with patients.

 

Tomer Libal, University of Luxembourg 

Tomer Libal is a researcher in the Department of Computer Science at the American University of Paris and principal investigator of several projects at the University of Luxembourg. He mainly focuses on symbolic and hybrid approaches to AI and law.

Photo © University of Luxembourg

As a researcher in ethical AI, what I like most is the attempt to put some order in the rather complex world of AI. By focusing on the intended use of an AI application (art. 5) or on the domain in which the AI application is applied (art. 6), the act somewhat clarifies when AI can have a negative impact on the lives of European citizens.  

This is far from a perfect classification, as the many opinions by stakeholders have shown. But it is an important first step and the Act contains mechanisms for regularly reviewing and updating the applications falling under the above articles. 

It is important especially when considering the alternative, which is to allow the market to regulate itself with the help of guidelines and recommendations. The European AI ethical guidelines have tried to create a voluntarily based approach and some companies have employed ethical officers to supervise the ethical aspects of AI. Nevertheless, when ethical constraints have contradicted economic targets, the latter ones usually won. 

With the Act, providers of certain AI applications now must prove, via assessments, monitoring and reporting, that the systems minimize negative effects to the rights of European citizens. 

In my opinion, the Act is still falling short of regulating AI. It tries to balance innovation and fundamental rights by focusing on certain application domains or intentions only. In a way, it tries to highlight greater violations of rights but ignores smaller ones. For example, online advertising does not seem to fall within the higher risk domains (unless it falls under art 5(1)(a) and may cause physical or psychological harm). But, the use of AI can change human economic behavior, which seems to violate the right to freedom of thought. 

While the AI Act mentions fines and other remedies in case of violations, the compliance itself is mostly expected to be done by the AI providers. This is similar to the obligations in the GDPR. On the one hand, data protection has greatly improved since the introduction of the GDPR in 2018. But violations are still common. Many of the violations are a result of a lack of knowledge or resources and it does not seem that compliance with the AI Act will be any different. 

On the other hand, the GDPR gives citizens specific rights, such as the right to be informed or the right of access. Within these rights, there are mechanisms for citizens to protect their data. The rights in the AI Act, on the other hand, are less of a personal nature. For example, the right for transparency to know when a user interacts with AI and the right to human oversight, which demands in some situations the involvement of humans in decision processes, do not directly help a specific person to avoid possible harm. A personal right, such as the right to ask a provider to not be subjected to the application of specific AI algorithms, does not seem to be included in the AI Act. 

Among the AI properties deemed most controversial are possible inaccuracy and lack of explain ability. If a user could trust decisions by AI applications to be fully accurate, and in addition, to know exactly what affected those decisions, the risk of the application having negative effects is greatly reduced. 

Nevertheless, AI applications can be made fully accurate and highly explainable. The technology is there and is proven. In my opinion, it is only a matter of cost. 

This balance between risk and cost seems to drive the focus on specific intentions and application domains. If the cost would decrease, more AI applications might fall under greater regulation, thus increasing the benefit of the general public. 

Still, most research and technological advancements keep focusing on increasing accuracy and explain ability instead of guaranteeing them. The reason for that, besides cost, is that there is a misconception among the public, and even among some researchers, about what AI really is. A common error is to synonym AI with Machine Learning, which is based on statistical inference and can therefore always make mistakes and is hard to explain by nature.  

Another form of AI - Machine Reasoning - has been used for decades in high-risk domains such as train scheduling. The reasons for its use in such domains are that it is more accurate than humans and more explainable, thus increasing the safety of such systems. 

Since AI is here to stay, I believe that an important thing right now is to acknowledge that AI can be used with low risk to the rights of European citizens, even in domains deemed currently of high risk. This acknowledgement can lead to investing resources in developing efficient ethical AI solutions that can compete with less ethical, but more economically viable ones. 

 

Francesco Ferrero, Luxembourg Institute of Science and Technology 

Francesco FERRERO (male) has been Director of the IT for Innovative Services (ITIS) department at the Luxembourg Institute of Science and Technology (LIST) since 2021. ITIS is home to more than 100 IT scientists and engineers who carry out research, development and innovation activities in the field of AI, data and software, with the aim of supporting the digital transformation of private and public organisations. Francesco started his R&D career in 2005 as part of the eSecurity Lab of Istituto Superiore Mario Boella, an Italian RTO, and has been active in the creation and execution of applied research, development and technology projects in the ICT field, particularly in the transport, logistics and smart cities sectors. 

Before joining LIST in 2016 as Lead Partnership Officer for the Mobility, Logistics and Smart Cities markets, he held several positions in research and development, management and partnership development. He has been recognised as an international expert in R&D and smart cities and smart mobility for many years, and is co-editor of a major research handbook in the field. He has also been a highly successful Horizon 2020 project coordinator, a keynote speaker at major international events and a member of high-level working groups in Luxembourg and abroad. He sits on the boards of ICT Luxembourg and the Luxembourg Media & Digital Design Centre. 

Photo © LIST

I like the fact that the EU AI Act is expected to include provisions exempting AI systems developed or used exclusively for research and development purposes from some of the more stringent requirements. This will allow researchers to experiment and test AI technologies without the full burden of compliance associated with high-risk applications, provided that appropriate safeguards are in place to manage the risks.  

I also like the fact that the Act may encourage the use of regulatory sandboxes, which are controlled environments where innovative technologies can be tested under regulatory oversight. This is a relatively new approach that will allow researchers to develop and test AI systems with more flexibility in terms of regulatory compliance. Yet it still ensures oversight and safety and allows regulators to adapt the existing regulatory framework to let new innovative technologies emerge. The latter will be an important element of the AI regulation debate for years to come.  

This is very much in the nature of LIST, a public research and technology organization with a strong track record of working with different regulators to develop technologies to assess the compliance and risks associated with new regulatory frameworks. 

While it is important for Europe to be a "regulatory superpower", it should also show the ambition to become a "knowledge and technology superpower". Today, Europe (and Luxembourg) are lagging behind the US and China in AI RDI. President Macron famously said that while the US has GAFA (Google, Apple, Facebook and Amazon) and China has BATX (Baidu, Alibaba, Tencent and Xiaomi), Europe has the GDPR (General Data Protection Regulation). A similar joke could be made about the AI Act. That's why I believe our government should push for "moonshot initiatives" to promote AI excellence.  

This should be done both at the national level, starting with the recognition of AI as a national research priority, and at the European level, where Luxembourg should join forces with other EU countries to reach the critical mass needed to compete globally. The example of the Chips Act is instructive: this is not a law to regulate chip production, but a political initiative to stimulate European RDI in the semiconductor sector with a public investment of more than 43 billion euros, thus reducing our strategic dependence on other countries. As AI has clearly become a strategic technology, the same should be done for it. 

First, we need to democratize access to AI. On the technological side, LIST is contributing with its BESSER project, supported by the FNR. This project creates an open-source low-code and no-code platform that allows people with little or no programming skills to build software that embeds AI solutions more quickly. On the education side, we need to work along two dimensions: In schools, where AI needs to be used to improve learning, for example through personalized tutoring, and where students need to be taught how to use new tools like ChatGPT responsibly. And in continuing education, where we need to up-skill and re-skill workers and citizens to use AI. 

Second, we need to address the public's fears and concerns about AI. I believe that civil society will have an important role to play in this, and that a public centre full of technical talent like the LIST should be part of a civil society movement to monitor the use of AI. In fact, we are already working in this direction and have developed the first prototype of an AI sandbox that will allow large language models, such as Open AI's GPT-4 or Mistral AI’s Mistral 7B, to be tested against a range of ethical concerns (race, age, gender bias, etc.), thus allowing the public to understand the limits of these technologies. 

Questions: Britta Schlüter
Editor: Jean-Paul Bertemes (FNR)

Aussi dans cette rubrique

FACTCHECK Quels sont les avantages et inconvénients de l’heure d’été ?

Changer d’heure deux fois par an est censé économiser de l’énergie, mais aurait un impact négatif sur la santé. Le point sur ce qu’en dit la science.

FNR
29 février
Histoire du calendrier Quelles sont les origines de notre calendrier ?

Le 29 février n'a lieu que tous les quatre ans – pendant ce qu'on appelle les années bissextiles. Mais quelle en est l’explication et comment nos ancêtres ont-ils mis au point notre système calendair...

FNR
Science-Check Voiture électrique vs voiture thermique: qui a la meilleure empreinte carbone ?

Les voitures électriques émettent-elles réellement moins de CO2 que les voitures thermiques ? Un épisode « Ziel mir keng ! » en collaboration avec le chercheur Thomas Gibon du LIST.

FNR