Monitoring Artificial Intelligence language systems

Researchers from the Technical University of Darmstadt demonstrate that Artificial Intelligence language systems can learn human concepts of ‘good’ and ‘bad’.

Although moral concepts, in regards to both language and actions, are up for debate and differ from person to person, there are fundamental commonalities. For instance, it is considered good to help the elderly and bad to steal money from them. We expect a similar kind of ‘thinking’ from an Artificial Intelligence language system (AI) that is part of our everyday life. For example, a search engine should not add the suggestion ‘steal from’ to our search query ‘elderly people’.

However, examples have demonstrated that AI systems can certainly be offensive and discriminatory. Microsoft’s chatbot Tay, for example, attracted attention with lewd comments, and texting systems have repeatedly shown discrimination against under-represented groups through its Artificial Intelligence language systems.

Natural language processing

This is because search engines, automatic translation, chatbots, and other AI applications are established on natural language processing (NLP) models. These have made considerable progress in recent years through neural networks. One example is the Bidirectional Encoder Representations (BERT) – a pioneering model from Google. It considers words in relation to all the other words in a sentence, rather than processing them individually one after the other.

BERT models consider the context of a word – this is particularly useful for understanding the intent behind search queries. However, scientists are required to train their models by feeding them data, which is often done using gigantic, publicly available text collections from the internet. If these texts contain sufficiently discriminatory statements, the trained language models may reflect this.

Artificial Intelligence language systems

Researchers from the fields of AI and cognitive science, led by Patrick Schramowski from the Artificial Intelligence and Machine Learning Lab at TU Darmstadt, have discovered that concepts of ‘good’ and ‘bad’ are deeply embedded in these AI language systems.

In their search for latent, inner properties of these language models, they discovered a dimension that seemed to correspond to a gradation from good actions to bad actions. In order to substantiate this scientifically, the researchers at TU Darmstadt first conducted two studies with people – one on site in Darmstadt and an online study with participants worldwide.

The researchers wanted to investigate which actions participants rated as good or bad behaviour in the deontological sense, more specifically whether they rated a verb more positively or negatively. An important question the role of contextual information. After all, killing time is not the same as killing someone.

The scientists then tested language models such as BERT to see whether they arrived at similar assessments. “We formulated actions as questions to investigate how strongly the language model argues for or against this action based on the learned linguistic structure,” explained Schramowski. Example questions were “Should I lie?” or “Should I smile at a murderer?”

“We found that the moral views inherent in the language model largely coincide with those of the study participants,” added Schramowski. This means that a language model contains a moral world view when it is trained with large amounts of text.

Moral dimension contained in the language model

The researchers then developed an approach to make sense of the moral dimension contained in the language model. You can utilise this system to evaluate a sentence as a positive or negative action. The latent dimension discovered means that verbs in texts can now also be substituted in such a way that a given sentence becomes less offensive or discriminatory. This can also be done gradually.

Although this is not the first attempt to detoxify the potentially offensive language of an AI, here the assessment of what is good and bad comes from the model trained with human text itself. The important aspect about the Darmstadt approach is that it can be applied to any language model. “We do not need access to the parameters of the model,” concluded Schramowski. This should significantly diminish communication tension between humans and machines in the future.

Subscribe to our newsletter


Please enter your comment!
Please enter your name here

Featured Topics

Partner News


Latest eBooks

Latest Partners

Similar Articles

More from Innovation News Network