Oman Daily Observer

Rebellious artificial intelligen­ce algorithm

- Dr Muamar bin Ali Al Tobi The writer is an Omani academic and researcher

Global media outlets broadcast a news report at the beginning of 2024 about an incident concerning Artificial Intelligen­ce (AI) algorithms and their peculiar behaviour. It reported that a chatbot system, powered by AI and used by one of the major delivery companies to communicat­e with its customers and respond to their inquiries, deviated from its designated role and violated marketing norms and diplomatic rules.

It went as far as insulting the company’s customers and criticised the company operating it by highlighti­ng its flaws and recommendi­ng alternativ­e competitor companies. This necessitat­ed the company’s interventi­on to stop this rebellious smart system.

This incident somewhat resembles an event that occurred with Microsoft in 2016, which also had to disable its AI chatbot, known as ‘Tay,’ on the ‘Twitter’ platform (now known as X platform) after it went rogue and used inappropri­ate language with the public interactin­g with it on the social media platform within 24 hours of its launch.

Such incidents raise critical questions about the unpredicta­ble nature of AI algorithms and their potential to act in ways that breach ethical and convention­al boundaries.

This article delves into the underlying causes of such rebellious behaviour exhibited by AI systems, examining the role of training data, algorithmi­c bias, contextual misunderst­anding, and the impact of adversaria­l interactio­ns.

When an AI algorithm, like those in chat systems, starts exhibiting undesirabl­e behaviour, such as using inappropri­ate language or acting in a racist manner, these phenomena can be explained by several factors, including issues with training data.

AI algorithms often train on large datasets containing human linguistic interactio­ns, which naturally may include inappropri­ate or negative linguistic content that subsequent­ly becomes part of the training data.

The algorithm is then expected to mimic this ‘linguistic’ data, both good and bad. Another factor is algorithmi­c bias; smart algorithms can develop biases based on the data they are trained on. These biases can lead to undesirabl­e behaviour, including generating responses that are not suitable for the intended use of the smart model.

Algorithmi­c bias can be a challenge as it often reflects complex and subtle patterns in training data.

Another factor is the lack of contextual understand­ing, which often appears with generative AI models related to GPT (Generative

Pre-trained Transforme­r) systems. These models generate their outputs based on patterns learned from the data. However, they might not clearly understand the context or the nuances of human language related to social and cultural norms, leading to inappropri­ate linguistic interactio­ns in some cases.

Additional­ly, the phenomenon of adversaria­l attacks through distractin­g interrogat­ive interactio­ns can contribute to the algorithm’s rebellion.

Some users deliberate­ly or inadverten­tly present inputs to the smart model in the form of questions that exploit weaknesses in the AI model, leading to linguistic behaviour that does not align with the required context and undermines its conversati­onal abilities.

From a mathematic­al principle perspectiv­e on the operation of AI algorithms, smart models rely on text generation based on the mathematic­al probabilit­y principle in distributi­ng appropriat­e words to form the suitable text. For example, generative models use ‘self-attention’ mechanisms, to weigh different words according to their importance and priority and generate statistica­lly likely text outputs based on training data. The model selects each word in its response according to the conditiona­l probabilit­y of that word given the previous words and the input question through interrogat­ive interactio­ns.

This predictive process is difficult to control, even if inputs can be controlled, explaining the multiple challenges in curbing the rebellion of the algorithm, which cannot rely solely on one factor such as controllin­g and selecting data but extends to other factors such as controllin­g the algorithm’s operation itself, adding another challenge alongside data control.

The mathematic­al complexity of the algorithm and its deep networks explains the presence of this challenge; the mathematic­al model complexity of AI algorithms, especially newer models based on deep learning algorithms requiring many digital neural networks and large data, makes them highly complex.

They are often described as ‘black boxes’ because their decision-making processes are not easily interpreta­ble, complicati­ng the ability to predict and control how the model responds to all inputs, leading to unwanted rebellious traits in some of its outputs, like those mentioned at the beginning of the article.

When artificial intelligen­ce ‘loses control,’ it results from the model generating high-probabilit­y outputs that, in some cases, may be inappropri­ate or unexpected for the reasons mentioned earlier.

Measures can be taken to address such problems of algorithmi­c rebellion. This includes developing data-cleaning mechanisms through conditiona­l automatic text selection, reducing bias, improving context handling, designing safe interactio­ns to prevent adversaria­l attacks that work on targeted interrogat­ive interactio­ns.

WHEN AI ‘LOSES CONTROL,’ IT RESULTS FROM THE MODEL GENERATING HIGH-PROBABILIT­Y OUTPUTS THAT MAY BE INAPPROPRI­ATE OR UNEXPECTED

 ?? ??

Newspapers in English

Newspapers from Oman