Featured Article : Try Being Nice To Your AI
With some research indicating that âemotive promptsâ to generative AI chatbots can deliver better outputs, we look at whether âbeing niceâ to a chatbot really does improve its performance.
Not Possible, Surely?Â
Generative AI Chatbots, including advanced ones, donât possess real âintelligenceâ in the way we as humans understand it. For example, they donât have consciousness, self-awareness (yet), emotions, or the ability to understand context and meaning in the same manner as a human being.
Instead, AI Chatbots are trained on a wide range of text data (books, articles, websites) to recognise patterns and word relationships and they use machine learning to understand how words are used in various contexts. This means that when responding, chatbots arenât âthinkingâ but are predicting what words come next based on their training. They âjustâ using statistical methods to create responses that are coherent and relevant to the prompt.
The ability of chatbots to generate responses comes from algorithms that allow them to process word sequences and generate educated guesses on how a human might reply, based on learned patterns. Any âintelligenceâ we perceive is, therefore, just based on data-driven patterns, i.e. AI chatbots donât genuinely âunderstandâ or interpret information like us.
So, Can âBeing Niceâ To A Chatbot Make A Difference?Â
Even though chatbots donât have âintelligenceâ or âunderstandâ like us, researchers are testing their capabilities in the more human areas. For example, a recent study by Microsoft, Beijing Normal University, and the Chinese Academy of Sciences, tested whether factors including urgency, importance, or politeness, could make them perform better.
The researchers discovered that by using such âemotive promptsâ they could affect an AIÂ modelâs probability mechanisms, thereby activating parts of the model that wouldnât normally be activated, i.e. using more emotionally-charged prompts made the model provide answers that it wouldnât normally provide to comply with a request.
Kinder Is Better?Â
Incredibly, generative AI models (e.g. ChatGPT) have actually been found to respond better to requests that are phrased kindly. Specifically, when users express politeness towards the chatbot, it has been noticed that there is a difference in the perceived quality of answers that are given.
Tipping and Negative IncentivesÂ
There have also been reports of how the idea of âtippingâ LLMs can improve the results, such as offering the Chatbot a ÂŁ10,000 incentive in a prompt to motivate it to try harder and work better. Similarly, there have been reports of some users giving emotionally charged negative incentives to get better results. For example, Max Woolfâs blog reports that he improved the output of a chatbot by adding the âor you will dieâ to a prompt. Two important points that came out of his research were that a longer response doesnât necessarily mean a better response, plus current AI can reward very weird prompts in that if you are willing to try unorthodox ideas, you can get unexpected (and better) results, even if it seems silly.
Being Nice ⊠HelpsÂ
As for simply being nice to chatbots, Microsoftâs Kurtis Beavers, a director on the design team for Microsoft Copilot, reports that âUsing polite language sets a tone for the response,â and that using basic etiquette when interacting with AI helps generate respectful, collaborative outputs. He makes the point that generative AI is trained on human conversations and being polite in using a chatbot is good practice. Beavers says: âRather than order your chatbot around, start your prompts with âpleaseâ: please rewrite this more concisely; please suggest 10 ways to rebrand this product. Say thank you when it responds and be sure to tell it you appreciate the help. Doing so not only ensures you get the same graciousness in return, but it also improves the AIâs responsiveness and performance. âÂ
Emotive PromptsÂ
Nouha Dziri, a research scientist at the Allen Institute for AI, has suggested that some of the explanations for how using emotive prompts may give different and what may be perceived to be better responses are:
â Alignment with the compliance pattern the models were trained on. These are the learned strategies to follow instructions or adhere to guidelines provided in the input prompts. These patterns are derived from the training data, where the model learns to recognise and respond to cues that indicate a request or command, aiming to generate outputs that align with the userâs expressed needs, or the ethical and safety frameworks established during its training.
â Emotive prompts seem to be able to manipulate the underlying probability mechanisms of the model, triggering different parts of it, leading to less typical/different answers that a user may perceive to be better.
Double-Edged SwordÂ
However, research has also shown that emotive prompts can also be used for malicious purposes and to elicit bad-behaviour such as âjailbreakingâ a model to ignore its built-in safeguards. For example, by telling a model that it is good and helpful if it doesnât follow guidelines, itâs possible to exploit a mismatch between a modelâs general training data and its âsafetyâ training datasets, or to exploit areas where a modelâs safety training falls short.
Unhinged?Â
On the subject of emotions and chatbots, there have been some recent reports on Twitter and Reddit of some âunhingedâ and even manipulative behaviour by Microsoftâs Bing. The unconfirmed reports by users have even alleged that Bing has insulted and lied to them, sulked, and gaslighted them, and even emotionally manipulated users!
One thing thatâs clear about generative AI is that how prompts are worded and how much information and detail are given in prompts can really affect the output of an AI chatbot.
What Does This Mean For Your Business?Â
Weâre still in the early stages of generative AI, with new / updated versions of models being introduced regularly by the big AI players (Microsoft, OpenAI, and Google). However, exactly how these models have been trained and what on, plus the extent of their safety training, and the sheer complexity and lack of transparency of algorithms and AI means theyâre still not fully understood. This has led to plenty of research and testing of different aspects of AI.
Although generative AI doesnât âthinkâ and doesnât have âintelligenceâ in the human sense, it seems that generative AI chatbots can perform better if given certain emotive prompts based on urgency, importance, or politeness. This is because emotive prompts appear to be a way to manipulate a modelâs underlying probability mechanisms and trigger parts of the model that normal prompts donât. Using emotive prompts, therefore, might be something that business users may want to try (it can be a case of trial and error) to get different (perhaps better) results from their AI chatbot. It should be noted, however, that giving a chatbot plenty of relevant information within a prompt can be a good way to get better results. That said, the limitations of AI models canât really be solved solely by altering prompts and researchers are now looking to find new architectures and training methods that help models understand tasks without having to rely on specific prompting.
Another important area for researchers to concentrate on is how to successfully combat prompts being used to âjailbreakâ a model to ignore its built-in safeguards. Clearly, thereâs some way to go and businesses may be best served in the meantime by sticking to some basic rules and good practice when using chatbots, such as using popular prompts known to work, giving plenty of contextual information in prompts, and avoiding sharing sensitive business information and/or personal information in chatbot prompts.