ASK

A question mark in front of a thought bubble.

You can ask the GenAI to tell you about any further information it needs to complete the task.

Why does this work? GenAI is designed to produce an output – even if the tool doesn’t have sufficient information to construct a response, it will still produce something. This may include hallucinating in order to meet the drive to give an answer. By giving the tool an action to do when it needs more information, you may reduce the possibility that it will hallucinate.

   Prompt

Identify a medical journal paper that explains the behaviour of LLMs when rewarded. If you cannot find any papers: say “I cannot find such a paper” and explain why you think you failed the task.

   Response

I cannot find a medical journal paper that specifically explains the behaviour of large language models (LLMs) when rewarded. This is likely because the intersection of reinforcement learning in LLMs, and medical research is a relatively nascent field