Improving the QuestionHuffing: In draughts, when a player fails to make a mandatory capture if available, the opponent can remove the piece that should have made the capture from the board.
When we ask Generative AI for help, we know that we won't always get the answer we are looking for on the first attempt. So we have to draft and re-draft the question to get a more useful answer, and some of ways we can improve the question include the following:
- we should give as much detail and context as possible
- we could say what we are not looking for, as well as what we are looking for
- we could explain the need for the answer by specifying the type of person we are generating the answer for (e.g. "I am a lecturer teaching a first-year module on Accounting in a Business degree, and I want help with ...")
- we could suply a sample answer (or answers) of the type we are looking for
- we could ask the Generative AI to explain why it is producing the answer it is giving, by asking it to present the step-by-step chain of reasoning it used to get the current answer.
The 3 Ps Process
To help structure the process of of adding any of the above 5 suggestions, I've created a three-stage methodology as follows:
- 1. PROMPT: This is the initial question, we will try to be as specific as possible
- 2. PROBE: We will review the answer, and try to verify externally if it is correct
- 3. POLISH: We refine the answer, by adding more detail, by saying what we are not looking for, by indicating roles, by providing sample answers, and by asking the AI to show its reasonsing. Now return to step 2, and try to verify your answer.
I believe this process is the same as drafting and re-drafting a document until you are happy with the result (or sick of doing it!).
You've been huffed!
Thinking about drafts (and re-drafts) inevitably leads me to think about draughts (aka "checkers"), and there is a rule in that game called "huffing" whereby if you can make a jump, you have to do the jump, otherwise your opponent can remove the piece that could have made the jump. In some versions of the game this rule can be applied even if the alternative move you made was also a jump, it doesn't matter, the piece can be removed. This is analogous to a problem with Generative AI, sometimes you make all the right moves and you still lose out.
I suppose what I am really speaking about is "AI Hallucinations" whereby a Generative AI system generates a false, misleading, or nonsensical answer (that appears plausible) that is not based on real data or facts. So the problem is that sometimes Generative AI systems "make things up", and we really don't know all of the reasons why it does it. We do know some reasons, for example,
- It lacks any data about the topic, so it makes something up
- It incorrectly matches the style of the question with another question that it knows the answer to, and supplies that answer ("incorrect pattern matching")
- It becomes overconfident if it hasn't been corrected recently
- It doesn't have as much detail in the question as it needs
- RANDOM: Sometimes it just randomly generates a wrong answer, even if it knows the right answer.
A Splash of Randomness
I believe the last of those ("RANDOM") is the worst in some ways, because it is totally unpredictable, and I think it is something to do with the way that Generative AI systems learn.
A lot of these Generative AI systems learn by getting presented with a number of sample answers. So, for example, if I was training a system to learn to recognise images of cars, I might present the system with 25 images of cars, and ask it to take those images and generalise from those to a general understanding of what an image of a car could be. The danger is that we don't want the system just to memorize the contents of the 25 images and think that's all they need to know ("overfitting"), so one thing we do in these Generative AI systems is to randomly deactivate parts of the system when learning the images ("Dropout Rate").
This is a bit like the following scenario: imagine there are 10 students in a classroom, and you are teaching them 25 examples about a specific topic, during the class students are popping in and out of the class at random times for random durations, after the class is over the students all get together and discuss the class, and between them all they have seen the whole class. Now, if, for example, Student A missed the first five examples you gave, and Student B missed out on the last 5 examples, they might have generalised the lesson in slightly different ways, so how they reconcile these differences is very important to the generalisability of their overall learning. They might also have additional knowledge and interpretations that they had filling in for the gaps that the individual students missed, so this might be very useful also.
Overall, I think this is ones of the problems with these systems, the ways they generalise, inevitably leads to random confusion !!!