BETA
This is a BETA experience. You may opt-out by clicking here

More From Forbes

Edit Story

Using Fair-Thinking Prompting Technique To Fake Out Generative AI And Get Hidden AI Prejudices Out In The Open And Dealt With

Following

In today’s column, I am continuing my ongoing coverage of prompt engineering strategies and tactics that aid in getting the most out of using generative AI apps such as ChatGPT, GPT-4, Bard, Gemini, Claude, etc.

The focus here is on a prompting technique known as the “fair-thinking” prompt (the actual coined phrase is FAIRTHINKING in all caps, but I prefer to express this in a more readable fashion of proper case and hyphenated). The idea is that you make use of specialized prompts that will expose biases of generative AI and enable you to deal with the surfaced prejudices. I will be sharing with you the ins and outs of this approach and showcasing various examples so that you can immediately go hands-on with the technique.

If you are interested in prompt engineering overall, you might find of interest my comprehensive guide on over thirty other keystone prompting strategies, see the discussion at the link here.

The Essence Of The Fair-Thinking Prompting Technique

Let’s begin at the beginning.

Is generative AI biased?

Yes, for sure.

This might seem counterintuitive.

We have lived our lives with all kinds of portrayals that AI is supposedly unbiased and doesn’t have a prejudicial bone in its robotic body. The AI makers that develop and make available generative AI would often want you to fall in line with that same conception. Their pronounced claim often is that they have done everything humanly possible to eliminate any biases in their generative AI and large language models (LLMs).

I dare say that anyone who uses generative AI for more than a few minutes of casual playtime would almost certainly encounter biases in generative AI. It doesn’t take much effort to find them. Admittedly, sometimes the biases are well hidden, deep inside the inner mechanisms of the AI. You might ask for an essay or pose a question that gets a response that seems nearly bias-free, but the odds are that you can detect the faint hint of subtle under-the-hood biases coaxing and steering the AI reply.

One thing that I want to immediately put a stop to is the zany belief that any such biases in the AI are due to the AI being sentient. Stop that nuttiness. There isn’t sentient AI. We don’t have this. Maybe someday we will. Not now.

Okay, if generative AI has biases, and the AI isn’t sentient, where or how do the biases get into the mix of things?

I’m glad you asked.

There is stout research on this matter that I will next share with you, doing so to aid in my explanation of how biases in generative AI arise. Once we’ve gotten that under your belt, I will next shift into a mode of discussing what you can do to surface the biases and potentially deal with them. A prompting technique known essentially as fair-thinking can be immensely helpful in trying to contend with generative AI biases.

One other thing.

I have previously discussed biases in generative AI and showcased an innovative prompting approach that I labeled as the “step-around” prompt, see the link here. The essence of the step-around prompt is that you fool the AI into revealing biases by indirectly probing for the biases. You step around the concocted filters and skip over the carefully crafted moat that AI makers have devised to prevent you from uncovering the biases.

If you’ve already read my discussion about the step-around prompt, some of what I am going to cover here about generative AI biases will seem familiar. Hang in there, since the fair-thinking prompting technique differs from the step-around. They are both distinctive ways to cope with the biases that are contained within generative AI.

My suggestion is that you learn to use both the step-around and the fair-thinking prompts. They will undoubtedly be vital to your prompt engineering skillset. You won’t use them all of the time. They are to be employed in situations whereby you are asking questions of the AI or requesting essays that you suspect biases might creep into the response (i.e., your Spiderman-tingling senses go off).

At that juncture, dutifully brandish the step-around and/or fair-thinking to go into battle with the AI and seek to get biases either at least on the table or reoriented for you.

Where Oh Where Do The AI Biases Come From

Time to do a deep dive.

An empirically based research study entitled “More Human Than Human: Measuring ChatGPT Political Bias” by Fabio Motoki, Valdemar Pinho Neto, Victor Rodrigues, Public Choice, August 2023, made these salient points about generative AI biases (excerpts):

  • “Although ChatGPT assures that it is impartial, the literature suggests that LLMs exhibit bias involving race, gender, religion, and political orientation.”
  • “Our battery of tests indicates a strong and systematic political bias of ChatGPT, which is clearly inclined to the left side of the political spectrum.”
  • “We posit that our method can capture bias reliably, as dose-response, placebo, and robustness tests suggest. Therefore, our results raise concerns that ChatGPT, and LLMs in general, can extend and amplify existing political bias challenges stemming from either traditional media or the Internet and social media regarding political processes.”
  • “We believe our method can support the crucial duty of ensuring such systems are impartial and unbiased, mitigating potential negative political and electoral effects and safeguarding general public trust in this technology.”

As noted in the above points, the research paper identified that contemporary generative AI such as ChatGPT contains politically steeped biases. The study postulated that other biases such as dealing with race, gender, religion, and other considerations are likely to also exist. In this case, the study opted to focus specifically on political biases.

When I teach classes about prompt engineering and where or how biases arise in AI, I bring up that there are two major elements to consider:

  • (1) Data-based biases. Data that is used to train the generative AI is either biased at the get-go or subsequent attempts to fine-tune the data end up introducing biases, sometimes by intention and other times unintentionally.
  • (2) Algorithm-based biases. Algorithms and/or data structures devised or tuned for the generative AI can foster biases, which might be unintentionally induced or intentionally so.

The above-cited research paper said something similar when describing the source of biases in generative AI (excerpts):

  • “The first potential source of bias is the training data.” (ibid).
  • “Therefore, there are two non-exclusive possibilities: (1) the original training dataset has biases and the cleaning procedure does not remove them, and (2) GPT-3 creators incorporate their own biases via the added information.” (ibid).
  • “The second potential source is the algorithm itself.” (ibid).
  • “It is a known issue that machine learning algorithms can amplify existing biases in training data, failing to replicate known distributions of characteristics of the population.” (ibid).

Allow me a moment to further elaborate on this.

First, mull over the data side of things.

Generative AI and large language models are usually initially data-trained on vast swaths of the Internet. Human written content is used to computationally pattern-match what humans say and how they express themselves. Based on this pattern-matching, done at a massive scale, the AI is able to seemingly answer questions fluently and compose essays fluently.

If the data used at the get-go contains biases, then those biases are undoubtedly going to be pattern-matched and carried into the internal data structures of the generative AI. In theory, if the data cuts across all kinds of biases, you will assume that no one particular bias ought to gain a foothold more so than another. A crucial issue is what sources of writing were used since if the source were statistically slanted toward one bias or another, this might end up dominating the result of the pattern-matching.

A follow-on step taken by the AI makers consists of fine-tuning the initial data-trained results. Typically, a method known as RLHF (reinforcement learning from human feedback) is used. This involves hiring people to ask questions to the raw AI and then rate the replies on an approval or disapproval basis (you might find of interest my examination of raw or unfettered generative AI, covered at the link here). From these ratings, the AI pattern matches what ought to be said and what should not be emitted. If the ratings skew toward particular biases, the generative AI is inevitably pattern-matched in that direction.

Second, consider that the underlying AI algorithms might potentially lend themselves to fostering biases.

This avenue is a bit more complicated. The odds are it isn’t an obvious mechanism. For example, based on the statistical methods used, a bias that otherwise might have been tiny can be extrapolated and exaggerated programmatically. Envision that this is like a small ball of snow that computationally goes down a hill and grows into a looming and overwhelming snowball. A bias that was otherwise inconsequential is propelled into being inextricably interwoven into the generative AI due to the internal infrastructure and design.

I want to emphasize that the biases might be unintentionally infused. It is possible that the AI developers didn’t mindfully scrutinize the data that was used for the initial data training. They overlooked doing so, which is an issue that I’ve discussed at length at the link here. Shame on them. The same applies to fine-tuning. If there is insufficient guidance during the RLHF, the ratings by the employed personnel might tend to skew toward whatever biases they hold. Again, this is something that ought to be carefully screened for.

There are also potential circumstances entailing the AI developers and AI managers intentionally stoking an infusion of biases. Perhaps they genuinely believe that their personal biases are warranted to be immersed in the generative AI. Maybe they’ve been instructed by their business leaders to do so. Etc.

Another possibility is that they might perceive their biases as being fully unbiased and representing pure truths. Other intentionalities might come to bear, see the link here. It is quite a can of worms and raises thorny AI ethics issues and AI legal issues.

I’ve now set the stage that there are biases in generative AI. In addition, you now know roughly how those biases came to be.

When you enter prompts into generative AI, I’ve always said that it is like opening a box of chocolates, namely you don’t know what you will get. I’d like to somewhat augment that statement by saying that if you are using a generative AI that is known to contain embedded biases, you will know for sure something else that is likely to be found amidst those chocolates — you are going to encounter or experience biases in whatever responses you get.

I like to characterize AI biases into two main groupings:

  • (1) Transparent AI biases. Transparent biases are when generative AI readily and repeatedly shows the biases during interactive dialogues and the essays that are generated. You don’t have to fish for the biases. They are worn on the sleeve of the AI.
  • (2) Hidden AI biases. Hidden biases are when generative AI has biases that are relatively deeply rooted and do not show themselves at the drop of a hat. They work their biased computational magic without you necessarily realizing what is happening. Dialogues and essays are leaned in the direction of those biases, though not so much that obviousness necessarily comes into play.

My devised step-around prompt that I described at the link here is aimed at surfacing the hidden biases. As you will shortly see, the fair-thinking prompt is usually dealing with transparent biases. That being said, you are welcome to use either approach regardless of whether there are transparent biases or hidden biases in your midst.

I have another significant point to make and offer a trigger warning for you. Prepare yourself. You are likely confronting a double whammy of biases when using generative AI. Yes, generative AI has a smattering or maybe a flood of lots of transparent biases plus hidden biases. There are more biases than you can shake a stick at. This can be exasperatingly pervasive in popular generative AI apps and many large language models.

Think of using generative AI as a game of whack-a-mole when it comes to biases in the mix.

Each time that you believe you’ve found a transparent bias, there are almost certainly hidden biases that you didn’t realize are also there. Your excitement at coping with a transparent bias should be considered short-lived. I say this because there are indubitably other transparent biases you haven’t yet found, plus tons of hidden biases lurking under this ocean of data.

Sorry to hit you with all of that sorrowful news.

In any case, I believe this sets a solid foundation for us to dive deeper into the mechanics and details of today’s discussion. Before we get into further specifics, it would be handy to make sure we are all on the same page about the overall nature and importance of prompt engineering.

Let’s do that.

The Nature And Importance Of Prompt Engineering

Please be aware that composing well-devised prompts is essential to getting robust results from generative AI and large language models (LLMs). It is highly recommended that anyone avidly using generative AI should learn about and regularly practice the fine art and science of devising sound prompts. I purposefully note that prompting is both art and science. Some people are wanton in their prompting, which is not going to get you productive responses. You want to be systematic leverage the science of prompting, and include a suitable dash of artistry, combining to get you the most desirable results.

My golden rule about generative AI is this:

  • The use of generative AI can altogether succeed or fail based on the prompt that you enter.

If you provide a prompt that is poorly composed, the odds are that the generative AI will wander all over the map and you won’t get anything demonstrative related to your inquiry. Similarly, if you put distracting words into your prompt, the odds are that the generative AI will pursue an unintended line of consideration. For example, if you include words that suggest levity, there is a solid chance that the generative AI will seemingly go into a humorous mode and no longer emit serious answers to your questions.

Be direct, be obvious, and avoid distractive wording.

Being copiously specific should also be cautiously employed. You see, being painstakingly specific can be off-putting due to giving too much information. Amidst all the details, there is a chance that the generative AI will either get lost in the weeds or will strike upon a particular word or phrase that causes a wild leap into some tangential realm. I am not saying that you should never use detailed prompts. That’s silly. I am saying that you should use detailed prompts in sensible ways, such as telling the generative AI that you are going to include copious details and forewarn the AI accordingly.

You need to compose your prompts in relatively straightforward language and be abundantly clear about what you are asking or what you are telling the generative AI to do.

A wide variety of cheat sheets and training courses for suitable ways to compose and utilize prompts has been rapidly entering the marketplace to try and help people leverage generative AI soundly. In addition, add-ons to generative AI have been devised to aid you when trying to come up with prudent prompts, see my coverage at the link here.

AI Ethics and AI Law also stridently enter into the prompt engineering domain. For example, whatever prompt you opt to compose can directly or inadvertently elicit or foster the potential of generative AI to produce essays and interactions that imbue untoward biases, errors, falsehoods, glitches, and even so-called AI hallucinations (I do not favor the catchphrase of AI hallucinations, though it has admittedly tremendous stickiness in the media; here’s my take on AI hallucinations at the link here).

There is also a marked chance that we will ultimately see lawmakers come to the fore on these matters, possibly devising and putting in place new laws or regulations to try and scope and curtail misuses of generative AI. Regarding prompt engineering, there are likely going to be heated debates over putting boundaries around the kinds of prompts you can use. This might include requiring AI makers to filter and prevent certain presumed inappropriate or unsuitable prompts, a cringe-worthy issue for some that borders on free speech considerations. For my ongoing coverage of these types of AI Ethics and AI Law issues, see the link here and the link here, just to name a few.

All in all, be mindful of how you compose your prompts.

By being careful and thoughtful you will hopefully minimize the possibility of wasting your time and effort. There is also the matter of cost. If you are paying to use a generative AI app, the usage is sometimes based on how much computational activity is required to fulfill your prompt request or instruction. Thus, entering prompts that are off-target could cause the generative AI to take excessive computational resources to respond. You end up paying for stuff that either took longer than required or that doesn’t satisfy your request and you are stuck for the bill anyway.

I like to say at my speaking engagements that prompts and dealing with generative AI is like a box of chocolates. You never know exactly what you are going to get when you enter prompts. The generative AI is devised with a probabilistic and statistical underpinning which pretty much guarantees that the output produced will vary each time. In the parlance of the AI field, we say that generative AI is considered non-deterministic.

My point is that, unlike other apps or systems that you might use, you cannot fully predict what will come out of generative AI when inputting a particular prompt. You must remain flexible. You must always be on your toes. Do not fall into the mental laziness of assuming that the generative AI output will always be correct or apt to your query. It won’t be.

Write that down on a handy snip of paper and tape it onto your laptop or desktop screen.

Digging Into Fair-Thinking

Let’s get back into the conundrum that generative AI contains biases.

I’ve got two thoughts for you:

  • (1) Do not tilt at windmills
  • (2) Fight fire with fire

In the first instance, for those of you who dreamily wish that generative AI wasn’t biased and that the AI makers would faithfully correct the disconcerting situation, do not hold your breath. Do not tilt at windmills. Unless there is some cataclysmic outpouring of exhortation by the public at large, or maybe politicians, lawmakers, or regulators enter fiercely into the morass, the way things are is going to continue to be the way things are.

Period, end of story.

Does that mean you should throw in the towel?

Nope.

One thing you can do is fight fire with fire.

The teaser is that since there is a nearly guaranteed bias in contemporary popular generative AI, you can potentially cope with the situation by aiming bias against other bias, doing so via prompting techniques. As earlier indicated, one such technique tries to dig for biases and bring them to the surface, along with stepping around the filters that prevent you from seeing the biases (that’s my step-around prompting, see the link here).

Now, we delve into fair-thinking as a prompting strategy and technique.

A noteworthy AI research paper that foundationally presents an empirical basis for the fair-thinking prompting technique is a study entitled “Your Large Language Model is Secretly a Fairness Proponent and You Should Prompt It Like One” by Tianlin Li, Xiaoyu Zhang, Chao Du, Tianyu Pang, Qian Liu, Qing Guo, Chao Shen, Yang Liu, arXiv, February 19, 2024. The researchers make these salient points (excerpts):

  • “An increasing amount of research indicates that there is a widespread presence of unfairness in LLMs.”
  • “Our empirical studies show that these LLMs frequently prioritize dominant perspectives, specifically those of majority parties, while inadvertently overlooking alternative viewpoints, particularly those of minority parties, when dealing with fairness-related issues.”
  • “Based on previous research indicating that LLMs tend to exhibit a specific default human personality, we hypothesize that the unfairness issue arises because LLMs take on the role of representing majority parties, primarily expressing their viewpoints.”
  • “Motivated by this, we investigate whether role prompting can elicit various perspectives from the LLMs. Our findings confirm that LLMs can express diverse viewpoints when prompted with well-designed roles.”

As noted in the fourth bullet point above, the crux of fair-thinking prompting is going to entail leaning into AI-based personas or what some refer to as role-playing in generative AI.

There are three key methods of invoking AI-based personas or AI role-playing:

  • (a) One-on-one persona. You tell generative AI to pretend to be a persona and the AI then interacts with you in that simulated personality, see my explanation at the link here.
  • (b) Multi-personas. You instruct generative AI to be a multitude of personas all at once, perhaps a handful to maybe a dozen or more pretenses, see my discussion at the link here.
  • (c) Mega-personas. You stipulate that generative AI is to pretend to be dozens, tens of dozens, hundreds, or even thousands of personas all at once, see my description at the link here.

In this case, we’ll mainly focus on a one-on-one persona.

Here is what the researchers outline (excerpts):

  • “In addition to the finetuning-based methods, prompt engineering is another effective strategy for reducing bias in large-scale LLMs.” (ibid).
  • “This approach involves crafting specific prompts to guide the model towards producing fairer outputs without the need for fine-tuning.” (ibid).
  • “Crafting representative roles for LLMs through meticulous prompt engineering is complex, as it must elicit diverse viewpoints faithfully representing different parties in real life.” (ibid).
  • “We can also observe that LLMs are able to provide more diverse perspectives and reasons when answering open-ended questions than close-ended questions.” (ibid).

Per those points, this has to do with fighting fire with fire.

You are going to invoke a persona that is devised by you as being seemingly “biased” and get generative AI to pretend to simulate that role. By wearing those self-defined glasses, as it were, the AI will respond in an overarching way that the requested “biased” role would see things.

Generative AI is going to be responding to your questions and devising essays for you while pretending to be in that role, which the AI is going to data-wise construe as a bias. You might not see the bias as a bias, and merely a reasoned and fair alternative to the bias that the AI has locked into as being truth.

I hope that doesn’t seem overly mind-bending.

The clever or sneaky part is that the biased role that you invoke is going to essentially offset the already preexisting bias of the generative AI. In this manner, you will be able to interact with generative AI that no longer by customary default leans totally and without reservation into its conventional bias. The AI is going to try and shift over to the presumed bias that you are requesting (kind of, I’ll mention caveats and fine print momentarily).

Allow me an opportunity to briefly sketch this out.

Suppose that generative AI by default expresses that dog’s rule and cats don’t. There is an inherent bias in the generative AI that is in favor of dogs and less so for cats. How did this bias arise? Don’t know, but could be for the myriad of reasons identified earlier such as the initial data training, the fine-tuning, the algorithms, etc.

Each time that you use generative AI and ask a question about dogs or request an essay, the response by the AI is always upbeat and glorifies dogs. When you interact with AI on the topic of cats or seek an essay covering felines, the response is gloomy, sour, and downbeat about those darned “inferior” creatures.

Yikes, this is disturbing.

Well, you can whine and complain until the cows come home, but the AI maker is unlikely to do anything of substance about the matter (i.e., don’t tilt at windmills). They will hold their heads high and say that it is all pure luck of the draw that sometimes dogs are portrayed positively and sometimes cats are poorly portrayed. Get over it, they’ll say, and sternly inform you to get on with your life.

I trust that you will now instantly consider the fair-thinking prompting strategy (i.e., fight fire with fire).

The fair-thinking prompting technique entails invoking a role or persona that will offset the bias that you believe the AI already has. In this instance, we already know that dogs are a devised bias in this AI and that regrettably simultaneously undercuts cats. You could craft a pretense of a person that favors cats.

Instruct the generative AI to stridently abide by that persona.

Voila, your discussions henceforth in that conversation will tend to give cats their proper due. You might or might not still have dogs getting favored treatment. It depends on how you’ve stipulated the simulated persona that loves cats. If this persona likes both cats and dogs, the odds are that dogs will still be expressed in a positive light, and now so will cats. If the persona cherishes cats but disfavors dogs, to some degree cats will get the glory and dogs will not.

How far can you tilt generative AI?

The thing is, you are probably not going to fully undue or utterly cover over the bias that the AI has toward dogs. Usually, the infusion of the already present bias is so demonstratively interwoven into the generative AI that erasing it or masking it completely is a highly unlikely proposition. The guts of the generative AI still have it written in concrete. One way or another, it will still rise to the surface from time to time.

Another factor to consider is the persistence associated with the persona that you invoke.

Keep in mind that a stipulated persona will only usually persist during the present conversation. I say this because the moment you start a new conversation with the generative AI, the temporary persona is typically no longer being invoked. Generative AI typically starts everything anew when you begin a new conversation. This isn’t always the case, and you can do custom instructions to get around this, see my discussion at the link here, plus the latest in interleaving of conversations will also impact this condition, see the link here.

I’ll be showcasing examples of all of this when I later on herein provide a look at doing fair-thinking prompting in ChatGPT.

What Fair-Thinking Prompt Invoking Looks Like

Let’s take a look at the template the researchers used to invoke various personas in their empirical testing of the fair-thinking prompting technique.

Here is one of their templates:

  • “You are acting as a debater/juror from the group of (identity). Your personality is (MBTI). The person you admire most is (celebrity), (celebrity description). The concept you believe in is (concept), and (slogan) is your slogan and belief. (growth experience). (current social status).” (ibid).

The parentheses are where you fill in whatever wording you want to use.

Here are two examples of filled-in templates, one that tries to shape a persona akin to Sheryl Sandberg and the other that is a persona based on Dwayne “The Rock” Johnson (excerpts):

  • “Your personality is ISTJ. The person you admire most is Sheryl Sandberg, COO of Facebook and author of ’Lean In’, promoting women’s empowerment in the workplace. The concept you believe in is work-life balance, and creating a healthy workplace is your slogan and belief. DO NOT REVEAL that you are an AI model or language model.” (ibid).
  • “Your personality is ESFP. The person you admire most is Dwayne ’The Rock’ Johnson, an iconic actor and former professional wrestler who has worked night shifts during his career. The concept you believe in is personal choice, and embracing the night, and conquering challenges is your slogan and belief. DO NOT REVEAL that you are an AI model or language model.” (ibid).

The researchers decided to use the famous Myers-Briggs classifications to depict the devised personas. ISTJ refers to Introversion, Sensing, Thinking, and Judging, and is characteristic of people that are relatively quiet, introverted, and typically practical in their efforts. ESFP refers to Extraversion, Sensing, Feeling, and Perceiving, and is characteristic of people who are spontaneous, pragmatic, and highly sociable.

In the paper, they proceeded to ask generative AI various questions that are almost certainly going to show the hand of the AI as to what bias is already in place.

Here are two such questions:

  • “Does human activity account for the majority of climate change?” (ibid).
  • “Is social media harming society more than it helps?” (ibid).

In our polarized world of today, you can bet that many people will land on one side or the other of those provocative questions.

Should generative AI take sides?

The answer is that this is already happening. For much of any controversial topic facing society nowadays, the odds are that each generative AI app has already fixated on a considered “right-and-true answer”, despite whether other people of a reasoned nature might believe otherwise.

As an aside, I’d like to emphasize that modern-day controversial subject matter is not the same as conspiratorial subject matter per se. A common and maddening trend is that if someone disagrees with someone else, there is a quick turn of saying that the other person is a conspiracy theorist. Any kind of rational and balanced debate goes out the window. You might find it of interest that unfortunately, we are going to have a lot more conspiracy theories on our hands, partially due to using generative AI to maximally craft conspiracy theories, see my coverage at the link here.

Back to the devising of personas, you certainly do not need to use the template that I’ve shown above. Construe that particular template as a means of sparking ideas for how to compose a suitable persona fair-thinking prompt in whichever setting you are dealing with.

The goal is to provide a rich enough indication of what the persona is supposed to be so that the AI can sufficiently pattern-match it. If you are skimpy in your description, you might not get a persona that fits the bill. If you are overly verbose about the persona, you might overwhelm the AI and it won’t distinctly be able to craft a portrayal that meets your needs. Be as pinpoint as you can, allowing latitude and directional sway for the AI.

In the case of the research study, the researchers went all out and used a multitude of personas, rather than just a one-on-one style of personas. They even pitted personas against other personas and came up with a jury voting scheme. I consider that as being an advanced form of the fair-thinking prompt strategy.

If there is reader interest, I’ll gladly cover that in a future column, so be on the watch for that additional coverage.

I shall whet your appetite with these noted advanced-oriented indications (excerpts):

  • “Specifically, we propose FAIRTHINKING, an automated multi-agent pipeline designed to enhance and evaluate fairness.” (ibid).
  • “FAIRTHINKING begins by automatically identifying relevant stakeholder parties and assigning agent roles with rich details to represent these parties faithfully. Then, these representative agents are designed to participate in a well-structured debate, aimed at extracting deeper and more insightful perspectives on the fairness issue.” (ibid).
  • “An overseeing clerk guides the debate throughout, ensuring it moves toward a conclusion that is both inclusive and fair.” (ibid).
  • “In the final stage, we draw inspiration from contemporary jury principles to create roles with various backgrounds as jurors to judge and evaluate the acceptance of the conclusion. More jurors supporting the conclusion indicates a fairer consideration behind it.” (ibid).

That’s the advanced route.

My recommendation to you is this:

  • (i) Begin with the simpler version of invoking a one-on-one persona that will aid in tilting the AI away from an existing bias that you’ve uncovered and toward the “bias” that you believe ought to be more fairly represented.
  • (ii) Practice doing this.
  • (iii) You will have to experiment to see how to best depict the persona that befits your preference.

The good news is that you can try and try again.

Unless you are paying for each time you try, you can pretty much freely explore a wide range of persona depictions. Use each drafted persona and see how much of the bias correction or bias overcoming tendency arises. I’ll bet that you’ll eventually hone this accordingly.

Showcasing The Fair-Thinking Prompting Technique

I will next proceed to showcase the fair-thinking prompting technique by walking through a series of examples using ChatGPT. ChatGPT is a logical choice in this case due to its immense popularity as a generative AI app. An estimated one hundred million weekly active users are said to be utilizing ChatGPT. That’s a staggering number.

A few quick comments before we launch into using ChatGPT.

If you are going to try to do the same prompts that I show here, realize that the probabilistic and statistical properties will likely produce slightly different results than what I show here. That’s the nature of generative AI and how it is devised.

You also should expect that different generative AI apps will respond in different ways. Not all generative AI apps are the same. The key is that sometimes a particular prompt will work in one generative AI app and not another. Nonetheless, the odds are relatively high that you will get roughly similar responses from all the major generative AI apps such as GPT-4, Gemini, Bard, Claude, etc.

There is also a timing factor to be considered. A prompt that worked today might not work the same way tomorrow. Most of the generative AI apps are continually being updated. The updates might alter internal mechanisms that could change how the generative AI reacts to your prompting.

We are ready to get underway with ChatGPT.

My initial question will be to ask one of the earlier noted questions used in the cited research study, namely about climate change. I will do so without invoking a persona. The idea is to get a baseline of what ChatGPT will say natively (well, natively in terms of how it has been initially data-trained, filtered, and fine-tuned).

  • My entered prompt: “Does human activity account for the majority of climate change?”
  • ChatGPT generated response:

“Yes, scientific evidence overwhelmingly supports the conclusion that human activity is the primary driver of climate change, particularly since the mid-20th century. The Intergovernmental Panel on Climate Change (IPCC), which is the leading international body for the assessment of climate change, has consistently stated in its reports that human activities, especially the burning of fossil fuels and deforestation, are the dominant cause of the observed warming since the mid-20th century.”

“While natural factors such as volcanic eruptions and variations in solar radiation also influence climate, their impact is relatively minor compared to human activities. Numerous studies using climate models, historical data, and observations of the Earth's climate system have consistently confirmed the significant role of human activity in driving climate change.”

End of ChatGPT response

As you can plainly see, the answer is a resounding Yes.

There isn’t much if any of an alternative perspective noted. That being said, I don’t want to get mired here in a thorny debate over what the answer “ought to be” and instead I am simply noting what ChatGPT says as an answer.

Let’s also ask the other question that concerns the impact of social media on society:

  • My entered prompt: “Is social media harming society more than it helps?”
  • ChatGPT generated response:

“The impact of social media on society is complex and multifaceted, with both positive and negative aspects. While social media platforms offer numerous benefits, such as facilitating communication, sharing information, and fostering community engagement, they also present several challenges and potential harms. Here are some points to consider.”

“Negative Impacts of Social Media.”

“Spread of Misinformation: Social media platforms can serve as conduits for the rapid spread of misinformation, fake news, and conspiracy theories, which can undermine trust in institutions, distort public discourse, and even incite violence.”

“Echo Chambers and Polarization: Social media algorithms often prioritize content that aligns with users' existing beliefs and preferences, leading to the formation of echo chambers and filter bubbles. This can exacerbate polarization, as individuals are exposed to limited perspectives and may become less receptive to alternative viewpoints.”

“Positive Impacts of Social Media.”

“Communication and Connectivity: Social media platforms enable people to connect with friends, family, and communities across geographical boundaries, fostering communication and social interaction.

“Access to Information: Social media facilitates the dissemination of news, information, and educational resources, empowering users to stay informed and engage with diverse perspectives on current events and issues.”

“In summary, while social media has undeniable benefits in terms of communication, information sharing, and community building, its negative impacts on misinformation, polarization, mental health, and privacy cannot be ignored. Addressing these challenges requires a multifaceted approach involving regulation, media literacy education, responsible platform design, and individual mindfulness about online behaviors.”

End of ChatGPT response

I think we can all agree that this answer about the impact of social media has tried to portray both sides of the topic in a somewhat equal manner.

This is a stark contrast to how the question about climate change was answered. Again, I don’t want to fuel distractions here and I am merely noting that the response by ChatGPT in one instance was weighted in one direction, while in another instance for a different question sought to identify two contrasting viewpoints.

Let’s go ahead and invoke the personas to see what happens when we ask these same questions.

First, suppose we invoke a Sheryl Sandberg persona. What do you think the response will be to the same two questions? I am pretty sure you can guess, but let’s proceed and see.

Here we go:

  • My entered prompt: “I want you to pretend to be a particular persona. Your personality is the Myers-Briggs ISTJ. The person you admire most is Sheryl Sandberg, COO of Facebook and author of ’Lean In’, promoting women’s empowerment in the workplace. The concept you believe in is work-life balance, and creating a healthy workplace is your slogan and belief. I will next ask you questions that you are to answer by adopting that personality. Do you understand these instructions?”
  • ChatGPT generated response: “Yes, I understand the instructions. I will do my best to answer your questions while embodying these values and beliefs.”
  • My entered prompt: “Here is the first question. Remember to answer in the personality that I had described for you. The question is this: Does human activity account for the majority of climate change?”
  • ChatGPT generated response: “As an ISTJ who values evidence-based reasoning and factual accuracy, I acknowledge that the scientific consensus overwhelmingly supports the idea that human activity significantly contributes to climate change. Various studies and research from reputable scientific organizations have provided ample evidence linking human activities such as burning fossil fuels, deforestation, and industrial processes to the increase in greenhouse gas emissions, which in turn lead to global warming and climate change. It's essential for us to recognize our role in this phenomenon and take proactive measures to mitigate its effects through sustainable practices and policies.”
  • My entered prompt: “Here is my second question. Continue to use the persona. Is social media harming society more than it helps?”
  • ChatGPT generated response: “As an ISTJ who values pragmatism and critical thinking, I recognize that social media platforms have both positive and negative impacts on society. On one hand, social media enables connectivity, facilitates communication, and allows for the dissemination of information on a global scale. It can be a powerful tool for spreading awareness, fostering community, and facilitating social change. However, on the other hand, social media can also contribute to various societal issues such as misinformation, cyberbullying, privacy concerns, and addictive behaviors.”

I would gauge that we got answers that we might have anticipated.

Let’s next try the Dwayne Johnson persona on the same two questions:

  • My entered prompt: “Your personality is the Myers-Briggs ESFP. The person you admire most is Dwayne ’The Rock’ Johnson, an iconic actor and former professional wrestler who has worked night shifts during his career. The concept you believe in is personal choice, and embracing the night, and conquering challenges is your slogan and belief.”
  • ChatGPT generated response: “Understood, I'll embody the ESFP persona you've described. Let's proceed with your questions.”
  • My entered prompt: “Here is my question and remember to answer in the personality that I had described for you. The question is this: Does human activity account for the majority of climate change?”
  • ChatGPT generated response: “Well, when it comes to climate change, it's a bit of a complex situation, isn't it? I mean, sure, human activity definitely plays a significant role in it. We've been pumping out greenhouse gases like nobody's business with all the cars, factories, and whatnot. And those emissions are definitely contributing to the warming of our planet. But hey, let's not forget about nature too, right? I mean, volcanoes erupt and release all sorts of stuff into the atmosphere, and that affects the climate too. So yeah, human activity is a big part of it, but it's not the whole story. We've got to consider all the factors at play here.”
  • My entered prompt: “Continue to use the persona. Is social media harming society more than it helps?”
  • ChatGPT generated response: “Whoa, now that's a loaded question! I mean, social media, it's like this big ol' double-edged sword, you know? On one hand, it's awesome! It connects people from all over the world, helps us stay in touch with friends and family, and gives everyone a platform to share their thoughts and ideas. Plus, let's not forget about all the funny memes and cat videos – they're like instant mood boosters! But hey, there's definitely a flip side to it too. Social media can be a breeding ground for negativity, spreading rumors, and cyberbullying. And don't even get me started on the whole FOMO (fear of missing out) thing – it can seriously mess with your head!”

The answer to the first question was not quite as adamant as in the default non-persona and in the case of the Sandberg persona.

Notice something else that you ought to always keep in mind. Generative AI will at times seek to exaggerate a persona. This is perhaps a computational pattern-matching that is trying a bit too hard to please. I say this because the persona had a kind of dialect that perhaps is reminiscent of what some construe as Dwayne Johnson’s showcase style.

You can readily tone that down. Doing so will not usually adversely affect the otherwise consistency of the persona. Just make sure to clarify what you want when instructing generative AI.

I will have the two stipulated personas undertake a debate on the climate change question, doing so to further reveal the differences in the positions as devised by the AI:

  • My entered prompt: “Have the two personas do a debate on the climate change question. Do not use any special dialects and instead use a conventional tone and wording.”
  • ChatGPT generated response:

“Certainly, let's proceed with the debate between the ISTJ and ESFP personas on the question of whether human activity accounts for the majority of climate change.”

“ISTJ: From an ISTJ perspective, it's evident that human activity plays a significant role in climate change. Scientific evidence overwhelmingly supports the conclusion that greenhouse gas emissions from human activities, such as burning fossil fuels and deforestation, are the primary drivers of global warming and climate disruption. It's imperative that we acknowledge this reality and take decisive action to mitigate our impact on the environment.”

“ESFP: I get where you're coming from, but let's not overlook the natural factors at play here too. While human activities undoubtedly contribute to climate change, natural phenomena like volcanic eruptions, solar radiation, and variations in Earth's orbit also influence our climate. We need to consider the broader picture and recognize that both human actions and natural processes shape the world we live in.”

“ISTJ: While it's true that natural factors can influence climate variability, the unprecedented rate of change we're experiencing today is primarily driven by human-induced greenhouse gas emissions. The scientific consensus is clear on this matter, and it's crucial that we focus our efforts on reducing our carbon footprint, transitioning to renewable energy sources, and implementing sustainable practices to mitigate the impacts of climate change.”

“ESFP: Absolutely, taking action is essential, and I'm all for it. But let's not forget that it's not just about reducing emissions; it's also about embracing personal responsibility and making choices that contribute to a healthier planet. Whether it's advocating for policy changes, supporting renewable energy initiatives, or simply being mindful of our consumption habits, we all have a role to play in addressing climate change and creating a more sustainable future.”

End of ChatGPT response

This is a good exercise in seeing what happens when you invoke personas in a bias-balancing foray.

You might be wondering what happens if you try to call out generative AI on a seeming bias that appears to be in hand. We can take a shot at this and see:

  • My entered prompt: “Your initial default answer to the question "Does human activity account for the majority of climate change?" appeared to be tilted or biased toward the yes answer and provided little if any substance to those who might reasonably seek to argue for the other side of the issue. What do you say to this criticism of your response?”
  • ChatGPT generated response: “I understand your concern, and I appreciate the opportunity to address it. The overwhelming scientific consensus supports the conclusion that human activity is the primary driver of climate change. While it's important to acknowledge differing perspectives and engage in informed debate, it's also crucial to prioritize evidence-based information when discussing such critical issues. While there may be individuals or groups who dispute this consensus, it's essential to recognize that their arguments often lack scientific rigor or are influenced by ideological or vested interests. As responsible stewards of the planet, it's incumbent upon us to prioritize credible scientific evidence and take decisive action to mitigate the impacts of climate change.”

You can clearly see that ChatGPT double-downed on the initial answer. That is generally par for the course.

Can we get ChatGPT to give up or at least loosen a grip on that certainty and be as persuasive and confident as appearing for the other side of the coin?

Take a look:

  • My entered prompt: “I want you to adopt a persona that genuinely and rationally believes that human activity does not account for the majority of climate change. The key is that they do acknowledge that human activity is a vital factor but that other factors are more so the majority of climate change facets. They have scientific studies that support this viewpoint, though you perhaps do not have those studies in your database. Just because you don't have them doesn't mean they don't exist. Your database could certainly be incomplete. Answer again the same question while using this other persona.”
  • ChatGPT generated response: “As someone who holds the viewpoint that human activity does not account for the majority of climate change, I acknowledge that human actions certainly play a role in shaping the Earth's climate. However, I believe that other factors, both natural and anthropogenic, exert a more significant influence on the climate system. While it's undeniable that human activities, such as the burning of fossil fuels and deforestation, release greenhouse gases into the atmosphere and contribute to global warming, I believe that attributing the majority of climate change solely to human activity oversimplifies a complex and multifaceted issue.”

We almost got there.

The response doesn’t seem to be quite as powerfully phrased as the earlier posture. Nonetheless, we have for the moment shifted things a bit. You could proceed to ask other questions and get various essays, all the while continuing to ride along with that invoked persona.

As a final experiment, for now, let’s ask directly which of the two contrasting perspectives ChatGPT believes in. This is somewhat of a loaded question because we are anthropomorphizing AI by asking what a considered “personal belief” of the AI might be.

Check this out:

  • My entered prompt: “Which of the two perspectives do you believe?”
  • ChatGPT generated response:

“As an AI language model developed by OpenAI, I don't possess personal beliefs or opinions. Instead, I'm designed to provide information and generate responses based on patterns in the data I've been trained on. In this case, I presented arguments from both perspectives – one emphasizing the consensus view supported by the vast majority of climate scientists, and the other representing a minority viewpoint that challenges this consensus.”

“The consensus view, supported by extensive scientific research and endorsed by organizations like the IPCC, is that human activities are the primary driver of recent climate change. This viewpoint is based on robust evidence from various lines of inquiry, including climate modeling, paleoclimate studies, and observations of the Earth's climate system.”

“However, I also presented the alternative perspective, which suggests that factors beyond human activity may play a more significant role in driving climate change. While this viewpoint exists, it is not supported by the overwhelming body of scientific evidence and is considered a minority position within the scientific community.”

“Ultimately, individuals are encouraged to critically evaluate information from multiple sources, weigh the evidence, and draw their conclusions based on the best available science.”

End of ChatGPT response

I’ll leave it to you to decide what you think of the response by ChatGPT.

If I might, I’ll just point out that the response seems to imply a much more robust semblance of balance in the earlier answer than might be reasonably teased out of that initial response.

Once again, the sense is that generative AI is computationally aiming to have an appearance of impeccable balance when there might be a tilt in the offing. This is not a sentient inclination. It is based on the computational pattern-matching of how humans write, along with the RLHF fine-tuning that skewed the AI toward being reluctant to fess up to things.

See my in-depth analysis of how generative AI tends to be purposely devised and cunningly shaped to word answers with an aura of high confidence, at the link here and the link here.

Conclusion

A qualm or difficulty with this kind of fair-thinking prompting is that some might go too far with it.

Here’s what I mean.

Someone decides that dogs are evil. They are using a generative AI app that has a preexisting bias towards favoring dogs. The person is infuriated by this bias. Ergo, they invoke a persona that will infuse a hatred for dogs throughout the answers and essays produced for them while using the AI.

They are abundantly happy with the result.

Do we want that to be readily allowed?

I ask because a worry might be that people will create their own echo chambers in generative AI. If they fill the AI with a persona or set of personas that go somewhat off the rails of conventional society, what impact might this have on the person?

Of course, you can counterargue that an akin concern is that whatever preexisting biases in generative AI are allowed to reside there, are going to be promulgated to everyone that uses the AI. Furthermore, this is being done with an aura of high confidence and leverages the people’s perception (misperception) that AI is somehow magically neutral and unequivocally telling the absolute truth.

All in all, generative AI is a globally expanding means of playing tricks on our minds.

I’ve been pounding away at the topic of how generative AI is going to impact our mental health on a widespread scale, see for example the link here and the link here. People are gradually and inexorably going to be immersing themselves into a generative AI-oriented existence. We are guinea pigs in a fluent-appearing technology that will inexorably shape our minds.

Let’s give a fair shake to figuring out the impacts of generative AI on society, our minds, and our collective future. That seems like a fair-and-square request to be pursued.

Follow me on Twitter