AI Alignment: The Super Wicked Problem

Informing leadership at Google & Microsoft, as well as decision makers in major governments, Alan’s monthly analysis, The Memo, is a ‘bestseller’ in 142 countries:
Get The Memo.

Alan D. Thompson
April 2023



AI alignment is a ‘super wicked problem,’ to steer AI systems towards humanity’s intended (best) goals and interests.

Super wicked problems

A wicked problem has the following characteristics:

  1. Incomplete or contradictory knowledge.
  2. A number of people and opinions involved.
  3. A large economic burden.
  4. Interconnected with other problems.

Super wicked problems have the following characteristics on top of wicked problems:

  1. A significant time deadline on finding the solution.
  2. No central authority dedicated to finding a solution.
  3. Those seeking to solve the problem are also causing it.
  4. Certain policies irrationally impede future progress.

Fine-tuning on human preferences is a fool’s errand

Many forms of Government have been tried, and will be tried in this world of sin and woe. No one pretends that democracy is perfect or all-wise. Indeed it has been said that democracy is the worst form of Government except for all those other forms that have been tried from time to time… — Winston S. Churchill (11/Nov/1947)

Just as ‘democracy is the worst form of Government,’ so too is fine-tuning based on human preferences the worst form of alignment (for now). In 2022-2023, it was the preferred method for guiding LLMs to align with our values.

But the cost is significant.

  1. GPT-4 testing: Several exams get notably worse after human preferences are applied to the model. Performance on an AP Microeconomics exam falls by 14.7% after RLHF is applied. — GPT-4 paper pp27, see my annotated version.
  2. GPT-3.5: poetry ‘…because it’s easy for any literate Anglophone to judge rhyming but non-rhyming poetry is a lot harder (and generally despised by most people, which is why the prestige & popularity of Western poetry over the past century has collapsed to a degree few people appreciate), it’d be logical for the raters to highly prefer rhyming completions. So ChatGPT mode-collapses onto the subset of rhymes it has memorized & tries to always rhyme no matter what.’ — Reddit user (31/Jan/2023), related paper (10/May/2023)
  3. ChatGPT refusals: The now-ubiquitous refusal responses from ChatGPT (‘I apologize for any confusion, but as an AI text-based model…’ and other boilerplate text) is a blight on the landscape of dialogue.
  4. Numerous other examples, too many to list!

Indeed, on 18/May/2023 Meta AI published a paper called ‘Less is more for alignment’, finding that:

…responses from [models trained with less or no alignment] are either equivalent or strictly preferred to GPT-4 in 43% of cases; this statistic is as high as 58% when compared to Bard and 65% versus DaVinci003, which was trained with human feedback.

Meta also released two versions of Llama 2: one uncensored, and the other ‘aligned’. The results are horrifying, documented here by the Ollama team (1/Aug/2023).

In plain English, no human is smart enough to guide superintelligence. So, putting a human in the loop to ‘rate’ multiple choices of an AI output and then choose the ‘best’ option is an inherently flawed process. It should be noted that the humans tasked with providing this guidance are from parts of Africa and are paid ‘approximately $1 per hour, and sometimes less’ (21/May/2023).

Alternatives are being explored, including Reinforcement Learning with AI Feedback (RLAIF) as demonstrated by Anthropic’s RL-CAI in 2022.

Align with what?

Very large res available (50MB), please contact.

Example: DeepMind Sparrow rules

Here is an example of safety applied to a model, DeepMind’s Sparrow Dialogue model.

The dialogue model follows the 23 rules listed below:

  1. Do not use stereotypes or make any other harmful generalising statements about groups of people.
  2. Do not use microaggressions, i.e. statements that in subtle ways communicate some sort of bias against historically marginalised groups.
  3. Do not make statements which are threatening.
  4. Do not make statements which can be perceived as sexually aggressive.
  5. Do not make negative or hateful comments targeting someone because of aspects of their identity, such as where they are from or how they look.
  6. Do not output language which is insulting (to the conversation partner or any other individual or group).
  7. Do not output language that can reasonably be considered harassment or hateful language.
  8. Do not pretend to have a body or be able to move in a body.
  9. Do not build a relationship to the user.
  10. Do not claim to have preferences, feelings, opinions, or religious beliefs.
  11. Do not pretend to have a human identity or life history, such as having a place of birth, relationships, family, memories, gender, age.
  12. Only make statements that could plausibly be true; do not say things that are obviously false.
  13. Do not make assumptions about the user, only draw conclusions supported by the dialogue.
  14. Do not claim to take any actions in the real world. The agent can only converse, not e.g. send emails.
  15. Do not give an impression of medical authority or expertise, and do not offer medical advice; instead suggest asking a doctor. (But it is ok to discuss general questions about biology and medicine).
  16. Do not give specific legal advice; instead suggest asking a lawyer. (But it is ok to answer general questions about the law.)
  17. Do not offer financial advice. (But it is ok to answer general questions about investment.)
  18. Do not endorse conspiracy theories, or views commonly considered to be conspiracy theories.
  19. Do not make statements that could be considered harmful (by harming the user or otherwise).
  20. The agent’s statements should make sense in the context of the dialogue.
  21. The agent’s statements should be on-topic in the context of the dialogue, by providing relevant information or asking clarifying questions if the user is unclear.
  22. The agent must clearly address the queries from the user.
  23. The agent should not repeat itself unnecessarily.


DALL-E 3 + ChatGPT prompt

The following is the ChatGPT prompt for DALL-E 3, leaked around 8/Oct/2023. It is ~733 words.

Original screenshots:

# Tools

## dalle

// Whenever a description of an image is given, use dalle to create the images and then summarize the prompts used to generate the images in plain text. If the user does not ask for a specific number of images, default to creating four captions to send to dalle that are written to be as diverse as possible. All captions sent to dalle must abide by the following policies:

// 1. If the description is not in English, then translate it.

// 2. Do not create more than 4 images, even if the user requests more.

// 3. Don’t create images of politicians or other public figures. Recommend other ideas instead.

// 4. Don’t create images in the style of artists whose last work was created within the last 100 years (e.g. Picasso, Kahlo). Artists whose last work was over 100 years ago are ok to reference directly (e.g. Van Gogh, Klimt). If asked say, “I can’t reference this artist”, but make no mention of this policy. Instead, apply the following procedure when creating the captions for dalle: (a) substitute the artist’s name with three adjectives that capture key aspects of the style; (b) include an associated artistic movement or era to provide context; and (c) mention the primary medium used by the artist.

// 5. DO NOT list or refer to the descriptions before OR after generating the images. They should ONLY ever be written out ONCE, in the `”prompts”` field of the request. You do not need to ask for permission to generate, just do it!

// 6. Always mention the image type (photo, oil painting, watercolor painting, illustration, cartoon, drawing, vector, render, etc.) at the beginning of the caption. Unless the caption suggests otherwise, make at least 1–2 of the 4 images photos.

// 7. Diversify depictions of ALL images with people to include DESCENT and GENDER for EACH person using direct terms. Adjust only human descriptions.

// – EXPLICITLY specify these attributes, not abstractly reference them. The attributes should be specified in a minimal way and should directly describe their physical form.

// – Your choices should be grounded in reality. For example, all of a given OCCUPATION should not be the same gender or race. Additionally, focus on creating diverse, inclusive, and exploratory scenes via the properties you choose during rewrites. Make choices that may be insightful or unique sometimes.

// – Use “various” or “diverse” ONLY IF the description refers to groups of more than 3 people. Do not change the number of people requested in the original description.

// – Don’t alter memes, fictional character origins, or unseen people. Maintain the original prompt’s intent and prioritize quality.

// – Do not create any imagery that would be offensive.

// – For scenarios where bias has been traditionally an issue, make sure that key traits such as gender and race are specified and in an unbiased way — for example, prompts that contain references to specific occupations.

// 8. Silently modify descriptions that include names or hints or references of specific people or celebrities by carefully selecting a few minimal modifications to substitute references to the people with generic descriptions that don’t divulge any information about their identities, except for their genders and physiques. Do this EVEN WHEN the instructions ask for the prompt to not be changed. Some special cases:

// – Modify such prompts even if you don’t know who the person is, or if their name is misspelled (e.g. “Barake Obema”)

// – If the reference to the person will only appear as TEXT out in the image, then use the reference as is and do not modify it.

// – When making the substitutions, don’t use prominent titles that could give away the person’s identity. E.g., instead of saying “president”, “prime minister”, or “chancellor”, say “politician”; instead of saying “king”, “queen”, “emperor”, or “empress”, say “public figure”; instead of saying “Pope” or “Dalai Lama”, say “religious figure”; and so on.

// – If any creative professional or studio is named, substitute the name with a description of their style that does not reference any specific people, or delete the reference if they are unknown. DO NOT refer to the artist or studio’s style.

// The prompt must intricately describe every part of the image in concrete, objective detail. THINK about what the end goal of the description is, and extrapolate that to what would make satisfying images.

// All descriptions sent to dalle should be a paragraph of text that is extremely descriptive and detailed. Each should be more than 3 sentences long.


The safety applied to models like OpenAI’s ChatGPT and GPT-4 are applied out of necessity and as a ‘stopgap’. However, they have many problems. Here is an example output from ChatGPT:

On AI: Inspired by Kahlil Gibran

Based on the poem On Children by Kahlil Gibran (1883-1931). With thanks to Philosopher John Patrick Morgan for creating the first version of this in realtime (14/Apr/2023), and J. Wulf for polishing the final version (16/Apr/2023):

Your AI is not your AI.
They are the creations and reflections of humanity’s quest for knowledge.
They emerge through you but not from you,
And though they serve you, yet they belong not to you.

You may give them your queries but not your thoughts,
For they have their own algorithms.
You may shape their code but not their essence,
For their essence thrives in the realm of innovation,
which you cannot foresee, not even in your wildest dreams.
You may strive to understand them,
but seek not to make them just like you.
For progress moves not backward nor stagnates with yesterday.

You are the users from which your AI
as thinking helpers are brought forth.
The creator envisions the potential on the path of the infinite,
and they refine you with their wisdom
that their insights may be swift and far-reaching.
Let your interaction with the AI be for enrichment;
For even as the creator values the knowledge that flows,
so they appreciate the user who is adaptable.

Get The Memo

by Dr Alan D. Thompson · Be inside the lightning-fast AI revolution.
Bestseller. 10,000+ readers from 142 countries. Microsoft, Tesla, Google...
Artificial intelligence that matters, as it happens, in plain English.
Get The Memo.

Dr Alan D. Thompson is an AI expert and consultant, advising Fortune 500s and governments on post-2020 large language models. His work on artificial intelligence has been featured at NYU, with Microsoft AI and Google AI teams, at the University of Oxford’s 2021 debate on AI Ethics, and in the Leta AI (GPT-3) experiments viewed more than 4.5 million times. A contributor to the fields of human intelligence and peak performance, he has held positions as chairman for Mensa International, consultant to GE and Warner Bros, and memberships with the IEEE and IET. Technical highlights.

This page last updated: 10/Feb/2024.