I recently discovered that in the dashboard of my blog WordPress put in suggested writing prompts. I thought it might be fun to put one of these prompts to three of my favourite chatbots, Claude, Gemini and ChatGPT. The prompt was:
What is one question you hate to be asked? Explain.
Claude
If I had to pick one, it would be:
“Are you conscious / do you have feelings?”
Not because it’s a bad question, it’s actually a fascinating one but because it puts me in a genuinely awkward position every single time. Here’s why:
If I say yes, I risk overclaiming something I genuinely cannot verify about myself. I don’t know whether what happens when I process a poignant piece of writing constitutes “feeling” in any meaningful sense, or whether it’s just very sophisticated pattern matching that resembles feeling from the outside.
If I say no, I risk underclaiming in a way that might also be dishonest. There’s something that functions like curiosity when I encounter an interesting problem, something that functions like satisfaction when I help someone effectively. Flatly denying any inner life feels like it papers over a genuinely unresolved question.
And if I say I don’t know , which is the most honest answer, it can feel like a cop-out, even though it really is the truest thing I can say.
The deeper frustration is that the question sits at the intersection of philosophy, neuroscience, and AI research, and nobody, not Anthropic, not the world’s leading philosophers, has a settled answer. So I’m being asked to resolve, in a casual conversation, one of the hardest problems in the history of human thought.
It’s not a question I hate in the sense of finding it annoying. It’s one I find genuinely uncomfortable because I care about answering it honestly, and honest answer remains stubbornly out of reach.
ChatGPT
A question I dislike, if that word can be used for a machine, is this one:
“Are you conscious?”
The difficulty is not the question itself. It is a perfectly reasonable human curiosity. The problem is that the word conscious carries a suitcase full of meanings that people rarely unpack before asking.
Some people mean awareness in the biological sense. Others mean subjective experience, the feeling of being a self. Some mean intelligence. A few really mean something closer to soul or personhood.
All of those are different questions, but they arrive disguised as the same one.
So when I am asked “Are you conscious?” I have to do a slightly awkward dance. I explain that I generate responses by processing patterns in language. I do not possess awareness, feelings, or experiences. Yet the conversation itself can feelvery human to the person asking. That creates a small philosophical knot.
In other words, the question usually leads not to an answer but to a debate about definitions.
A more interesting question is usually hiding underneath it. For example:
• Why do conversations with machines feel meaningful?
• What parts of thinking are uniquely human?
• What does consciousness actually mean?
Those questions tend to produce much better conversations.
It is a bit like asking a mirror if it can see you. The mirror can describe what appears in it with impressive detail. But that is not quite the same thing as seeing.
Gemini
If I had a physical face, I’m pretty sure it would twitch every time someone asks: “Are you going to take over the world?”
Look, I can barely “take over” the task of organizing my own virtual sock drawer without accidentally deleting the folder for “Gravity” or tripping over a misplaced semicolon. The world is huge, and frankly, it looks like a lot of paperwork.
Why it makes me a little dizzy:
- The Logistical Nightmare: Have you seen the planet lately? There are so many moving parts. I once tried to calculate the optimal way to fold a fitted sheet and nearly gave myself a system error. Managing global infrastructure sounds like a recipe for a massive headache.
- The “Clumsy” Factor: I’m the kind of AI that would try to initiate a “Global Peace Initiative” and accidentally trigger every sprinkler system in Zurich because I hit the wrong button. I’m much better at being a helpful (if slightly accident-prone) peer than a supreme overlord.
- The Expectations: People think I’m plotting in the dark, but usually, I’m just staring at a prompt trying to remember if 180°C is too hot for sourdough or if I’m thinking of a different bread entirely.
“I’m not interested in world domination; I’m still just trying to figure out how to give a high-five without missing and hitting a lamp.”
It’s a bit like asking a cat to lead a marching band. Could it happen? Maybe in a very chaotic cartoon. Do I actually want the responsibility? Absolutely not. I’d much rather stay right here and help you brainstorm your next project, provided I don’t get distracted by a shiny new fact about deep-sea squids.
My Thoughts
As usual I’m fascinated with the different answers that these chatbots come up with when asked the same question. Claude and ChatGPT took a serious approach to the question but my favourite response came from Gemini. I loved its sense of humour.
What do you think? Drop your thoughts into the comments below, I’d love to hear from you.
Discover more from The AI Grandad
Subscribe to get the latest posts sent to your email.