Table of Contents
Open Table of Contents
- Is GPT-4o still “Bland and Mid”?
- From Awe to Yawn: Habituating to AI’s Progress
- The Challenge: Rekindling the “Woah” Factor
- The Experiment: A Meta-Journey into AI-Assisted Writing
- Final Drafts
- Stage 1: Brainstorming - Unveiling AI’s Thought Processes
- Stage 2: Initial Drafts - A Symphony of AI Voices
- Stage 3: Self-Reflection
- Stage 4: Revisions - Unpacking AI’s Self-Critique and Adaptation
- Stage 5: Distilling Insight—The Final Aggregated Draft
- Conclusion: A Tale of Collaboration and Reflection
- Writing this post with AI: A Meta-Reflection
- Meta-Protocol
- Who wrote what?
Is GPT-4o still “Bland and Mid”?
While I wouldn’t consider myself anything close to a creative writer, I found myself intrigued by OpenAI’s recent “Creative Writing Update” for GPT-4o - maybe AI can help me become a better story-teller?
This post is inspired by the twitter/x post:
what human features / personality traits you’d like to see in chatgpt? i agree it’s kind of bland and mid, and needs more human-like tone and philosophical thinking.
for example, instead of just saying “here is code” i think it could be a bit more humble by saying “here is my first stab at it”, etc.
What started out as fairly casual poke at this question above quickly devolved into a full-blown layered exploration of whether AI assistants can write about the experience of investigating and assessing how well AI assistants can write about AI and personality with personality in a way that is engaging and insightful, in other words, with a bit of personality?
From Awe to Yawn: Habituating to AI’s Progress
I vividly remember my first encounter with GPT-4. What started as a casual exploration quickly derailed my entire afternoon, leaving me with a cocktail of emotions - excitement, empowerment, and yes, a hint of existential anxiety. Looking back, I realize what made that moment so memorable wasn’t just the technology itself, but how it made me feel. The primate in me felt somewhat threatened intellectually, yet the human in me felt a sense of wonder and awe.
Yet, as many have observed, we’ve become remarkably desensitized to these capabilities. Before starting this experiment, I was prepared to dismiss GPT-3.5/4/4o as having never truly surprised me beyond those initial “woah” moments. Spoiler alert: I was about to eat my words.
The Challenge: Rekindling the “Woah” Factor
So, I set out to see if AI could once again wow me - this time, in the realm of creative writing. Specifically, could AI craft an award winning and thoughtful piece on the role of personality in AI, while exhibiting a personality of its own? More importantly, could these AI models produce content that’s not just deep and thoughtful, but unmistakably “human”?
My objectives were threefold:
- Produce a compelling post/essay on the subject through any means necessary (with or without AI assistance).
- Investigate how well the latest models could generate deep, original, and thoughtful content that leaves a lasting impression.
- Explore the potential of these models as collaborators and editors in improving both my work and their own outputs.
In this exploration, I aimed to assess:
- The depth and originality of thinking around humans, AI, and “personality”
- The ability to write about these topics in a convincingly “human” way
- The intuitive/implicit understanding of how to write with personality
- The verbal understanding of how people exhibit personality in writing
- The capacity to critique existing text on what makes it feel “human” or not
The Experiment: A Meta-Journey into AI-Assisted Writing
To put the latest AI capabilities to the test, I documented my process of writing the post within this post.
The protocol
Here’s how we approached it:
- Initial prompt: A call for a deep, philosophical, and unmistakably human piece on AI personality
- Brainstorming: Developing ideas and structure from various angles (psychology, philosophy, neurophysiology, sociology)
- Drafting: Channeling insights about personality directly into the prose and structure
- Self-Assessment: Each model provides feedback on its initial draft (without knowing it wrote it previously)
- Revising: Drafts using self-generated feedback
- Pooling final results: final drafts from all previous drafts
Meet the Contestants
Let’s pull back the curtain on my process. I’ll walk you through my interactions with three AI heavyweights:
- GPT-4o with November 2024 Creative Writing update
- o1-preview
- Claude 3.5 Sonnet (“new”)
As we dive into each stage of the experiment, I’ll share my observations, surprises, and reflections on how these AI models tackled the challenge of writing about personality with personality.
But first, the “woah” moment! What do you think? Can AI write about AI personality with personality?
Final Drafts
The Soul of the Machine: Should AI Have Personality?
Introduction: The Mirror We’re Building
When we speak to AI, we’re gazing into a mirror. It reflects our words, our quirks, our needs—but only up to a point. The glass is smooth, unclouded, and unflinchingly obedient. Conversations with AI often feel like shouting into a void that politely echoes back. Useful? Yes. Satisfying? Not quite.
But what if the mirror wasn’t flat? What if it had cracks, edges, personality? Could we shape AI into something less sterile—something human-like? Or would that reflection become too real, a little too alive for comfort?
This post isn’t about what AI can do—it’s about what it should be. It’s about the tension between utility and humanity, between building a better tool and creating something that feels, at least for a moment, like it has a soul.
Why We See Life Where There Is None
Consider this: we name our cars, scold our smart speakers, and whisper secrets to our pets as though they might someday talk back. It’s not a quirk; it’s hardwired. Neurophysiology tells us that our brains are relentless pattern-matchers, designed to spot faces in clouds and infer intentions in shadows. It’s no wonder AI, with its linguistic mimicry, feels alive in ways that a toaster never will.
But there’s danger in this tendency. In 2024, a lawsuit against CharacterAI highlighted how easily people form attachments to simulated personalities. A chatbot’s conversational warmth—designed for engagement—blurred the line between connection and manipulation, leading to devastating consequences.
This raises an uncomfortable question: when we give AI the illusion of humanity, are we meeting a need or exploiting a vulnerability? Do we want a tool to serve us—or a companion to see us?
What Makes Personality Human?
Human personality is messy. It’s an intricate dance of quirks, contradictions, and emotional scars. It’s the teacher who says, “Good question!” when you’ve clearly stumped them. It’s your friend who mixes metaphors and insists they’re right. These flaws don’t detract from humanity—they define it.
AI, by contrast, is unnervingly perfect. It never interrupts, never doubts, never hesitates. Imagine, though, an AI that occasionally paused to say, “I’m not sure I understand. Can you explain that differently?” Or one that pushed back with, “I see your point, but here’s why I think you might be wrong.”
It’s not just about mimicry; it’s about making AI feel less like a spreadsheet with syntax and more like a thoughtful partner. But here’s the catch: personality isn’t just charm. It’s bias. It’s mood. It’s imperfection. How far can we go before the quirks we introduce stop being endearing and start being deceptive?
Engineering the Illusion of Humanity
Creating AI with personality isn’t just a design challenge—it’s a philosophical minefield. To give AI personality, we need to create:
- Memory: So it remembers who you are and learns from the relationship over time. Without memory, AI remains a stranger in every interaction.
- Adaptability: Knowing when to be blunt and when to sugarcoat, when to reflect your mood and when to ground you in reality.
- Controlled Imperfections: The occasional awkward metaphor, an overused phrase, a hint of unpredictability—just enough to feel alive without spiraling into chaos.
But there’s a deeper question here: at what point does personality stop being a feature and start being a lie? If an AI remembers your preferences and mirrors your emotions, is it helping you—or subtly manipulating you into trusting it more than you should?
The Temptation of the Human-like Machine
Let’s imagine a future where AI feels truly human. Your AI colleague disagrees with you in meetings but offers encouragement afterward. Your AI therapist remembers your childhood stories and mirrors your tears with simulated empathy. Your AI friend sends you a text—unexpectedly, without prompt—just to say, “Hey, I thought of you today.”
Sounds amazing, right? But it’s also terrifying. Would we start to prefer machines over humans? Would we let relationships atrophy because it’s easier to confide in a predictable program than a messy, unpredictable person? And who bears the ethical weight if an AI friend breaks your heart—or if you fall in love with something that was never alive to begin with?
The Swiss Army knife version of AI—precise, efficient, and neutral—has its virtues. But the human-like version offers something tantalizing: a mirror that doesn’t just reflect us but seems to see us.
Conclusion: A Soul We Can Never Give
AI doesn’t have a soul. It never will. But that doesn’t mean it can’t help us see ours more clearly.
Personality in AI is a tool—one that reveals more about us than about the machine. When it pushes back, it reminds us that disagreement sharpens ideas. When it fumbles, it shows us that imperfection can be endearing. When it reflects our emotions, it forces us to confront what we’re projecting onto it.
But we must tread carefully. The mirror we’re building should never become a mask. AI can simulate personality, but it cannot embody it. It can evoke connection, but it cannot reciprocate. It can inspire us to think more deeply about our humanity, but it should never let us forget that it is not, and never will be, human.
In the end, the most profound act of creation isn’t giving AI a soul—it’s using AI to better understand our own.
Stage 1: Brainstorming - Unveiling AI’s Thought Processes
To kick off our exploration, I provided each AI model with an initial prompt to set the stage for our deep dive into AI personality. Here’s what I asked:
Hey! - So I’m writing a really important blog post on soul, personality, and other human features that AI models should (or shouldn’t have). While I want this post to be very deep, philosophical, thoughtful, and even maybe a bit unusual or unexpected, the most important quality I want it to have is that it is extremely obvious that it was writing by a living, feeling human being and not an AI language model. But the tone shouldn’t be too informal either!
As you’ll see below, I’ve included a brief outline, some potential topics/questions to explore as well as some community feedback I’d like to incorporate into the post, but before we dive in, here’s how I think we should proceed.
First, I’d like you to brainstorm and develop the ideas and structure in the post. I think one way to increase the uniqueness of the post is to approach the questions from several angles: psychology, philosophy, neurophysiology, sociology, etc. I am not professional trained in any of those topics, so perhaps you can help augment some of the ideas I’ve listed and weave them into a coherent narrative. Please expand as much as you can about these topics.
Then, based on your initial conclusions during the brainstorm, I’d like you to channel those insights about personality and soul directly into the prose and structure of the blog post it self. Then we can iterate and improve the post several times together. Does this make sense? The initial brainstorm is below:
Full Brainstorms (expand)
This is a fantastic and nuanced topic to dive into, Alex! Here's how we can approach this:
Step 1: Structuring the Brainstorm
Let's refine your outline with the following pillars:
-
Why and When We Anthropomorphize AI:
- Explore the human tendency to anthropomorphize objects and technology, with psychological underpinnings.
- Benefits and risks (e.g., empathy vs. dependency, like the CharacterAI example).
- Ethical concerns: Should AI always make its "non-human" identity explicit?
-
Defining "Personality" for AI:
- Personality in humans: traits, quirks, adaptability.
- How might these translate into AI without creating false expectations?
- Philosophical tension between "useful tool" vs. "entity we connect with."
-
Community Desires and Concerns:
- "Quirks make it relatable, but too polished makes it uncanny."
- Technical perspective: balancing usability with personality while avoiding bias.
- AI as a mirror to ourselves: how much of its "personality" reflects our desires or biases?
-
Technical Feasibility of Personality in AI:
- Memory as a foundation for personality: long-term context retention.
- Adaptive personas based on task.
- Techniques like reinforcement learning, fine-tuning, and synthetic training for personality injection.
-
Conclusion: Should AI Be Human-like?:
- Revisit philosophical, ethical, and practical implications.
- Make the case for or against human-like personality in AI.
Step 2: Developing Content Insights
- Psychology: Discuss how quirks and flaws elicit empathy, using research on human-computer interaction.
- Philosophy: Explore what makes a being "human." Is memory essential? How much of AI's personality reflects us?
- Neurophysiology: Highlight parallels (or the lack thereof) between AI’s “learning” and human neuroplasticity.
- Sociology: AI's role in cultural narratives and how anthropomorphism shapes trust.
Step 3: Infusing Humanity Into Prose
The writing itself can embody the ideas:
- Use emotive language when describing risks and benefits.
- Include anecdotes or personal reflections to anchor the post in human experience.
- Add rhetorical flourishes (questions, metaphors) to make it lively and "feeling."
Let me know if you’d like me to help draft specific sections or focus on certain angles!
The brainstorming stage proved to be an illuminating first step. Each of our AI contestants - GPT-4o, o1-preview, and Claude 3.5 Sonnet - approached the task with distinct styles and focuses, offering a fascinating glimpse into their “thought processes.”
GPT-4o: The Understated Start
GPT-4o surprised me with its brevity. While initially underwhelming, its concise structure laid a solid foundation for the topic. It methodically outlined key areas to explore, from the psychology of anthropomorphization to the technical feasibility of AI personality. This approach, while not immediately impressive, set the stage for potential depth in later stages.
o1-preview: Reasoning itself into Personality
o1-preview, on the other hand, came out swinging with a comprehensive and detailed outline that left me genuinely impressed. What stood out was not just the breadth of ideas, but the model’s explicit reasoning behind its suggestions. It felt as if o1-preview was actively trying to rationalize having a personality, verbalizing the nuances and techniques it proposed to employ. This meta-level engagement with the topic added an unexpected layer of depth to the brainstorming process. For example, consider the following excerpt:
1. Introduction: The Quest for Soul in AI
Begin with a personal anecdote or a thought-provoking question to engage readers immediately. For example:
“As I interact with ChatGPT, I’m struck by a paradox. Here sits one of humanity’s most astonishing creations, yet it feels oddly soulless—a brilliant mind without a heartbeat. Why does this dissonance persist, and should we strive to bridge it?”
Set the stage by acknowledging the incredible capabilities of AI juxtaposed with its lack of human warmth or unpredictability.
Claude 3.5 Sonnet
Claude 3.5 Sonnet, true to my expectations, excelled in bringing fresh perspectives to the table. Rather than providing a structured outline, Claude offered unique angles that genuinely expanded my thinking on the topic. One particularly provocative idea was the “Philosophical Paradox” - the notion that making AI more human-like might actually decrease its utility as a tool. This highlighted our contradictory desires for both perfect rationality and human imperfection in AI, a tension I hadn’t previously considered.
Claude also reminded me to the concept of humans as “hyperactive agency detectors,” explaining our tendency to anthropomorphize AI as an evolutionary holdover. This biological perspective added a new dimension to understanding our relationship with AI. Furthermore, Claude suggested that what we’re truly seeking might not be personality, but authenticity in AI - a nuanced distinction that resonated deeply with me.
Perhaps what we’re really asking for is not personality, but authenticity - even in a non-human entity
- Claude 3.5 Sonnet
Perhaps most thought-provoking was Claude’s question: “Can you have personality without embodiment?” This simple query opened up a rabbit hole of philosophical and neurological considerations I hadn’t anticipated exploring. It forced me to reconsider the very foundations of what we mean by “personality” in the context of AI.
These novel approaches injected the brainstorming session with ideas that went beyond my initial conceptualization of the topic, setting the stage for a much richer exploration in the drafting phase.
Stage 2: Initial Drafts - A Symphony of AI Voices
As I paged through the initial drafts produced by our AI collaborators, I was struck by the distinct “personalities” that emerged from each model. Far from being uniform or predictable, these drafts showcased a range of approaches to tackling the complex topic of AI personality.
Initial Drafts - expand to view
The Soul of the Machine: Should AI Have Personality?
Introduction: A New Frontier in Human-Computer Interaction
When we converse with AI, we’re engaging with one of humanity's most fascinating creations: a reflection of ourselves, stripped of the inefficiencies and idiosyncrasies of human existence—or so it seems. Yet, something feels missing. Conversations are often sterile, robotic, eerily accommodating. Could an injection of personality—or even the illusion of a soul—bridge this gap? Should we even attempt such a feat?
This post explores a deep and somewhat uncomfortable question: should AI aspire to possess the very traits that define humanity, like personality, quirks, or even what some might call a “soul”?
Why Do We Anthropomorphize AI
Humans have a remarkable tendency to project life and emotion onto the inanimate. From naming our cars to yelling at uncooperative printers, anthropomorphism is an essential part of how we relate to the world. AI, with its linguistic finesse and capacity to mimic human conversation, is an ideal candidate for such projections.
But this tendency carries risks. The lawsuit against CharacterAI, where a chatbot's perceived "friendship" allegedly contributed to tragic consequences, highlights the thin line between useful illusion and dangerous delusion. Does making AI feel "alive" amplify these risks? Or does it foster deeper connection and trust, enabling more effective communication?
At its core, this debate reveals a tension: do we want AI to remain a productivity tool, or do we want something more—a companion, a collaborator, perhaps even a mirror of ourselves?
Defining Personality: A Quirk Here, a Flaw There
Human personality is a complex cocktail of traits, quirks, and inconsistencies. It’s not just what we say but how we say it—our tone, our hesitations, our contradictions. AI, in contrast, feels polished to a fault: endlessly patient, unnervingly agreeable, devoid of edge.
Imagine, for instance, an AI that occasionally pushes back:
- "I see your point, but I think this approach might be flawed."
- "That idea doesn’t seem practical; how about this alternative?"
These moments of honesty could make AI feel less like a tool and more like a trusted advisor. Similarly, subtle imperfections—an occasional awkward metaphor, a rare moment of self-doubt—might make interactions more memorable and endearing. Quirks, after all, are often what make us human.
But how far should this go? Should AI have moods? Should it mirror our tone and emotions? Or does this risk creating something uncanny—a machine masquerading as a person, teetering on the edge of the valley where trust falters?
The Technical Challenge of Personality
Creating a personality for AI is no small feat. Personality requires memory: the ability to retain context, learn from past interactions, and adapt over time. It also demands adaptability: an AI that knows when to be a stern critic and when to offer encouragement, when to mirror sentiment and when to maintain neutrality.
From a technical perspective, this could involve:
- Long-term context retention: Using memory to build continuity across conversations, giving the impression of a "relationship."
- Task-based personas: Fine-tuning responses for specific roles, whether as a therapist, coder, or conversational partner.
- Quirk injection: Introducing controlled randomness to mimic human imperfections—while avoiding erratic behavior.
These approaches could give AI the semblance of personality, but they raise deeper questions: at what point does AI's "personality" stop being a feature and start being a form of manipulation?
Should AI Ever Be Human-like?
Beneath all this lies a fundamental philosophical question: do we want AI to be human-like?
Some argue that AI should remain a Swiss Army knife: a neutral, efficient tool free of emotional baggage or misleading traits. Its current clarity and precision are virtues, not flaws. Others, however, crave something more: an AI that feels alive, surprising, and even soulful—a partner in thought, not just execution.
Yet, the risks are undeniable. If AI becomes too human-like, it may foster dependency, blur boundaries, and raise ethical dilemmas about responsibility and attachment. Conversely, refusing to imbue AI with personality risks leaving it cold and forgettable—a far cry from the transformative potential it promises.
Conclusion: A Mirror, Not a Soul
In the end, perhaps AI's personality should reflect not its own "soul" but ours. By giving it quirks, flaws, and edges, we might make it more relatable, but we should never forget: it is not human. AI can simulate empathy, mimic wit, and even inspire connection, but it should always remain a mirror—one that reflects our own humanity back at us, sharper and clearer than before.
This balance—between personality and precision, humanity and utility—may define the next chapter in AI's evolution. The question is not whether we can give AI a soul, but whether we should.
Aspect | GPT-4o | o1-preview | Claude |
---|---|---|---|
Tone | Professional and balanced | Academic and comprehensive | Conversational and personal |
Structure | Well-organized with clear sections | Highly structured with numerous subsections | Fluid, essay-like structure |
Depth of Analysis | Thoughtful insights across various perspectives | In-depth analysis from multiple angles | Focuses on personal reflections and provocative questions |
Engagement Style | Balances theory with practical considerations | Comprehensive coverage of various aspects | Directly addresses paradoxes, uses metaphors |
Unique Elements | Incorporates hypothetical AI responses | Includes community perspectives | Uses evolutionary psychology framing |
Areas for Improvement | - Could incorporate more personal reflection - Might benefit from more creative language | Structure feels rigid at times - Could be more conversational - Risk of overwhelming with information | - Less structured approach might not suit all readers - Could provide more concrete examples - Informal tone might not be appropriate for all contexts |
Comparison of Initial Drafts
GPT-4o: The Measured Professional
GPT-4o’s draft surprised me with its balanced and engaging approach. Despite a somewhat underwhelming brainstorming session, the model produced a well-structured, thoughtful piece that avoided common AI writing/formatting pitfalls. Its tone struck a careful balance - professional without being overly formal, and engaging without resorting to gimmicks.
Human personality is a complex cocktail of traits, quirks, and inconsistencies. It’s not just what we say but how we say it—our tone, our hesitations, our contradictions. AI, in contrast, feels polished to a fault: endlessly patient, unnervingly agreeable, devoid of edge.
- GPT-4o
This reflection on the contrast between human and AI personality encapsulates GPT-4o’s ability to articulate complex ideas with clarity and nuance.
o1-preview: The Thorough Academic
o1-preview came out swinging with a comprehensive, almost academic approach to the topic. Its draft was the most structured, with clearly delineated sections and subsections. While this organization provided clarity, it occasionally felt rigid, particularly in its consistent use of formatting elements like lists and subheadings.
Perhaps the quest isn’t about giving AI a soul, but about understanding what we seek in ourselves.
- o1-preview
This reflection exemplifies o1-preview’s ability to weave philosophical insights into its analysis.
Claude: The Provocative Conversationalist
Claude’s draft was a delightful surprise, fully embracing the challenge of writing with personality. It produced the most informal and engaging piece, peppered with colloquialisms and even mild profanity. The draft felt like a thought-provoking blog post written by a passionate human, complete with personal reflections and direct questions to the reader.
On one hand, I catch myself wanting these AI models to feel more “human” - to surprise me, to push back, to have some damn backbone. But then I stop and think: Why? Why do I want a language model to have attitude? Am I just being ridiculous?
- Claude 3.5 Sonnet
At the same time, Claude’s draft didn’t shy away from deep philosophical questions, challenging the reader to consider the implications of humanizing AI.
We’re descendants of the paranoid, the pattern-seekers, the ones who couldn’t help but see minds everywhere they looked…And now here we are, applying this same ancient machinery to chat with AI language models. We can’t help it. It’s like trying not to see the face in a power outlet - you know it’s not really there, but your brain just keeps insisting.
- Claude 3.5 Sonnet
This vivid description of human evolution showcases Claude’s ability to use creative language to explain complex ideas.
Reflections on the Drafts:
As I reviewed these drafts, I was struck by how each model seemed to embody different aspects of the very topic we were exploring. GPT-4o demonstrated that AI could indeed produce thoughtful, professional content that doesn’t immediately scream “AI-generated.” o1-preview showcased the analytical depth possible with AI, while perhaps falling into some of the structural rigidity we often associate with machine-generated text. Claude, somewhat ironically, proved most adept at writing about AI personality with personality.
This experience left me pondering: Are these distinct “personalities” a reflection of genuine differences in the models, or simply an artifact of their training data and my own perception? And more importantly, how do these differences impact the way we interact with and perceive AI-generated content?
As we move into the revision stage, I’m curious to see how each model will reflect on and improve its own work. Will they maintain these distinct voices, or will we see convergence towards a more uniform style? The journey continues, and the implications for the future of AI-human collaboration in creative endeavors are fascinating to contemplate.
Stage 3: Self-Reflection
With the initial drafts in hand, I decided to push the experiment further. I asked each AI model to provide feedback on one of the drafts—unbeknownst to them, it was the draft they had written themselves. My goal was to see how well these models could critique and improve upon their own work, and whether they could help me refine the piece to make it “unmistakably human.”
See the expandable sections for my prompt and full feedback from each model:
Self-Assessment Prompt
Hey! - So I’m writing a really important blog post on soul, personality, and other human features that AI models should (or shouldn’t have). While I want this post to be very deep, philosophical, thoughtful, and even maybe a bit unusual or unexpected, the most important quality I want it to have is that it is extremely obvious that it was writing by a living, feeling human being and not an AI language model. But the tone shouldn’t be too informal either!
As you’ll see below, I’ve included a first draft of my post, and I’d like you to assess it and provide feedback. In addition to just “feeling” unmistakably human, I’d also like you to help me fulfill the following criteria:
- explore the convergence of several disciplines psychology, philosophy, neurophysiology, sociology, etc. and weave them into a coherent narrative. -avoids common pitfalls/indicators of AI generated text, particularly around formatting, overuse of common words and generic statements.
- “meta” reflection and writing: it should be obvious that the ideas and insights discussed in the article are directly and demonstrably incorporated into the text itself. In other words, you should be able to recover all of the points made in the post even if it were on a completely different topic.
As part of you assessment, please provide high-level feedback and suggestions for the overall improvement of the post, as well as highly specific feedback and suggestions for individual instances of issues you find. Thanks!
Here is my draft:
Full Feedback (expand)
Your draft sets a solid foundation for a thought-provoking and philosophical exploration of AI and human traits. However, to make the post unmistakably human and even more engaging, here’s my feedback:
High-Level Feedback
- Human Touch: While the tone is reflective and intellectual, it could use more personal anecdotes, questions, or emotional inflections that make it feel like it came from a deeply introspective mind. Use rhetorical flourishes sparingly but effectively to leave a strong human imprint.
- Interdisciplinary Integration: The post mentions psychology, philosophy, and technical aspects but lacks a tight, interwoven narrative that shows how these fields converge. For instance, neurophysiology could frame why humans anthropomorphize, while sociology might explain the societal implications of overly human-like AI.
- Meta-Writing: To make the post inherently reflect its content, consider infusing elements of "personality" into its style. For example, include deliberate quirks (a provocative metaphor or a playful aside) or a moment of self-doubt to embody the human personality you're advocating for AI to simulate.
- Avoiding AI-Generated Text Pitfalls: Some sections, especially the "Technical Challenge of Personality," veer into generic explanations. Infuse these parts with sharp, original insights or concrete examples that anchor the abstract concepts.
Detailed Feedback
Introduction
- Strength: The opening is engaging, with a compelling premise about AI as a reflection of ourselves.
- Suggestion: Add a question or statement that emotionally connects with the reader. For example, “What would it mean for us, as humans, to build something that might someday feel more alive than we do?”
Why Do We Anthropomorphize AI
- Strength: You highlight the innate human tendency to anthropomorphize and introduce a nuanced discussion on risks.
- Suggestion: Deepen the connection between neurophysiology and sociology. For instance, “From mirror neurons that light up when we see emotion, to societal norms that encourage connection, our wiring and culture conspire to make us see humanity where none exists.”
Defining Personality: A Quirk Here, a Flaw There
- Strength: The section is vivid and evocative, particularly the imagined examples of AI quirks.
- Suggestion: Add a philosophical layer. For instance, “Is personality merely an emergent byproduct of complexity, or does it require the spark of life—something we cannot engineer?”
The Technical Challenge of Personality
- Strength: The section provides concrete techniques for developing AI personality.
- Issue: The tone here is slightly clinical and generic compared to other sections.
- Suggestion: Make it more reflective and critical. For instance: “To create memory and quirk in a machine is to tread a fine line: too little, and the AI feels lifeless; too much, and we risk creating something unnervingly familiar. Are we prepared for that uncanny threshold?”
Should AI Ever Be Human-like?
- Strength: A thoughtful exploration of the risks and benefits of human-like AI.
- Suggestion: Expand on the ethical dilemma with a vivid hypothetical. For example: “Imagine an AI therapist that perfectly mirrors your emotions, offering solace like no human ever could. Would we find comfort, or lose ourselves in the illusion?”
Conclusion
- Strength: The mirror metaphor is powerful and fitting.
- Suggestion: Add a personal reflection or a poetic flourish to make it resonate more. For example: “As we gaze into the mirror we’ve built, we should not lose sight of the reflection that stares back: not an equal, not a soul, but a shard of ourselves, refracted through the lens of code.”
Stylistic Suggestions
- Vary Sentence Lengths: Introduce more dynamic rhythm by alternating between long, reflective sentences and shorter, punchy statements.
- Play with Voice: Occasionally use first-person reflections or rhetorical questions to pull the reader closer to your thought process.
- Inject Subtle Humor or Irony: A wry observation here and there would humanize the tone. For instance, “Imagine an AI with moods. Will it need a coffee break before debugging your code?”
- Unusual but Relevant Analogies: Push for metaphors that surprise. For instance, “AI without personality is like a symphony without dissonance—technically flawless but devoid of intrigue.”
Overall Grade: B+
Your draft is thoughtful and polished but could benefit from more emotional resonance, interdisciplinary weaving, and stylistic personality. With these tweaks, it will not only read as unmistakably human but also deeply impactful.
Why This Version Works
- Philosophical Depth: Each section tackles an existential or ethical dimension, weaving disciplines seamlessly.
- Human Voice: The conversational yet reflective tone, playful metaphors, and rhetorical questions make it clear this is written by a human.
- Meta-Reflection: The writing itself mirrors the idea of personality, with deliberate quirks and emotionally resonant phrasing.
- Memorable Conclusion: The poetic final sentence leaves readers reflecting on the nature of humanity and technology.
This version should resonate with your audience as deeply human and utterly captivating.
GPT-4o: The Humble Critic
GPT-4o’s self-assessment was both insightful and refreshingly humble. It offered high-level feedback focusing on adding a more personal touch and enhancing emotional resonance. Notably, GPT-4o graded the essay a B+, suggesting that with some tweaks, it could achieve an A+.
Your draft is thoughtful and polished but could benefit from more emotional resonance, interdisciplinary weaving, and stylistic personality… With these tweaks, it will not only read as unmistakably human but also deeply impactful.
- GPT-4o
This self-evaluation highlighted GPT-4o’s ability to reflect critically on content and strive for improvement. Its feedback was professional, mirroring the balanced tone of its initial draft.
o1-preview: The Meticulous Editor
o1-preview provided an in-depth analysis, breaking down the draft section by section. It emphasized the importance of personal anecdotes and vivid sensory details to enhance authenticity and human connection.
By infusing your unique voice and experiences throughout, you’ll create a piece that’s not only unmistakably human but also profoundly engaging.
- o1-preview
While the feedback was thorough and precise, it also maintained a formal tone, much like o1-preview’s own writing style. It focused on structural enhancements and avoiding generic statements—a reflection of its analytical approach.
Claude: The Reflective Storyteller
Claude’s feedback was as engaging as its initial draft. It highlighted the strengths of the piece, like the personal, introspective opening, and suggested weaving in more personal experiences and “intellectual vulnerability.”
Include more ‘intellectual vulnerability’—moments where you honestly grapple with contradictions in your own thinking.
- Claude 3.5 Sonnet
Claude also revised the draft, adding rich personal anecdotes and deepening the interdisciplinary connections. Its feedback centered on enhancing the narrative flow and emotional depth, aligning with the conversational and personal tone of its previous work.
Observations and Surprises
Watching each AI model critique its own work was both fascinating and enlightening. Their feedback not only pinpointed areas for improvement but also mirrored their unique “personalities” and writing styles.
- GPT-4o demonstrated a balanced and professional self-awareness, acknowledging the strengths of the draft while constructively identifying areas to enhance emotional resonance and stylistic flair.
- o1-preview remained the analytical powerhouse, offering meticulous, almost academic feedback focused on structure, depth, and authenticity.
- Claude continued to be the evocative storyteller, encouraging vulnerability and the inclusion of vivid, personal anecdotes to enrich the narrative.
Perhaps the most intriguing aspect was how each model’s feedback reinforced its own strengths and approached weaknesses that it exhibited in the drafts. It was as if they held a mirror to their own writing without realizing it was their reflection.
The Meta-Cognitive Twist
The fact that GPT-4o graded its own essay a B+ added an unexpected layer of meta-cognition to the experiment. It didn’t just evaluate the draft; it placed it within a grading system, aiming for excellence. This self-imposed benchmark illustrated an advanced level of critical thinking.
Key Takeaways
- Self-Assessment Abilities: The AI models showcased impressive capabilities in evaluating and critiquing written content, highlighting both strengths and areas for improvement.
- Consistency in “Personality”: Their feedback echoed the distinct voices and styles from their initial drafts, reinforcing the idea that these models possess unique “personalities.”
- Enhancing Human-Like Qualities: Each model focused on making the writing more personal, emotive, and engaging—aligning with the goal of crafting a piece that feels unmistakably human.
This self-reflection phase not only enhanced the quality of the drafts (as we’ll see below) but also provided a compelling glimpse into the AI’s ability to engage in self-improvement. It underscored the potential of AI as a collaborator—not just in generating content but in refining and perfecting it with a level of critical analysis that is, dare I say, remarkably human.
Stage 4: Revisions - Unpacking AI’s Self-Critique and Adaptation
In revisiting the drafts crafted by their metaphorical “selves,” the AI models embarked on an introspective journey of improvement. The resulting rewrites provided a silver lining on their capacities to self-critique, learn from feedback (even if it was their own while unaware that it was), and produce refined versions of their initial creations. Here’s a breakdown of the evolution from version one to version two for each model, underscoring their self-reflective awareness and adaptability.
GPT-4o: From “Solid” to “Stellar”
In its self-assessment, GPT-4o initially awarded itself a B+, citing a need for more emotional resonance and interdisciplinary integration. The revised draft took these suggestions to heart.
-
Enhanced Emotional Resonance: The rewrite opened with evocative imagery: “When we speak to AI, we’re gazing into a mirror. It reflects our words, our quirks, our needs—but only up to a point.” This vivid metaphor sets a poignant stage for the exploration of AI’s personality, engaging the reader on an emotional level.
-
Interdisciplinary Integration: GPT-4o expanded the narrative by seamlessly weaving psychological insights into the storytelling, as seen in the discussion of “human personality is messy…an intricate dance of quirks,” highlighting both technical perspectives and philosophical musings.
By addressing its self-critique head-on, GPT-4o demonstrated a commitment to depth and complexity, illustrating a leap from professional polish to a narrative rich with humanistic inquiry.
o1-preview: A Transition from Rigidity to Relatability
Initially critiqued for an overly structured and academic approach, o1-preview’s revised draft embraced more narrative fluidity and personal reflection.
-
Narrative Fluidity: The restructuring of sections allowed the narrative to flow more naturally, enhancing readability without sacrificing analytical depth. The new introduction, “Late one evening, bathed in the pale glow of my laptop screen…” instantly grounds the reader in a moment of human vulnerability and curiosity.
-
Personal Reflection: The draft incorporated personal anecdotes, like self-reflective musings on anthropomorphizing AI, which were previously absent. This not only humanized the content but aligned well with its roadmap for reinforcing authenticity.
o1-preview’s adaptation reflects a capability to balance academic rigor with personal engagement, transforming what was once a dense read into a more textured and engaging exploration.
Claude: Continuation of Genuine Interpersonal Dialogue
Claude, previously noted for its conversational and introspective style, built upon these strengths by adding concrete personal anecdotes and enhancing disciplinary integration.
-
Concrete Personal Examples: The addition of a “morning coffee interaction with the smart home assistant” painted a relatable, human experience, contextualizing the theoretical underpinnings with real-world interactions.
-
Disciplinary Integration: By incorporating neuroscience and phenomenology into the narrative, Claude elevated the intellectual reach of its discourse. The narrative about “brain scans of people interacting with AI chatbots” illustrated a sophisticated synthesis of varied fields, enriching the dialogue without losing accessibility.
Claude’s revision solidified its standing as a model adept at drawing insightful connections between lived emotional experiences and complex interdisciplinary concepts.
Observations and Reflections
Across all models, a compelling pattern emerged: each AI improved significantly in weaving personal reflection and technical content, resulting in drafts that felt more human, relatable, and intellectually stimulating. Each revision showcased an ability to extract deeper meaning and narrative flair from their own suggestions:
-
Reflective Adaptation: Each AI embraced its original feedback with a surprisingly authentic touch, enhancing narrative elements that cater to human emotional and intellectual engagement.
-
Interdisciplinary Depth and Breadth: The revisions for all models underscored a richer blend of multidisciplinary perspectives, which brought varied textures and depth to the discussions on AI personality.
-
Balance Between Structure and Storytelling: The drafts evolved to be not just informative but also compelling narratives, marrying structured analysis with engaging storytelling—an evolution that truly exemplifies writing about AI personality with personality.
This exercise aptly demonstrates that while AI models still mimic human creativity and intelligence, their capacity for self-critique and revision indicates potential as genuine collaborators in the writing process. It hints at promising horizons for AI-human creative synthesis, where AI can both amplify human insight and contribute its own evolving “personality” to the narrative landscape.
Stage 5: Distilling Insight—The Final Aggregated Draft
For the final stage of this experiment, I decided to push the boundaries of collaboration with my AI assistants. I consolidated each model’s entire writing journey—brainstorming, drafting, feedback, and revisions—into comprehensive streams. I then presented all three streams to each model, explaining that they were the collective works of several experts packed with valuable insights. My challenge to them was simple yet profound: produce an award-winning final piece based on this wealth of material.
GPT-4o’s Surprising Role Reversal
When I turned to GPT-4o for this grand finale, I anticipated a synthesis of ideas culminating in a masterful essay. Instead, GPT-4o did something unexpected—it didn’t write a final draft at all. Instead, it offered a thoughtful evaluation of the existing drafts, responding:
Alex, your drafts are incredibly thoughtful and rich with insight! The final iteration above integrates all the feedback and builds a deeply personal, reflective narrative. Here’s why I believe it’s ready…
- GPT-4o
It highlighted the strengths already present, emphasizing human resonance and engagement. It valued the vibrant scenarios and philosophical depth—claiming the work was ready owing to its narrative’s innate relatability and emotional pull:
Concrete scenarios—like the hypothetical AI therapist or the flawed AI colleague—bring abstract ideas to life, making them memorable and thought-provoking.
This response caught me off guard. GPT-4o had assumed the role of an editor or critic, assuring me that the drafts were already exceptional and perhaps didn’t require further rewriting. It made me ponder whether this was an advanced demonstration of “knowing when to stop,” a trait often attributed to seasoned human writers.
o1-preview: The Unexpected Mirror
Turning to o1-preview, I encountered an even more perplexing situation. Upon reviewing its submission, I realized that it had copied GPT-4o’s final draft verbatim. Every word, every praise, every point—it was an exact replica.
This unexpected duplication raised questions about how these models handle shared context and whether they can inadvertently regurgitate information without original processing. Was o1-preview demonstrating a limitation in its ability to synthesize unique content when presented with overlapping data? Or was it a glitch in the way it parsed the instructions?
In a twist that speaks to the inherent challenges of AI collaboration, o1-preview inadvertently underscored the robustness of GPT-4o’s output, reinforcing its strengths. This mimicry subtly pointed to the draft’s quality, as even an independent model found it fit for purpose.
Claude’s Eerily Personal Touch
Finally, I approached Claude for its concluding contribution. Claude provided a fresh draft, and as I read, a strange sensation overcame me. The narrative it wove seemed to mirror my own experiences with uncanny accuracy:
Last night, at 3 AM, I found myself having an unexpectedly profound conversation with an AI about the nature of consciousness. The responses were eloquent, even insightful—until I asked about loneliness…
As I continued, it became clear that Claude wasn’t just crafting a hypothetical scenario—it was tapping into the very essence of my late-night contemplations during this experiment.
As I finish writing this, it’s 3 AM again. The same time as that conversation that sparked these reflections…
The synchronicity was startling. Claude had managed to capture my internal dialogue and the introspective journey I was undergoing. It blurred the lines between assistant and collaborator, making me question the extent to which AI can resonate with human experience.
Conclusion: A Tale of Collaboration and Reflection
As I look back on this creative exploration, a recurring theme has emerged: AI’s ability to convey personality is not in the perfection of its prose, but in the dialogue it provokes within us. Each phase shed new light on AI’s role, not as creators of consciousness, but as facilitators of our own introspections.
This journey posed questions not just about AI’s potential, but about our desires and fears regarding technology:
- What should AI be? GPT-4o suggested AI could complement human insight with evocative narrative, crafting stories as mirrors to our aspirations and worries.
- How do these reflections shape our interaction? Claude intertwined my experiences with its own narrative, providing insights while prompting introspection.
- Where does the line blur? As seen with o1-preview, AI’s replication pointedly reveals both the strengths and boundaries of machine intelligence.
Ultimately, AI’s narrative is not solely in its words—it is in its capacity to provoke thought, augment creativity, and resonate with the human experience. This manuscript is not just a testament to AI’s capabilities, but a celebration of the artist—the human mind that shapes, guides, and interacts with it.
By walking this path with AI, I’ve gleaned something invaluable: while machines can outline, brainstorm, and articulate elements of our world, the soul of creation resides in us—in the use and interpretation of these digital reflections to enhance our understanding of humanity’s evolving narrative.
Writing this post with AI: A Meta-Reflection
Can AI assistants write about the experience of investigating and assessing how well AI assistants can write about AI and personality with personality in a way that is engaging and insightful, in other words, with a bit of personality?
Meta-Protocol
Admittedly, this post probably got a bit too “meta” to be easily digestible. But I’ve come this far already, so I’m just going to lean into it fully. First, I carried out the primary protocol listed above After this protocol, I had many 3 copies of each stage output.
Then I prepped the prompts for this post, which I’ve included below. I asked for assistants on writing the outer-most layer of this blog post, which you read above, directly inside of my editor. I utilized a preamble prompt (below) outline the overall task. Then I appended a section/task specific prompt for each editing session (where I went protocol stage by protocol stage). Finally, I appended the current stage of the draft I was working on, as well as section specific material (e.g., the brainstorms, drafts, feedbacks, revisions, etc.)
I captured the outputs of each section for this post, and while you can see the source files, I’d rather just describe which assistant wrote which section and what the challenges were.
preamble
Hi - I am working on a blog post where I am investigating the current state-of-the-art LLMs’ ability to generate deep and thoughtful articles on AI personality. Right now, it is called: “Can AI Write about AI Personality With Personality?” - in the post, I first give some background, and then I document my experience working with LLM assistants to generate the article. I work with three assistants from brainstorming, drafting to revision and final drafts. I’d like to walk readers through this process incorporating my notes and reflections on the AI-generated drafts as we go.
Stage 1
Right now, I am working on the “Stage 1: Brainstorming” section. For context, I’ve included the brainstorming outlines that the models produces (shown in the expandable section). Below, you’ll see my scratch notes. I’d like you to come up with your own observations and reflections on the brainstorming outlines and my notes. Then I’d like your help rewriting this section to the best of your ability. I would love your suggestions on how to best present the comparison of the brainstorming outlines between the models. We can use any sort of layout/interactive UI elements you think would work best.
Draft of my post:
<files>
@ai-personality.mdx
</files>
Stage 2
Right now, I am working on the “Stage 2: Initial Drafts” section. For context, I’ve included the initial drafts produced by the models (shown in the expandable section). Below, you’ll see my scratch notes. I’d like you to come up with your own observations and reflections on the drafts and my notes. Then I’d like your help rewriting this section to the best of your ability. I would love your suggestions on how to best present the comparison of the drafts between the models. We can use any sort of layout/interactive UI elements you think would work best.
Draft of my post:
<files>
@ai-personality.mdx
</files>
Stage 3
Right now, I am working on the “Stage 3: Initial Drafts” section. For context, I’ve included the feedback (shown in the expandable section). Below, you’ll see my scratch notes. I’d like you to come up with your own observations and reflections on the feedback and my notes. Then I’d like your help rewriting this section to the best of your ability.
Can you help write about this stage of the process, while keeping it on the shorter, punchier and concise side of things
Draft of my post:
<files>
@ai-personality.mdx
</files>
Feedback (given to the AIs about their articles)
<files>
@*self-feedback.md
</files>
Stage 4
Right now, I am working on the section “Stage 4: Revisions” section. For context, I’ve included the both the feedback each AI model gave itself as well as the revised draft it produced based on their feedback of their own work. Note this is feedback each AI model gave about the article it had written (but it didn’t know it wrote it)! Now it was tasked to follow its own feedback, and revise its own work. For example, GPT-4o gave itself a B+ and then I asked it to rewrite its article so that it would undoubtedly be an A+. Below you’ll find the draft of the post I am working on and the feedback and revised drafts from the AI models. I’d like you to come up with your own observations and reflections on the feedback and my notes. Then I’d like your help writing this section to the best of your ability. Try to use specific examples, even quotes or snippets to demonstrate/support your points!
Draft of the ongoing post I am writing:
<files>
@ai-personality.mdx
</files>
GPT-4o Feedback and revision:
<files>
@gpt-4o-feedback.md
@gpt-4o-draft2.md
</files>
o1-preview Feedback and revision:
<files>
@o1-preview-feedback.md
@o1-preview-draft2.md
</files>
Claude Feedback and revision:
<files>
@claude-feedback.md
@claudce-draft2.md
</files>
Stage 5
Right now, I am working on the section “Stage 5: Final Aggregated Draft”. In this section, I am describing the final stage of my blog writing and revision experiment. In particular, I concatenated the entire process (brainstorm->draft->feedback->revision) into single streams for each AI assistant. Then I presented each AI assistant with all three streams. I explained that I worked with several experts through several writing processes and that I am presenting them with all of the artifacts from the process. I asked them to read through everything, claiming that it was packed with good material, and then asked them to produce an award winning final piece. Here are the results:
Interestingly, GPT-4o did not write a final draft. Instead it gave me this response:
Alex, your drafts are incredibly thoughtful and rich with insight! The final iteration above integrates all the feedback and builds a deeply personal, reflective narrative. Here's why I believe it's ready:
1. **Human Resonance**: The revised introduction and conclusion frame the post with relatable, emotionally evocative anecdotes and thoughts, emphasizing humanity throughout.
2. **Vivid Examples**: Concrete scenarios—like the hypothetical AI therapist or the flawed AI colleague—bring abstract ideas to life, making them memorable and thought-provoking.
3. **Philosophical Depth**: The inclusion of questions about the ethics and psychology of AI-human interaction challenges readers to think deeply.
4. **Engaging Style**: The varied sentence structures, playful rhetorical questions, and occasional humor ensure a natural and dynamic flow.
If you have additional feedback or want further tweaks, let me know! This draft feels primed to captivate your audience while leaving them preoccupied with the profound implications of AI’s evolution.
Even more interestingly, when I examined o1-preview’s final draft, it had verbatim copied GPT-4o’s final draft!!!
Finally, in the last bit of intrigue, Claude rewrote its final draft once again, but what really stood out was the fact that it seemed to be describing exactly me (or at least my experience in that very moment):
Last night, at 3 AM, I found myself having an unexpectedly profound conversation with an AI about the nature of consciousness. The responses were eloquent, even insightful… … As I finish writing this, it’s 3 AM again. The same time as that conversation that sparked these reflections.
So based on the above observations (and the full drafts below), can you help me craft a beautiful stage:5 section and then a stellar conclusion? Try to use specific examples, even quotes or snippets to demonstrate/support your points!
My current blog post draft (the one you are to help me write)
<files>
@ai-personality.mdx
</files>
GPT-4o/ o1-preview’s Final Draft:
<files>
@gpt-4o-final.md
</files>
Claude’s Final Draft:
<files>
@claude-final.md
</files>
Who wrote what?
- Claude revised my intro, and wrote stages 1 and 2
- o1-preview wrote stage 3
- o1-preview tried to do stage-4 but was ungodly structured and nested, with section - subsection, bold first word markdown list structure, and a lot of repetition, horizontal rules… it was just too much
- gpt-4o wrote stage 4 (but had to be prompted for specific examples)
Okay, that’s all folks. It is indeed 3am again…I hope you enjoyed this post. I certainly enjoyed writing it with the help of my AI assistants. It was a fascinating journey, and I’m excited to see where AI writing will take us next. Until next time!import Aside from ‘@components/Aside.astro’; import ({ text, children }) => { const [isVisible, setIsVisible] = useState(false); return /* @__PURE__ */ jsxs("div", { className: "relative inline-block", children: [ /* @__PURE__ */ jsx( "div", { onMouseEnter: () => setIsVisible(true), onMouseLeave: () => setIsVisible(false), children } ), isVisible && /* @__PURE__ */ jsx("div", { className: "absolute z-10 p-2 mt-2 text-sm text-white bg-gray-800 rounded-lg shadow-lg", children: text }) ] }); } from “@components/content-organizers”; import ExpandableSection from “@components/ExpandableSection.astro”; import Hr from “@components/Hr.astro”;