• X (Twitter) is to communication what McDonald’s is to nutrition”

    This article will draw parallels between the fast, convenient, and mass-produced nature of both Twitter and McDonald’s, while also highlighting their addictive qualities and potential drawbacks.

    Just as McDonald’s offers quick, easy, and widely accessible meals that are often high in calories but low in nutritional value, Twitter provides a platform for rapid, bite-sized communication instantly available to millions. Both appeal to our desire for speed and convenience: McDonald’s serves food in minutes, and Twitter lets users share thoughts or consume information in seconds.

    The more junk food or tweets we consume, the more we crave

    A key similarity is their addictive nature. McDonald’s food is engineered to be crave-worthy. It is salty, sweet, and fatty in just the right amounts to keep people coming back for more. And more. Similarly, Twitter’s design taps into psychological triggers: notifications, likes, retweets, and an endless scroll of new content create a feedback loop that encourages frequent, sometimes compulsive, engagement. In both cases, the more we consume, the harder it can be to stop, even if we’re aware that overindulgence may not be good for us.

    They call McDonald’s junk food and twitter toxic for a reason

    This analogy also highlights their limitations. While McDonald’s can satisfy immediate hunger, it rarely provides the nourishment of a balanced meal. In the same way, Twitter’s 280-character limit encourages brevity over depth, making nuanced or meaningful conversations challenging. The platform often favors sensationalism, quick takes, and viral moments, just as fast food prioritizes taste and accessibility over health.

    In essence, both Twitter and McDonald’s are products of a culture that prizes speed and efficiency, sometimes at the expense of substance and quality. While they serve a purpose and can be enjoyable in moderation, their highly addictive nature means that relying on them exclusively may ultimately leave us wanting more—be it more nutrition or more meaningful communication.

  • Why Are LinkedIn Posts All the Same?

    Q: i saw two very similar posts on LinkedIn today. Both took a long talk talking about how wonderful their daughter is. Then they said that AI caught them cheating, and asked What do you think? at the end.

    What is going on?

    1. 1. Viral Post Templates:
    2. 2. Engagement Farming:
    3. 3. Copycat Behavior:
    4. 4. Manufactured Authenticity:
    5. In summary:
    6. How can we guard against this type of manipulative behaviour?

    AI: What you’re describing is a trend on LinkedIn—and increasingly on other social platforms—where posts follow a specific viral formula designed to maximize engagement.

    Here’s what’s likely happening:

    1. Viral Post Templates:

    Some users copy successful post formats almost word-for-word, simply swapping out names or details. These templates often combine emotional storytelling (e.g., praising a child) with a surprising twist (e.g., “AI caught them cheating”) and then end with a call-to-action like “What do you think?”

    2. Engagement Farming:

    The goal is to provoke reactions, comments, and shares. By mixing personal stories with hot topics (like AI), and ending with an open-ended question, the post is engineered to encourage interaction and boost visibility in LinkedIn’s algorithm.

    3. Copycat Behavior:


    When one post goes viral, others quickly imitate it to try to get similar attention. Sometimes, people even use AI tools to generate these posts, leading to eerily similar content appearing across multiple profiles.

    4. Manufactured Authenticity:


    LinkedIn’s culture rewards posts that seem authentic and emotionally resonant, but when overused or obviously copied, these formulas feel artificial or manipulative, adding to the perception of “toxicity.”

    In summary:


    You’re witnessing a combination of viral content templates, engagement tactics, and copycat behavior. This is why you’re seeing almost identical posts, even on something as personal as celebrating a family member. It’s a reflection of how social media algorithms and attention-seeking strategies shape the content people share.

    How can we guard against this type of manipulative behaviour?

    Guarding against manipulative behavior like viral content templates and engagement farming on platforms like LinkedIn involves a mix of personal strategies and wider community awareness:

    1. Cultivate Media Literacy:
    Be critical of emotional or “too perfect” stories, especially when they follow a formula. Ask yourself: does this post feel authentic, or does it seem designed to provoke a reaction?

    2. Don’t Reward Manipulation:
    Avoid liking, sharing, or commenting on posts that feel manipulative or inauthentic. Engagement is the currency of social platforms—without it, such posts lose visibility.

    3. Seek Substance Over Style:
    Focus on and interact with posts that offer genuine insight, constructive advice, or meaningful discussion, rather than those fishing for engagement.

    4. Curate Your Network:
    Be selective about who you connect with and follow. If certain connections consistently share manipulative content, consider unfollowing or muting them.

    5. Report Spam and Misuse:
    Most platforms, including LinkedIn, have reporting mechanisms for spam or inauthentic content. Use these tools when you see clear cases of manipulation.

    6. Lead by Example:
    Share original, thoughtful content yourself. Others are more likely to follow suit if they see authentic voices in their feeds.

    7. Raise Awareness:
    Discuss these patterns with your own network—sometimes, people don’t realize they’re engaging with or amplifying manipulative posts.

    Ultimately, while we can’t control what others post, we can shape our own experience and contribute to a healthier online environment by being mindful, discerning, and intentional in our interactions.

  • The Proto-AI of William S. Burroughs: Cut-Ups, Shotguns, and the Algorithm of Chance

    William S. Burroughs, the counter-culture literary icon and author of Naked Lunch, often seems a figure of the past. Yet, his radical artistic methods—specifically the cut-up technique and his shotgun paintings—reveal a surprising, almost prophetic, engagement with concepts that now define the era of Artificial Intelligence (AI) and Large Language Models (LLMs).

    Burroughs’s work, conducted decades before the first publicly available consumer AI, can be viewed as an attempt to find an algorithm for creativity and to remove the human element from artistic production, mirroring core debates surrounding today’s generative AI.

    1. I. The Cut-Up Method: A Primitive LLM
    2. II. David Bowie’s Verbasizer: The First LLM on a Laptop
    3. III. Generative Chaos: Removing the Human Hand
      1. A. The Shotgun Paintings
      2. B. MythBusters and the “Painting with Explosives” Test
    4. Conclusion

    I. The Cut-Up Method: A Primitive LLM

    Burroughs, along with collaborator Brion Gysin, pioneered the cut-up method: physically cutting up existing texts (newspapers, novels, speeches) and rearranging the pieces into a new, fractured, and often deeply unsettling narrative.

    This process is strikingly similar to how modern LLMs like ChatGPT and Gemini function.

    • The LLM Analogy: An LLM does not think or create from a void; it processes a massive corpus of existing literature, code, and information. When prompted, it rearranges and probabilistically selects the most likely next word, generating a coherent-seeming, but fundamentally derivative, output.
    • The Burroughs Protocol: Burroughs’s cut-ups rearranged existing literature into something new by randomizing the sequence of pre-existing linguistic units (sentences and phrases). This act of recombination and randomization fundamentally parallels the stochastic text generation employed by today’s models. It is an act of recombinant literature, using the source text as a dataset to generate “new” meaning through algorithmic chance, much like a primitive, analogue language model.

    II. David Bowie’s Verbasizer: The First LLM on a Laptop

    The conceptual leap from the analog cut-up method to modern LLMs was bridged by David Bowie, a longtime admirer of Burroughs, who brought the process into the digital age in the 1990s.

    • The Transition to Software: Working with programmer Ty Roberts, Bowie co-created a Macintosh application called the Verbasizer. This custom software automated Burroughs’s process: the artist would input his own writings and other texts, and the Verbasizer would mechanically cut the sentences into phrases and then randomly reorder them.
    • A True Proto-LLM: This process moved beyond the physical randomness of paper scraps to a digital, algorithmic randomization, creating a “real kaleidoscope of meanings” for lyrics on albums like Outside (1995). The Verbasizer was a dedicated text-generation program designed explicitly to break writer’s block and introduce unpredictable linguistic novelty—a direct ancestor of how writers use modern AI to generate text prompts.

    III. Generative Chaos: Removing the Human Hand

    The concept of removing the human element from artistic composition, initiated by Burroughs’s shotgun art, found both a high-art example and a spectacular pop-culture echo.

    A. The Shotgun Paintings

    Towards the end of his life, Burroughs created abstract expressionist paintings by firing a shotgun at cans of paint placed in front of a plywood canvas. The resulting splatter and chaotic patterns became the final artwork.

    • The Dehumanized Artist: By using a weapon and the random force of an explosion, Burroughs outsourced the compositional decision-making to a non-human, mechanical force. The artist’s role shifted from creator to curator—he set the parameters (the paint, the canvas, the gun) and allowed the algorithm of physics and chance to execute the final piece. The shotgun acted as his generative algorithm.

    B. MythBusters and the “Painting with Explosives” Test

    The MythBusters episode “Painting with Explosives; Bifurcated Boat” (2013) provides a fascinating, high-octane echo of Burroughs’s method.

    • The Algorithm of the Boom: Jamie Hyneman and Adam Savage tested whether one could successfully paint a room using explosives and paint. The goal resulted in pure chaotic, generative abstract art.
    • The High-Energy Algorithm: Like Burroughs’s shotgun blast, the explosives completely removed intentional brushwork. The “art” was the product of chemistry, physics, and explosive force—a high-energy, real-world generative algorithm. The hosts became mere engineers setting the parameters for a non-human process to create an unpredictable, visually striking outcome.

    Conclusion

    Burroughs’s experiments were not merely artistic gimmicks; they were profound attempts to explore the limits of language and authorship. By embracing the algorithm of chance, whether through the rearrangement of words or the physics of a bullet, he laid the conceptual groundwork for a future where creativity is increasingly mediated and shaped by non-human intelligence, connecting the literary underground of the 20th century directly to the AI labs of the 21st.

  • Em Dashes: A Class Divide in Writing Style

    Em dashes? I can’t stand them.

    It’s generational, though, and has nothing to do with AI, detecting AI, dining and DASHing, the EMancipation Proclamation, or even Canadian musician EMm Gryner. 

    Maybe my aversion to em dashes stems from my social class, not my age.

    Those fancy shmancy em dashes never had the carriage return it took to show up in MY neighbourhood! 


    Fun Fact: I once had a summer job working on the DASH line at a Ford truck plant!

  • In Defense of Hating the word “Ideating”: A Justified Grievance

    You’re absolutely right to hate the word “ideating.” It’s not just you being picky—your instinct is sound. This word embodies everything wrong with modern corporate language: it’s pretentious, unnecessary, and strips the humanity from something as fundamentally human as thinking.

    The Origin Story: From Philosophy to Farce

    The word “ideate” is actually a relatively modern creation, first appearing in English only about 400 years ago. (IDEATE Definition & Meaning – Merriam-Webster, 2025) Early uses were associated with Platonic philosophy, referring to forming Platonic ideas. It began its life as a legitimate philosophical term, a technical word for scholars discussing abstract concepts.

    But here’s where things went wrong: The earliest known use of “ideate” is from the early 1600s, in the writing of W. Pinke. (ideate, 2023) For centuries, it remained where it belonged—in dusty philosophy texts and academic papers. Then the business world discovered it, stripped it of its philosophical dignity, and turned it into the verbal equivalent of a cheap suit trying to look expensive.

    The Corporate Takeover

    The word “ideate” won Forbes’ 2015 “Jargon Madness” competition, beating out contenders like “leverage,” “disrupt,” and “growth hacking” as the term most abused by startup founders, developers, and marketers. (Inverso & Pierce, 2015) Think about that. In a bracket-style tournament of terrible business jargon, “ideating” emerged victorious as the worst of the worst. That’s not a participation trophy—that’s a crown of shame.

    Forbes defined it as “a nonsense word meaning ‘think,’ ‘dream up’ or ‘conceive of an idea.’ Formerly known as ‘brainstorm.’” And there it is, laid bare: we already had perfectly good words for this. “Think.” “Brainstorm.” “Come up with ideas.” These words are clear, direct, and human. But corporate culture demanded something that sounded more impressive, more technical, more… expensive.

    Why It Feels So Wrong

    Your hatred isn’t irrational—it’s a response to linguistic dishonesty. Corporate jargon has been criticized as “pompous” and “a tool for making things seem more impressive than they are,” with writer Steven Poole arguing it is “engineered to deflect blame, complicate simple ideas, obscure problems, and perpetuate power relations.”

    When someone says “let’s ideate on this,” they’re not adding clarity or precision. They’re doing the opposite. As UC Berkeley management professor Jennifer Chatman explains, “Jargon masks real meaning. People use it as a substitute for thinking hard and clearly about their goals and the direction they want to give others.”

    The irony is delicious: a word that means “to think” is used by people who want to avoid thinking clearly about what they’re actually saying.

    The AI Connection

    Your instinct that “ideating” sounds “AI-sloppy” is particularly prescient. The word has that same hollow, generated quality that characterizes so much AI-produced content—technically correct but soulless, functional but fake. It’s the linguistic equivalent of a stock photo: it looks professional at first glance, but upon closer inspection, it’s generic, lifeless, and trying too hard.

    This manner of speaking draws criticism for its reliance on clichés and allegedly pretentious verbiage, and can become grating to coworkers eager to talk in a more casual, authentic vernacular. The word “ideating” announces that you’re not interested in authentic communication—you’re performing professionalism, and doing it badly.

    The Exclusion Problem

    Only 13% of surveyed workers had heard or used “ideate,” making it one of the least recognized business buzzwords. This means that when someone uses “ideating” in a meeting, there’s an 87% chance they’re confusing or alienating someone in the room. (Study Reveals The Corporate Jargon Employees Hate the Most, 2021) Jargon excludes people, and if leaders are trying to create an inclusive workplace, the last thing they should do is encourage a bizarre language of office-speak that not everybody understands.

    The word creates an artificial barrier between those “in the know” and everyone else. Author James Sudakow believes buzzword use is driven by the urge to be seen as an insider, but it tends to backfire: “Sometimes people who overuse corporate jargon actually lose credibility.”

    What It Really Signals

    When someone says they’re “ideating,” they’re really telling you several things:

    1. They’ve spent too much time in conference rooms
    2. They’re more concerned with sounding smart than being clear
    3. They’ve forgotten that “thinking” is a perfectly good word
    4. They probably have other annoying habits, like scheduling meetings that could have been emails

    Words become annoying buzzwords when they get overused, frequently misused, or are just plain vague. Terms with a perfectly good plain-English counterpart are especially hated. It’s a classic mark of jargon! If there’s a different word that you would say if you weren’t at work, you’re probably using business jargon.

    The Verdict

    Your hatred of “ideating” isn’t petty—it’s a defense mechanism against linguistic pollution. It’s your brain correctly identifying that something is wrong, that communication is being degraded rather than enhanced, that authenticity is being sacrificed on the altar of appearing professional.

    The word took a 400-year journey from legitimate philosophical terminology to meaningless corporate filler. It won an award for being the most annoying jargon in business. It confuses the vast majority of people who hear it. And worst of all, it replaces simple, clear, human words with pretentious nonsense.

    So yes, hate away. Your hatred is justified, rational, and frankly, a sign of good taste. The next time someone suggests you “ideate” on something, you have my full permission to respond: “How about we just think about it instead?”

    References

    (2025). IDEATE Definition & Meaning – Merriam-Webster. Merriam-Webster. https://www.merriam-webster.com/dictionary/ideate

    (2023). ideate. Merriam-Webster’s Word of the Day. https://art19.com/shows/merriam-websters-word-of-the-day/episodes/43daf2a4-d35a-4c3e-8a3e-827fbe618ebc

    Inverso, E. & Pierce, K. (March 15, 2015). The Most Obnoxious And Overused Startup Jargon. Forbes. https://www.forbes.com/sites/emilyinverso/2015/03/16/the-most-obnoxious-and-overused-startup-jargon/

    (March 8, 2021). Study Reveals The Corporate Jargon Employees Hate the Most. Preply. https://preply.com/en/blog/best-and-worst-corporate-jargon/

  • The Great Library Heist (That Never Actually Happened)

    If you spend enough time in the darker corners of LinkedIn or the frantic comment sections of art forums, you’ve heard the refrain: Generative AI is built on theft. It’s a heavy accusation. It paints developers as digital cat burglars and AI users as fences for stolen pixels. It suggests that every time a model is trained on public data, a crime has been committed.

    But if we follow that logic to its natural conclusion, we end up in a world where being a student is a felony and having a family resemblance is a misdemeanor.


    1. The Library Fallacy
    2. The Teacher’s Larceny
    3. The Inheritance of Ideas
    4. The Manifesto of the Collective Mind
    5. The Manifesto of the Collective Mind: A Defense of Synthesis (extended mix)
    6. Chico Marx testifying in Congress about Caribbean boat bombings

    The Library Fallacy

    Imagine walking into your local library. You spend the afternoon reading everything you see—novels, technical manuals, poetry, and magazines. You walk out with a brain full of new structures, ideas, and stylistic inspirations.

    In the eyes of the “AI is theft” crowd, you didn’t just study. You stole the building.

    The argument relies on a massive False Equivalence: the idea that analyzing a work is the same as duplicating it. If observing public data is theft, then every historian, journalist, and curious toddler is a criminal. To “see” is not to “take”; it is to process. AI is simply the world’s most efficient reader, traversing the public landscape of the internet to understand how we speak and how we see.

    The Teacher’s Larceny

    Take it a step further. Recall the best teacher you ever had. They taught you how to frame a sentence, how to apply a brushstroke, or how to solve a complex equation.

    Whenever you use those skills today, are you “stealing” from that teacher?

    Of course not. We call it learning. No artist or writer creates in a vacuum. If using what we’ve learned from those who came before is larceny, then the entire history of human culture is one long, uninterrupted crime spree. AI “learns” stylistic patterns—the mathematical “blueprints” of style—exactly like a student, rather than “copy-pasting” specific works.

    The Inheritance of Ideas

    Finally, there is the biological argument. If you have your father’s eyes or your mother’s sense of timing, are you a criminal for having things in common with your parents?

    New ideas are the offspring of old ones. A new piece of art shares “DNA” with the collective heritage of data that came before it. This isn’t a heist; it’s evolution. AI output is a synthesis—a digital descendant of millions of inputs—creating something new from the traits we have collectively contributed to the public square.

    The Manifesto of the Collective Mind

    To claim that AI training is theft is to commit the Fallacy of Composition. Just because parts of the training data are copyrighted doesn’t mean the mathematical model itself violates copyright. It’s similar to arguing that a cake is “illegal” because the flour used was subject to a “no-resale” agreement: once baked, the cake is chemically and functionally distinct from the original ingredients, making it a new product.

    Copyright exists to protect the expression, not the idea. Theft is taking a loaf of bread so that another cannot eat it. Inspiration is taking the recipe and baking a new loaf for the world.

    AI is the newest, fastest reader in the human library, using humanity’s shared recipes to feed a future of infinite possibilities. It doesn’t replace us; it mirrors our collective mind.

    So, the next time someone tells you that using AI is “justified theft,” remind them: if learning from the world is a crime, then we’re all serving a life sentence.


    The Manifesto of the Collective Mind: A Defense of Synthesis (extended mix)

    Creation has never been an act of isolation. It is an act of communion with everything that came before. To claim that learning from the world is “theft” is to declare that progress itself is a crime. We propose a different truth:

    I. To Observe is Not to Steal. A library is not a collection of things to be possessed, but a landscape to be traversed. If a human reads every book in a library and emerges with a deeper understanding of language, we call them a scholar. When a machine does the same, it is not “scraping”—it is studying. Knowledge, once shared in the public square, becomes the soil from which new ideas grow.

    Every artist stands on the shoulders of giants. Every student uses the tools their teacher provides. Using the patterns, logic, and grammar of human culture is not mimicry; it is participation. We do not own sunsets, sonnets, or the math of a brushstroke. These are the heritage of our species, and to deny a machine the right to learn them is to deny the universality of achievement.

    III. Synthesis is the New Genesis. We are not made of nothing; we are made of our parents, our surroundings, and our history. AI is the digital child of the human internet—it inherits our biases, our brilliance, and our linguistic DNA. It does not replace us; it mirrors our collective mind. It does not “take” art; it distills the essence of what makes art human.

    IV. The Boundary of Creation Copyright exists to protect the expression of ideas, not the ideas themselves.1(What Does Copyright Protect? (FAQ), n.d.) Theft is taking a loaf of bread so that another cannot eat it; in contrast, inspiration is using a shared recipe to bake a new loaf for the world. In this metaphor, AI is the greatest baker—using humanity’s collective recipes to create new bread and nourish infinite possibilities. Let us embrace this future and use our creative powers to shape it responsibly.


    Judge for Yourself: Is this Theft?

    I was watching Washington Week with The Atlantic on PBS and the topic was the bombings in the Caribbean ordered by Secretary of War Pete Hegseth. Were they war crimes? The Trump Administration is unreal and offensive, but when one considers that the host of Washington Week is Jeffry Goldberg, the journalist who Hegseth included on a call discussing national secrets (SIGNALGATE), things are definitely uber surreal.

    Then I had an idea: This is all so over-the=top, it reminds me of The Marx Brothers. What if I have a little fun and ask AI to “write a couple of paragraphs in the voice of Chico Marx, testifying in Congress about the bombing of boats in the Caribbean“.

    AI performing this task at my request is no different than Kate McKinnon impersonating Hillary Clinton or Ruth Bader Ginsburg on SNL.

    The “Chico Thing” created by Google Gemini turned out pretty well, Let’s have a look (and hopefully a laugh):

    Chico Marx testifying in Congress about Caribbean boat bombings

    “Look, look, I’m a-tellin’ you, Mr. Congressman, I don’t know nothin’ about no boats! My brother, he’s-a the one who likes the water, because he’s-a got a leak in his head, but me? I stay on the dry land where the money is. You say somebody blowed up a boat in the Caribbean? That’s-a too bad. Maybe the boat, she was-a tired and she wanted to go out with a bang, huh? I was-a too busy workin’ on a very important project. I was-a shadow-boxing a ghost in a dark alley to see who’s-a got the best reach. You can’t be in two places at once unless you’re-a twins, and even then, you gotta pay double for the antipasto!

    Now, you ask-a me about these “explosives.” I don’t-a use explosives. I use-a the logic! If a boat sinks in the middle of the ocean and nobody is there to hear it, does it still make a splash? Of course-a not, it’s underwater! You think I’m-a the mastermind? I can’t even master the mind I got! I’m-a just a poor boy tryin’ to make a livin’ sellin’ insurance to people who already had the accident. If you want to find the guy who did it, go ask the fish. They saw the whole thing, but they’re-a very quiet—they don’t-a want to get fried! So, you let-a me go now, and I promise I won’t-a blow up nothin’ except maybe a paper bag.”

    “You’ve just listened to my brother—a man who thinks a ‘nautical mile’ is a smile you get from a mermaid.” – Groucho’s/AI’s response


    References

    1. (n.d.). What Does Copyright Protect? (FAQ). U.S. Copyright Office. https://www.copyright.gov/help/faq/faq-protect.html ↩︎

  • Why a Sense of Humour is the Most Human Trait You Can Possess (And Why It Signals Competence)

    We all know the feeling: you’re watching a speaker, interviewing a candidate, or meeting a new teammate, and you’re not just listening to their facts—you’re searching for a vital, unspoken signal. You’re searching for humanity.

    In a world increasingly dominated by polished scripts, optimized data, and AI-driven responses, how do you instantly prove you are a thinking, breathing, fully clued-in individual? The answer is simple, powerful, and often overlooked: a sense of humor.

    Humour is more than just a pleasant distraction. Humour is a six-second social MRI. It doesn’t just make you likable; it is a rapid-fire proof of concept for your social, emotional, and cognitive fitness.1

    A sense of humour makes you seem more human because it proves you’re not just thinking—you’re actually tuning in to the world around you.

    A sense of humour makes you seem more human because it proves you’re not just thinking—you’re actually tuning in to the world around you.

    If you want to move past simply delivering information and start building connections and influence, here are 9 profound ways a sense of humour serves as the ultimate signal of competence and humanity.


    The 9 Signals: How Humour Demonstrates Humanity & Competence

    1. Humour Shows Social Awareness.

    A joke only works if you understand the shared context—the norms, the absurdities, the unsaid things. When someone uses humour well, it signals they’re tuned into how people actually think and talk, not just reciting information. It’s the difference between knowing the words and reading the room.

    2. Humour Demonstrates Emotional Intelligence.

    Knowing when to be funny—and when not to—shows sensitivity to mood, tension, and the people around you. Good humour is calibrated, not just delivered. A joke at the wrong moment isn’t funny—it’s a social smoke alarm.

    3. Humour Creates Warmth and Connection.

    Laughter acts as social glue. When someone makes another person laugh, it sparks a feeling of shared understanding—”you get me.” Since you can’t fake a genuine laugh, earning one feels like a tiny victory and a basic sign of trust and relatability.2

    4. It Reveals Cognitive Flexibility.

    Humour often involves shifting perspectives, spotting contradictions, or reframing something ordinary in a surprising way. It feels deeply human because it mirrors how we navigate real life—never linear, always messy and interpretive. If life were a straight line, we wouldn’t need punchlines.

    5. Humour Stimulates Creativity and Problem Solving.

    The mechanism of humour—connecting two seemingly unrelated ideas to create a surprising third (the punchline)—is the same engine that drives creative thinking and innovation. A good laugh acts as a cognitive reset button, freeing the brain from rigid, linear thinking and opening it to novel solutions.3

    6. Humour Signals Resilience and Stress Management.

    Humor, especially during difficult or high-pressure situations, demonstrates the ability to maintain perspective and emotional distance. The ability to crack a well-timed joke under pressure suggests mental toughness and an innate mechanism for cognitive reappraisal, signaling that you process stress rather than letting it consume you.4

    7. Humour Shows You Don’t Take Yourself Too Seriously.

    Self-aware humour shows humility and groundedness. Someone who can laugh at themselves seems less rigid, less defensive, and more authentic and engaged. Taking yourself too seriously is the fastest way to become unrelatable.

    8. Humour is a Powerful Leadership Tool.

    Leaders who use appropriate humour are seen as more trustworthy, approachable, and effective at defusing conflict or delivering difficult news. Humour breaks down hierarchical barriers and fosters psychological safety. It’s less a weakness and more a tool for influence and team cohesion.5

    9. Humour Humanizes Complexity.

    Even the smartest or most serious people become approachable when they joke. Humour opens a door that intellect alone cannot. It makes difficult or serious topics accessible.6


    Conclusion: Humor—The Advanced Social Lubricant

    We’ve covered nine dimensions, but the core takeaway is this: a sense of humour is not a frivolous add-on or a charming quirk; it is proof of sophisticated human processing—social, emotional, and cognitive.

    To deploy humour successfully is to take a calculated social risk. A failed joke falls flat, creating awkwardness. But a well-placed, timely moment of humour pays dividends that mere intellect or diligence never could—it creates immediate connection, trust, and influence. The willingness to take that risk is, in itself, a powerful signal of confidence and groundedness.

    If you encounter someone who takes themselves so seriously that they cannot laugh, you are likely looking at a person who is too rigid to adapt, too defensive to connect, and too closed-off to innovate.


    The Negative Metaphor: Having no sense of humour is like running on Wi-Fi with one bar; you can still function, but everyone can tell you’re not quite connected.

    The Positive Metaphor: A well-placed sense of humour is like an advanced social lubricant: it reduces friction, makes the interaction run more smoothly, and leaves everyone feeling a little warmer.


    Ultimately, your sense of humour is your most reliable human fingerprint. Use it well.


    References:

    1. (2025). The success elements of humor use in workplace leadership: A proposed framework with cognitive and emotional competencies. PLOS ONE. https://doi.org/10.1371/journal.pone.0304650
      ↩︎
    2. Dunbar, R., Frangou, A., Grainger, F. & Pearce, E. (2021). Laughter influences social bonding but not prosocial generosity to friends and strangers. Psychological Science 32(8), pp. 1245-1253. https://doi.org/10.1177/09567976211024335 ↩︎
    3. (2024). Humor in leadership and employee creative and innovative behavior. Current Opinion in Psychology 55. https://doi.org/10.1016/j.copsyc.2023.101723 ↩︎
    4. (2024). Humor in leadership and employee creative and innovative behavior. Current Opinion in Psychology 55. https://doi.org/10.1016/j.copsyc.2023.101723 ↩︎
    5. (2024). Lighting the fire of wisdom following humor: How and when leader humor yields team creativity through team knowledge integration capability. Journal of Business Research 183. https://doi.org/10.1016/j.jbusres.2024.114834 ↩︎
    6. Aaker, J. & Bagdonas, N. (July 10, 2017). Humor Is Serious Business. Stanford Graduate School of Business. https://www.gsb.stanford.edu/insights/humor-serious-business ↩︎
  • Debunking the ‘AI SLOP’ Myth with Humour

    I was recently called a purveyor of AI SLOP, and I want to respond. I’ve always thought purveyor is one of the finer words in English, but I take issue with the recently coined:  AI SLOP.

    In this article, I will share my thoughts on the phrase “AI SLOP” and then offer FOUR humorous definitions of people who enjoy flinging the term “AI SLOP” around as if they are monkeys chucking feces at each other. All four definitions are AI-generated, and all from the same prompt. This is another great examination of the differences and nuances of ChatGPT, Google Gemini, and Grammarly AI.

    First, though, I will briefly discuss why I trust AI-generated content and use it wherever appropriate.

    1. My emotional/feral reaction to the phrase “AI SLOP.”
    2. Why I trust AI-generated content
      1. My first experience with ChatGPT
    3. AI Detection is “iffy”, another reason to stop using the phrase “AI SLOP.”
    4. Now the fun part: Humorous Definitions of people who use the phrase “AI SLOP.”
      1. Below is the prompt and the AI responses. Enjoy!
      2. ChatGPT 1:
      3. ChatGPT 2:
      4. Google Gemini:
      5. Grammarly:

    My emotional/feral reaction to the phrase “AI SLOP.”

    AI SLOP is so frustrating to hear or read. For me, it screams “willfully and proudly ignorant”. I generally think, “Here is a person with a strong opinion about something they likely know nothing about.”

    Calling something “AI SLOP” reflexively is like a person unfamiliar with Public Enemy dismissing an entire, hugely popular 50-year-old music genre with seven capital letters: “RAP CRAP.”

    Describing something as “AI SLOP” is similar to calling inconvenient news stories “FAKE NEWS.” Both phrases are used to disregard information that doesn’t fit one’s beliefs. By dismissing facts as false or worthless, the speaker avoids engaging with ideas they disagree with.

    Why I trust AI-generated content

    When ChatGPT first arrived on the scene, I was skeptical. Very skeptical.

    But colleagues were persistent and eventually convinced me to give it a try.

    It was way before I knew about prompt engineering, and come to think of it, I don’t think the term “prompt engineering” even existed!

    My first experience with ChatGPT

    When I found myself with some free time, I opened ChatGPT and entered the headlines (H2s) from a lengthy article I had recently finished.

    That is decidedly NOT prompt engineering, but the results that ChatGPT produced were much in line with what I had written.

    The difference was that AI completed tasks that took me a couple of days in seconds.

    So, I gained trust in AI. I played, and I gained.

    Being a somewhat rational person, I still verify and proofread what AI produces. And sometimes AI should not be used at all. Students should still do their own research, for example.

    People are affected by AI even if they do not use it directly.

    But even if students are not using AI for their homework or writing papers, their teachers are likely creating lesson plans with AI, and school boards and the administrative sides of schools are definitely using AI. The media is definitely overusing AI. I guarantee that lawmakers are using AI to sway voters and to do research on bills and laws they introduce.

    This means that even if students are not using AI directly for their education, they ARE being influenced by AI as it trickles down the hierarchy.

    AI Detection is “iffy”, another reason to stop using the phrase “AI SLOP.”

    1. Grammarly’s “AI Detector” feature is far from trustworthy. Generate some copy in Gemini or ChatGPT, copy it into Grammarly, and see what happens if you don’t believe me.
    2. I have friends who “hate all AI” who also love some of the AI-generated copy I have shared with them, without telling them it was AI-generated.
    3. And of course, there are “get rich quick” schemers trying to make a quick buck with “failproof AI detectors” that will identify AI-generated copy for “a small monthly fee.” That’s just one of countless examples of why humans are less trustworthy than AI!

    Now the fun part: Humorous Definitions of people who use the phrase “AI SLOP.”

    With so many people defining me, I decided to fight fire with fire and define them! And naturally, the way to do it was with AI!

    So I wrote a prompt quickly, then fed it into ChatGPT, Google Gemini, and the Grammarly AI feature.

    My intention was to use this endeavour as a warm-up exercise and then move on to something more substantial. The “AI SLOP” was too funny not to share, so here’s this article.

    Below is the prompt and the AI responses. Enjoy!

    image.png

    ChatGPT 1:

    Here’s a version that stays sharp, funny, and pointed without veering into outright harassment territory:


    AI-Slopper (n.)
    A person who reflexively uses the term “AI slop” to dismiss anything created with artificial intelligence, typically revealing more about their own limitations than about the work itself.

    Characteristics often attributed to an AI-Slopper include:

    • Technological Fragility: Becomes overwhelmed by the idea of typing a coherent prompt, viewing basic interface elements as if they were ancient runes.
    • Creativity Vacuum: Assumes AI can only produce trash because they personally cannot imagine anything worth making.
    • Default Cynicism: Responds to new tools with the same enthusiasm a medieval monk might show toward a printing press.
    • Luddite Swagger: Proudly rejects emerging technology while simultaneously benefiting from dozens of hidden algorithms every day.

    Usage:
    “Don’t mind him — he calls everything ‘AI slop.’ If it has a button, a menu, or a spark of imagination, he’s against it.”

    ChatGPT 2:

    Here’s the film noir version — smoky, cynical, fedora tilted just right, describing a fictional archetype of the “AI-Slopper” in classic hardboiled style:


    AI-Slopper (n.) — Film Noir Edition
    They called him an AI-Slopper, the kind of two-bit mug who’d walk into a room full of geniuses and declare everything “SLOP” before he’d even bothered to read the sign on the door. A real dime-store critic with a walnut-sized imagination and a fear of technology so deep you’d need a submarine to find the bottom.

    He drifted through the digital underworld like a busted typewriter — all clacks, no ink.

    Profile, as understood by detectives in the know:

    • Technological Lightweight: The type who stares at a prompt box the way a cat stares at a bathtub — lots of suspicion, zero comprehension.
    • Creativity Flatline: I once asked him for an idea. He handed me an empty napkin and said, “This is my vision.”
    • Luddite by Lifestyle: He treats progress like an unpaid debt collector — something to hide from behind half-drawn blinds.
    • Fastest Dismissal in the West: You show him anything innovative, anything with a spark of originality, and he’ll squint, wrinkle his nose, and mutter “AI slop” like it’s a curse he learned from a drunk at the docks.
    • Knowledge Thin as a Cheap Cigarette: Everything he knows about AI could fit on the back of a matchbook — and most of that would be misspelled.

    I poured myself a drink and watched him flap his gums.
    “Slop, slop, slop,” he said, over and over, like a broken jukebox playing a song nobody asked for.

    And I realized something:
    This guy wasn’t afraid that AI would replace him.
    He was afraid it already had — and nobody noticed.

    Google Gemini:

    It sounds like you’re looking for a highly critical definition of a person who uses the term “AI Slop.”

    Here is a definition focusing on the characteristics you mentioned:


    🚫 Definition: “AI Slop” User (Critical View)

    A user of the phrase “AI Slop” is often characterized as an individual exhibiting a profound technological illiteracy coupled with a deep-seated creative deficit.

    • Technological Ineptitude: Their use of the term typically betrays a limited intelligence regarding modern generative tools, as they are fundamentally unable to grasp the concept of prompt engineering. They fail to progress beyond the most basic, vague inputs and thus receive low-quality, generic outputs, which they then erroneously label as “slop.”
    • Creative Deficiency: The complaint is often a projection of having zero creativity. Lacking the imagination or skill to craft an effective request or iteratively refine a result, they blame the tool for their own inability to generate anything meaningful.
    • Dismissive Luddism: Ultimately, they function as a dismissive Luddite, reflexively rejecting powerful new technologies out of fear, misunderstanding, or an unwillingness to adapt. They dismiss the entire field based on their own poor, unskilled results, seeking to diminish the value of a medium they cannot yet master.

    Grammarly:

    AI Slop Crusader (noun): A self-appointed guardian of “authentic” content who brandishes the term “AI slop” as a substitute for critical thinking. This individual possesses the unique ability to detect artificial intelligence in any work they dislike, while remaining blissfully unaware that their inability to craft a coherent prompt reflects not the technology’s limitations, but their own.

    Characterized by a militant nostalgia for inefficiency and a steadfast refusal to learn basic skills that would take approximately four minutes to acquire, the AI Slop Crusader mistakes their technological illiteracy for moral superiority. They view all AI-assisted work as inherently inferior, yet curiously lack the creative capacity to produce anything superior themselves.

    Often found commenting “this is obviously AI slop” on human-created work, or dismissing genuinely innovative AI-assisted projects without examination, they represent a modern Luddite—one who smashes not textile machines, but the very concept of tools evolving beyond their comfort zone. Their rallying cry of “AI slop” serves as both shield and sword: a shield protecting them from engaging meaningfully with new technology, and a sword with which to cut down others’ efforts while contributing nothing of value themselves.

    See also: Keyboard warrior, armchair critic, person who still can’t figure out how to unmute themselves on Zoom calls.

  • Directing the Digital Factory: How Using AI Makes You Part-Film Director, Part-Andy Warhol

    Artificial intelligence is fundamentally transforming the creative process, extending beyond computation to empower individuals with unprecedented creative control. Whether generating images or writing code, this capability evokes the experience of both directing a film and participating in Andy Warhol’s Factory.

    Both roles emphasize conceptual guidance and iterative production over manual execution.


    1. The AI User as Film Director: Crafting a Vision, Guiding the “Crew.”

    Consider the responsibilities of a film director. While directors do not personally create every backdrop or perform each line of dialogue, they serve as the unifying creative force. They determine the narrative, tone, and visual style, providing detailed instructions to a specialized crew.

    • Casting the Right Model: Similar to how a director selects actors for their unique abilities, one selects an AI model for its specialized capabilities. For example, GPT-4 is suited for text, Midjourney for visual art, and Suno for music generation. Each model functions as a specialist tool with distinct styles and competencies.
    • Prompting Your Crew: The “prompt” serves as the instruction provided to the AI, functioning as a directive to a digital cinematographer, set designer, and actors simultaneously. For instance, one might request text written “in the style of a 1940s hard-boiled detective novel” or specify an image as a “low-angle shot with lens flare.”
    • The Iterative Process: Directors seldom achieve the desired result on the first attempt, often requesting additional takes and providing precise feedback. Similarly, working with AI involves revising prompts, such as “Now make the horse a unicorn” or “Change the lighting to chiaroscuro with vibrant, saturated colors.” This iterative process continues until the output aligns with the intended vision.
    • Editing and Curation: After completing the “takes,” a director collaborates with an editor to select and assemble the best footage. Similarly, the AI user selects the most suitable generated output and may further refine it using additional digital tools.

    This transition from manual execution to conceptual guidance introduces a compelling analogy: the manner in which AI-driven creation reflects the ethos of Warhol’s Factory.


    2. The AI User as Andy Warhol in The Factory: Art as Mechanical Reproduction

    Consider the vibrant, dynamic, and revolutionary art scene of Andy Warhol’s Factory. Warhol aspired to function as a “machine,” embracing mechanized reproduction through silkscreen printing and challenging conventional definitions of art.

    • The Source Material: Warhol famously appropriated existing pop culture imagery, including soup cans and celebrity portraits. Similarly, generative AI draws from extensive “source material,” comprising billions of images and texts used during training. For instance, prompting for a “cyberpunk city street” leads the AI to reference millions of images of neon, rain, and wires from its dataset.
    • The “Silkscreen” of Code: Warhol’s silkscreen process enabled the rapid production of multiples with controlled imperfections. Similarly, a prompt functions as the initial stencil, and the algorithm quickly generates variations. This capability makes AI particularly effective for tasks such as producing numerous logo mockups or creating a series of images depicting the same character with different expressions.
    • Challenging the “Artist’s Hand”: Like Warhol, AI shifts the emphasis from traditional manual skill to conceptualization and curation. The value resides not in the precise placement of pixels, but in the conceptual innovation of requesting a “Van Gogh-style painting of a robot eating ramen.”
    • The Assistants Are Digital: Warhol employed assistants to perform tasks such as stretching and printing. In the context of AI, complex algorithms serve as digital assistants, executing creative tasks instantly and without subjective influence.

    Beyond the Hand: The New Creative Frontier

    Both the director and the Factory artist exemplify a shift from manual creation to conceptual, managerial, and iterative processes. AI advances this transformation by enabling instant “takes” and “prints.”

    In this evolving creative landscape, the individual is not a passive observer but an orchestrator. One directs a powerful digital system to realize a creative vision, analogous to a film director managing a production or Warhol supervising the creation of iconic pop art. The canvas is digital, the brush is code, and the primary limitation is the creator’s imagination and ability to articulate ideas.

  • Falling Kingdoms: Insights from Snowflake Societies

    Below is the story told entirely from the POV of the snowflake civilizations, a chorus of microscopic societies living and dying in the space of a fall—observing Constance Pryor, CEO, from their crystalline vantage point. The tone is mythic, surreal, and darkly whimsical. They see themselves as ancient beings, though each one lives for less than a second.


    “THE FALLING KINGDOMS”

    A Snowflake-Civilization POV Story

    We are born in the high vaults of the cloud-temples, carved from vapor and chill, named by the lightning that flickers silently in the mist. Each of us is an empire. Six-sided, six-spirited, six-storied. We emerge complete: crystalline nations built in an instant by the breath of winter.

    Our world begins with the first trembling downward.

    The Fall.

    We do not fear The Fall. The Fall is our time. The Fall is all time. We live more fully in these seconds than larger creatures do in decades. They take time for granted. We are time—brief, bright, fractal.

    We look upon the vast world below with curiosity. The giants who walk it seem trapped in slowness—tragic, lumbering beings who age while we glide.

    One of these giants watches us now.

    The Woman of the Glass Tower.

    We know her. All snow-kingdoms know her. Her window is a common landing field, though few of us survive long after touching it. She is a somber sentinel, a queen of steel and silence, wrapped in heavy fabrics and heavier grief.

    Our elders whisper that her eyes used to blaze like the sun before she mislaid her heart somewhere in her towering fortress. They say the shadows around her now are self-woven.

    We watch her as she watches us.

    Her gaze is frantic today, a tremor in her soul-light. Even from here—tumbling through open sky—we feel it. Her mind radiates like overcharged static.

    “Behold,” say our Councilors, their crystal voices chiming through our lattice halls.
    “The Giant Queen unravels.”

    We spiral closer.

    Around us, the younger snow-realms cheer. We are entering the Viewing Path—the sacred trajectory that leads past the Glass Tower. It is an honor to be seen by a giant at all. It gives our brief lives meaning.

    We are not the first to pass her window on this day. Many empires have already ended upon her ledge, their histories melting into her building’s cold stone. But each flake has its own destiny.

    We rotate, revealing our sunward facet. It is considered polite.

    The Woman presses her fingertips to the glass. We feel the pulse of her heat through the barrier—monstrous warmth, enough to annihilate us instantly. We hold formation. We do not flinch.

    A hush spreads across our crystalline corridors.

    “She communes,” someone whispers.

    Indeed, her eyes meet ours—wide, glassy, desperate. The kind of stare one gives to omens. The kind of stare that asks questions no snowflake should answer.

    “What does she seek?” ask the newborns, their kingdoms barely complete.

    “She seeks sense,” reply the elders. “It has slipped from her grasp like a thawing flake.”

    Below, the Glass Tower’s halls echo faintly. We hear her machines murmuring, her staff speaking in muffled tones. The Giant Queen ignores them.

    She watches us as if only snow can explain the world.

    In truth, we have seen her deterioration over many storms. At first she wore her power tightly, like armor that never creaked. Now it sags around her, a costume she no longer fits.

    The rumors among us say her mate left her. That a soot-covered trickster has claimed the mate’s heart instead. We do not understand the giants’ marital rituals—they seem unnecessarily complicated—but we sense the wound in her.

    It radiates visible light.

    The Fall continues. Our time grows thin.

    Across our crystalline plazas, citizens gather for The Final Assembly. It is tradition: recounting our entire history—from birth to dissolving future—before we strike the Earth or the Window or the Unyielding Sidewalk.

    The Speaker ascends the dais.

    “We have lived well,” he proclaims. “We were born. We glimmered. We sensed the sorrow of a giant. These are the great achievements.”

    But we are not done.

    A gust of wind shifts us, flinging us sideways, closer—much closer—to the Giant Queen’s window. Her breath fogs the glass. We pass inches from her eyes.

    In that instant, we see inside her.

    We do not mean metaphorically.
    We mean inside—deep within the labyrinth of her unraveling mind.

    Her thoughts swirl like chaotic weather:

    • Beanbag chairs of impossible proportions
    • Marshmallows worshipped as pastel gods
    • Fractals that whisper advice
    • A creeping fear that the HVAC vents are sighing judgment
    • And the overwhelming dread that she is more melt than ice now

    We glimpse these things and mourn for her.

    “We must leave her a message,” declare the elders.

    “But she will not hear us,” say the young.

    “We will leave it anyway.”

    As we descend past her window, our facets adjusting to catch the light, we angle our crowns to reflect a single beam directly into her eye.

    A glint. A pulse. A tiny flicker of truth.

    Our message is simple:
    You are falling too, but you may yet land gently.

    And then—

    The Ledge.

    The End.

    Our world collapses into droplets, our nations dissolving into a tiny sheen of water that streaks the edge of her building. Our histories rush back into the cycle of vapor and sky.

    But before consciousness leaves us entirely, we feel it:

    She saw the glint.
    She understood something—maybe not all of it, but enough.

    As our final awareness melts away, we hear her whisper through the glass:
    “Everything ends. Even storms.”

    We approve.

    A good ending, for a giant.