Who Am I This Time?
New developments in Generative AI promise that we'll all have digital BFFs to help us live our best lives, but is this really possible? (Issue #116)
Before we get to today's main topic, some miscellaneous goodies and things worth your attention…
Free Shipping Today (Sunday) from Bookshop.org in honor of Independent Bookstore Day!
Followup to last week's piece about Wordle: my college roomie Peter Sarrett objected that it's fine to use Google to solve Wordle because there is no stated rule against it. Peter is correct. My sense that using Google is cheating, and I still feel is way, is idiosyncratic based on both my desire to solve it myself and also not to misrepresent my results to the folks with whom I share daily scores. However, when I posted this question to the Facebook Hive Mind, the results were mixed. I saw many short "yes, it's cheating" answers but also several longer "well, it depends" answers. Thank you, Peter, for forcing me to think more clearly.
What the heck is a Birkin bag, and why are they so expensive? Plus, why are organized criminal gangs stealing them? See this fascinating Los Angeles Times ($) article. (H/T Mom and Dad.)
Deepfakes are a real problem: see this NYT ($) piece about a disgruntled high school athletic director in Baltimore using Gen AI to create audio of his principal saying racist and antisemitic things. The lesson here? Don't reflexively forward things that come to you online. Check them out first.
Along the same lines, if you want to see just how scarily convincing a deepfake can be, a Technology Review reporter shares her experience with a "hyperrealistic" deep fake of herself in this compelling piece ($).
Marketing Fail: Longtime readers might remember that I'm a fan of David Weber's Honor Harrington novels (I read 11 of them in 2022 alone), so I was excited to learn about a new novel in the series. However, the new one, Toll of Honor, is what's called a "paraquel"—an old story told from different points of view. I'm still reading it, of course, but the dust jacket and "About David Weber" page should have made clear that this is not a continuation, not a new story. I had to do serious research (e.g., I found and read the transcript of a podcast) before I realized why the plot seemed so familiar...
The title of this week's main piece, "Who Am I This Time?" comes from a Vonnegut short story. There's also a fantastic 1982 TV adaptation starring Susan Sarandon and Christopher Walken, directed by Jonathan Demme. You can watch the whole thing on YouTube or on Roku.
Hire me to speak at your event! Get a sample of what I'm like onstage here.
Please follow me on Bluesky, Instagram, LinkedIn, Post and Threads (but not X) for between-issue insights and updates.
On to our top story...
Who Am I This Time?
Last week, I read an intriguing Psychology Today piece about the next wave of Generative AI powered digital assistants. In "The Emergence of Private LLMs," John Nosta argues:
The role of Large Language Models (LLMs) is about to change. Two groundbreaking advancements are set to redefine the way we interact with personal assistants: the rise of private LLMs and the expansion of prompt memory. This powerful combination of memory and privacy in LLMs is poised to create the most sophisticated and influential personal assistants ever conceived, offering unprecedented insights while safeguarding user confidentiality. In the simplest of terms, you just might be finding a new BFF—best friend forever.
To translate: Generative AI will change in two powerful ways. First, instead of folks logging into ChatGPT or Gemini (etc.), we'll each own programs that work with our data to help us do what we want to do. Second, instead of those programs forgetting that they've ever spoken to us after each interaction (digital amnesia), our new digital pals will remember our interactions (Ars Technica has a handy explanation), becoming ever more personalized helpers.
Nosta's conclusion is optimistic:
As LLMs continue to evolve, we can anticipate a future where AI becomes a seamless extension of ourselves, helping us navigate life's challenges, optimize our potential, and discover new frontiers of personal growth. The synergy between memory and privacy will give rise to remarkably powerful tools that not only support us but also help us uncover hidden aspects of our own capabilities.
I'm skeptical. My doubts are not about the ever-increasing power of Gen AI (more on this towards the end). Instead, I'm skeptical because of how incoherent we humans are.
How can even the most powerful algorithm help us "optimize our potential" (such a bloodless phrase) when we don't have a stable sense of who we are or what we want?
Sam Harris, in Waking Up: A Guide to Spirituality without Religion (2014), has a terrific account of how our sense of a consistent self behind the wheel of our ongoing consciousness is an illusion. But, if you don't want to read 200+ pages, Walt Whitman has a succinct, three-line version in his poem "Song of Myself" from Leaves of Grass (1892):
Do I contradict myself?
Very well then I contradict myself,
(I am large, I contain multitudes.)
Many thinkers (e.g., Bernard Beckerman, Daniel Kahneman, Charan Ranganath) have made the distinction between the experiencing self (what's happening right now) and the remembering self (what happened before). I believe we should also consider a distinct predicting self (what we think is about to happen).
So... which self is our new digital BFF supposed to (ick) optimize? Right now, I'm trying to lose a few pounds, but I also love vanilla ice cream with fresh blueberries. Should my digital pal be loyal to my future-focused, predicting self and lock the freezer after dinner or somehow distract me from dessert ("Look, Brad! Comic books!")? Or should my pal be loyal to my right-now self and remind me that I also like Nestlé Carnation malted milk powder in ice cream and that a fresh container is in the pantry?
Which of our wishes should get priority?
This sort of dilemma isn't only about dessert. When my beloved grandmother, Edith Berens, died in 2010, we discovered a problem. Shortly before her death, Grandma told my parents that she did not want a funeral service, particularly one conducted by a rabbi. She confessed that she had always secretly been an atheist. However, as Mom and Dad started to make the arrangements with the cemetery, they realized something. When my grandfather died seven years before, Grandma had prepaid for an hour of chapel time after her death. (It was a twofer.)
Which wish should we respect? The 2010 wish not to have a funeral or the 2003 prepaid wish to have a funeral? My grandmother was of sound mind until the moment of her death, no dementia, so we couldn't explain away the "no funeral" request.
It was like a moment in Oscar Wilde's The Importance of Being Earnest (1895) when Gwendolyn and Cecily both believe themselves to be engaged to Ernest Worthing. "I have the prior claim," Gwendolyn says because the man she thinks is Ernest proposed to her the day before. Cecily, to whom the man she thinks is Ernest has proposed just a few minutes earlier, ripostes, "since Ernest proposed to you he clearly has changed his mind!"
With my grandmother, in the end we decided both a) funerals exist to help the living grapple with the death of a loved one, and b) my grandmother, a child of The Great Depression, would have flinched to waste a perfectly good hour of chapel time. So we had the funeral.
"But she didn't want a rabbi," I said to my parents. "So who is going to do the service?"
There was a pause on the phone, then Dad said, "you are." (A story for another column.)
Practically speaking, it's unlikely that our coming-real-soon Gen AI BFFs will have complete access to everything we do in the world—everywhere we go, everybody we interact with, everything we see and hear—that's a lot of sensors, cameras, and microphones!
But even if that terrifying total information awareness happens, that won't give an algorithm access to our internal landscapes: the shifting, improvisational, multitudinous throngs of ideas, emotions, impulses, and considered thoughts in our heads. How is an algorithm going to optimize that?
Last thoughts: Nosta's optimistic conclusion also ignores practical issues. What individual will be able to afford and program a personal LLM? The skyrocketing power consumption of Generative AI alone means that most mere mortals wouldn't be able to pay the electricity bill for their new digital BFFs.
Instead, I predict that Gen AI will make digital assistants like Siri, Alexa, and the Google Assistant more powerful and helpful. But their loyalties will be to the trillion dollar companies that create them, Apple, Amazon, Alphabet, not to us.
Thanks for reading. See you next Sunday.
This is good stuff and fits nicely with your recent story, Fleeing the Emerald City. Maybe the personal A.I. will simply present some options to us that our "normal" thoughts wouldn't usually go to. Perhaps a voice of reason when we are angry, or that little voice saying, "This too shall pass."