The future is here: Learning to live with conversational AI chatbots like ChatGPT
A chatbot sounds like it can do anything - building realistic expectations is important so you're not frustrated, disappointed, or misinformed. This is "Everything Starts Out Looking Like a Toy" #123.
Hi, I’m Greg 👋! I write essays on product development. Some key topics for me are system “handshakes”, the expectations for workflow, and the jobs we expect data to do. This all started when I tried to define What is Data Operations?
This week’s toy: one of the stranger things I’ve seen someone do with ChatGPT is to teach it to invent and translate a made-up language. This is how you teach an AI to speak the language of slime. Maybe my dreams of Klingon fluency are closer than I thought.
Edition 123 of this newsletter is here - it’s December 12, 2022.
The Big Idea
A short long-form essay about data things
⚙️ What goals should we have for Conversational-style AI bots like ChatGPT?
Imagine you can ask for anything you want. A picture of a unicorn eating a hamburger. An essay on the relative merits of a hot dog and whether it is a sandwich. A recipe that hallucinates a new combination of Frito pie.
Perhaps your mind turns to the process of solving problems. Thinking about potential solutions, you might want suggestions for the best way to introduce a new product feature. You recently read an article on Michael Porter’s competitive forces and you need a succinct 5 paragraph essay explaining these ideas. Maybe you are stuck looking at an empty screen while you try to compose a persuasive 50 word email that fills only one screen of a mobile phone.
Conversational-style chatbots like ChatGPT can help you break this creative block. All of these suggestions above could be realized with this machine learning tool that uses a generative process to find the combination of text that matches text (characters) nearby the text that you provide. I’m not a machine learning engineer, so I probably got the description a bit off. The point is that this technology behaves a bit like a magical genie that can take your prompt, examine its corpus of data for content that seems similar, and deliver human-readable content that is really quite good.
What is “good” for a chatbot?
Asking the question: “what is a chatbot good at” might be the wrong question. Since we’ve never had a chatbot quite like the ones that have emerged over the last month I’m not sure what questions to ask. To start, I feel we need to understand the current limitations of this technology to better understand how to use it responsibly.
What are chatbots good at doing?
Chatbots like ChatGPT are good at creating generalized summaries of information that has a lot of detail in the body of documents indexed by the model. For example, you can ask ChatGPT to summarize the article I mentioned earlier on Michael Porter’s competitive forces.
It does a pretty good job of describing Porter’s forces, and can also compare this data against other well-known management theories.
These Chatbots are also solid at imaginative templating of content with a lot of training data. For example, Sitcom television has lots of screenplays with dialog that describe how individual characters speak, so it’s possible to get unusual situations overlaid onto cultural icons such as Star Trek, Seinfeld, or other pop culture that has generated interest on the Internet.
One goal for conversational chat should be summarizing information that is relatively unstructured while creating a linked bibliography of sources. If we ask an agent to build a summary, we ought to know why it chose the summary items it did. Then, you can commercialize this sort of technology on a very narrow set of training data, e.g. Deepscribe for medical notation or Sybill to track emotional intelligence on sales calls.
Without a trained master model, AI chatbots might hallucinate stuff.
What are chatbots lousy at doing?
Extending logic beyond a single instruction set - they can string together prompts but it’s unpredictable
In the current iteration, they need a human operator to follow tasks (this is probably good)
Chatbots struggle to take intention from a few sets of instructions. The fact that we can even suggest this would be possible means there is a huge market here. But be cautious. I’m not suggesting yet that an untrained operator will be able to use chatbots to solve unsolvable problems. I’m suggesting that improving (even by a small amount) the ability of people to follow most workflow processes alone using faceted conversation makes Chatbots inevitable.
It just saves work everywhere you put it, even though the experience on average for some people will be much much worse when they don’t know how to answer the bot.
A better idea: make conversational chat smarter
Conversational chat is a concept that emerged about 5 years ago as sort of a next-level IVR (interactive voice response, for those who never experienced the horror of an unending phone tree). The goal of that interaction was to take a conversation and reduce it to a series of closed questions (facets) making it easier to draw a graph of potential outcomes from a process.
For example, when you answer a form the prompts are often organized in sequence. This sequence is partly built as a trick to progressively reveal information (validate your phone number, look up information about you, etc) and partly because these systems are not great at figuring out what you need when you don’t deliver the prompts in the right sequence for the process.
Smarter versions can interpret data types and take information slightly out of sequence. Natural Language Processing can cluster information into a likeliness score to known prompts and direct people along a tree of known answers. ChatGPT and its children and grandchildren will change all of this drastically.
My first instinct upon understanding that GPTChat can mimic templates was to think about the kind of templates that would make responses fun. And then, to think about the kind of templates that represent grids or tables we often build in a rote fashion but are kind of a pain (like building all of the test cases for the future, considering all of the combinations and permutations.)
Instead of rote templates, we should focus on the line of questioning that will help chatbots improve and reflect the intent of the questioner.
Some skills and features that need to be trained into bots like these include:
Building Empathy - tuning AI to respond when people are upset and helping to use language and patterns that can calm down emotionally charged situations
Training the user - helping the conversation bot to know how you normally like to chat. This also presents privacy challenges, so here are a few ideas how to handle this at different exposure levels
Answer a few questions, forget after every interaction
Build a list of public preferences that can be indexed by every bot
Train the bot on a local corpus of information and do not allow your questions to reach the outside world (this one’s interesting but risky (imagine a public key/private key handshake to allow this information to be used sparingly by an API)
How would you know whether any of these methods are successful? The same way you test any A/B experiment. Track the outcomes and determine whether they are successful based on the customer’s stated intent.
What are the downsides of this technology?
Any new technology that is sufficiently advanced has the capability to be misunderstood as being sentient. Benn Stancil wrote a great piece this week where he talked about some of the dangers of a chatbot that sounds really impressive even though it has the possibility of spouting incorrect information easily.
I'm starting to think that we need defensive AI trained both on our personal preferences and delivery preferences. This might mean helping me to reduce the sort of content that doesn’t have highly ranked sources or comes from people I don’t know, avoiding sourcing content from certain sites, and implementing a block list of content.
In addition, it would be very valuable to train ChatGPT on my own private corpus of data. The intent here is not just to be able to protect searches (whether they are sensitive or not, they involve lots of metadata that is personally identifiable), but also to create a version of ChatGPT trained for the way I write and process data, not the average outcome. If this worked I also might trust it to deliver new content on a schedule (almost like a newsletter).
The cost to serve an AI result is high today (10x-100x a regular search result) and too slow. But it’s going to get faster, and soon.
Without this layer of filtering, it will be easy to be overwhelmed by a layer of AI-generated BS that hallucinates something that sounds familiar to us. If you’re not familiar with Clay Shirky’s 2008 classic talk on Information Overload/Filter Failure, here it is. It seems even more prescient now.
What’s the takeaway? The technology in generative chatbots like GPTChat is too attractive to ignore. Even if all it ever does is make relatively dumb processes incrementally better, the improvement potential is vast. Conversational chatbots are going to be everywhere, so we need to learn how to use them well.
Links for Reading and Sharing
These are links that caught my 👀
1/ The real metaverse - the trend of people on the Internet to invent personas, build AI-driven selfies, and create themselves online might be the real metaverse. We might not need headsets, VR, or anything else to put us in invented spaces where we prefer to control the information around us.
2/ Real-time de-aging and aging - The future of generated video is going to get even weirder than we think. Combine the hallucinatory powers of GPT tools with this tech from Disney that can de-age and age actors in real time and videos to create an astonishing array of deep fakes, new roles for existing actors, and perhaps roles for made-up ones.
3/ On the paradox of speed - Sahil Bloom shares some excellent observations on speed. It can both speed you up and slow you down.
What to do next
Hit reply if you’ve got links to share, data stories, or want to say hello.
Want more essays? Read on Data Operations or other writings at gregmeyer.com.
The next big thing always starts out being dismissed as a “toy.” - Chris Dixon