Categories
Misinformation and AI

Chatbots generate the illusion of freshness

AI platforms promise instant answers, generated in real-time as you ask a question. Chatbots seem poised to render stale, out-of-date webpages a relic of the past. 

Yet chatbot users aren’t getting fresh answers. They are getting “chatbot theater.”

Previously in this series, I have looked at how AI platforms can mislead users.  In this post, I focus on a specific dimension of chatbot misinformation: how they present the recency of their answers.

The urgency for current information 

Online readers more and more prioritize real-time information. Given the rapid pace and unpredictability of events in business and society, information published online risks being overtaken by events.

People require fresh information – old information can be misleading. 

First-hand news is often fresh and immediate compared to third-hand accounts. For example, social media posts announce what people just saw or things that just happened to them. Traditional online publishers like corporate customer service departments or newspapers are slower to reflect changes, if they ever acknowledge them at all.

Despite the appeal of real-time information, most of us don’t receive information exactly when it’s published because we don’t need it at that moment. Only certain types of information (sports scores, stock prices) are suitable for a real-time feed.  In most cases, our goal is to minimize the time gap between when we need information and when it is produced. We aim to maximize the recency of information that matches our interests. 

Online forums are a place for breaking news

Online forums can be an ideal source for recent information in many situations.

Forums don’t just host consumer rants and raves.  Forums have become the frontline of service for many businesses, serving as an essential platform for both customers and employees to get answers. 

Customers rely on forums for product and service recommendations, as well as post-purchase support and self-service.   

Businesses rely on forums to collect comments from customers and employees. Tools like ServiceNow and Slack capture the first-hand experiences users post online. Enterprises are exploring ways to integrate forums with AI for issue monitoring and resolution.

Forums play an overlooked role in online content ecosystems. They deal with unplanned and emergent information — the very kind of recent information people might need to know.

Forums disclose disruptions associated with change. Planned announcements introducing a new product or benefit can be broadcast in a press release. Forums, by contrast, tend to deal with unplanned changes.  

For example, a software update might fix a problem – or create a new one. Supply chain changes might impact product reliability. New management might alter service support.  Customers encounter many unannounced changes — changes that will only generate content once customers notice them.

Now, chatbots seek to replace online forums by offering real-time information generated as soon as people pose a question. It’s an enticing prospect, but unfortunately a deceptive one.

Disaggregating information origins and delivery

Information needs to come from somewhere. Let’s refer to the original source of information as first-hand news. It reflects what an individual with intimate knowledge of an event posts online.  

First-hand news is not always accurate or complete, but when it’s first posted, at least it’s fresh. 

But how do we get first-hand news (fresh information), and at what point does it become third-hand news (stale information)?  The delivery channel shapes this process of revelation.

First, let’s look at how news arrives in an immediate delivery channel such as a feed or a notification.  Here, people receive information as soon as it is posted.  There is no difference between how fresh people perceive the information and the actual age of the information.

Next, let’s consider how a forum works.  Some people post fresh news in forums, and readers may get a notification, which delivers a nearly real-time experience.

But more often, forums deal with questions and answers rather than announcements. One person asks a question, and another responds based on their experience and knowledge.   Even if the question and answer exchange happens quickly and have the same posting dates, the timeframes of the question and answer can be quite different. The questioner typically will ask a question relevant to their current needs, while the answer reflects a past experience. The answer conveys a first-hand experience in the past: a situation the person encountered previously that seems relevant to the current question. 

Q&A forums are hosted in an archive, which can be searched. In this situation, we introduce a third party, the searcher, who is drawing on the previous exchange between the question poster and the person answering. Rather than ask a question themselves (if they have that privilege), they try to determine if someone has already done so. It’s generally good practice (and socially expected behavior) not to ask a question in a forum that’s already been raised and answered. 

When searching for answers in a forum, the searcher encounters two timeframes.  They see a past Q&A exchange and tend to view the posting date “timestamp” as indicating when the information was current.  

But in reality, the basis of the answer posted may be an even earlier experience. If someone asked how to do something, the answer may refer to the process used the last time it was done.  If someone asks whether something is possible, the answer might note that the respondent tried it once in the past and how it worked out for them. 

Each party (seacher, question poster, respondent) is associated with a different point of time.  For example:

  • (Now) The current information seeker looks for answers by doing a search
  • (Last year) A similar question was posted in a forum in the past. It appears to the searcher as if this timestamp is the date of the information. But the answer is based on an earlier experience. 
  • (Two years ago) Respondent had a similar experience related to the question posed in the forum. The far past is the actual basis of the information. 

We can see that the date the answer was posted is not the true age of the information.

Finally, let us consider how chatbots use this information.

Chatbots don’t (yet) have the privilege of asking questions directly of people in forums – they can only answer questions and often rely on previous forum answers to do so.   

Chatbots generate answers that are essentially rewrites of previous answers.

From the questioner’s perspective, the chatbot appears to be generating real-time, up-to-date information.  But in reality, the answer reflects old Q&A conversations. The underlying information could be based on first-hand experiences from the distant past.  Yet, because the questioner does not see the provenance of the information, they are inclined to perceive it as current. 

AI platforms obscure the age of information

AI platforms depend on the answers people have contributed in the past.  But regrettably, they commonly fail to reveal the basis of their answers. 

AI platforms confuse the picture by emphasizing an LLM’s “cutoff” date (they won’t know about events after the cutoff date). They imply that the crawl date is the primary factor in determining content recency. 

Yet, bots now crawl the web frequently to update LLMs, unlike when they first launched. The crawl frequency creates a false impression that a chatbot will provide only the latest information. 

Chatbots struggle to indicate a clear date as of when the information was current.  Clarity of time depends not on the date of the last crawl, but whether LLMs can understand the temporal context of the information they crawl.  Unfortunately, they can’t.  

The root problem is that AI platforms position themselves as the source of information rather than the referrer of information from other sources. They conceal the source of the facts and thwart users from seeing the context of the original information.  

Being dependent on legacy web content, chatbots are unable to generate fresh information. They are stuck rewriting existing information. But they make this rewriting seem as if it provides real-time information. In doing so, they undercut the credibility of the information they offer. 

– Michael Andrews

Categories
Misinformation and AI

AI platforms make crowd-sourced information less intelligent

Chatbots effortlessly answer unremarkable questions.  But we can’t trust them to answer unexpected ones.

I’ve been exploring the complex roots of misinformation in AI. So far in this series, I have noted that online content is full of crowd-sourced information, and that AI platforms depend on crawling online content that’s full of crowd-sourced information.  These dependencies have created a deep vulnerability for AI platforms toward misinformation.

This post looks at how platforms are changing due to AI, and why these changes make crowd-sourced information less useful and informative.  AI platforms are perversely making essential online information less intelligent.

The evolution of platforms 

Platforms emerged to solve the problem of how to access information written by different parties. Platforms aspire to be a one-stop source of information. To deliver that promise, they host information from many sources.

No single organization can publish everything anyone would need to know and provide comprehensive information online that covers every contingency its customers or stakeholders might face. Even when they should ideally be the authoritative source of information about a topic, organizations face the realities of resource constraints. They can only publish about those issues that are most frequently sought or have the highest business impact.  

Outside parties, such as partners or customers, contribute advice and information that authoritative sources don’t have the capacity or inclination to provide.  Long tail information may cover lesser-known details, considerations associated with resolving issues, edge cases, and sometimes, problems companies would prefer not to publicize prominently, such as known problems.  Platforms recognize that users seek such information and aggregate it to make it more easily available. 

Platforms emerged that specialized in offering access to a range of online content. Search platforms collect and rank relevant links to any web page from any website. Ratings platforms like Rotten Tomatoes or Angie’s List collect and host user comments from any commentator. Marketplace platforms such as eBay or Amazon collect customer reviews of products and vendors. GitHub emerged as a platform for discussions about all kinds of code, hosting bug reports, feature requests, and proposed solutions.

Platforms live on content contributed by outside sources. Crowd-sourced content is often characterized as “user-generated”, implying customers write it. Yet platforms also aggregate content from other parties that contribute online content, such as partners, distributors, journalists, critics, and influencers. Some platforms aggregate or syndicate machine-generated data (such as prices, inventories, or schedules) from different sources. 

Platforms aggregate details that no single contributor could develop. The platform assembles a mosaic from many individual pieces. Sometimes the mosaic is complete, though often it’s not. 

Platforms have taken advantage of – and benefited from – the web’s open contribution model.  Anyone can post their views online, and it’s up to readers to decide what’s useful.  Readers vote their preferences by clicking links, which signals the value of the content, which algorithms in turn rank.  Such a ranking is never a perfect process, but at least individual readers played a role in shaping the process.   

AI platforms alter the utility of crowd-sourced information

AI platforms such as ChatGPT and Claude are the latest stage in the evolution of platforms.  Like their predecessors, they pull together information originally contributed by diverse sources and present themselves as a one-stop destination for answers.  But they change what value readers get from the sources. 

Readers value crowd-sourced information according to whether it’s efficient and informative for them.

When there are many contributions, it can be inefficient to read them all.  

But in many situations, each contribution will provide extra perspectives that make reading additional contributions informative. For example, it’s often informative to compare different sources of information about a topic, such as from a vendor and its competitor. It is rarer to rely on a single source of information confidently.

Sifting through numerous postings is an inefficient way of determining undisputed truths, because there’s a lot of redundancy in them. Yet the collective voice of the crowd is informative for complex situations where distinct perspectives contribute to a fuller picture, though the process is still inefficient.  

AI platforms make the process of assessing crowd-sourced content more efficient. But by doing so, it makes the information less informative.

When aggregated, individual insights can be flattened into anodyne statements.  For example, we learn from an AI summary of customer reviews of a bookstore chain branch that the store offers a variety of books – an obvious observation. But AI summaries won’t tell us if the store has many books about philosophy or learning an instrument. We expect computers to produce “intelligence” but find it missing. 

AI platforms damage the quality of crowd-sourced information

Before LLMs, platforms encouraged users to view the original posts. The platform’s role was to act as a clearinghouse that indexes contributed content.

Now, clearinghouse-oriented platforms are morphing into AI platforms. Forums like StackExchange have been replaced by tools such as Copilot and ChatGPT.  

The AI platform transforms the content developed by others, a role I refer to as third-party AI.  The AI platform does not originate the source information nor does it take responsibility for its accuracy.  It operates on the assumption that relevant and accurate information exists within the corpus of content it has crawled. 

AI platforms utilize open web content that is “freely available” (not blocked by paywalls) and repurposable (easily scraped and tokenized). Bots harvest online content and transform it enough to avoid copyright infringement. For AI platforms, online content is a cost-free resource on which to build services.

But source content can only be bent so far before it deforms. 

Long tail information – highly specific information that’s not common knowledge – is most likely to be crowd-sourced.  It is also least likely to be fact-checked, qualified, or maintained.   Crowd-sourced information is incomplete in both its coverage of issues and the scope it addresses for each issue.  An answer you seek may never have been written about.

Imagine you are troubleshooting a software glitch, which could be caused by many factors, such as your hardware, other software you run, the version of software you are using, and so on.  The software vendor doesn’t offer clear information about solving your specific problem, so you turn to an online forum for answers. Others have posted similar problems and offered a range of diverging solutions. Some solutions don’t seem to make sense in your situation, while others don’t work. As far as you can tell, none of the suggestions relates to the exact setup or circumstances you have.   

With crowd-sourced information, it can be challenging to figure out what answers are similar enough to a question. Some problems are perennial, and some are novel. Solutions can be routine or idiosyncratic. Rebooting your computer or clearing your browser cache is common advice that may be helpful sometimes, but often isn’t.

These examples highlight the challenges of matching queries with information in long tail scenarios. Until recently, people needed to vet all the answers one by one to decide which were useful. Now, LLMs promise to do this.

The folly of the crowd in AI platforms

Crowd-sourced content provides vital information not otherwise available, though it is not reliable. Individual contributions can be informative, though they are rarely definitive. When summarized collectively, they become both unspecific and prone to collective biases. 

LLMs are reliable when summarizing ubiquitous, stable knowledge with a high degree of consensus and agreement. A chatbot will confidently tell us the year of US independence from Britain because there’s little controversy about it. 

When everyone knows the same facts or has identical experiences, all crawled text says the same things. There is little need to consult many sources. After all, if everyone has the same opinion or says the same thing, each person’s view adds no new information. 

When bots crawl content and encounter the same information repeated in multiple sources, they infer that the information is likely accurate. Yet, the ubiquity of a statement is not always a reliable proxy for its presumed accuracy. 

Instead of leveraging the “wisdom of the crowd,” bots can fall prey to the “tragedy of the commons”: collective ignorance embedded in past online content. 

Bot answers are anchored in eclectic and unvetted sources that are blended together into a vast corpus. Bots have trouble surfacing information that is not widely known, especially if it is at variance with more common explanations.  

Bot behavior can perpetuate a bias toward legacy content and ideas. Much of the content that bots crawl may contain dated information or unreliable folk knowledge that is widely repeated but misleading. 

Bots misuse content from online forums. Readers find forums useful as places of discovery, not for their past history. Forums are often where new issues first surface. A freshly discovered problem, or an alternative viewpoint, starts as a weak signal that could emerge into a more significant piece of information. But until new issues are widely discussed (and noticed by bot crawlers), they aren’t likely to show up in bot answers. 

The rationale for online platforms is being upended. In the pre-bot era, platforms offered the convenience of gathering different, sometimes diverging perspectives in a single place. Readers could scan for the most relevant or recent information. Contributors had an incentive to post if they felt their statements would be read and noticed. 

Now, bots become the audience and the judge of the value of contributions.  Bots read posts and decide if and how to summarize them for human readers. They are hungry for any content they can access. They can’t be caught not knowing an answer.

Yet chatbots have limited powers of discrimination. They rely on vast quantities of legacy content that may no longer be current.

AI platforms depend on crowd-sourced content to generate answers, but make crowd-sourced content less informative.  

– Michael Andrews