Categories
Misinformation and AI

Chatbots must consider the role of sources, but don’t

Many chatbot problems stem from their inability to understand context. In earlier posts in this series, I’ve discussed how aggregation flattens nuances in individual statements and how scraped content can disregard the timeframes for which the original source statements applied.  

This post explores how user context affects the statements LLMs use to generate answers. It argues that essential context is routinely omitted from statements crawled by AI platforms and, as a result, is not included in chatbot responses. Notably, chatbots don’t consider the point of view of sources expressed in statements they draw upon.

AI platforms harvest online information that’s been stripped of its original context. Bots omit essential context by ignoring the role of the source posting the information.

Information accuracy is often highly contingent on its circumstances. While most online information was reasonably accurate at some point, it may be accurate only in specific circumstances. It can be described as “yes, it is (or was) true, but only if or when a particular circumstance is true.” These qualifications extend to who is making an assertion and what their role is. Although people do lie online, the bigger problem is that they misunderstand and miscommunicate. Bots struggle even more than humans with these ambiguities.   

When assessing the credibility of information, readers must consider the circumstances of the person providing the information. They are interested not only in who said something, but also in their role.

We are accustomed to distinguishing between primary and secondary sources from years of schooling. We separate direct statements by people from indirect ones, where they are quoted or summarized.  We focus on who said something. 

Google recommends users search for information about sources they find online.

It’s important to look beyond the naive idea that sources have either a good or a bad reputation.  Many platforms make simplistic assumptions about whether a source is trustworthy, without regard to the scope or domain of the topic.  Contrary to SEO folklore, authority online isn’t an attribute of a website; it is intrinsically related to the topic of the content itself.

People and platforms should look more broadly at how information originates.  

First-party and third-party information are similar to primary and secondary sources in that both concepts distinguish different categories of sources. But the concepts are slightly different. Instead of focusing only on who said something (the source), we also consider their authority to speak about what is said (the information).  

In online forums, that rich source of advice, reviews and updates, first-person observations can be third-party information – someone’s interpretation.  For example, John might post in an online forum that the IRS doesn’t allow a certain deduction because he wasn’t able to take it himself. But John doesn’t work for the IRS (which isn’t noted for posting helpful advice in online forums). He is only conveying his personal experience. The issue is not necessarily John’s credibility or knowledge – he’s candid about what he knows, as far as he knows it.  And read carefully, John’s post may offer useful information for understanding how some taxpayers are able to take deductions or not.  But John’s post can’t be taken as the universal truth. 

First-hand statements are not first-party information unless they are made by someone who works for the organization that decides the information. An individual’s views can be first-hand and appear credible but not authoritative, as they involve interpretations, opinions, or experiences. Statements can be true as they relate to the individual’s circumstances, but not be correct if taken as global statements that apply to all situations. 

Information provenance leads to an important qualification: eyewitness accounts are not the absolute truth. 

This scepticism challenges the widely cherished idea that first-hand experiences provide the unvarnished truth.  But in reality, experiences expressed online offer at best a limited truth that’s constrained by the circumstance of when, where, and who said it.   

Chatbots can’t discern the context of the information they crawl. Even Google’s Gemini chatbot doesn’t follow Google’s guidelines for humans to investigate “why it’s sharing that info.”  Gemini offers a blanket disclaimer, “AI responses may include mistakes.”  It’s up to the human to figure out if the chatbot made mistakes and what those mistakes might be. 

Chatbots have trouble distinguishing between third-hand and first-hand information. I’ll return to an example I raised in an earlier post in this series about finding a vegetarian restaurant while on vacation. Platforms scrape reviews, which can be misleading when someone mentions the word “vegetarian” in passing, even if it’s just a general comment.  That’s an example of the unreliability of third-party information. The restaurant never made this claim. 

Every time third-party information is used, someone else’s assumptions are being applied.

If platforms were scraping restaurants’ menus and could decipher which dishes were vegetarian, they would be relying on first-party information.  If, however, the platform were deciding if the dish was vegetarian based solely on its name, we would be back to third-party information. The bot interprets menu names using third-party information to determine whether a dish is vegetarian. But many vegetable dishes have bacon or chicken stock in them, which won’t be apparent from the name of the dish. So even with first-hand information, the full context may be missing. 

Textual declarations seldom explicitly qualify the limitations of a statement – the reader is expected to infer any limitations from the context in which the declaration is made.  Bots, however, tend to decontextualize statements and make them into universal ones.  Bot-generated statements derived from crowd-contributed content are often misleading. 

Your experience may vary

The source’s identity will reflect their role: what matters to them and what they know about a situation. Various people can make statements that are inconsistent but nonetheless valid for them individually.

Online forums are where people share stories about themselves. A person will write in a forum about “what I did, and what worked for me”, with little initial consideration of how readers might be in different circumstances. Such egocentricity reflects the incentives and motivations of crowd-contributed forums.  People enjoy talking about themselves and believe they are influencing others to emulate them. They enjoy getting praise and recognition when they post something deemed notable that hasn’t been seen before.   

The individual posts that bots crawl contain sampling biases (the advice in each post is a sample of one).  People write about what they did – what they considered and tried. Rarely do they write about having tried all possibilities and evaluated them. The information is selective. 

When all parties view communication as a point-to-point exchange, each party strips out the context they deem unnecessary. They emphasize what they want to know rather than spending much time discussing what others may know. The information tends to be personal.

The writer of advice and the seeker of advice can have different preference profiles.  The “best way” to do something depends heavily on the situation and individual preferences. For many tasks, determining the best approach can be challenging without understanding who, when, and why someone wants to undertake the task.  

The challenges of human communication are magnified online, where distance in time and space makes clarification and qualification of statements much harder.

Even with these challenges, many forum participants want to help and may clarify statements in subsequent threads, especially when questions arise.

But bots crawl online forums with a more acquisitive agenda.  They are indifferent to the discussion’s context.  They simply want to harvest statements made.  Whereas humans may engage in a close reading of the discussion, bots engage in a distant reading of it.

The problem is that much of the context shaping what’s said online is never explicitly stated, and if it is revealed, it may be noted later in the discussion. 

Where context is omitted, gaps in understanding emerge. The writer’s context may not be transparent (even to the writer).  The reader’s context – their preferences and circumstances – may be unknown to the writer.  The bot, driven by its mission to scrape the discussion, is indifferent to the context. 

The phantom of contextual AI 

The omission of context in crawled online content poses a formidable challenge to the growth and development of AI.  

The latest wave of AI development is focused on agents that use the Model Context Protocol. Context is essential for AI, but chatbots can’t supply the context needed.

There’s no simple fix for the omission of context in online information.

Content professionals often champion the importance of context in supplying relevant information.  Many argue that contextual metadata should be added to source statements to enable bots to provide high-quality answers. Approaches such as GraphRAG are having a moment. Although commendable in principle, applying context to online content after it’s been written is difficult in practice. 

Online content, particularly forum discussions, is not written for machines. People are writing for each other – in some cases, telling stories to themselves. The writer may be blissfully unaware of the limitations of their pronouncements and how those pronouncements reflect their personal biases.  

Bots can’t detect the possibility that the facts of the matter may be specific to what the individual experienced in a given context.  Omitted context can’t be auto-magically restored.  

Yes, some context can be applied after the fact with automated tags.  Yet, realistically, much of the context of online content requires close human reading to infer.  Bots process text superficially, relying on relatively crude tools such as keyword and entity recognition, which are no match for the inherent ambiguity of most online discussions. 

– Michael Andrews

Categories
Misinformation and AI

Chatbots generate the illusion of freshness

AI platforms promise instant answers, generated in real-time as you ask a question. Chatbots seem poised to render stale, out-of-date webpages a relic of the past. 

Yet chatbot users aren’t getting fresh answers. They are getting “chatbot theater.”

Previously in this series, I have looked at how AI platforms can mislead users.  In this post, I focus on a specific dimension of chatbot misinformation: how they present the recency of their answers.

The urgency for current information 

Online readers more and more prioritize real-time information. Given the rapid pace and unpredictability of events in business and society, information published online risks being overtaken by events.

People require fresh information – old information can be misleading. 

First-hand news is often fresh and immediate compared to third-hand accounts. For example, social media posts announce what people just saw or things that just happened to them. Traditional online publishers like corporate customer service departments or newspapers are slower to reflect changes, if they ever acknowledge them at all.

Despite the appeal of real-time information, most of us don’t receive information exactly when it’s published because we don’t need it at that moment. Only certain types of information (sports scores, stock prices) are suitable for a real-time feed.  In most cases, our goal is to minimize the time gap between when we need information and when it is produced. We aim to maximize the recency of information that matches our interests. 

Online forums are a place for breaking news

Online forums can be an ideal source for recent information in many situations.

Forums don’t just host consumer rants and raves.  Forums have become the frontline of service for many businesses, serving as an essential platform for both customers and employees to get answers. 

Customers rely on forums for product and service recommendations, as well as post-purchase support and self-service.   

Businesses rely on forums to collect comments from customers and employees. Tools like ServiceNow and Slack capture the first-hand experiences users post online. Enterprises are exploring ways to integrate forums with AI for issue monitoring and resolution.

Forums play an overlooked role in online content ecosystems. They deal with unplanned and emergent information — the very kind of recent information people might need to know.

Forums disclose disruptions associated with change. Planned announcements introducing a new product or benefit can be broadcast in a press release. Forums, by contrast, tend to deal with unplanned changes.  

For example, a software update might fix a problem – or create a new one. Supply chain changes might impact product reliability. New management might alter service support.  Customers encounter many unannounced changes — changes that will only generate content once customers notice them.

Now, chatbots seek to replace online forums by offering real-time information generated as soon as people pose a question. It’s an enticing prospect, but unfortunately a deceptive one.

Disaggregating information origins and delivery

Information needs to come from somewhere. Let’s refer to the original source of information as first-hand news. It reflects what an individual with intimate knowledge of an event posts online.  

First-hand news is not always accurate or complete, but when it’s first posted, at least it’s fresh. 

But how do we get first-hand news (fresh information), and at what point does it become third-hand news (stale information)?  The delivery channel shapes this process of revelation.

First, let’s look at how news arrives in an immediate delivery channel such as a feed or a notification.  Here, people receive information as soon as it is posted.  There is no difference between how fresh people perceive the information and the actual age of the information.

Next, let’s consider how a forum works.  Some people post fresh news in forums, and readers may get a notification, which delivers a nearly real-time experience.

But more often, forums deal with questions and answers rather than announcements. One person asks a question, and another responds based on their experience and knowledge.   Even if the question and answer exchange happens quickly and have the same posting dates, the timeframes of the question and answer can be quite different. The questioner typically will ask a question relevant to their current needs, while the answer reflects a past experience. The answer conveys a first-hand experience in the past: a situation the person encountered previously that seems relevant to the current question. 

Q&A forums are hosted in an archive, which can be searched. In this situation, we introduce a third party, the searcher, who is drawing on the previous exchange between the question poster and the person answering. Rather than ask a question themselves (if they have that privilege), they try to determine if someone has already done so. It’s generally good practice (and socially expected behavior) not to ask a question in a forum that’s already been raised and answered. 

When searching for answers in a forum, the searcher encounters two timeframes.  They see a past Q&A exchange and tend to view the posting date “timestamp” as indicating when the information was current.  

But in reality, the basis of the answer posted may be an even earlier experience. If someone asked how to do something, the answer may refer to the process used the last time it was done.  If someone asks whether something is possible, the answer might note that the respondent tried it once in the past and how it worked out for them. 

Each party (seacher, question poster, respondent) is associated with a different point of time.  For example:

  • (Now) The current information seeker looks for answers by doing a search
  • (Last year) A similar question was posted in a forum in the past. It appears to the searcher as if this timestamp is the date of the information. But the answer is based on an earlier experience. 
  • (Two years ago) Respondent had a similar experience related to the question posed in the forum. The far past is the actual basis of the information. 

We can see that the date the answer was posted is not the true age of the information.

Finally, let us consider how chatbots use this information.

Chatbots don’t (yet) have the privilege of asking questions directly of people in forums – they can only answer questions and often rely on previous forum answers to do so.   

Chatbots generate answers that are essentially rewrites of previous answers.

From the questioner’s perspective, the chatbot appears to be generating real-time, up-to-date information.  But in reality, the answer reflects old Q&A conversations. The underlying information could be based on first-hand experiences from the distant past.  Yet, because the questioner does not see the provenance of the information, they are inclined to perceive it as current. 

AI platforms obscure the age of information

AI platforms depend on the answers people have contributed in the past.  But regrettably, they commonly fail to reveal the basis of their answers. 

AI platforms confuse the picture by emphasizing an LLM’s “cutoff” date (they won’t know about events after the cutoff date). They imply that the crawl date is the primary factor in determining content recency. 

Yet, bots now crawl the web frequently to update LLMs, unlike when they first launched. The crawl frequency creates a false impression that a chatbot will provide only the latest information. 

Chatbots struggle to indicate a clear date as of when the information was current.  Clarity of time depends not on the date of the last crawl, but whether LLMs can understand the temporal context of the information they crawl.  Unfortunately, they can’t.  

The root problem is that AI platforms position themselves as the source of information rather than the referrer of information from other sources. They conceal the source of the facts and thwart users from seeing the context of the original information.  

Being dependent on legacy web content, chatbots are unable to generate fresh information. They are stuck rewriting existing information. But they make this rewriting seem as if it provides real-time information. In doing so, they undercut the credibility of the information they offer. 

– Michael Andrews