Categories
Intelligent Content

AI won’t fix organizational problems

Content can’t do its job if executives and employees don’t support it. A lack of support signals organizational problems relating to process, ownership, and accountability. Such problems are sometimes invisible or willfully ignored because they are gnarly to deal with. The problems manifest when content becomes difficult to manage and yields disappointing results.

AI has emerged as a quick fix to organizational problems. If employees aren’t supporting enterprise content, maybe tireless, uncomplaining automation can. It’s a seductive proposition, promising more free time for all concerned. Yet bots can’t solve human-created problems.

We see surveys reporting enterprise disappointment with AI initiatives. Of AI consuming a lot of investment but having little impact. Much of this disappointment stems from expecting AI to fix organizational problems. Successful AI implementations require sound processes and governance. AI fails when content isn’t well-maintained.

Organizational problems related to content come in many forms:

  • Unclear ownership of content
  • Missing or inconsistent rules
  • Incomplete quality standards
  • Uncertainty about when content is “done”
  • Missing processes to review draft or published content
  • Conflicting priorities within the organization
  • Vague goals or objectives for content items

Content needs maintenance before, during, and after publication.

I just finished reading Stewart Brand’s recent book, Maintenance of Everything. Brand’s fascinating book is about the maintenance of complex physical objects like sailboats, motorcycles, and tanks. But it has numerous insights relevant to AI implementation.

Brand created the Whole Earth Catalog in the 1960s as a guide to fixing and improving things. Some Brandian insights can help us understand why content maintenance is important for AI implementation.

Never paint rust.

Paint protects steel from rusting. But if the steel is already rusting, the paint covers over the problem rather than improving it. The same is true of AI. It might help maintain content’s health, but it can’t magically fix content with hidden quality problems, such as incorrect, irrelevant, or out-of-date information. You need to first understand how those problems arose to begin with.

Repair is nearly always a disruptive intervention in an intricate system.

Fixing content problems isn’t effort-free, which is why they can’t be offloaded to bots. This reality is a rude shock to anyone expecting AI to make work instantly easier.

People need to care about the problems and spend time thinking about how they want to improve things. It requires a commitment, which necessitates spending less time on other priorities.

“‘Repair’ represents a subset of maintenance activities that occurs after a failure. Maintenance includes repair, but also activities associated with keeping the system from failing.”

Don’t expect to fix your content after starting an AI project. It will be too late. You have already set yourself up for failure. A lack of clarity about what needs to be done — and by whom — will become more evident once you try to automate tasks.

“Read the fucking manual!” By which they mean: Part of taking proper ownership of something is to study its manual first

If a process isn’t documented, it isn’t owned by anyone. If the documentation isn’t read or used, then the process is offline and out of service. If an AI agent can’t read a manual on what to do, it will either do nothing or make something up. AI agents can’t own tasks because they aren’t accountable — only people can be held accountable.

Pirsig [in Zen and the Art of Motorcycle Maintenance] proposes that to become expert at keeping anything in good repair, you need to understand it in two ways: how it works and how it’s made.

In large organizations, employees can be unclear about how their content is developed and published. They just do something and pass the baton to the next person. If a problem comes up, they ask for help on Slack. Pity the bot trying to take over that process. How is it supposed to know what to do?

On second thought, don’t pity the bot. The employee needs empathy because the AI tooling is a black box and unfathomable. If it gets things wrong, how is the employee expected to fix something they don’t understand?

Accountability depends on transparency. A spaghetti trail of agents and tools won’t provide that.

Maintaining a horse is different from maintaining a car or a bicycle.

A horse, a car, and a bicycle all provide transport, but some rely on muscle power and some on mechanical power. But when lumped together, we conflate them as units of horsepower.

The same dynamic is happening with AI. Bots are being anthropomorphized, much the same way engineers skeuomorphized UI elements to mimic physical objects. The AI industry has adopted human resources vocabulary to describe bots. Bots are now customer service agents, seemingly indistinguishable from human agents. AI gets training, just like employees. But remember: bots aren’t people. They have different needs. They don’t learn skills; they need their instructions retold each time.

The power to maintain is the power to improve.

Content maintenance may sound tedious, but it is the foundation for improving outcomes. Maintenance needs to be an explicit goal, but it is an intermediary one, not an end goal. It enables digital transformation and supports AI implementation. It allows for organizational growth.

Sustainability is merely a goal, whereas sustainment is a plan, a program, a set of actions

AI promises to scale everything. But already, we see this is a false hope because it presumes a missing foundation. And the more an organization expects to scale using AI, the more important it is that its content infrastructure is reliable.

AI won’t be successful if employees are set up to reactively fix problems surfaced by a fragile AI implementation. The implementation needs to be sustainable. And sustainability requires its own process and maintenance, which is known as sustainment.

While AI technologies remain imperfect and immature, they can be useful — but only if they are given a fair chance. AI can be set up to fail.

But AI can perform an unexpected miracle too. AI implementations can force organizations to fix long-festering problems that would never otherwise get the attention they need.

— Michael Andrews

Categories
Content Engineering

How will bots see your content?

Your customers aren’t that into your website anymore. Most websites have noticed a drop in traffic as users query bots and bots supply answers. Bots generate few clicks to web pages, and the proportion of referral clicks seems to be falling.

Web publishers are aware of the existential threat they face. So far, they’ve tried to make themselves more lovable for bots. They scheme to get noticed by bots (GEO – generative engine optimization). Or they try to make their pages “friendlier” for bots (Google’s WebMCP is the latest example). The legacy thinking still frames the problem as one of visibility — getting noticed in a crowd.

Yet bots aren’t people, and don’t need to be wooed. The old psychology of wooing is no longer relevant. If bots need something, they will take it from your website, whether you invite them or not. In many cases, they will take content even if you don’t want them to.

The problem websites must solve now is how to ensure bots extract the right content from your site. And your existing HTML content, built for web browsers and surfers, isn’t what bots need, if your organization cares about ensuring the accuracy and relevance of what bots provide. JavaScript, the foundation of websites, is a liability for bots.

AI platforms are evolving quickly. They are pivoting away from indiscriminate web scraping for “training” and towards RAG, where they search first for information before generating answers. AI platforms have also embraced the Model Context Protocol (MCP) standard, which, when enabled, allows them to access enterprise content directly. Already, third-party MCP platforms such as Scite and Tollbit have emerged to connect content publishers with AI platforms.

Publishers will continue to publish webpages for human readers, but they need to ensure that AI platforms access the right content for bot users. The best practices for doing this are still emerging, and several initiatives are underway to define protocols and standards.

What’s becoming apparent is that MCP will play an important role in controlling bot access and content governance. The diagram below illustrates a potential content pipeline for a scholarly publisher. A similar pipeline might be adopted by a website publisher — but some additional steps are needed to transform HTML-centric content into bot-ready content.

Example pipeline. Source: Scholarly Kitchen

How are publishers getting ready? Let’s look at how Tollbit helps web publishers. Tollbit works with the Associated Press and other publishers to make their content ready for AI platforms.

The first task is to “clean” the web content to remove material that’s not relevant or canonical. This can be done through DOM filtering to exclude certain classes of content, such as navigation text, promotional assets, or customer comments.

Additional filtering can be done by excluding pages or directories that are procedural or administrative rather than substantive in focus.

Next, the content should be transformed by removing clunky HTML tags to convert the content into a bot-readable format. Many organizations opt to convert content into Markdown, which preserves heading hierarchies (useful for bots) while striping away extraneous markup that bots don’t need.

Bots benefit from metadata, but need help identifying it. The content transformation process should address metadata that’s not visible to human readers. This includes descriptive metadata (such as schema.org) about the content for external systems like search engines, and internal administrative and technical metadata (such as geolocation coordinates) used for web page delivery. This conversion, known as re-serialization, makes the metadata queriable. The metadata can be “hydrated” into the bot’s payload.

AI platforms, ever motivated to increase the sophistication of their products, will take advantage of these content enhancements.

Getting content “bot-ready” will become crucial as AI platforms expand their agentic capabilities. Publishers will need to define access rights and permissions. What materials can bots read, re-publish, or process?

Publishers will shape these affordances through both explicit statements and implicit decisions that influence the ease with which bots can perform actions.

— Michael Andrews