Categories
Content Experience

Should Information be Data-Rich or Content-Rich?

One of the most challenging issues in online publishing is how to strike the right balance between content and data.  Publishers of online information, as a matter of habit, tend to favor either a content-centric, or a data-centric approach.  Publishers may hold deep seeded beliefs about what form of information is most valuable.  Some believe that compelling stories will wow audiences. Others expect that new artificial intelligence agents, providing instantaneous answers, will delight them. This emphasis on information delivery can overshadow consideration of what audiences really need to know and do. How information is delivered can get in the way of what the audience needs. Instead of delight, audiences experience apathy and frustration. The information fails to deliver the right balance between facts, and explanation.

The Cultural Divide

Information on the web can take different forms. Perhaps the most fundamental difference is whether online information provides a data-rich or content-rich experience. Each form of experience has its champions, who promote the virtues of data (or content).  Some go further, and dismiss the value of the approach they don’t favor, arguing that content (or data) actually gets in the way of what users want to know.

  • A (arguing for data-richness): Customers don’t want to read all that text!  They just want the facts.  
  • B (arguing for content-richness): Just showing facts and figures will lull customers to sleep!

Which is more important, offering content or data?  Do users want explanations and interpretations, or do they just want the cold hard facts?  Perhaps it depends on the situation, you think.  Think of a situation where people need information.  Do they want to read an explanation and get advice, or do they want a quick unambiguous answer that doesn’t involve reading (or listening to a talking head)?  The scenario you have in mind, and how you imagine people’s needs in that scenario, probably reveals something about your own preferences and values.  Do you like to compare data when making decisions, or do you like to consider commentary?  Do your own PowerPoint slides show words and images, or do they show numbers and graphs? Did you study a content-centric discipline such as the humanities in university, or did you study a data-centric one such as commerce or engineering? What are your own definitions of what’s helpful or boring?

Our attitudes toward content and data reflect how we value different forms of information.  Some people favor more generalized and interpreted information, and others prefer specific and concrete information.  Different people structure information in different ways, through stories for example, or by using clearly defined criteria to evaluate and categorize information.  These differences may exist within your target audience, just as they may show up within the web team trying to deliver the right information to that audience.  People vary in their preferences. Individuals may shift their personal  preferences depending on topic or situation.  What form of information audiences will find most helpful can elude simple explanations.

Content and data have an awkward relationship. Each seems to involve a distinct mode of understanding.  Each can seem to interrupt the message of the other. When relying on a single mode of information, publishers risk either over-communicating, or under-communicating.

Content and Data in Silhouette

To keep things simple (and avoid conceptual hairsplitting), let’s think about data as any values that are described with an attribute.  We can consider data as facts about something.  Data can be any kind of fact about a thing; it doesn’t need to be a number. Whether text or numeric, data are values that can be counted.

Content can involve many distinct types, but for simplicity, we’ll consider content as articles and videos — containers  where words and images combine to express ideas, stories, instructions, and arguments.

Both data and content can inform.  Content has the power to persuade, as sometimes data can possess that power as well.  So what is the essential difference between them?  Each has distinct limitations.

The Limits of Content

In certain situations content can get in the way of solving user problems.  Many times people are in a hurry, and want to get a fact as quickly as possible.  Presenting data directly to audiences doesn’t always mean people get their questioned answered instantly, of course.  Some databases are lousy answering questions for ordinary people who don’t use databases often.  But a growing range of applications now provide “instant answers” to user queries by relying on data and computational power.  Whereas content is a linear experience, requiring time to read, view or listen, data promises instant experience that can gratify immediately.  After all, who wants to waste their customer’s time?  Content strategy has long advocated solving audience problems as quickly as possible.  Can data obviate the need for linear content?

“When you think about something and don’t really know much about it, you will automatically get information.  Eventually you’ll have an implant where if you think about a fact, it will just tell you the answer.”  Google’s Larry Page, in Steven Levy’s  “In the Plex”.

The argument that users don’t need websites (and their content) is advanced by SEO expert Aaron Bradley in his article “Zero Blue Links: Search After Traffic”.   Aaron asks us to “imagine a world in which there was still an internet, but no websites. A world in which you could still look for and find information, but not by clicking from web page to web page in a browser.”

Aaron notes that within Google search results, increasingly it is “data that’s being provided, rather than a document summary.”  Audiences can see a list of product specs, rather than a few sentences that discuss those specs. He sees this as the future of how audiences will access information on different devices.  “Users of search engines will increasingly be the owners of smart phones and smart watches and smart automobiles and smart TVs, and will come to expect seamless, connected, data-rich internet experiences that have nothing whatsoever to do with making website visits.”

In Aaron’s view, we are seeing a movement from “documents to data” on the web. “The evolution of search results in terms of the gradual supplanting of document references by data than it is to infer that direction through the enumeration of individual features.”  No need to read a document: search results will answer the question.  It’s an appealing notion, and one that is becoming more commonplace.  Content isn’t always necessary if clear, unambiguous data is available that can answer the question.

Google, or any search engine, is just a channel — an important one for sure, but not the end-all and be-all.  Search engines locate information created by others, but unless they have rights to that information, they are limited in what they do with it. Yet the principles here can apply to other kinds of interactive apps, channels and platforms that let users get information instantly, without wading through articles or videos.  So is content now obsolete?

There is an important limitation to considering SEO search results as data.  Even though the SEO community refers to search metadata as “structured data”, the use of this term is highly misleading.  The values described by the metadata aren’t true data that can be counted.  They are values to display, or are links to other values.  The problem with structured data as currently practiced is that is doesn’t enforce how the values need to be described.  The structured data values are never validated, so computers can’t be sure if two prices appearing on two random websites are both quoting the same currency, even if both mention dollars.  SEO structured data rarely requires controlled vocabulary for text values, and most of its values doesn’t include or mandate data typing that computers would need to aggregate and compare different values.  Publishers are free to use most any kind of text value they like in many situations.   The reality of SEO structured data is less glamorous than its image: much of the information described by SEO structured data is display content for humans to read, rather than data for machines to transform.  The customers who scan Google’s search results are people, not machines.  People still need to evaluate the information, and decide its credibility and relevance.  The values aren’t precise and reliable enough for computers to make such judgements.

When an individual wants to know what time a shop closes, it’s a no brainer to provide exactly that information, and no more. The strongest cases for presenting data directly is when the user already knows exactly what they want to know, and they will understand the meaning and significance of the data shown.  These are the “known unknowns” (or “knowns but forgotten”) use cases.  Plenty of such cases exist.  But while the lure of instant gratification is strong, people aren’t always in a rush to get answers, and in many cases they shouldn’t be in a rush, because the question is bigger than a single answer can address.

The Limits of Data

Data in various circumstances can get in the way of what interests audiences.  At a time when the corporate world increasingly extols the virtues of data, it’s important to recognize when data can be useless, because it doesn’t answer questions that audiences have.  Publishers should identify when data is oversold, as always being what audiences want.  Unless data reflects audiences priorities, the data is junk as far as audiences are concerned.

Data can bring credibility to content, though has the potential to confuse and mislead as well.  Audiences can be blinded by data when it is hard to comprehend, or is too voluminous. Audiences need to be interested in the data for it to provide them with value.  Much of the initial enthusiasm for data journalism, the idea of writing stories based on the detailed analysis of facts and statistics, has receded.  Some stories have been of high quality, but many weren’t intrinsically interesting to large numbers of viewers.  Audiences didn’t necessarily see themselves in the minutiae, or feel compelled to interact with raw material being offered to them.  Data journalism stories are different from commercially oriented information, which have well defined use cases specifying how people will interact with data.  Data journalism can presume people will be interested in topics simply because public data on these topics is available.  However, this data may be collected for a different purpose, often for technical specialists.  Presenting it doesn’t transform it into something interesting to audiences.

The experience of data journalism shows that not all data is intrinsically interesting or useful to audiences.  But some technologists believe that making endless volumes of data available is intrinsically worthwhile, because machines have the power to unlock value from the data that can’t be anticipated.

The notion that “data is God” has fueled the development of the semantic web approach, which has subsequently been  rebranded as “linked data”.  The semantic web has promised many things, including giving audiences direct access to information without the extraneous baggage of content.  It even promised to make audiences irrelevant in many cases, by handing over data to machines to act on, so that audiences don’t even need to view that data.  In its extreme articulation, the semantic web/linked data vision considers content as irrelevant, and even audiences as irrelevant.

These ideas, while still alive and championed by their supporters, have largely failed to live up to expectations.  There are many reasons for this failure, but a key one has been that proponents of linked data have failed to articulate its value to publishers and audiences. The goal of linked data always seems to be to feed more data to the machine.  Linked data discussions get trapped in the mechanics of what’s best for machines (de-referencable URIs,  machine values that have no intrinsic meaning to humans), instead of what’s useful for people.

The emergence of schema.org (the structured data standard used in SEO) represents a step back from such machine-centric thinking, to accommodate at least some of the needs of human metadata creators by allowing text values. But schema.org still doesn’t offer much in the way of controlled vocabularies for values, which would be both machine-reliable and human-friendly.  It only offers a narrow list of specialized “enumerations”, some of which are not easy-to-read text values.

Schema.org has lots of potential, but its current capabilities get over-hyped by some in the SEO community.  Just as schema.org metadata should not be considered structured data, it is not really the semantic web either.  It’s unable to make inferences, which was a key promise of the semantic web.  Its limitations show why content remains important. Google’s answer to the problem of how to make structured data relevant to people was the rich snippet.  Rich snippets displayed in Google search results are essentially a vanity statement. Sometimes these snippets answer the question, but other times they simply tease the user with related information.  Publishers and audiences alike may enjoy seeing an extract of content in search results, and certainly rich snippets are a positive development in search. But displaying extracts of information does not represent an achievement of the power of data.  A list of answers supplied by rich snippets is far less definitive than a list of answers supplied by a conventional structured query database — an approach that has been around for over three decades.

The value of data comes from its capacity to aggregate, manipulate and compare information relating to many items.  Data can be impactful when arranged and processed in ways that change an audience’s perception and understanding of a topic. Genuine data provides values that can be counted and transformed, something that schema.org doesn’t support very robustly, as previously mentioned.  Google’s snippets, when parsing metadata values from articles, simply display fragments  from individual items of content.  A list of snippets doesn’t really federate information from multiple sources into a unified, consolidated answer.  If you ask Google what store sells the cheapest milk in your city, Google can’t directly answer that question, because that information is not available as data that can be compared.  Information retrieval (locating information) is not the same as data processing (consolidating information).

“What is the point of all that data? A large data set is a product like any other. It must be maintained and updated, given attention. What are we to make of it?”  Paul Ford in “Usable Data

But let’s assume that we do have solid data that machines can process without difficulty.  Can that data provide audiences with what they need?  Is content unnecessary when the data is machine quality?  Some evidence suggests that even the highest quality linked data isn’t sufficient to interest audiences.

The museum sector has been interested in linked data for many years.  Unlike most web publishers, they haven’t been guided by schema.org and Google.  They’ve been developing their own metadata standards.  Yet this project has had its problems.  The data lead of a well known art museum complained recently of the “fetishization of Linked Open Data (LOD)”.  Many museums approached data as something intrinsically valuable, without thinking through who would use the data, and why.  Museums reasoned that they have lots of great content (their collections) and that they needed to provide information about their collections online to everyone, so that linked data was the way to do that.  But the author notes: ‘“I can’t wait to see what people do with our data” is not a clear ROI.’  When data is considered as the goal, instead of as a means to a goal, then audiences get left out of the picture.  This situation is common to many linked data projects, where getting data into a linked data structure becomes an all consuming end, without anchoring the project in audience and business needs.  For linked data to be useful, it needs to address specific use cases for people relying on the data.

Much magical thinking about linked data involves two assumptions: that the data will answer burning questions audiences have, and these answers will be sufficient to make explanatory content unnecessary.  When combined, these assumptions become one: everything you could possibly want to know is now available as a knowledge graph.

The promise that data can answer any question is animating development of knowledge graphs and “intelligent assistants” by nearly every big tech company: Google, Bing, LinkedIn, Apple, Facebook, etc.  This latest wave of data enthusiasm again raises questions whether content is becoming less relevant.

Knowledge graphs are a special form of linked data.  Instead of the data living in many places, hosted by many different publishers, the data is instead consolidated into a single source curated by one firm, for example, Bing. A knowledge graph combines millions of facts about all kinds of things into a single data set. A knowledge graph creator generally relies on other publisher’s linked data. But it assumes responsibility for validating that data itself when incorporating the information in its knowledge graph.  In principle, the information is more reliable, both factually and technically.

Knowledge graphs work best for persistent data (the birth year of a celebrity) but less well for high velocity data that can change frequently (the humidity right now).   Knowledge graphs can be incredibly powerful.  They can allow people to find connections between pieces of data that might not seem related, but are.  Sometimes these connections are simply fun trivia (two famous people born in the same hospital on the same day). Other times these connections are significant as actionable information.  Because knowledge graphs hold so much potential, it is often difficult to know how they can be used effectively.   Many knowledge graph use cases relate to open ended exploration, instead of specific tasks that solve well defined user problems.   Few people can offer a succinct, universally relevant reply to the question: “What problem does a knowledge graph solve?” Most of the success I’ve seen for knowledge graphs has been in specialized vertical applications aimed at researchers, such as biomedical research or financial fraud investigations.  To be useful to general audiences, knowledge graphs require editorial decisions that queue up on-topic questions, and return information relevant to audience needs and interests.  Knowledge graphs are less useful when they simply provide a dump of information that’s related to a topic.

Knowledge graphs combine aspects of Wikipedia (the crowdsourcing of data) with aspects of a proprietary gatekeeping platform such as Facebook (the centralized control of access to and prioritization of information).  No one party can be expected to develop all the data needed in a knowledge graph, yet one party needs to own the graph to make it work consistently — something that doesn’t always happen with linked data.   The host of the knowledge graph enjoys a privileged position: others must supply data, but have no guarantee of what they receive in return.

Under this arrangement, suppliers of data to a knowledge graph can’t calculate their ROI. Publishers are back in the situation where they must take a leap of faith that they’ll benefit from their effort.  Publishers are asked to supply data to a service on the basis of a vague promise that the service will provide their customers with helpful answers.  Exactly how the service will use the data is often not transparent. Knowledge graphs don’t reveal what data gets used, and when.   Publisher also know their rivals are also supplying data to the same graph.  The faith-based approach to developing data, in hopes that it will be used, has a poor track record.

The context of data retrieved from a knowledge graph may not be clear.  Google, Siri, Cortana, or Alexa may provide an answer.  But on what basis do they make that judgment?  The need for context to understand the meaning of data leads us back to content.   What a fact means may not be self-evident. Even facts that seem straightforward can depend on qualified definitions.

“A dataset precise enough for one purpose may not be sufficiently precise for another. Data on the Web may be wrong, or wrong in some context—with or without intent.” Bernstein, Hendler & Noy

The interaction between content and data is becoming even more consequential as the tech industry promotes services incorporating artificial intelligence.  In his book Free Speech, Timothy Garton Ash shared his experience using WolfamAlpha, a semantic AI platform that competes with IBM Watson, and that boldly claims to make the “world’s knowledge computable.”  When Ash asked WolfamAlpha “How free should speech be?”, it replied: “WolframAlpha doesn’t understand your query.”   This kind of result is entirely expected, but it is worth exploring why something billed as being smart fails to understand.  Conversational interfaces, after all, are promising to answer our questions.  Data needs to exist for questions to get answers.  For data to operate independently of content, an answer must be expressible as data. But many answers can’t be reduced to one or two values.  Sometimes they involve many values.  Sometimes answers can’t be expressed as a data value at all. This actuality means that content will always be necessary for some answers.

Data as a Bridge to Content

Data and content have different temperaments.  The role of content is often to lead the audience to reveal what’s interesting.  The role of data is frequently to follow the audience as they indicate their interests. Content and data play complementary roles.  Each can be incomplete without the other.

Content, whether articles, video or audio, is typically linear.  Content is meant to be consumed in a prescribed order.   Stories have beginnings and ends, and procedures normally have fixed sequences of steps.  Hyperlinking content provides a partial solution to making a content experience less linear, when that is desired.  Linear experiences can be helpful when audiences need orientation, but they are constraining when such orientation isn’t necessary.

Data, to be seen, must first be selected. Publishers must select what data to highlight, or they must delegate that task to the audience. Data is non-linear: it can be approached in any order.  It can be highly interactive, providing audiences with the ability to navigate and explore the information in any order, and change the focus of the information.  With that freedom comes the possibility that audiences get lost, unable to identify information of value.  What data means is highly dependent on the audience’s previous understanding.  Data can be explained with other data, but even these explanations require prior  knowledge.

From an audience perspective, data plays various roles.  Sometimes data is an answer, and the end of a task.  Sometimes data is the start of a larger activity.  Data is sometimes a signal that a topic should be looked at more closely.  Few people decide to see a movie based on an average rating alone.  A high rating might prompt someone to read about the film.  Or the person may be already be interested in reading about the film, and consults the average rating simply to confirm their own expectation of whether they’d like it.  Data can be an entryway into a topic, and a point of comparison for audiences.

Writers can undervalue data because they want to start with the story they wish to tell, rather than the question or fact that prompts initial interest from the audience.   Audiences often begin exploration by seeking out a fact. But what that fact may be will be different according to each individual.  Content needs facts to be discovered.

Data evangelists can undervalue content because they focus on the simple use cases, and ignore the messier ones.  Data can answer questions only in some situations.  In an ideal world, a list of questions and answers get paired together as data. Just match the right data with the right question.  But audiences may find it difficult to articulate the right question, or they may not know what question to ask. Audiences may find they need to ask so many specific questions to develop a broad understanding.  They may find the process of asking questions exhausting.  Search engines and intelligent agents aren’t going to Socratically enlighten us about new or unfamiliar topics.  Content is needed.

Ultimately, whether data or content is most important depends on how much communication is needed to support the audience.  Data supplies answers, but doesn’t communicate ideas.  Content communicates ideas, but can fail to answer if it lacks specific details (data) that audiences expect.

No bold line divides data from content.  Even basic information, such as expressing how to do something, can be approached either episodically as content, or atomically as data.  Publishers can present the minimal facts necessary to perform a task (the must do’s), or they can provide a story about possibilities of tasks to do (the may do’s).  How should they make that decision?

In my experience, publishers rarely create two radically alternative versions of online information, a data-centric and content-centric version, and test these against each other to see which better meets audience needs.  Such an approach could help publishers understand what the balance between content and data needs to be.  It could help them understand how much communication is required, so the information they provide is never in the way of the audience’s goals.

— Michael Andrews

Categories
Content Experience

Designing Glanceable Content for Partial Attention

Captive audiences are the exception, not the norm.  Audiences only rarely offer their undivided attention. How to design content for audiences who are only half paying attention is a growing challenge.

Despite wide recognition that most folks have too much on their minds, content producers continue to hope they’ll get the full attention of audiences.  According to conventional advice, if content designers make content simple and relevant, the content will earn the attention of audiences.  The messy, everyday reality that audiences face interferes with their attention, no matter how well considered and crafted the content.  The stark truth is that attention cannot always be earned. It must be bargained with.

Mounting evidence indicates audiences have less attention to dedicate to consuming content actively. Even when they want content, they don’t necessarily want the content to dominate what they are doing.  People want content they can multitask with, content that supports them as they do things like chores, exercise, or driving — instead of being the focus of what they are doing.  The rising popularity of audio books and podcasts are indications of the desire for content that supports multitasking or continuous partial attention. Streaming content doesn’t require active interaction.

Another tactic to compensate for our fragmenting attention is to make content smaller.  Many publishers seek to turn content into bite sized chunks that don’t take long to read.  Audiences nibble bite sized content when they are focused on other things, such as a conversation or a physical activity such as walking or waiting, or when they are mentally preoccupied with matters that may not be directly related to the content.

Bite sized content appears as informational cards displayed on smartphones, on the screens of smart watches, or as the snippets of audio available on voiced controlled audio platforms.  Contrary to common belief, making content bite sized does not mean it demands little attention from audiences, even if the item itself requires little time to read, watch or hear.  Attention is not the same as time to read.  Attention is part of a broader concept of content engagement.

Transcending the Four Standard Categories of Content Engagement

Content typically falls into one of four categories of engagement, according to the attention and interaction they require:

  • Content that requires constant user interface (UI) interaction and full attention.  An Xbox game might fall in this category, where the game paces the player, requiring them to interact and give their full attention.
  • Content that doesn’t require much UI interaction, but requires full attention.  An involved murder mystery story, whether  delivered as a video, audio or text, would be an example.
  • Content that requires UI interaction, but only intermittent attention.  An example is a buzzing notification on a watch or smartphone.  You can’t avoid paying attention to the notification, because it’s in your face.
  • Content that requires no UI interaction, and only intermittent attention.  Glancing at a dull sports match on a TV in a sports bar might qualify.

But another category exists, where there is intermittent attention, and only limited UI interaction.  The content is not zombie-like, where no interaction happens, but neither is it reactive, where the machine is demanding interaction.  Instead, the content allows the user to pay attention electively, and interact electively.  This is glanceable content that offers some choice.

Finding the Threshold of Significance

Chunks of content should be significant to audiences. When audiences can give their full attention, a sequence of small content chunks can be woven into a narrative that provides detailed meaning.  But when audience attention is partial, a content chunk must stand alone to convey meaning.

In some cases, content designers know the context of the audience, and can design small items of content that matches that context.  Notifications, at their best, provide an example of small items of content that match the context of the user.  It is easy to get notifications wrong.  Notifications or alerts tend to poke the user.  They are often action focused, requesting or confirming an action.  Notifications either presume attention, or demand attention.  Notifications can be hectoring, where the machine sets the pace for humans, rather than letting humans set their own pace.

In contrast to notifications, glanceable content presumes that the audience will choose for themselves when they want to look, and for how long.  The content needs to tempt the user to look at it occasionally, without being a distraction.

What makes a small chunk of information significant to users, making it worth their while?  The chunk of content needs to be specific enough to satisfy the user’s curiosity, but not require a distracting amount of work to access. The content will fail if it doesn’t get this balance right.

Apple recognizes this conundrum in its guidelines for the Apple Watch. It advocates “glanceable” content with “lightweight interactions”.  Not all interaction guidelines of the Watch meet those goals, however. Because only a small amount of content is presented, designers compensate by asking users to interact with it to get exactly the bit they want.

Compact content can be complex as well (Images from Apple Watch guidelines.)

Google has tried to get around the distraction of interaction by requiring little of it from audiences. Google Now cards provide bite sized content, where information within smartphone apps gets surfaced to a main screen.  Google Now curates information through a combination of machine learning to anticipate likely user preferences (implicit feedback), and getting explicit user feedback on these preferences.   While very clever, this approach can’t easily be extended to other publishers who lack the technical resources and embedded positioning of Google.  Not all Google Now cards are easily glanceable, either.

Bite sized content can get busy.  Many layers of information can get crammed on a display, using hieroglyphic cues.  Such displays can vainly call attention to themselves, requiring both sustained gazing and mental interpretation.  They are compact, but not truly glanceable in the sense that someone can notice, at a glance, a key nugget of interest without effort.  Audio content runs less risk of this, since it is harder to compress content at a rate faster than the brain can process (though advertisers sneakily do such things).

Content for Multitasking

Multitasking is a myth: humans don’t really focus on several things simultaneously.  We slice our attention between different targets, switching back and forth between them rapidly.  But multitasking is a powerful metaphor for how people relate to content: they commonly process content from two more sources at the same time.  To be successful, glanceable content needs to respect what the audience wants to focus on.

It’s hard to multitask when everything demands your full attention, and your active interaction.   It’s hard watch a video or listen to audio if one needs to fiddle with screen settings, if one needs to tap and swipe incessantly to hunt down a piece of information, or if one feels rattled by surprise alerts that jump out at you.  All these annoyances are the product of interaction designers who expect and demand attention and interaction from the audience.  Interaction designers often fail to consider the broader content needs of the user.  They fail to think outside the UI in which they are boxed.

Before interaction design became largely associated with Javascript tricks, there was a movement to study how to best design “ambient displays” that provided supplemental, contextual information that was easily glanceable.  A well known example was the “Ambient Orb,” a small light that changed color according to how the stock market was doing (or, if you preferred, the status of the weather).  The product provided a primitive example of how to make content glanceable and non-distracting. But the product treated the user as a passive vessel who could express little choice in what content to follow, and consequently had little substance to keep them interested.  Multitasking wasn’t a challenge, but the secondary task wasn’t very meaningful either.  Ambient displays can come up short because they only focus on ensuring content doesn’t require much attention. They don’t provide any interaction.

A better example of content that is both glanceable and requires only light interaction comes from a digital content display for radio broadcasts known as Journaline.  I’ve only recently learned about this platform, even though it has nearly a decade of history behind it.  I want to spend time discussing concepts  related to Journaline because it provides a good example of how content can be used to support different scenarios of use.  A major use case for Journaline is to provide text-based content on the screens of radios (though the same technology can be applied to video as well).  Its premise is that the text content is secondary to the broadcast audio, and should not interfere with the primary attention of the audience, who may be washing dishes or waiting at a stop light.  Journaline provides guidelines for presenting content in conditions involving low attention.  I am less concerned with whether Journaline itself will become successful commercially than I am in the lessons it can provide content designers more generally.

Designing Elective Supplemental Content

Journaline is a full content schema with its own XML-based markup language. Developed by Germany’s Fraunhofer IIS, “Journaline® is a data application for the DAB and DRM digital radio systems that provides hierarchically structured textual information. The user can easily and immediately access the topics he is currently interested in.”  Alternatively, it has been described as “teletext for digital radio.”

Journaline offers straight text-based content, without distracting elements such as gestures, animations, switches, sliders, or activity ring displays.  The simplicity of its approach makes it a powerful example to learn from.  Because of its XML foundation, the schema can reuse existing content delivered from other sources such as RSS feeds.

Imagine listening to audio content such as a radio broadcast.  You may be doing other activities while listening.  The audio content itself might make reference to other things.  So you may have a need to access supplementary content either relating to the activity you are doing (such as getting traffic information), or supplemental content relating to the audio you are listening to (such as the name of the guest being interviewed).  Journaline provides such information, which can be accessed and seen simply.  Some models of BMWs in Germany have Journaline content displays embedded in their digital radios.

Journaline delivers supplemental baseline information that is changeable, and subject to updating.  It provides a feed of content, but allows some audience control over specific items of interest. The paradigm balances the amount of content offered, and the interaction required.  Sufficient new content is provided to make the display interesting enough to glance at occasionally when convenient to do so, while enough control is offered so that audiences can choose what to get updates on without requiring a distracting amount of interaction.

Journaline identifies five content types associated with informational updates:

  1. News
  2. Sports events and results
  3. Financial information and stock market values
  4. Airport departure and arrival times
  5. Games and lottery

Many of these types involve awareness information that is not so essential that people would actively seek it.  The information is useful when presented, but not worth investing special effort.  The airport info could provide useful pre-arrival information while driving, but won’t provide the depth of details available on a smartphone once one has arrived.

Supplemental information is often related to broadcast programs.  Journaline provides the following scenarios for program related information:

  • Show background information (with optional link to a website)
  • Direct phone link to participate in chat show
  • Song info and purchases, and podcast downloads
  • Captions (radio for the hearing impaired)

I’m not sure all these are necessarily compelling use cases in the context of listening to a broadcast, but they do illustrate the range of uses of supplemental content.  Sometimes the content has a call to action that is thematically tied the primary content, and thus does not compete with it.   But the actions presented in the supplemental content are always elective — the  primary content isn’t promoting the secondary content; rather, the secondary content offers deeper support for the primary content.  Elective actions include ones that cannot be anticipated or expected (such as a desire to call the station), and  options that allow the audience to defer tasks for later, such as the ability to order something to check out later.  When done properly, low-key options let the audience choose the right time to take action, instead of pestering them to act when it may be inconvenient.

Dynamic Information Architecture

Journaline is effective in how it manages information, a kind of dynamic information architecture.  Like traditional information architecture, it provides elements that give audiences choices to navigate and select.  But it is also dynamic, where the elements themselves change as new information is available.  It blurs what’s navigation and what’s content.  Items in a list can update continuously, and can be navigational links to more details about the item.

Journaline works on a feed model, but it doesn’t simply provide a chronological stream.  Instead, content items are pushed out when updates happen, and are stored and available to view when no updates are required.  What’s available to view will be the most up-to-date content about preselected topics.  These topics can be selected (pulled) either through navigation, or by following topics based on simple metadata components such as keywords or geotags (such as content related to things near me).

Journaline uses simple structures for content.  Items are broken into two parts: an initial summary and detail, which could be a headline followed by an explanatory statement. Summaries can be collected by topic.  This simple structuring allows for three basic layers of information: a consolidated view of several items, a general view of a specific item, and a more detailed statement about the item.

The below figure shows how different levels of information relate to each other.

Levels of Glanceable Content
Levels of Glanceable Content

There are three basic levels of content, and three options to reveal content.  Conceptually, they provide different windows for glancing at content.

The highest level provides an overview or digest.  Screens may contain several headlines that relate to a common keyword, or may provide a table of information on a common theme that is updated continually, such as a scoreboard showing teams and scores.

At the mid level, the screen will show a summary of information related to an event or topic. This will typically be a headline, or a caption relating to the audio or to a displayed image.

At the lowest level, items will present details supporting the headline.

Content is revealed in different ways.  Some content is static and fully visible at a glance.  Such content will never require more than a single glance.  But other content is disclosed incrementally over time.

Overview or digest content will refresh frequently. Each headline displayed may serve a predefined role.  A digest on a topic may be composed of different headlines that are presented in a list, where each item is assigned a slot.  A slot— a fixed position on the list —  shows a headline message that changes.  For example, people might track several related items, perhaps headlines relating to an election.  The first slot in the list relates to the keyword “candidate X” and will change throughout the day according to whatever the most important event is relating to candidate X.  A scorecard is a list in table form, where the left column is fixed with the names of teams currently playing, while the right side automatically updates with the most current score.  When the content item refreshes in a predefined spot, it makes it easy for the audience to glance at the content and notice what’s current and new.

Scorecard table, with scores updated continuously
Scorecard table, with scores updated continuously

A different behavior for refreshing content occurs when messages are revealed in stages.  Perhaps several ideas need to be shown that won’t all fit in a single short headline.  Several techniques can be used to reveal such content.  One technique is the news ticker, which displays a running headline.  Text captions for audio may be presented karaoke style.  Another technique is to rotate short headline snippets every few seconds.  These techniques can be useful for background content.  They can be digested with a quick glance, and will change enough to encourage curiosity.  If, however, they contain essential information that must be monitored continually, then they will become distracting.  The staged revelation of content provides a wide frame for audiences to follow, but can demand more attention from them.

Sometimes the audience will want more information than can be presented in a glance.  They want a detail to complete the information hinted at in the headline.  A full message might fit on a single small screen, but if not, then scrolling is required.  Even gesture-based scrolling entails interaction by the user, and requires longer attention spans as well.  This option exists to provide audiences with choice.  They choose this degree of distraction, and will hopefully be in a position to focus away from other activities and content.

Short-burst Content: A Challenge for Content Designers

Audiences routinely experience short bursts of content.  Events are chronicled with live tweets.  Siri and Alexa provide short answers to queries.  All these options offer great benefits to audiences, but they aren’t always the best option in a given situation.  Sometimes short messages require too much attention to follow, or too much work to specify what one wants.

According to the cliché, timing is everything.  When audiences face so many demands on their attention, content timing matters significantly.  Content designers should consider ways content can work in harmony with the activities of audiences, instead of competing with these activities.  Disruption, while cool for startups, is annoying to audiences.  Supplemental content can be a new category of content that supports other activities, instead of being its own activity.  To realize this possibility requires the embracing a radical idea: accepting that the secondary content presented may not be the primary motivation of your audience.  Because audiences in many circumstance are only willing and able to offer partial attention, the secondary content needs to be sufficiently interesting and timely to merit glancing.

The content design challenge is discovering how to get audiences to glance at content that they don’t feel they need to see, but will want to view, if only briefly.  Content that’s useful, without being demanding.  How can we create glanceable moments?

Journaline offers two ideas worth exploring more.  First, it shows the value of creating a hub for content that features topics that change, and that are of interest to audiences.  People will revisit the hub to glance at what changes have happened.

Second, it shows the possibilities of using a secondary channel (in this case text) to augment the primary content channel (audio) to support elective, secondary tasks.  Audiences are primarily interested in the content in the primary channel, and only some will be interested in the secondary channel content at any given time, since they may be preoccupied with other matters.  This technique can easily be used with video, by providing additional real-time or more personalized secondary information relating to the video content that some people will be interested in glancing at, and which is short enough not to distract long from the primary video content.

Designing for partial attention can seem like a loss of control, since there are no guarantees a specific message will reach a specific person.  But it also represents an opportunity to reach audiences who may not be otherwise available.  And it encourages us to think about the fundamentals of how to attract audiences and build their interest in content.

— Michael Andrews

Categories
Content Experience

What is Content Design?

The growing interest in content design is a welcome development.  Such interest recognizes that content decisions can’t be separated from the context in which the content will be used.  Consideration of content design corrects two common misperceptions: the notion that content presentation is simply visual styling, and the belief that because content may need to exist in many contexts, the context in which content is displayed becomes irrelevant.  Direct collaboration between writers and UI designers is now encouraged.  Content must fit the design where it appears — and conversely, UI designs must support the content displayed.  Content has no impact independently of a container or interaction platform for which it has been designed, and is being relied upon by users.  Content depends on context.  And context frames the content experience.

Yet content design is more than a collaborative attitude. What content design actually entails is still not well understood. Content design requires all involved to consider how different elements should work together as a system.

“Content and Design Are Inseparable Work Partners”  — Jared Spool

Current Definitions of Content Design

There is no single accepted definition of content design.  Two meanings are in use, both of which are incomplete.

The first emphasizes layout and UI decisions relating to the presentation of content.  It looks at such questions as will the text fit on the screen, or how to show and hide information.  The layout perspective of content design is sometimes referred to as the application of content patterns.

The second, popularized by the Government Digital Service (GDS) in Britain, focuses on whether the words being presented in an article support the tasks that users are trying to accomplish.  The GDS instructs: “know your users’ needs and design your content around them” and talks about “designing by writing great content.”  The GDS’ emphasis on words reflects the fixed character of their content types —a stock of 40 formats.  These structures provide ready templates for inserting content, but don’t give content creators a voice in how or what to present apart from wording.

Content design encompasses much more than wording and layout.

The design of content, including printed media, has always involved layout and wording, and the interaction between the two. Comprehensive content design today goes further by considering behavior: the behavior of the content, and the behavior of users interacting with the content.  It designs content as a dynamic resource.  It evaluates and positions content within a stream of continuous interaction.

Most discussion of content design approaches content from a “one size fits all” perspective.  What’s missing in current discussions is how to design content that can serve multiple needs.  User needs are neither fixed, nor uniform.  Designs must be able to accommodate diverse needs.  Formulaic templates generally fall short of doing this.  Content must be supported by structures that are sophisticated enough to accommodate different scenarios of use.

Breaking Free from the Static Content Paradigm

Content creators typically think about content in terms of topics.  Topics are monolithic.  They are meant to be solid: to provide the answers to questions the audience has. In an ideal scenario, the content presented on the topic perfectly matches the goals of the audience.

The problem with topics is that they too often reflect a publisher-centric view of the world.  Publishers know people are seeking information about certain topics — their web logs tell them this.  They know the key information they need to provide on the topic.  They strive to provide succinct answers relating to the topic.  But they don’t consider the wide variation of user needs relating to the topic.  They can’t imagine that numerous people all reading the same content might want slightly different things.

Consider the many factors that can influence what people want and expect from content:

  • Their path of arrival — where they have come from and what they’ve seen already
  • Their prior knowledge of the topic
  • Their goals or motivations that brought them to the content
  • The potential actions they might want to take after they’ve seen the content

Some people are viewing the content to metaphorically “kick the tires,” while others approach the content motivated to take action.  Some people will choose to take action after seeing the content, but others will defer action.  People may visit the content with one goal, and after viewing the content have a different goal. Regardless of the intended purpose of the content, people are prone to redefine their goals, because their decisions always involve more than what is presented on the screen.

In the future, content might be able to adjust automatically to accommodate differences in user familiarity and intent.  Until that day arrives (if it ever does), creators of content need to produce content that addresses a multitude of users with slightly varying needs.  This marks the essence of content design: to create units of content that can address diverse needs successfully.

A common example of content involving diverse needs relates to product comparison.  Many people share a common task of comparing similar products.  But they may differ in what precisely they are most interested in:

  • What’s available?
  • What’s best?
  • What are the tradeoffs between products?
  • What options are available?
  • How to configure product options and prices?
  • How to save options for use later?
  • How to buy a specific configuration?

A single item of content providing a product comparison may need to support many different purposes, and accommodate people with different knowledge and interests.  That is the challenge of content design.

Aspects of Content Design

How does one create content structures that respond to the diverse needs of users in different scenarios? Content design needs to think beyond words and static informational elements.  When designs include features and dynamic information, content can accomplish more.  The goal is to build choice into the content, so that different people can take away different information from the same item of content.

Design of Content Features

A feature in content is any structural element of the content that is generated by code.  Much template-driven content, in contrast, renders the structure fixed, and makes the representation static.  Content features can make content more “app-like” — exhibiting behaviors such as updating automatically, and offering interactivity.  Designing content features involves asking how functionality can change the representation of content to deliver additional value to audiences and the business.  Features can provide for different views of content, with different levels of detail or different perspectives.

Consider a simple content design decision: should certain information be presented as a list, in a table, or as a graph?  Each of these options are structures.  The same content can be presented in all three structures.  Each structure has benefits. Graphs are easy to scan, tables allow more exact information, while lists are better for screen readers.  The “right” choice may depend on the expected context of use — assuming only one exists.  But it is also possible that the same content could be delivered in all three structures, which could be used by different users in different contexts.

Design of Data-driven Information

Many content features depend on data-driven information.  Instead of considering content as static — only reflecting what was known at the time it was published — content can be designed to incorporate information about activities related to the content that have happened after publication of the article.

Algorithmically-generated information is increasingly common.  A major goal is to harvest behavioral data that might be informative to audiences, and use that data to manage and prioritize the display of information.  Doing this successfully requires the designer to think in terms of a system of inter-relationships between activities, needs, and behavioral scenarios.

Features and data can be tools to solve problems that words and layout alone can’t address.  Both these aspects involve loosening the control over what the audience sees and notices.  Features and data can enrich the content experience.  They can provide different points of interest, so that different people can choose to focus on what elements of information interest them the most.   Features and data can make the content more flexible in supporting various goals by offering users more choice.

Content Design in the Wild

Real-world examples provide the best way to see the possibilities of content design, and the challenges involved.

Amazon is famous for both the depth of its product information, and its use of data.  Product reviews on Amazon are sometimes vital to the success of a product.  Many people read Amazon product reviews, even if they’ve no intention of buying the product from Amazon.  And people who have not bought the product are allowed to leave reviews, and often do.

Amazon’s product reviews illustrate different aspects of content design.  The reviews are enriched with various features and data that let people scan and filter the content according to their priorities.  But simply adding features and data does not automatically result in a good design.

Below is a recent screenshot of reviews for a book on Amazon.  It illustrates some of the many layers of information available.  There are ratings of books, comments on the books, identification of the reviewers, and reactions to the ratings.  Seemingly, everything that might be useful has been included.

Design of product review information for a book
Design of product review information for a book

The design is sophisticated on many levels. But instead of providing clear answers for users trying to evaluate the suitability of a book, the design raises various questions.  Consider the information conveyed:

  • The book attracted three reviews
  • All three reviewers rated the book highly, either four or five stars
  • All the reviewers left a short comment
  • Some 19 people provided feedback on the reviews
  • Only one person found the reviews helpful; the other 18 found the reviews unhelpful

Perhaps the most puzzling element is the heading: “Most Helpful Customer Reviews.”  Clearly people did not find the reviews helpful, but the page indicates the opposite.

This example illustrates some important aspects of content design.  First, different elements of content can be inter-dependent.  The heading depends on the feedback on reviews, and the feedback on reviews depend on the reviews themselves. Second, because the content is dynamic, what gets displayed is subject to a wide range of inputs that can change over time.  Whether what’s display makes sense to audiences will depend on the design’s capacity to adapt to different scenarios in a meaningful way. Content design depends on a system of interactions.

Content Design as Problem Solving

Content design is most effective when treated as the exploration of user problems, rather than as the fulfillment of user tasks.  Amazon’s design checks the box in terms of providing information that can be consulted as part of a purchase decision.  A purely functional perspective would break tasks into user stories: “Customer reads reviews”, etc.  But tasks have a tendency to make the content interaction too generic. The design exploration needs to come before writing the stories, rather than the reverse. The design needs to consider various problems the user may encounter.  Clearly the example we are critiquing did not consider all these possibilities.

An examination of the content as presented in the design suggests the source of problems readers of the reviews encountered.  They did not find the comments helpful.  The comments are short, and vague as to what would justify the high rating.  A likely reason the comments are vague is that the purchasers of the product were not the true endusers of the product, so they refrained from evaluating the qualities of the product, and commented on their purchase experience instead.  The algorithms that prioritize the reviews don’t have a meaningful subroutine for dealing with cases where all the reviews are rated as unhelpful.

 Critique as the Exploration of Questions

Critiquing the design of content allows content creators to consider the interaction of content as seen from the audience perspective.  As different scenarios are applied to various content elements, the critique can ask more fundamental questions about audience expectations, and in so doing, reconsider design assumptions.

Suppose we shift the discussion away from the minutiae of screen elements to consider the people involved.  The issue is not necessarily whether a specific book is sold.  The lifetime value of customers shopping on Amazon is far more important.  And here, the content design is failing in a big way.

Customers want to know if a book, which they can’t look at physically or in extensive detail, is really what they want to purchase.  Amazon counts on customers to give other customers confidence that what they purchase is what they want.  Returned merchandise is a lose-lose proposition for everyone.  Most customers who leave reviews do so voluntarily, without direct benefit — that is what makes their reviews credible.   So we have buyers of a book altruistically offering their opinion about the product.  They have taken the trouble to log-in and provide a review, with the expectation the review will be published, and the hope it will be helpful to others.  Instead, potential buyers of the book are dinging the reviews.  The people who have volunteered their time to help others are being criticized, while people who are interested in buying the book are unhappy they can’t get reliable information.  Through poor content design, Amazon is alienating two important customer constituencies at once: loyal customers who provide reviews on which Amazon depends, and potential buyers considering a product.

How did this happen, and how can it be fixed?  Amazon has talented employees, and unrivaled data analytics.  Despite those enviable resources, the design of the product review information nonetheless has issues.  Issues of this sort don’t lend themselves to A/B testing, or quick fixes, because of the interdependencies involved.  One could deploy a quick fix such as changing the heading if no helpful reviews exist, but the core problems would remain.  Indeed, the tendency in agile IT practices to apply incremental changes to designs is often a source of content design problems, rather than a means of resolving them.  Such patchwork changes mean that elements are considered in isolation, rather than as part of a system involving interdependencies.

Many sophisticated content designs such as the product review pages evolve over time.  No one person is directing the design: different people work on the design at different stages, sometimes over the course of years.  Paradoxically, even though the process is trumpeted as being agile, it can emulate some of the worst aspects of a “design by committee” approach where everyone leaves their fingerprints on the design, but no holistic concept is maintained.

News reports indicate Amazon has been concerned with managing review contributors.  Amazon wants to attract known reviewers, and has instituted a program called Vine that provides incentives to approved reviewers.  At the same time, it wants to discourage reviewers who are paid by outside parties, and has sued people it believes provide fake reviews.  To address the issue of review veracity, reviews use badges indicating the reviewer’s status as being a verified purchaser, a top reviewer, or a Vine reviewer.  The feedback concerning whether a review is helpful is probably also linked to goals of being able to distinguish real reviews from fake ones.  It would appear that the issue of preventing fake reviews has become conflated with the issue of providing helpful reviews, when in reality they are separate issues.  The example clearly shows that real reviews are not necessarily helpful reviews.

The content design should support valid business goals, but it needs to make sure that doing so doesn’t work at cross-purposes with the goals of audiences using the design.  Letting customers criticize other customers may support the management of review content, but in some cases it may do so at the cost of customer satisfaction.

A critique of the design also brings into focus the fact that the review content involves two distinct user segments: the readers of reviews, and the writers of reviews.  The behavior of each affects the other.  The success of the content depends on meeting the needs of both.

The design must look beyond the stated problem of how to present review information.  It must also solve second-order problems.  How to encourage useful reviews?  What to do when there are no useful reviews?  Many critical design issues may be lurking behind the assumptions of the “happy path” scenario.

Re-examining Assumptions

A comprehensive content design process keeps in mind the full range of (sometimes competing) goals the design needs to fulfill, and the range of scenarios in which the design must accommodate.  From these vantage points, it can test assumptions about how a design solution performs in different situations and against different objectives.

When applied to the example of product reviews, different vantage points raise different core questions.   Let’s focus on the issue of encouraging helpful reviews, given its pivotal leverage.  The issue involves many dimensions.

Who is the audience for the reviews: other customers, or the seller or maker of the product?  Who do the reviewers imagine is seeing their content, and what do they imagine is being done with that information?  What are the expectations of reviewers, and how can the content be designed to match their expectations — or to reset them?

What are the reviewers supposed to be rating?  Are they rating the product, or rating Amazon?  When the product is flawed, who does the reviewer hold accountable, and is that communicated clearly?  Do raters or readers of ratings want finer distinctions, or not?  How does the content design influence these expectations?

What do the providers of feedback on reviews expect will be done with their feedback?  Do they expect it to be used by Amazon, by other customers, or be seen and considered by the reviewer evaluated?  How does the content design communicate these dimensions?

What is a helpful review, according to available evidence?  What do customers believe is a helpful review?  Is “most helpful” the best metric?  Suppose long reviews are more likely to be considered helpful reviews. Is “most detailed” a better way to rank reviews?

What kinds of detail are expected in the review comments?  What kinds of statements do people object to?  How does the content design impact the quality of the comments?

What information is not being presented?  Should Amazon include information about number of returns?  Should people returning items provide comments that show up in the product reviews?

There are of course many more questions that could be posed.  The current design reflects a comment moderation structure, complete with a “report abuse” link.  The policing of comments, and voting on reviews hits on extrinsic motivators — people seeking status from positive feedback, or skirting negative feedback. But it doesn’t do much to address intrinsic motivators to participate and contribute.  A fun exercise to shift perspective would be to try imagining how to design the reviews to rank-order them according to their sincerity. Because people can be so different in what they seek in product information, it is always valuable to ask what different people care about most, and never to assume to know the answer to that with certainty.

Designing Experiences, Not Tasks

Tasks are a starting point for thinking about content design, but are not sufficient for developing a successful design.  Tasks tend to simplify activities, without giving sufficient attention to contextual issues or alternative scenarios.  A task-orientation tends to make assumptions about user motivations to do things.

Content design is stronger when content is considered experientially.  Trust is a vitally important factor for content, but it is difficult to reduce into a task.  Part of what makes trust so hard is that it is subjective.  Different people value different factors when assessing trustworthiness — rationality or emotiveness, thoroughness or clarity.  For that reason, content designs often need to provide a range of information and detail.

Designing for experiences frees us from thinking about user content needs as being uniform. Instead of focusing only on what people are doing (or what we want them to be doing), the experiential perspective focuses on why people may want to take action, or not.

People expect a richly-layered content experience, able to meet varied and changing needs.  Delivering this vision entails creating a dynamic ecosystem that provides the right kinds of details. The details must be coordinated so that they are meaningful in combination. Content becomes a living entity, powered by many inputs.  Dynamic content, properly designed, can provide people with positive and confidence-inducing experiences. Unless people feel comfortable with the information they view, they are reluctant to take action.  Experience may seem intangible, and thus inconsequential.  But the content experience has real-world consequences: it impacts behavior.

— Michael Andrews