Categories
Content Engineering

What’s the value of content previews?

Content previews let you see how your content will look before it’s published.  CMSs have long offered previews, but preview capabilities are becoming more varied as content management is increasingly decoupled from UI design and channel delivery. Preview functionality can introduce unmanaged complexity to content and design development processes.  

Discussions about previews can spark stong opinions.

Are content previews:

  1. Helpful?
  2. Unnecessary?
  3. A crutch used to avoid fixing existing problems?
  4. A source of follow-on problems?
  5. All of the above?

Many people would answer previews are helpful because they personally like seeing previews. Yet whether previews are helpful depends on more than individual preferences. In practice, all of the above can be true.  

It may seem paradoxical that a feature like previews can be good and bad. The contradiction exists only if one assumes all users and preview functionality are the same. Users have distinct needs and diverging expectations depending on their role and experience. How previews are used and who is impacted by them can vary widely. 

Many people assume previews can solve major problems authors face. Previews are popular because they promise to bring closure to one’s efforts. Authors can see how their content will look just before publishing it. Previews offer tangible evidence of one’s work. They bring a psychic reward. 

Yet many factors beyond psychic rewards shape the value of content previews. 

What you see while developing content and how you see it can be complicated. Writers are accustomed to word processing applications where they control both the words and their styling. But in enterprise content publishing, many people and systems become involved with wording and presentation. How content appears involves various perspectives. 

Content teams should understand the many sides of previews, from the helpful to the problematic.  These issues are becoming more important as content becomes uncoupled from templated UI design. 

Previews can be helpful 

Previews help when they highlight an unanticipated problem with how the content will be rendered when it is published. Consider situations that introduce unanticipated elements. Often, these will be people who are either new to the content team or who interact with the team infrequently. Employees less familiar with the CMS can be encouraged to view the preview to confirm everything is as expected.  Such encouragement allows the summer intern, who may not realize the need to add an image to an article, to check the preview to spot a gap.  

Remember that previews should never be your first line of defense against quality problems. Unfortunately, that’s often how previews are used: to catch problems that were invisible authors and designers when developing the content or the design.

Previews can be unnecessary 

Previews aren’t really necessary when writers create routine content that’s presented the same way each time.  Writers shouldn’t need to do a visual check of their writing and won’t feel the need to do so provided their systems are set up properly to support them. They should be able to see and correct issues in their immediate work environment rather than seesaw to a preview. Content should align with the design automatically. It should just work.

In most cases, it’s a red flag if writers must check the visual appearance of their work to determine if they have written things correctly. The visual design should accommodate the information and messages rather than expect them to adapt to the design. Any constraints on available space should be predefined rather than having writers discover in a preview that the design doesn’t permit enough space. Writers shouldn’t be responsible to ensuring the design can display their content properly. 

The one notable exception is UX writing, where the context in which discrete text strings appear can sometimes shape how the wording needs to be written. UX writing is unique because the content is highly structured but infrequently written and revised, meaning that writers are less familiar with how the content will display. For less common editorial design patterns, previews help ensure the alignment of text and widgets. However, authors shouldn’t need previews routinely for highly repetitive designs, such as those used in e-commerce.

None of the above is to say a preview shouldn’t be available; only that standard processes shouldn’t rely on checking the preview. If standard content tasks require writers to check the preview, the CMS setup is not adequate. 

Previews can be a crutch 

Previews are a crutch when writers rely on them to catch routine problems with how the content is rendered. They become a risk management tool and force writers to play the role of risk manager. 

Many CMSs have clunky, admin-like interfaces that authors have trouble using. Vendors, after all, win tenders by adding features to address the RFP checklist, and enterprise software is notorious for its bad usability (conferences are devoted to this problem).  The authoring UI becomes cluttered with distracting widgets and alerts.  Because of all the functionality, vendors use “ghost menus” to keep the interface looking clean, which is important for customer demos. Many features are hidden and thus easy for users to miss, or they’ll pop up and cover over text that users need to read.  

The answer to the cluttered UI or the phantom menus is to offer previews. No matter how confusing the experience of defining the content may be within the authoring environment, a preview will provide a pristine view of how the content will look when published.  If any problems exist, writers can catch them before publication. If problems keep happening, it becomes the writer’s fault for not checking the preview thoroughly and spotting the issue.

At its worst, vendors promote previews as the solution to problems in the authoring environment. They conclude writers, unlike their uncomplaining admin colleagues, aren’t quite capable enough to use UIs and need to see the visual appearance. They avoid addressing the limitations of the authoring environment, such as:

  • Why simple tasks take so many clicks 
  • Why the UI is so distracting that it is hard to notice basic writing problems
  • Why it’s hard to know how long text or what dimensions images should be

Writers deserve a “focus” mode in which secondary functionality is placed in the background while writers do essential writing and editing tasks. But previews don’t offer a focus mode – they take writers away from their core tasks. 

Previews can cause follow-on problems

Previews can become a can of worms when authors use them to change things that impact other teams. The preview becomes the editor and sometimes a design tool. Unfortunately, vendors are embracing this trend.

Potential problems compound when the preview is used not simply to check for mistakes but as the basis for deciding writing, which can happen when:

  1. Major revisions happen in previews
  2. Writers rely on previews to change text in UI components 
  3. Writers expect previews to guide how to write content appearing in different devices and channels 
  4. Writers use previews to change content that appears in multiple renderings
  5. Writers use previews to change the core design substantially and undermine the governance of the user experience 

Pushing users to revise content in previews. Many vendors rely on previews to hide usability problems with the findability and navigation of their content inventory. Users complain they have difficulty finding the source content that’s been published and want to navigate to the published page to make edits. Instead of fixing the content inventory, vendors encourage writers to directly edit in the preview. 

Editing in a preview can support small corrections and updates. But editing in previews creates a host of problems when used for extensive revisions, or multi-party edits because the authoring interface functionality is bypassed. The practices change the context of the task.  Revisions are no longer part of a managed workflow. Previews don’t display field validation or contextual cues about versioning and traceability.  It’s hard to see what changes have been made, who has made them, or where assets or text items have come from. Editing in context undermines content governance. 

Relying on previews to change text in UI components. Previews become a problem when they don’t map to the underlying content. More vendors are promoting what they call “hybrid” CMSs (a multi-headed hydra) that mix visual UI components with content-only components – confusingly, both are often called “blocks.” Users don’t understand the rendering differences in these different kinds of components. They check the preview because they can’t understand the behavior of blocks within the authoring tool. 

When some blocks have special stylings and layouts while others don’t, it’s unsurprising that writers wonder if their writing needs to appear in a specific rendering. Their words become secondary to the layout, and the message becomes less important than how it looks. 

Expecting previews to guide how to write content appearing in different devices and channels. A major limitation of previews occurs when they are relied upon to control content appearing in different channels or sites. 

In the simplest case, the preview shows how content appears on different devices. It may offer a suggestive approximation of the appearance but won’t necessarily be a faithful rendering of the delivered experience to customers. No one, writers especially, can rely on these previews to check the quality of the designs or how content might need to change to work with the design.

Make no mistake: how content appears in context in various channels matters. But the place to define and check this fit is early in the design process, not on the fly, just before publishing the content. Multi-channel real-time previews can promote a range of bad practices for design operations.

Using previews to change content that appears in multiple renderings. One of the benefits of a decoupled design is that content can appear in multiple renderings. Structured writing interfaces allow authors to plan how content will be used in various channels. 

We’ve touched on the limitations of previews of multiple channels already.  But consider how multi-channel previews work with in-context editing scenarios.  Editing within a preview will  focus on a single device or channel and won’t highlight that the content supports multiple scenarios. But any editing of content in one preview will influence the content that appears in different sites or devices. This situation can unleash pandemonium.

When an author edits content in a preview but that content is delivered to multiple channels, the author has no way of knowing how their changes to content will impact the overall design. Authors are separated from the contextual information in the authoring environment about the content’s role in various channels. They can’t see how their changes will impact other channels.

Colleagues may find content that appears in a product or website they support has been changed without warning by another author who was editing the content in a preview of a different rendering, unaware of the knock-on impact. They may be tempted to use the same preview editing functionality to revert to the prior wording. Because editing in previews undermines content governance, staff face an endless cycle of “who moved my cheese” problems. 

Using previews to substantially change the core design. Some vendors have extended previews to allow not just the editing of content but also the changing of UI layout and design. The preview becomes a “page builder” where writers can decide the layout and styling themselves. 

Unfortunately, this “enhancement“ is another example of “kicking the can” so that purported benefits become someone else’s problem. It represents the triumph of adding features over improving usability.

Writers wrestle control over layout and styling decisions that they dislike. And developers celebrate not having to deal with writers requesting changes.  But page building tries to fix problems after the fact.  If the design isn’t adequate, why isn’t it getting fixed in the core layout? Why are writers trying to fix design problems?

Previews as page builders can generate many idiosyncratic designs that undermine UX teams. UI designs should be defined in a tool like Figma, incorporated in a design system, and implemented in reusable code libraries available to all. Instead of enabling maturing design systems and promoting design consistency, page builders hurt brand consistency and generate long term technical debt.

Writers may have legitimate concerns about how the layout has been set up and want to change it. Page builders aren’t the solution. Instead, vendors must improve how content structure and UI components interoperate in a genuinely decoupled fashion. Every vendor needs to work on this problem.

Some rules of thumb

  • Previews won’t fix significant quality problems.
  • Previews can be useful when the content involves complex visual layouts in certain situations where content is infrequently edited. They are less necessary for loosely structured webpages or frequently repeated structured content.
  • The desire for previews can indicate that the front-end design needs to be more mature. Many design systems don’t address detailed scenarios; they only cover superficial, generic ones. If content routinely breaks the design, then the design needs refinement.
  • Previews won’t solve problems that arise when mixing a complex visual design with highly variable content. They will merely highlight them. Both the content model and design system need to become more precisely defined.
  • Previews are least risky when limited to viewing content and most risky when used to change content.
  • Preview issues aren’t new, but their role and behavior are changing. WYSIWYG desktop publishing metaphors that web CMS products adopted don’t scale. Don’t assume what seems most familiar is necessarily the most appropriate solution.

– Michael Andrews

Categories
Content Integration

Digital transformation for content workflows

Content workflows remain a manually intensive process. Content staff face the burden of deciding what to do and who should do it. How can workflow tools evolve to reduce burdens and improve outcomes? 

Content operations are arguably one of the most backward areas of enterprise business operations. They have been largely untouched by enterprise digital transformation. They haven’t “change[d] the conditions under which business is done, in ways that change the expectations of customers, partners, and employees” – even though business operations increasingly rely on online content to function. Compared with other enterprise functions, such as HR or supply chain management, content operations rely little on process automation or big data. Content operations depend on content workflow tools that haven’t modernized significantly.  Content workflow has become a barrier to digital transformation.

The missing flow 

Water flows seamlessly around any obstacle, downward toward a destination below.  Content, in contrast, doesn’t flow on its own. Content items get stuck or bounce around in no apparent direction. Content development can resemble a game of tag, where individuals run in various directions without a clear sense of the final destination.  Workflow exists to provide direction to content development.

Developing content is becoming more complex, but content workflow capabilities remain rudimentary. Workflow functionality has limited awareness of what’s happened previously or what should (or could) happen later. It requires users to perform actions and decisions manually. It doesn’t add value.

Workflow functionality has largely stayed the same over the years, whether in a CMS or a separate content workflow tool. Vendors are far removed from the daily issues the content creators face managing content that’s in development. All offer similar generic workflow functionality. They don’t understand the problem space.  

Vendors consider workflow problems to be people problems, not software problems. Because people are prone to be “messy” (as one vendor puts it), the problem the software aims to solve is to track people more closely. 

To the extent workflow functionality has changed in the past decade, it has mainly focused on “collaboration.” The vendor’s solution is to make the workflow resemble the time-sucking chats of social media, which persistently demand one’s attention. By promoting open discussion of any task, tools encourage the relitigation of routine decisions rather than facilitating their seamless implementation. Tagging people for input is often a sign that the workflow isn’t clear. Waiting on responses from tagged individuals delays tasks. 

End users find workflow tools kludgy. Workflows trigger loads of notifications, which result in notification fatigue and notification blindness. Individuals can be overwhelmed by the lists and messages that workflow tools generate. 

Authors seek ways to compensate for tool limitations. Teams often supplement CMS workflow tools with project management tools or spreadsheets. Many end users skirt the built-in CMS workflow by avoiding optional features. 

Workflow optimization—making content workflows faster and easier—is immature in most organizations. Ironically, writers are often more likely to write about improving other people’s workflows (such as those of their customers or their firm’s products and services) than to dedicate time to improving their own content workflows.  

Content workflows must step up to address growing demands.  The workflow of yesterday needs reimagining.

Deane Barker wrote in his 2016 book on content management: “Workflow is the single most overpurchased aspect of any CMS…I fully believe that 95% of content approvals are simple, serial workflows, and 95% of those have a single step.”

Today, workflow is not limited to churning out simple static web pages. Content operations must coordinate supply chains of assets and copy, provide services on demand, create variants to test and optimize, plan delivery across multiple channels, and produce complex, rich media. 

Content also requires greater coordination across organizational divisions. Workflows could stay simple when limited to a small team. But as enterprises work to reduce silos and improve internal integration, workflows have needed to become more sophisticated. Workflows must sometimes connect people in different business functions, business units, or geographic regions. 

Current content workflows are hindered by:

  • Limited capabilities, missing features, and closed architectures that preclude extensions
  • Unutilized functionality that suffers from poor usability or misalignment with work practices

Broken workflows breed cynicism. Because workflow tools are cumbersome and avoided by content staff, some observers conclude workflow doesn’t matter. The opposite is true: workflows are more consequential than ever and must work better. 

While content workflow tools have stagnated, other kinds of software have introduced innovations to workflow management. They address the new normal: teams that are not co-located but need to coordinate distinct responsibilities. Modern workflow tools include IT service management workflows and sophisticated media production toolchains that coordinate the preproduction, production, and postproduction of rich media.

What is the purpose of a content workflow?

Workflow isn’t email. Existing workflow tools don’t solve the right problems. They are tactical solutions focused on managing indicators rather than substance. They reflect a belief that if everyone achieves a “zero inbox” with no outstanding tasks, then the workflow is successful.  But a workflow queue shouldn’t resemble an email box stuffed with junk mail, unsolicited requests, and extraneous notices, with a few high-priority action items buried within the pile. Workflows should play a role in deciding what’s important for people to work on.

Don’t believe the myth that having a workflow is all that’s needed. Workflow problems stem from the failure to understand why a workflow is necessary. Vendors position the issue as a choice of whether or not to have a workflow instead of what kind of workflow enterprises should have.  

Most workflow tools focus on tracking content items by offering a fancy checklist. The UI covers up an unsightly sausage-making process without improving it. 

Many tools prioritize date tracking. They equate content success with being on time. While content should be timely, its success depends on far more than the publication date and time. 

A workflow in itself doesn’t ensure content quality. A poorly implemented workflow can even detract from quality, for example, by specifying the wrong parties or steps. A robust workflow, in contrast, will promote consistency in applying best practices.  It will help all involved with doing things correctly and making sound decisions.  

As we shall see, workflow can support the development of high-quality content if it:

  • Validates the content for correctness
  • Supports sound governance

A workflow won’t necessarily make content development more productive. Workflows can be needlessly complex, time-consuming, or confusing. They are often not empowering and don’t allow individuals to make the best choices because they constrain people in counterproductive ways.  

Contrary to common belief, the primary goal of workflow should not be to track the status of content items. If all a workflow tool does is shout in red that many tasks are overdue, it doesn’t help. The tool behaves like an airport arrival and departure board that tells you flights are delayed without revealing why.  

Status-centric workflow tools simply present an endless queue of tasks with no opportunity to make the workload more manageable. 

Workflows should improve content quality and productivity.  Workflow tools contribute value to the extent they make the content more valuable. Quality and productivity drive content’s value. 

Yet few CMS workflow tools can seriously claim they significantly impact either the quality or productivity of the content development process. Administratively focused tools don’t add value.

Workflow tools should support people and goals –  the dimensions that ultimately shape the quality of outcomes. Yet workflow tools typically delegate all responsibility to people to ensure the workflow succeeds. Administratively focused workflows don’t offer genuine support. 

A workflow will enhance productivity – making content more valuable relative to the effort applied – only if it: 

  • Makes planning more precise
  • Accelerates the completion of tasks
  • Focuses on goals, not just activities
Elements of content workflow

Generic workflows presume generic tasks

Workflow tools fail to be “fit for purpose” when they don’t distinguish activities according to their purpose. They treat all activities as similar and equally important. Everything is a generic task: the company lawyer’s compliance review is no different than an intern’s review of broken links.  

Workflows track and forward tasks in a pass-the-batton relay. Each task involves a chain of dependencies. Tasks are assigned to one or more persons. Each task has a status, which determines the follow-on task.

CMS workflow tools focus on configuring a few variables:

  • Stage in the process
  • Task(s) associated with a stage
  • Steps involved with a task
  • Assigned employees required to do a step or task
  • Status after completing a task
  • The subsequent task or stage

From a coding perspective, workflow tools implement a series of simple procedural loops. The workflow engine resembles a hampster wheel. 

Hamster wheel
Like a hamster wheel, content workflow “engines” require manual pushing. Image: Wikimedia

A simple procedural loop would be adequate if all workflow tasks were similar. However, generic tasks don’t reflect the diversity of content work.

Content workflow tasks vary in multiple dimensions, involving differing priorities and hierarchies. Simple workflow tools flatten out these differences by designing for generic tasks rather than concrete ones. 

Variability within content workflows

Workflows vary because they involve different kinds of tasks.  Content tasks can be:

  • Cognitive (applying judgment)
  • Procedural (applying rules)
  • Clerical (manipulating resources) 

Tasks differ in the thought required to complete them.  Workflow tools commonly treat tasks as forms for users to complete.  They highlight discrete fields or content sections that require attention. They don’t distinguish between:

  1. Reflexive tasks (click, tap, or type)
  2. Reflective tasks (pause and think)

The user’s goal for reflexive tasks is to “Just do it” or “Don’t make me think.” They want these tasks streamlined as much as possible.  

In contrast, their goal for reflective tasks is to provide the most value when performing the task. They want more options to make the best decision. 

Workflows vary in their predictability. Some factors (people, budget, resources, priorities) are known ahead of time, while others will be unknown. Workflows should plan for the knowns and anticipate the unknowns.

Generic workflows are a poor way to compensate for uncertainty or a lack of clarity about how content should proceed. Workflows should be specific the content and associated business and technical requirements.  

Many specific workflows are repeatable. Workflows can be classified into three categories according to their frequency of use:

  1. Routine workflows 
  2. Ad hoc, reusable workflows
  3. Ad hoc, one-off workflows 

Routine workflows recur frequently. Once set, they don’t need adjustment. Because tasks are repeated often, routine workflows offer many opportunities to optimize, meaning they can be streamlined, automated, or integrated with related tasks. 

Ad hoc workflows are not predefined. Teams need to decide how to shape the workflow based on the specific requirements of a content type, subject matter, and ownership. 

Ad hoc workflows can be reusable. In some cases, teams might modify an existing workflow to address additional needs, either adding or eliminating tasks or changing who is responsible. Once defined, the new workflow is ready for immediate use. But while not routinely used, it may be useful again in the future, especially if it addresses occasional or rare but important requirements.  

Even when a content item is an outlier and doesn’t fit any existing workflow, it still requires oversight.  Workflow tools should make it easy to create one-off workflows. Ideally, generative AI could help employees state in general terms what tasks need to be done and who should be involved, and a bot could define the workflow tasks and assignments.

Workflows vary in the timing and discretion of decisions.  Some are preset, and some are decided at the spur of the moment.  

Consider deadlines, which can apply to intermediate tasks in addition to the final act of publishing.  Workflow software could suggest the timing of tasks – when a task should be completed – according to the operational requirements. It might assign task due dates:

  • Ahead of time, based on when actions must be completed to meet a mandatory publication deadline. 
  • Dynamically, based on the availability of people or resources.

Similarly, decisions associated with tasks have different requirements. Content task decisions could be 

  • Rules-driven, where rules predetermine the decision   
  • Discretionary and dependent on the decisionmaker’s judgment.

Workflows for individual items don’t happen in isolation. Most workflows assume a discrete content item. But workflows can also apply to groups of related items.  

Two common situations exist where multiple content items will have similar workflows:

  • Campaigns of related items, where items are processed together
  • A series of related items, where items are processed serially

In many cases, the workflow for related items should follow the same process and involve the same people.  Tools should enable employees to reuse the same workflow for related items so that the same team is involved.

Does the workflow validate the content for correctness?

Content quality starts with preventing errors. Workflows can and should prevent errors from happening.  

Workflows should check for multiple dimensions of content correctness, such as whether the content is:

  • Accurate – the workflow checks that dates, numbers, prices, addresses, and other details are valid.
  • Complete – the workflow checks that all required fields, assets, or statements are included.
  • Specific – the workflow accesses the most relevant specific details to include.
  • Up-to-date – the workflow validates that the data is the most recent available.
  • Conforming – the workflow checks that terminology and phrasing conform to approved usage.
  • Compliant – the workflow checks that disclaimers, warranties, commitments, and other statements meet legal and regulatory obligations.

Because performing these checks is not trivial, they are often not explicitly included in the workflow.  It’s more expeditious to place the responsibility for these dimensions entirely on an individual.  

Leverage machines to unburden users. Workflows should prevent obvious errors without requiring people to check themselves if an error is present. They should scrutinize text entry tasks to prevent input errors by including default or conditional values and auto-checking the formatting of inputs. In more ambiguous situations, they can flag potential errors that require an individual to look at. But they should never act too aggressively, where they generate errors through over-correction.

Error preemption is becoming easier as API integrations and AI tools become more prevalent. Many checks can be partially or fully automated by:

  • Applying logic rules and parameter-testing decision trees
  • Pulling information from other systems
  • Using AI pattern-matching capabilities 

Workflows must be self-aware. Workflows require hindsight and foresight. Error checking should be both reactive and proactive.  They must be capable of recognizing and remediating problems.

One of the biggest drivers of workflow problems is delays. Many delays are caused by people or contributions being unavailable because:

  • Contributors are overbooked or are away
  • Inputs are missing because they were never requested

Workflows should be able to anticipate problems stemming from resource non-availability.  Workflow tools can connect to enterprise calendars to know when essential people are unavailable to meet a deadline.  In such situations, it could invoke a fallback. The task could be reassigned, or the content’s publication could be a provisional release, pending final input from the unavailable stakeholder.

Workflows should be able to perform quality checks that transcend the responsibilities of a single individual to ensure these issues are not so dependent on one person. Before publication, it can monitor and check what’s missing, late, or incompatible. 

Automation promises to compress workflows but also carries risks. Workflows should check automation tasks in a staging environment to ensure they will perform as expected. Before making automation functionality generally available, the workflow staging will monitor discrete automation tasks and run batch tests on the automation of multiple items. Teams don’t want to discover that the automation they depend on doesn’t work when they have a deadline to meet. 

Does the workflow support sound governance?

Governance, risk, and compliance (GRC) are growing concerns for online publishers, particularly as regulators introduce more privacy, transparency, and online safety requirements. 

Governance provides reusable guidelines for performing tasks. It promotes consistency in quality and execution. It enables workflows to run faster and more smoothly by avoiding repeated questions about how to do things.  It ensures compliance with regulatory requirements and reduces reputation, legal, and commercial risks arising from a failure to vet content adequately.  

Workflow tools should promote three objectives:

  • Accountability (who is supposed to do what)
  • Transparency (what is happening compared to what’s supposed to happen)
  • Explainability (why tasks should be done in a certain way)

These qualities are absent from most content workflow functionality.

Defining responsibilities is not enough. At the most elemental level, a generic workflow specifies roles, responsibilities, and permissions. It controls access to content and actions, determining who is involved with a task and what they are permitted to do.  This kind of governance can prevent the wrong actors from messing up work, but they don’t help people responsible for the work from making unintended mistakes.

Assigned team members need support. The workflow should make it easier for them to make the correct decisions.  

Workflows should operationalize governance policies. However, if guidance is too intrusive, autocorrecting too aggressively, or making wrong assumptions, team members will try to short-circuit intrusive it.  

Discretionary decisions need guardrails, not enforcement. When a decision is discretionary, the goal should be to guide employees to make the most appropriate decision, not enforce a simple rule.  

Unfortunately, most governance guidance exists in documentation that is separated from workflow tools. Workflows fail to reveal pertinent guidance when it is needed. 

Incorporate governance into workflows at the point of decision. Bring guidance to the task so employees don’t need to seesaw between governance documents and workflow applications.  

Workflows can incorporate governance guidance in multiple ways by providing:

  • Guided decisions incorporating decision trees
  • Screen overlays highlighting areas to assess or check
  • Hints in the use interface
  • Coaching prompts from chatbots

When governance guidance isn’t specific enough for employees to make a clear decision, the workflow should provide a pathway to resolve the issue for the future. Workflows can include Issue management that triggers tasks to review and develop additional guidelines.

Does the workflow make planning more precise?

Bad plans are a common source of workflow problems.  Workflow planning tools can make tasks difficult to execute.

Planning acts like a steering wheel for a workflow, indicating the direction to go. 

Planning functionality is loosely integrated with workflow functionality, if at all. Some workflow tools don’t include planning, while those that do commonly detach the workflow from the planning.  

Planning and doing are symbiotic activities.  Planning functionality is commonly a calendar to set end dates, which the workflow should align with. 

But calendars don’t care about the resources necessary to develop the content. They expect that by choosing dates, the needed resources will be available.

Calendars are prevalent because content planning doesn’t follow a standardized process. How you plan will depend on what you know. Teams know some issues in advance, but other issues are unknown.  

Individuals will have differing expectations about what content planning comprises.  Content planning has two essential dimensions:

  • Task planning that emphasizes what tasks are required
  • Date planning that emphasizes deadlines

While tasks and dates are interrelated, workflow tools rarely give them equal billing.  Planning tools favor one perspective over the other.  

Task plans focus on lists of activities that need doing. The plan may have no dates associated with discrete tasks or have fungible dates that change.  One can track tasks, but there’s limited ability to manage the plan. Many workflows provide no scheduling or visibility into when tasks will happen.  At most, they show a Kanban board showing progress tracking.  They focus on if a task is done rather than when it should be done.

Design systems won’t solve workflow problems. Source: Utah design system

Date plans emphasize calendars. Individuals must schedule when various tasks are due. In many cases, those assigned to perform a task are notified in real time when they should do something. The due date drives a RAG (red-amber-green) traffic light indicator, where tasks are color-coded as on-track, delayed, or overdue based on dates entered in the calendar.

Manually selecting tasks and dates doesn’t provide insights into how the process will happen in practice.  Manual planning lacks a preplanning capability, where the software can help to decide in advance what tasks will be completed at specific times based on a forecast of when these can be done. 

Workflow planning capabilities typically focus on setting deadlines. Individuals are responsible for setting the publication deadline and may optionally set intermediate deadlines for tasks leading to the final deadline. This approach is both labor-intensive and prone to inaccuracies. The deadlines reflect wishes rather than realistic estimates of how long the process will take to complete. 

Teams need to be able to estimate the resources required for each task. Preplanning requires the workflow to: 

  1. Know all activities and resources that will be required  
  2. Schedule them when they are expected to happen.  

The software should set task dates based on end dates or SLAs. Content planning should resemble a project planning tool, estimating effort based on task times and sequencing—it will provide a baseline against which to judge performance.

For preplanning to be realistic, dates must be changeable. This requires the workflow to adjust dates dynamically based on changing circumstances. Replanning workflows will assess deadlines and reallocate priorities or assignments.

Does the workflow accelerate the completion of tasks?

Workflows are supposed to ensure work gets done on schedule. But apart from notifying individuals about pending dates, how much does the workflow tool help people complete work more quickly?  In practice, very little because the workflow is primarily a reminder system.  It may prevent delays caused by people forgetting to do a task without helping people complete tasks faster. 

Help employees start tasks faster with task recommendations. As content grows in volume, locating what needs attention becomes more difficult. Notifications can indicate what items need action but don’t necessarily highlight what specific sections need attention. For self-initiated tasks, such as evaluating groups of items or identifying problem spots, the onus is on the employee to search and locate the right items. Workflows should incorporate recommendations on tasks to prioritize.

Recommendations are a common feature in consumer content delivery. But they aren’t common in enterprise content workflows. Task recommendations can help employees address the expanding atomization of content and proliferation of content variants more effectively by highlighting which items are most likely relevant to an employee based on their responsibilities, recent activities, or organizational planning priorities.

Facilitate workflow streamlining. When workflows push manual activities from one person to another, they don’t reduce the total effort required by a team. A more data-driven workflow that utilizes semantic task tagging, by contrast, can reduce the number of steps necessary to perform tasks by:

  • Reducing the actions and actors needed 
  • Allowing multiple tasks to be done at the same time 

Compress the amount of time necessary to complete work. Most current content workflows are serial, where people must wait on others before being told to complete their assigned tasks. 

Workflows should shorten the path to completion by expanding the integration of: 

  1. Tasks related to an item and groups of related items
  2. IT systems and platforms that interface with the content management system

Compression is achieved through a multi-pronged approach:

  • Simplifying required steps by scrutinizing low-value, manually intensive steps
  • Eliminating repetition of activities through modularization and batch operations  
  • Involving fewer people by democratizing expertise and promoting self-service
  • Bringing together relevant background information needed to make a decision.

Synchronize tasks using semantically tagged workflows. Tasks, like other content types, need tags that indicate their purpose and how they fit within a larger model. Tags give workflows understanding, revealing what tasks are dependent on each other.  

Semantic tags provide information that can allow multiple tasks to be done at the same time. Tags can inform workflows:

  • Bulk tasks that can be done as batch operations
  • Tasks without cross-dependencies that can be done concurrently
  • Inter-related items that can be worked on concurrently

Automate assignments based on awareness of workloads. It’s a burden on staff to figure out to whom to assign a task. Often, task assignments are directed to the wrong individual, wasting time to reassign the task. Otherwise, the task is assigned to a generic queue, where the person who will do it may not immediately see it.  The disconnection between the assignment and the allocation of time to complete the task leads to delays.

The software should make assignments based on:

  • Job roles (responsibilities and experience) 
  • Employee availability (looking at assignments, vacation schedules, etc.) 

Tasks such as sourcing assets or translation should be assigned based on workload capacity. Content workflows need to integrate with other enterprise systems, such as employee calendars and reporting systems, to be aware of how busy people are and who is available.

Workload allocation can integrate rule-based prioritization that’s used in customer service queues. It’s common for tasks to back up due to temporary capacity constraints. Rule-based prioritization avoids finger-pointing. If the staff has too many requests to fulfill, there is an order of priority for requests in the backlog.  Items in backlog move up in priority according to their score, which reflects their predefined criticality and the amount of time they’ve been in the backlog. 

Automate routine actions and augment more complex ones. Most content workflow tools implement a description of processes rather than execute a workflow model, limiting the potential for automation. The system doesn’t know what actions to take without an underlying model.

A workflow model will specify automatic steps within content workflows, where the system takes action on tasks without human prompting. For example, the software can automate many approvals by checking that the submission matches the defined criteria. 

Linking task decisions to rules is a necessary capability. The tool can support event-driven workflows by including the parameters that drive the decision.

Help staff make the right decisions. Not all decisions can be boiled down to concrete rules. In such cases, the workflow should augment the decision-making process. It should accelerate judgment calls by making it easier for questions to be answered quickly.  Open questions can be tagged according to the issue so they can be cross-referenced with knowledge bases and routed to the appropriate subject matter expert.

Content workflow automation depends on deep integration with tools outside the CMS.  The content workflow must be aware of data and status information from other systems. Unfortunately, such deep integration, while increasingly feasible with APIs and microservices, remains rare. Most workflow tools opt for clunky plugins or rely on webhooks.  Not only is the integration superficial, but it is often counterproductive, where trigger-happy webhooks push tasks elsewhere without enabling true automation.

Does the workflow focus on goals, not just activities?

Workflow tools should improve the maturity of content operations. They should produce better work, not just get work done faster. 

Tracking is an administrative task. Workflow tracking capabilities focus on task completion rather than operational performance. With their administrative focus, workflows act like shadow mid-level managers who shuffle paper. Workflows concentrate on low-level task management, such as assignments and dates.

Workflows can automate low-level task activities; they shouldn’t force people to track them.   

Plug workflows’ memory hole. Workflows generally lack memory of past actions and don’t learn for the future. At most, they act like habit trackers (did I remember to take my vitamin pill today?) rather than performance trackers (how did my workout performance today compare with the rest of the week?)

Workflow should learn over time. It should prioritize tracking trends, not low-level tasks.

Highlight performance to improve maturity. While many teams measure the outcomes that content delivers, few have analytic tools that allow them to measure the performance of their work. 

Workflow analytics can answer: 

  • Is the organization getting more efficient at producing content at each stage? 
  • Is end-to-end execution improving?  

Workflow analytics can monitor and record past performance and compare it to current performance. They can reveal if content production is moving toward:

  • Fewer revisions
  • Less time needed by stakeholders
  • Fewer steps and redundant checks

Benchmark task performance. Workflows can measure and monitor tasks and flows, observing the relationships between processes and performance. Looking at historical data, workflow tools can benchmark the average task performance.

The most basic factor workflows should measure is the resources required. Each task requires people and time, which are critical KPIs relating to content production, 

Analytics can:

  1. Measure the total time to complete tasks
  2. Reveal which people are involved in tasks and the time they take.

Historic data can be used to forecast the time and people needed, which is useful for workflow planning. This data will also help determine if operations are improving.  

Spot invisible issues and provide actionable remediation.  It can be difficult for staff to notice systemic problems in complex content systems with multiple workflows. But a workflow system can utilize item data to spot recurring issues that need fixing.  

Bottlenecks are a prevalent problem. Workflows that are defined without the benefit of analytics are prone to develop bottlenecks that recur under certain circumstances. Solving these problems requires the ability to view the behavior of many similar items. 

Analytics can parse historical data to reveal if bottlenecks tend to involve certain stages or people. 

Historical workflow data can provide insights into the causes of bottlenecks, such as tasks that frequently involve:

  • Waiting on others
  • Abnormal levels of rework
  • Approval escalations

The data can also suggest ways to unblock dependencies through smart allocation of resources.  Changes could include:

  • Proactive notifications of forecast bottlenecks
  • Re-scheduling
  • Sifting tasks to an alternative platform that is more conducive

Utilize analytics for process optimization. Workflow tools supporting other kinds of business operations are beginning to take advantage of process mining and root cause analysis.  Content workflows should explore these opportunities.

Reinventing workflow to address the content tsunami

Workflow solutions can’t be postponed.  AI is making content easier to produce: a short prompt generates volumes of text, graphics, and video. The problem is that this content still needs management.  It needs quality control and organization. Otherwise, enterprises will be buried under petabytes of content debt.

Our twentieth-century-era content workflows are ill-equipped to respond to the building tsunami. They require human intervention in every micro decision, from setting due dates to approving wording changes. Manual workflows aren’t working now and won’t be sustainable as content volumes grow.

Workflow tools must help content professionals focus on what’s important. We find some hints of this evolution in the category of “marketing resource management” tools that integrate asset, work, and performance management. Such tools recognize the interrelationships between various content items, and what they are expected to accomplish.  

The emergence of no-code workflow tools, such as robotic process automation (RPA) tools, also points to a productive direction for content workflows. Existing content workflows are generic because that’s how they try to be flexible enough to handle different situations. They can’t be more specific because the barriers to customizing them are too high: developers must code each decision, and these decisions are difficult to change later. 

No code solutions give the content staff, who understand their needs firsthand, the ability to implement decisions about workflows themselves without help from IT. Enterprises can build a more efficient and flexible solution by empowering content staff to customize workflows.

Many content professionals advocate the goal of providing Content as a Service (CaaS).  The content strategist Sarah O’Keefe says, “Content as a Service (CaaS) means that you make information available on request.” Customers demand specific information at the exact moment they need it.  But for CaaS to become a reality, enterprises must ensure that the information that customers request is available in their repositories. 

Systemic challenges require systemic solutions. As workflow evolves to handle more involved scenarios and provide information on demand, it will need orchestration.  While individuals need to shape the edges of the system, the larger system needs a nervous system that can coordinate the activities of individuals.  Workflow orchestration can provide that coordination.

Orchestration is the configuration of multiple tasks (some may be automated) into one complete end-to-end process or job. Orchestration software also needs to react to events or activities throughout the process and make decisions based on outputs from one automated task to determine and coordinate the next tasks.”  

Orchestration is typically viewed as a way to decide what content to provide to customers through content orchestration (how content is assembled) and journey orchestration (how it is delivered).  But the same concepts can apply to the content teams developing and managing the content that must be ready for customers.  The workflows of other kinds of business operations embrace orchestration. Content workflows must do the same. 

Content teams can’t pause technological change; they must shape it.  A common view holds that content operations are immature because of organizational issues. Enterprises need to sort out the problems of how they want to manage their people and processes before they worry about technology. 

We are well past the point where we can expect technology to be put on hold while sorting out organizational issues. These issues must be addressed together. Other areas of digital transformation demonstrate that new technology is usually the catalyst that drives the restructuring of business processes and job roles. Without embracing the best technology can offer, content operations won’t experience the change it needs.

– Michael Andrews

Categories
Content Engineering Personalization

Orchestrating the assembly of content

Structured content enables online publishers to assemble pieces of content in multiple ways.  However, the process by which this assembly happens can be opaque to authors and designers. Read on to learn how orchestration is evolving and how it works.

To many people, orchestration sounds like jargon or a marketing buzzword. Yet orchestration is no gimmick. It is increasingly vital to developing, managing, and delivering online content. It transforms how publishers make decisions about content, bringing flexibility and learning to a process hampered in the past by short-term planning and jumbled, ad-hoc decisions.  

Revealing the hidden hand of orchestration

Orchestration is both a technical term in content management and a metaphor. Before discussing the technical aspects of orchestration, let’s consider the metaphor.  Orchestration in music is how you translate a tune into a score that involves multiple instruments that play together harmoniously. It’s done by someone referred to as an arranger, someone like Quincy Jones. As the New Yorker once wrote: “Everyone knows Quincy Jones’s name, even if no one is quite sure what he does. Jones got his start in the late nineteen-forties as a trumpeter, but he soon mastered the art of arranging jazz—turning tunes and melodies into written music for jazz ensembles.”

Much like music arranging, content orchestration happens off stage, away from the spotlight. It doesn’t get the attention given to UI design. Despite its stealthy profile, numerous employees in organizations become involved with orchestration, often through small-scale A/B testing by changing an image or a headline. 

Orchestration typically focuses on minor tweaks to content, often cosmetic changes. But orchestration can also address how to assemble content on a bigger scale. The emergence of structured content makes intricate, highly customized orchestration possible.

Content assembly requires design and a strategy. Few people consider orchestration when planning how content is delivered to customers. They generally plan content assembly by focusing on building individual screens or a collection of web pages on a website. The UI design dictates the assembly logic and reflects choices made at a specific time.  While the logic can change, it tends to happen only in conjunction with changes to the UI design. 

Orchestration allows publishers to specify content assembly independently of its layout presentation. It does so by approaching the assembly process abstractly: evaluating content pieces’ roles and purposes that address specific user scenarios.

Assembly logic is becoming distributed. Content assembly logic doesn’t happen in one place anymore. Originally, web teams created content for assembly into web pages using templates defined by a CMS on the backend. In the early 2000s, frontend developers devised ways to change the content of web pages presented in the browser using an approach known initially as Ajax, a term coined by the information architect Jesse James Garrett. Today, content assembly can happen at any stage and in any place. 

Assembly is becoming more sophisticated. At first, publishers focused on selecting the right web page to deliver. The pages were preassembled – often hand-assembled. Next, the focus shifted to showing or hiding parts of that web page by manipulating the DOM (document object model).  

Nowadays, content is much more dynamic. Many web pages, especially in e-commerce, are generated programmatically and have no permanent existence.  “Single page applications” (SPAs) have become popular, and the content will morph continuously. 

The need for sophisticated approaches for assembling content has grown with the emergence of API-accessible structured content. When content is defined semantically, rather than as web pages, the content units are more granular. Instead of simply matching a couple of web page characteristics, such as a category tag and a date, publishers now have many more parameters to consider when deciding what to deliver to a user.

Orchestration logic is becoming decoupled from applications. While orchestration can occur within a CMS platform, it is increasingly happening outside the CMS to take advantage of a broader range of resources and capabilities. With APIs growing in coordinating web content, much content assembly now occurs in a middle layer between the back-end storing the content and the front-end presenting it. The logic driving assembly is becoming decoupled from both the back-end and front-end. 

Publishers have a growing range of options outside their CMS for deciding what content to deliver.  Tools include:

  • Digital experience, composition, and personalization orchestration engines (e.g., Conscia, Ninetailed)
  • Graph query tools (e.g., PoolParty)
  • API federation management tools (e.g., Apollo Federation)

These options vary in their aims and motivations, and they differ in their implementations and features. Their capabilities are sometimes complementary, which means they can be used in combination. 

Orchestration inputs that frame the content’s context

Content structuring supports extensive variation in the types of content to present and what that content says. 

Orchestration involves more than retrieving a predefined web page.  It requires considering many kinds of inputs to deliver the correct details. 

Content orchestration will reflect three kinds of considerations:

  1. The content’s intent – the purpose of each content piece
  2. The organization’s operational readiness to satisfy a customer’s need
  3. The customer or user’s intent – their immediate or longer-term goal

Content characteristics play a significant role in assembly. Content characteristics define variations among and within content items. An orchestration layer will account for characteristics of available content pieces, such as:

  • Its editorial role and purpose, such as headings, explanations, or calls to action
  • Topics and themes, including specific products or services addressed
  • Intended audience or customer segment
  • Knowledge level such as beginner or expert
  • Intended journey or task stage
  • Language and locale
  • Date of creation or updating
  • Author or source
  • Size, length, or dimensions
  • Format and media
  • Campaign or announcement cycle
  • Product or business unit owner
  • Location information, such as cities or regions that are relevant or mentioned
  • Version 

Each of these characteristics can be a variable and useful when deciding what content to assemble. They indicate the compatibility between pieces and their suitability for specific contexts.

Other information in the enterprise IT ecosystem can help decide what content to assemble that will be most relevant for a specific context of use. This information is external to the content but relevant to its assembly.

Business data is also an orchestration input. Content addresses something a business offers. The assembled content should link to business operations to reflect what’s available accurately.

The assembled content will be contextually relevant only if the business can deliver to the customer the product or services that the content addresses. Customers want to know which pharmacy branches are open now or which items are available for delivery overnight.  The assembled content must reflect what the business can deliver when the customer seeks it.

The orchestration needs to combine content characteristics from the CMS with business data managed by other IT systems. Many factors can influence what content should be presented, such as:

  • Inventory management data
  • Bookings and orders data
  • Facilities’ capacity or availability
  • Location hours
  • Pricing information, promotions, and discount rules
  • Service level agreement (SLA) rules
  • Fulfillment status data
  • Event or appointment schedules
  • Campaigns and promotions schedule
  • Enterprise taxonomy structure defining products and operating units

Business data have complex rules managed by the IT system of record, not the CMS or the orchestration layer.  For content orchestration, sometimes it is only necessary to provide a “flag,” checking whether a condition is satisfied to determine which content option to show.

Customer context is the third kind of orchestration input. Ideally, the publisher will tailor the content to the customer’s needs – the aim of personalization.  The orchestration process must draw upon relevant known information about the customer: the customer’s context.

The customer context encompasses their identity and their circumstances. A customer’s circumstances can change, sometimes in a short time.  And in some situations, the customer’s circumstances dictate the customer’s identity. People can have multiple identities, for example, as consumers, business customers at work, or parents overseeing decisions made by their children.

Numerous dimensions will influence a customer’s opinions and needs, which in turn will influence the most appropriate content to assemble. Some common customer dimensions include:

  • Their location
  • Their personal characteristics, which might include their age, gender, and household composition, especially when these factors directly influence the relevance of the content, for example, with some health topics
  • Things they own, such as property or possessions, especially for content relating to the maintenance, insurance, or buying and selling of owned things
  • Their profession or job role, especially for content focused on business and professional audiences
  • Their status as a new, loyal, or churned customer
  • Their purchase and support history

The chief challenge in establishing the customer context is having solid insights.  Customers’ interactions on social media and with customer care provide some insights, but publishers can tap a more extensive information store.  Various sources of customer data could be available:

  • Self-disclosed information and preferences to the business (zero-party data or 0PD)
  • The history of a customer’s interactions with the business (first-party data or 1PD) 
  • Things customers have disclosed about themselves in other channels such as social media or survey firms (second-party data or 2PD)
  • Information about a cohort they are categorized as belonging to, using aggregated data originating from multiple sources (third-party data or 3PD)

Much of this information will be stored in a customer data platform (CDP), but other data will be sourced from various systems.  The data is valid only to the extent it is up-to-date and accurate, which is only sometimes a safe assumption.

Content behavior can shape the timing and details assembled in orchestration. Users can signal their intent through their interaction with content. The user’s decisions while interacting with content can signal their intentions.  Some behavior variables include:

  • Source of referral 
  • Previously viewed content 
  • Expressed interest in topics or themes based on prior content consumed
  • Frequency of repeat visits 
  • Search terms used 
  • Chatbot queries submitted
  • Subscriptions chosen or online events booked
  • Downloads or requests for follow-up information
  • The timing of their visit in relation to an offer 

The most valuable and reliable signals will be specific to the context. Many factors can shape intent, so many potential factors will not be relevant to individual customers. Just because some factors could be relevant in certain cases does not imply they will be applicable in most cases. 

Though challenging, leveraging customer intent offers many opportunities to improve the relevance of content. A rich range of possible dimensions is available. Selecting the right ones can make a difference. 

Don’t rely on weak signals to overdetermine intent. When the details about individual content behavior or motivations are scant, publishers sometimes rely on collective behavioral data to predict individual customer intentions.  While occasionally useful, predictive inputs about customers can be based on faulty assumptions that yield uneven results. 

Note the difference between tailoring content to match an individual’s needs and the practice of targeting. Targeting differs from personalization because it aims to increase average uptake rather than satisfy individual goals. It can risk alienating customers who don’t want the proffered content.

Draw upon diverse sources of input. By utilizing a separate layer to manage orchestration, publishers, in effect, create a virtual data tier that can federate and assess many distinct and independent sources of information to support decisions relating to content delivery. 

An orchestration layer gives publishers greater control over choosing the right pieces of content to offer in different situations. Publishers gain direct control over parameters to select,  unlike many AI-powered “decision engines” that operate like a black box and assume control over the content chosen.

The orchestration score

If the inputs are the notes in orchestration, the score is how they are put together – the arrangement. A rich arrangement will sometimes be simple but often will be sophisticated. 

Orchestration goes beyond web search and retrieval. In contrast to a ordinary web search, which retrieves a list of relevant web pages, orchestration queries must address many more dimensions. 

In a web search, there’s a close relationship between what is requested and what is retrieved. Typically, only a few terms need matching. Web search queries are often loose, and the results can be hit or miss. The user is both specifying and deciding what they want from the results retrieved.

In orchestration, what is requested needs to anticipate what will be relevant and exclude what won’t be. The request may refer to metadata values or data parameters that aren’t presented in the content that’s retrieved. The results must be more precise. The user will have limited direct input into the request for assembled content and limited ability to change what is provided to them.

Unlike a one-shot web search process, in orchestration, content assembly involves a multistage process.  

The orchestration of structured content is not just choosing a specific web page based on a particular content type.  It differs in two ways:

  1. You may be combining details from two (or more) content types.  
  2. Instead of delivering a complete web page associated with each content type (and potentially needing to hide parts you don’t want to show), you select specific details from content items to deliver as an iterative procedure.

Unpacking the orchestration process. Content orchestration consists of three stages:

  1. FIND stage: Choose which content items have relevant material to support a user scenario
  2. MATCH stage: Combine content types that, if presented together, provide a meaningful, relevant experience
  3. SELECT and RETURN stage: Choose which elements within the content items will be most relevant to deliver to a user at a given point in time

Find relevant content items. Generally, this involves searching metadata tags such as taxonomy terms or specific values such as dates. Sometimes, specific words in text values are sought. If we have content about events, and all the event descriptions have a field with the date, it is a simple query to retrieve descriptions for events during a specified time period.

Typically, a primary content type will provide most of the essential information or messages. However, we’ll often also want to draw on information and messages from other content types to compose a content experience. We must associate different types of items to be able to combine their details.

Match companion content types. What other topics or themes will provide more context to a message? The role of matching is to associate related topics or tasks so that complementary information and messages can be included together.

Graph queries are a powerful approach to matching because they allow one to query “edges” (relationships) between “nodes” (content types.)  For example, if we know a customer is located in a specific city, we might want to generate a list of sponsors of events happening in that city.  The event description will have a field indicating the city. It will also reference another content type that provides a profile of event sponsors.  It might look like this in a graph query language like GQL, with the content types in round brackets and the relationships in square brackets.

MATCH (:Event WHERE location=”My City”) - [:SponsoredBy] -> (:SponsorProfile)

We have filtered events in the customer’s city (fictiously named My City) and associated content items about sponsors who have sponsored those events. Note that this query hasn’t indicated what details to present to users. It only identifies which content types would be relevant so that various types of details can be combined. 

Unlike in a common database query, what we are looking for and want to show are not the same. 

Select which details to assemble. We need to decide what information within a relevant content type which details will be of greatest interest to a user. Customers want enough details for the pieces to provide meaningful context. Yet they probably won’t want to see everything available, especially all at once – that’s the old approach of delivering preassembled web pages and expecting users to hunt for relevant information themselves.

Different users will want different details, necessitating decisions about which details to show. This stage is sometimes referred to as experience composition because the focus is on which content elements to deliver. We don’t have to worry about how these elements will appear on a screen, but we will be thinking about what specific details should be offered.

GraphQL, a query language used in APIs, is very direct in allowing you to specify what details to show. The GraphQL query mirrors the structure of the content so that one can decide which fields to show after seeing which fields are available. We don’t want to show everything about a sponsor, just their name, logo, and how long they’ve been sponsoring the event.  A hypothetical query named “local sponsor highlights” would extract only those details about the sponsor we want to provide in a specific content experience.

Query LocalSponsorHighlights {
… on SponsorProfile {
name
logo
sponsorSince
} }

The process of pulling out specific details will be repeated iteratively as customers interact with the content.

Turning visions into versions

Now that we have covered the structure and process of orchestration let’s look at its planning and design. Publishers enjoy a broad scope for orchestrating content. They need a vision for what they aim to accomplish. They’ll want to move beyond the ad hoc orchestration of page-level optimization and develop a scenario-driven approach to orchestration that’s repeatable and scaleable.

Consider what the content needs to accomplish. Content can have a range of goals. They can explicitly encourage a reader to do something immediately or in the future. Or they encourage a reader’s behavior by showing goodwill and being helpful enough that the customer wants to do something without being told what to do.

Content goalImmediate 
(Action outcome)
Consequent
 (Stage outcome)
Explicit 
(stated in the content)
CTA (call to action) conversionContact sales or visit a retail outlet
Implicit 
(encouraged by the content)
Resolve an issue without contacting customer supportRenew their subscription

Content goals must be congruent with the customer’s context. If customers have an immediate goal, then the content should be action-oriented. If their goal is longer-term, the content should focus on helping the customer move from one stage to another.

Orchestration will generate a version of the content representing the vision of what the pieces working together aim to accomplish.

Specify the context.  Break down the scenario and identify which contextual dimensions are most critical to providing the right content. The content should adapt to the user context, reflect the business context, and provide users with viable options. The context includes:

  • Who is seeking content (the segment, especially when the content is tailored for new or existing customers, or businesses or consumers, for example)
  • What they are seeking (topics, questions, requests, formats, and media)
  • When they are seeking it (time of day, day of the week, month, season, or holiday, all can be relevant)
  • Where they are seeking it (region, country, city, or facility such as an airport if relevant)
  • Why (their goal or intent as far as can be determined)
  • How (where they started their journey, channels used, how long have they pursuing task)

Perfecting the performance: testing and learning

Leonard Bernstein conducts the New York Philharmonic in a Young People’s Concert. Image: Library of Congress


An orchestral performance is perfected through rehearsal. The performance realized is a byproduct of practice and improvisation.

Pick the correct parameters. With hundreds of parameters that could influence the optimal content orchestration, it is essential that teams not lock themselves into a few narrow ones. The learning will arise from determining which factors deliver the right experience and results in which circumstances. 

Content parameters can be of two kinds:

  1. Necessary characteristics tell us what values are required 
  2. Contingency characteristics indicate values to try to find which ones work best
Specifies in the orchestrationDetermines in the contentOutcome expected
Necessary characteristics

(tightly defined scenarios)
What values are required in a given situationWhich categorical version or option the customer getsThe right details to show to a given customer in a given situation
Contingency characteristics

(loosely defined scenarios)
What values are allowed in a given situationWhich versions could be presentedCandidate options to present to learn which most effectively matches the customer’s needs

The two approaches are not mutually exclusive. More complex orchestration (sometimes referred to as “multihop” queries) will involve a combination of both approaches.

Necessary characteristics reflect known and fixed attributes in the customer or business context that will affect the correct content to show. For example, if the customer has a particular characteristic, then a specific content value must be shown. The goal should be to test that the orchestration is working correctly – that the assumptions about the context are correct.  For example, there are no wrong assumptions or missing ones. This dimension is especially important for aspects that are fixed and non-negotiable. The content needs to adapt to these circumstances, not ignore them. 

Contingency characteristics reflect uncertain or changeable attributes relating to the customer’s context. For example, if the customer has had any one of several characteristics now or in the past, try showing any one of several available content values to see which works best given what’s known. The orchestration will prioritize variations randomly or in some ranked order based on what’s available to address the situation.

You can apply the approach to other situations involving uncertainty. When there are information gaps or delays, contingency characteristics can apply to business operations variables and to the content itself.  The goal of using contingency characteristics is to try different content versions to learn what’s most effective in various scenarios.  

Be clear on what content can influence. We have mostly looked at the customer’s context as an input into orchestration. Customers will vary widely in their goals, interests, abilities, and behaviors. A large part of orchestration concerns adapting content to the customer’s context. But how does orchestration impact the customer? In what ways might the customer’s context be the outcome of the content?

Consider how orchestration supports a shift in the customer’s context. Orchestration can’t change the fixed characteristics of the customer. It can sway ephemeral characteristics, especially content choices, such as whether the customer has requested further information.  And the content may guide customers toward a different context. 

Context shifting involves using content to meet customers where they are so they can get where they want to be. Much content exists to change the customer’s context by enabling them to resolve a problem or encouraging them to take action on something that will improve their situation. 

The orchestration of content needs to connect to immediate and downstream outcomes.  Testing orchestration entails looking at its effects on online content behavior and how it influences interactions with the business in other areas. Some of these interactions will happen offline.  

The task of business analytics is to connect orchestration outputs with customer outcomes. The migration of orchestration to an API layer should open more possibilities for insights and learning. 

– Michael Andrews