Microsoft Project Cortex Ushers in the Age of Topic Computing
The announcement states: “Project Cortex is a new service in Microsoft 365 that … uses AI to automatically classify all your content into topics to form a knowledge network .... It empowers people and teams with knowledge, learning, and expertise, in the flow of their everyday work … [and it] improves individual productivity and organizational intelligence, helping identify experts on specific topics, and surfacing knowledge through interactive experiences across Microsoft 365, such as in the Office apps, Outlook, and Microsoft Teams.”
In short, Cortex purports to organize your emails, documents and meeting recordings, and group this information by topics, so you focus on your work; what I call “topic computing.” I have been predicting an age of topic computing for several years as a practical solution to acute information overload, namely organizing all of our "stuff" the way the brain works so we can focus on what really matters most. People intuitively think in terms of topics like projects, products, services and customers rather than documents, emails and tasks. So presenting information by topics not only increases productivity, it also reduces the stress associated with information overload. As a brain-like way of presenting information, Cortex is an appropriate name for this initiative.
A History of Topic Computing
Of course, Microsoft didn’t invent topic computing, nor did it come up with the idea of presenting information the way the brain works. The idea is as old as Aristotle.
In more modern times, FDR’s chief scientific advisor, Vannevar Bush, espoused such an approach back in 1945 as a way to deal with the information overload he experienced emanating from an explosion of professional articles and correspondence. Bush envisioned a device to organize information by topics, which he called Memex. Bush’s "invention" consisted of a desk that incorporated a book the user would use to build an index of interesting topics. The user would photocopy, record (initially voice and later video), or type text into the Memex interface, whereupon it would be captured into the record. The user would then assign each piece of content a set of topics (think metadata) and the topics would be added to the index. Later, when the user would select a topic from the index, they could locate related content which was stored on microfilm.
The Memex was never built, although several attempts have been made over the years to refine the device concept using newer technologies. Ultimately, its success relied upon a massive amount of user input and the user experience was too complicated to derive real value. Plus, the Memex was a personal device. There was no uniform or standard way to define or share common topics.
Related Article: Vannevar Bush Points the Way to a New Era of Computing
Factors Exacerbating Information Overload
As businesses began to operate as networked organizations, the sources of information overload shifted from journals, letters and memos, to shared content like emails, documents and tasks. So the need to assimilate collective information, in addition to individual content, exacerbated the problem of overload. In today’s workplace, productivity has largely become a collective goal rather than a personal achievement, so a solution to overload must address the group as well as the individual.
Different attempts to organize information by topics have been tried over the years but ultimately failed because of technical difficulties. Some of these hurdles include the following:
- Information is spread across many locations and is not accessible to a single processing engine. Aggregating information in an organized fashion has been challenging because it exists in disparate repositories and it is organized in various, non-standard ways.
- Different information sources use dissimilar terms to describe the same entities. For example, the term "module" and "unit" might be used interchangeably by people to refer to the same thing. It is difficult for machines to reconcile such discrepancies that are readily understood by humans.
- The meaning of terms is often ambiguous. For example, if the word apple is contained in a message, is the author referring to the fruit, the company or the record company? Similarly, the use of abbreviations and acronyms is usually specific to a particular group or organization. For example, Wikipedia defines 16 popular uses of the acronym, SMB. In your organization this term might mean something else entirely.
- Most organizational information is unstructured, often contained in freeform artifacts like as emails, documents, event descriptions and more. Ferreting out the important topics from freeform text has been an especially difficult task. Plus, a term that is critical in one organization might not be important to another. Trying to have a machine predict what is important without human guidance is very difficult. Over the years, many organizations have tried to manually classify information using metadata tags as a form of topics, but this method depends on the willingness of employees to enter metadata. Traditionally, employees have resisted tagging information because of the extra effort involved. Even when they do, the method is susceptible to human error and differences in judgement.
- The relationship between individual pieces of information and their relevance to people has been near impossible to predict with any degree of accuracy. Until now, there simply haven’t been any obvious signals to help explain why an email or document might be interesting to someone at a particular point in time. For example, a colleague in a different branch office might be struggling with a similar customer complaint, but it would be hard to pair the two work processes, since the two employees don’t work together, and they are dealing with different customers.
- The amount of information being created and stored is exploding, outstripping the ability for traditional systems to analyze them in near real-time.
Related Article: Information Overload Comes in 3 Flavors: Here's How to Combat It
New opportunities to manage information by topics have opened as new technologies made their way into the workplace. Specifically, the introduction of the cloud as an infrastructure for shared data across many applications, and the advent of practical AI — in the form of natural language processing and machine learning — create a new reality for information managers. For the first time, business content can be aggregated into a single place: the data cloud.
The aggregation of data into a single location also enables the exploitation of a data graph, which is a construct for capturing and detailing relationships between content and people. Facebook and Google have built data graphs for the consumer market, and now the Microsoft Graph captures the relationships between business people and their content, like email, documents, tasks, events, photos and more. It captures the authors of content, and it tracks who is editing and sharing the content. As a result, the graph can also register usage information like what content is popular.
Related Article: Will Microsoft Graph Deliver on the Promises of the Social Graph?
The Premise and the Promise of Project Cortex
Microsoft has developed the infrastructure components needed to address the technical challenges that previously inhibited topic computing. It has solved the aggregation and standardization of data storage challenge by storing each person’s documents, emails, tasks, events, photos, and all other pieces of business information into a single location, the Office 365 cloud. The Microsoft Graph provides the connections between people and content. And finally, Microsoft can tap the consumer AI experience harvested from 10 years of Bing search experience in the consumer market, experience it can now bring to bear to ferret out topics embedded in unstructured emails, documents and other artifacts.
Putting all the pieces together, the recently announced Project Cortex purports to build a knowledge graph of enterprise topics, showing how the topics people are working on are logically connected, as show in the figure below.
Using the intelligence of this graph, Cortex claims to tie content to topics like the name of a customer or product, and to present all the information related to that customer or product in a single "topic card," as seen in the graphic below, as well as to point out related topics.
Rediscovering Collaboration in the Office
How to create a plan that brings collaboration back into the office
Modernized Knowledge Management: The New Driving Force Behind a Great Employee and Customer Experience
How creating engaged workplaces drives business impact and delivers optimal customer experiences
Making the Employee Experience Empathetic to Frontline Workers
Learn how leading organizations use EX tools to connect people with the resources they need in the field or on the move.
Voice of the Employee: Aligning Employee Listening with User Experience Tools and Services
Keys to ensuring technology platforms are intuitive and accessible for all
Project Cortex was detailed in a number of keynote addresses and technical sessions at Ignite 2019. If it works as advertised — and that's a big IF — Cortex may forever change the way we work. Microsoft has announced general worldwide availability of this technology in the first half of 2020, which doesn’t seem like enough time to go from concept to fully baked service, especially one of this magnitude.
Project Cortex Raises a Few Big Questions
According to Microsoft spokespeople, Cortex has been under development as a secret project for the last two years. As per information presented at the conference, it has been tested within Microsoft and has begun testing at Mott McDonald, a global consulting company, which was the sole customer reference included in the product announcement.
Currently, no packaging or pricing details are available for Cortex.
One of the big questions for Cortex is how much human intervention will be needed to identify important topics within an organization? At one of the Ignite technical sessions, it was suggested that human topic curation (what Microsoft is calling "light editing") is helpful, but not necessary, for topic extraction and identification.
How well a fully automated system would work is unclear. Yet if the system depends on people in the organization to pinpoint interesting topic terms, adoption will be much slower. Asking users to identify topic terms, filter out bad topics, and augment details about the topic is challenging. It remains to be seen if employees will cooperate. While Microsoft may call the task light editing, it doesn't appear any lighter than applying metadata when capturing documents to shared repositories like SharePoint, which we've seen has been a futile effort. Workers are rarely eager to take the extra steps of attaching tags to content when saving documents while in the flow of a business process, regardless of the value they may derive later. From one of the demos presented at the conference, curating topics doesn’t appear any easier than applying metadata — so the user experience will be a critical factor in Cortex’s success.
Related Article: Managing Metadata: Any Volunteers?
What Does Project Cortex Mean for You?
Topic computing presents a paradigm shift in how we help people focus on what matters most in the workplace. Project Cortex offers enormous potential for dealing with information overload at work. As such, there is a huge incentive to get it right.
On the other hand, the 2020 release schedule seems a bit ambitious. It is reasonable to assume Microsoft will need additional experience with more diverse customer experiences to develop a satisfactory offering for the broad market.
About the Author
David is a product expert with extensive experience leading information-intensive technology organizations. His specialty is helping organizations “do it right the first time”— get to market quickly and successfully through a structured process of working closely with design partners from day one.