What is Watson?

Ed. Note: This is the second installment in a three-part series exploring IBM’s multi-year project to create the ultimate Big Data patient diagnostic and treatment plan tool using its unique Watson natural language cognitive system. The first part defined the scope of the problem. This part will look at what Watson is and what makes it a new generation of computing system and arguably the ultimate Big Data query engine. The third part will cover IBM’s vision for Watson in healthcare.

On Jeopardy Watson was billed as a supercomputer because it ran on a very large Power system, consumed a large amount of processing space, and used a tremendous amount of memory, Watson CTO, IBM Fellow and VP Robert High says. “But in reality Watson is at its heart a piece of software designed to answer questions against a knowledge base with incredibly high accuracy. So we refer to it as a cognitive system, not as a supercomputer or mainframe, because the level of fidelity and accuracy we can get from answering natural language questions is uncanny with respect to how much it captures the capabilities human beings have. So as a cognitive system, it is very much a natural language processing system that does what we refer to as deep natural language processing.”

Watson can take a question that someone literally asks out loud, and run that against hundreds of millions of pages of documents, and give a highly accurate answer. This is much more complicated than it sounds.

“For humans, the process of asking a question, understanding the question, & understanding the body of knowledge we are reading from to answer that question, is second nature,” High says. “We don’t think about disambiguating the question. In reality there is a tremendous amount of ambiguity.”

Contextualizing intelligence

.

Suppose, for instance, Ken Jennings had read the clue “Jodie Foster took this home for her role in Little Man Tate.” A person would quickly realize first that Jennings wants to be told what Foster took home, which means it must be something easily transportable and valuable. That implies that the question refers to an award of some sort. Since Jodie Foster is a movie actress and director, the person would quickly think of motion picture awards and the Oscars. With an Internet search at most that person would quickly discover that Jodie Foster won an Academy Award for Best Director for the movie. And a person would realize that the Ken Jennings reference means this is a Jeopardy clue, so the correct reply must be in the form of a question: “What is an Academy Award?”

Getting a computer to realize all that from a simple statement, however, is a major artificial intelligence challenge, High says. First, consider that Jennings made a statement, not a question. Then consider the many things that people take home – groceries, colds, babies, pets, frustrations from work. Then consider the things that people win that they do not take home – a contract, a corner office, a court case. The simple approach might be to create a rule that associates the phrase “take something home” with winning an award, but then you would need similar rules for all the other things that people take home, which quickly becomes an impossible task. Nor could any set of rules cover all the things people might do with things they win.

“So rather than building rules, we tried to follow the same processes that we as human beings take when we resolve ambiguities,” High said. “That is, we add context. There is no particular rule to it.” But in the context of Jody Foster, who directed Little Man Tate, one association would be an Academy Award, which includes a statuette of a size and value that winners would want to take home. “So we as human beings disambiguate by adding a lot of context. That’s essentially what Watson does, too. It takes a question like this and teases it apart, and then it goes out and seeks as much context as it can to associate with both temporal, spatial and juxtaposition aspects of the language used to ask the question, and, by the way, the language we use to find answers. And in the case of Jeopardy we had to do that in three seconds or less because on average you have three seconds to determine that you have an answer and be the first to buzz in. And we had to do that against about 60 million pages of documentation.”

Developing Watson

.

In the two years since Watson’s Jeopardy debut, High says, the development team has done a lot of work on engineering Watson implementations. Today it is much more compact, and it is about 240X faster. This is another reason that IBM does not refer to Watson as a supercomputer. Depending on the size of the problem, Watson can run on systems of a variety of sizes. In fact, in its recent announcement of the new smaller Power7 RISC computer systems for SMBs, IBM said that the environment running on them includes elements of Watson technology. So if you have ever dreamed of having that computer from Star Trek in your office, that dream may now be achievable.

In a sense, High said, this is software-led computing. However, the software has a very strong synergy with the IBM Power7 RISC systems, which “are uniquely architected to optimally run the kinds of workloads that are inherent in a Watson and other analytic systems. We have engineered Watson to exploit that unique systems architecture – machine instruction set, cache organization, bus-channel management, etc. We maintain a very strong relationship between the software and hardware to drive even more optimization around these kinds of systems architecture elements.

“In terms of go-to-market, we have been concentrating Watson in spaces where we can make an enormous difference with the kind of capability we just described. Which is why we’re in the healthcare industry.”

Bert Latamore is a freelance writer covering the intersection of IT and business for SiliconANGLE. He is a frequent contributor to CrowdChats focused on theCUBE coverage of major IT industry events and site editor at Wikibon.org. He has 35 years’ experience covering the IT industry including four with Gartner, five with Meta Group, and eight with Wikibon. He lives in the Virginia Blue Ridge Mountains with his wife, Moire, and their dog, cat and macaw. In his spare time he enjoys reading, hiking and photography.

Premium Research

Wikibon argues strongly against Revolution towards a 3rd platform. The conclusion from this analysis is that applications will evolve; conversion should be avoided like the plague. The greatest opportunity is to continuously adapt today's operational applications by the addition of real-time or near real-time analytics applied directly to the current organizational processes and applications that support these processes. This is likely to translate to the greatest value to most organizations, and where possible avoid the risks of converting systems. The study of organizations that have applied real-time analytics to their current operational systems have shown incredible improvements in lower costs and greater adaptability. Business and IT executives should understand the enormous potential for adding decision automation through real-time analytics to current operational applications in their organizations. New technologies should be judged by their ability to support real-time analytics applied to operational systems, and supporting incremental improvement over time.

In a recent web-based survey conducted by Wikibon, 300 North American enterprises whom had either been utilizing, or considering the adoption of public cloud, answered questions regarding IaaS (Infrastructure as a Service) perceptions and usages. These questions varied in topic but were centered around an examination of which workloads were best suited for usage in the public cloud. This research examines a few additional key insights that shed some light on the growing IaaS world.

Today's Technology infrastructure management is largely non-differentiated and wasteful. Technology executives must re-think the strategic role of human capital and begin to implement new ways to consume IT as a service. This post draws on the learnings of senior executive Alan Nance from Royal Philips who is dogmatic in its approach to transforming its infrastructure to a service model.

There have only been two successful volume introductions into the marketplace in the last 50 years - DRAM and NAND flash. There has to be a clear volume case with good economics for 3D XP to be able to gain a foothold in consumer products. Without volume in the consumer space, there is unlikely to be much volume traction in the enterprise space. CIOs, CTOs and enterprise professionals should take a wait and see stance, and monitor the adoption of 3D XP in the consumer and military spaces. If and when there is volume production for 3D XP, enterprise adoption should start about two years later.

The use of open source software continues to accelerate and expand in the marketplace, especially in areas where technology is significantly disrupting established business models. IT organizations should be actively seeking to understand how open communities operate, how different licensing models work, and how they can be more actively engaged with both the vendors and communities that are shaping open source software.

CIOs understand that a clear cloud strategy is critical for IT today. Wikibon believes the biggest mistake organizations can make is converting major applications into the public cloud (including SaaS) without thinking about the implications to their existing business process workflows. Wikibon recommends that IT develop and implement a hybrid cloud strategy using the existing management workflows and compliance processes for both the public and private cloud components in the hybrid cloud.

In 2014, Wikibon defined a new category "Server SAN" that sits at the intersection of software-defined storage, hyperscale methodologies and converged infrastructure. This article is the executive summary of primary research that gives the status of the market, examines the vendor ecosystem, lays forth the revenue and 10 year forecast and gives direction for expansion beyond simple "hyperconverged infrastructure". This information is available for public consumption, the full research is available to Wikibon clients.

In this research paper, Wikibon looks back at the introductory Server SAN research, adjusts the Server SAN definition to include System Drag, and increases the speed of adoption of Server SAN based on very fast adoption from 2012 to 2014. The overall growth of Server SAN is projected to be about a 23% CAGR from 2014 to 2026, with a faster growth from 2014 to 2020 of 38%. The total Server SAN market is projected to grow to over $48 billion by 2026. The traditional enterprise storage market is projected to decline by -16% CAGR, leading to an overall growth in storage spend of 3% CAGR through 2026. Traditional enterprise storage is being squeezed in a vice between a superior, lower cost and more flexible storage model with Enterprise Server SAN, and the migration of IT towards cloud computing and Hyperscale Server SAN deployments. Wikibon strongly recommends that CTOs & CIOs initiate Server SAN pilot projects in 2015, particularly for applications where either low cost or high performance is required.

If containers are at the center of a shift in how applications are developers and delivered, and their pace of growth and change is unprecedented in IT history, this could have a massive ripple effect on both suppliers and consumers of the ecosystem of IT technologies.