IBM Extends Its Goals For AI And Quantum Computing – Seeking Alpha

David Ramos

While no one ever doubts the heritage of technological advancements that NYSE:IBM has made over the last several decades, there are certainly some whove wondered recently whether the company is able to sustain those types of efforts into the future. At a recent analyst day at their historic Thomas J. Watson Research Center, IBM made a convincing argument that they are up to the task, especially in the fields of AI - generative AI, in particular - as well as quantum computing.

What was particularly notable was the fact that the company showed a much tighter connection between the work its research group is doing on advanced technologies and the rapid productization of this work into its commercial product organizations. In both prepared remarks and in response to questions, it was clear that theres a renewed focus to ensure that the two groups are in lockstep with regard to their future outlook and development priorities.

As with many other organizations, that hasnt always been the case with IBM. The result has been that some potentially interesting research efforts havent always made it to the market. Thanks to a very clear directive from CEO Arvind Krishna (who used to run IBM Research) about the companys need to focus on a few specific areas - hybrid cloud, AI and quantum - current research director Dario Gil said that the coordination between research and commercial products groups has never been stronger. The net result should be - and is starting to show - important new capabilities that are making it into commercial products at a much faster pace.

One real-world impact of this new strategic initiative is the companys very rapid development of its suite of AI tools they call watsonx. First unveiled at the companys Think conference earlier this year (see "IBM Unleashes Generative AI Strategy With watsonx" for more), watsonx continues to evolve, driven in large part by new capabilities first developed by the IBM research group. What was particularly impressive at the recent analyst event was the number of real-world applications and customer examples using watsonx that IBM was able to talk about. While admitting that many organizations are still in the exploratory and proof-of-concept phase when it comes to GenAI, there were still a solid set of company logos from real-world implementations that they shared. In addition, IBM had an impressively thorough taxonomy of applications for which companies are starting to use watsonx and genAI.

On the application front, IBM noted that the top applications that its starting to see companies leverage GenAI for fall into three main categories: Digital Labor or HR-related activities, Customer Care or customer support, and App Modernization or code creation. Within those categories, the company discussed content creation, summarization, classification, and coding applications. Given the long history of older mainframe-related software that run on IBM mainframes, IBM noted particular interest in companies who want to move from old COBOL code to modern programming languages with the help of GenAI-powered tools.

In addition to applications, IBM talked about a number of technologies its working on within its research group to improve its watsonx offerings. Specifically, IBM discussed its efforts in Performance and Scale, Model Customization, Governance and Application Enablement. For Performance, IBM said that its working on a variety of new ways to improve the efficiency of how large foundation models. Its doing that through various combinations of technologies that do things like shrink the model size via quantization, improve the ability to share limited compute resources with GPU fractioning, and more.

Given its open-source focus, IBM also provided more details on all the work its doing with AI application framework tool Pytorch, which Meta (META) made open-source back in 2017. By leveraging the open-source community as well as its own efforts, the company talked about how its making significant improvements in both optimizing model performance and opening up the possibility of running Pytorch-built models across a wide range of different computing chip architectures from multiple vendors. Adding a hardware abstraction layer like Pytorch opens up the potential for a much wider range of programmers to build or customize GenAI models. The reason is that models can be created with these tools using languages such as JavaScript that are much more widely known than the chip-specific tools and their lower-level language requirements. At the same time, these hardware abstraction layers often end up adding fairly significant performance penalties because of their high-level nature (an issue that Nvidias (NVDA) Cuda software tools dont suffer from). With the new Pytorch 2.0, however, IBM said they and others are making concerted efforts to reduce that impact by better organizing where various types of optimization layers need to be and, as a result, are getting closer to on the metal performance.

On the Model Customization front, its clear IBM is spending a great deal of effort because theyve recognized that very few companies are actually building their own models - most are simply customizing or fine-tuning existing ones. (To read more about that development and some of its potential industry implications, check out my recent column "The Rapidly Evolving State Of Generative AI".) To that end, they discussed foundation model tuning techniques such as LoRA (Low Rank Adaptation), parameter-efficient tuning, multi-task prompt tuning, and more, all of which are expected to be commercialized within watsonx in the not-too-distant future. They also described the need to provide educational guidance in the model-building process to help developers pick the right size model and data sets for a given task. While this may sound simplistic, its an absolutely essential requirement, as even basic knowledge about how GenAI models are built and function is much more limited than people realize (or are willing to admit!).

IBMs efforts on Governance - that is, the tracking and reporting of details around how a model is built and evolved, they data used to create it, etc. - look to be an extremely important and key differentiating capability for the company. This is particularly true in regulated industries and environments where the company has a large customer base. While more details on IBMs specific governance capabilities are expected shortly, they did share some of the work theyre doing on providing guardrails to prevent the inclusion of biases, social stigmas, obscene content, and personally identifiable information (PII) into datasets intended for model ingestion. In addition, they talked about some of the work on risk assessment and prevention that theyve done. IBM recently announced that they will be offering indemnification for customers who use their foundation models so that they can avoid any IP protection-related lawsuits. Together with this governance work, these two efforts clearly demonstrate that IBM is in a market-leading position when it comes to critical concerns that some companies have about the trust and reliability of GenAI technology in general.

In the area of Application Enablement, IBM talked a great deal about the work its doing around Retrieval Augmented Generation (RAG). RAG is a relatively new technique that supercharges the inferencing process, makes it significantly easier and more cost-efficient for companies to leverage their own data, and eases the process of fine-tune existing foundation models so that organizations dont have to worry about creating models of their own. IBM says it has already seen a number of its customers start to experiment with and/or adopt RAG techniques so its working on refining its capabilities there to make the creation of more useful GenAI applications much easier for its customers.

In the world of quantum computing, IBM is already seen as a leader, in large part because of the amount of time theyve already spent working on discussing the innovations theyve made there. What was particularly impressive at the analyst event, however, was that the company showed off a detailed technology roadmap that extends all the way out to 2030. While some tech companies are willing to share their plans a few years out, its virtually unheard of for a company to provide this much information so far in advance. In part, IBM recognizes that they need to do it because quantum computing is such a dramatic and forward-looking technology that many potential customers feel the need to know how they can plan for it. To put it simply, they want to understand whats coming in order to bet on the roadmap.

Full details of the specific IBM quantum computing developments will be unveiled at an event that the company will be hosting in early December. Suffice it to say, however, that the company continues to be at the cutting edge of this technology and is growing increasingly confident about its ability to eventually make it into mainstream enterprise computing.

Given the long, sad history of early technology companies that no longer exist, its certainly understandable why some harbor doubts about the 112-year-old IBMs ability to continue innovating. As it recently demonstrated, however, not only is that spirit of invention still alive, it looks to be gaining some serious steam.

Disclaimer: Some of the author's clients are vendors in the tech industry.

Disclosure: None.

Source: Author

Editor's Note: The summary bullets for this article were chosen by Seeking Alpha editors.

Read more:
IBM Extends Its Goals For AI And Quantum Computing - Seeking Alpha

Related Posts

Comments are closed.