- Unlocking AI for NHS: Now Assist in IT Service Management
- CompTIA launches expert-level cybersecurity certification
- How Visibility Affects the Warfighter
- What WordPress users need to know about the Automattic and WP Engine conflict
- US Government Issues Cloud Security Requirements for Federal Agencies
How IT pros can benefit from generative AI safely
The enterprise IT landscape is littered with supposedly paradigm-shifting technologies that failed to live up to the hype, and intil now, one could argue that AI fell into that category. But generative AI, which has taken the world by storm in the form of OpenAI’s ChatGPT chatbot, just might be the real deal.
Chris Bedi, chief digital information officer at ServiceNow, says the release of ChatGPT last November was “an iPhone moment,” an event that captured the public’s attention in a way that “changed everything forever.” He predicts that generative AI will become embedded into the fabric of every enterprise, and he recommends that CIOs and other IT leaders should begin now to develop their generative AI strategies.
Gartner is no less effusive, predicting that generative AI will become “a general-purpose technology with an impact similar to that of the steam engine, electricity and the internet.” Although generative AI is still in its infancy and there are many pitfalls that need to be navigated, Gartner says, “Generative AI provides new and disruptive opportunities to increase revenue, reduce costs, improve productivity and better manage risk. In the near future, it will become a competitive advantage and differentiator.”
AI has been around for a long time, but generative AI takes machine learning to the next level with a neural network architecture called transformer (the T in GPT), first described by Google researchers in 2017. So, this is new. Generative AI systems are built on pre-trained (the P in GPT) data sets (45 terabytes for ChatGPT) and are able to respond to queries in conversational language. Generative AI can produce text, images, and video, including software code and networking scripts.
We went right to source and asked ChatGPT itself how it can make life easier for enterprise IT. After a pause of no more than a couple of seconds, we got back a numbered list: 1) troubleshooting and issue resolution, 2) documentation and knowledge management, 3) automation and scripting, 4) training and onboarding, 5) security and compliance, 6) project management and planning, 7) stay updated on technology trends.
Without any prompting, the chatbot added, “It’s important to note that while ChatGPT can provide valuable guidance and support it should not be solely relied upon for critical decision-making. Human expertise and judgment should always be considered alongside AI-generated suggestions.”
After getting the chatbot’s perspective, we moved on to intelligent humans for their take on several key questions about generative AI: What exactly is it? What can it do for enterprise IT? What can’t it do? How do I get it? What are some of the potential pitfalls that I need to be aware of?
What is generative AI and how is it different from ‘traditional’ AI?
For the most part, traditional AI/ML technology sits in the background, looking to identify patterns in large data sets. It makes predictions and provides recommendations based on those predictions.
Generative AI is fundamentally different. It is a large language model (LLM) trained with vast amounts of data, including samples of human conversation. It is able to digest and summarize data and can interact with a human using natural language. ChatGPT is a super Siri that surprised even its creators when it racked up a million users in its first week after launch and 100 million after two months. It currently generates 1.8 billion visitors per month.
In general, when systems scale rapidly, they become more complex, harder to manage, less reliable and less efficient. With large language models, the more data, the more queries, the more interactions, the smarter the system becomes, and the more it begins to resemble human intelligence.
But, at least at this stage, these models are not the same as human intelligence. Forrester analyst Rowan Curran says, “What they are not doing is creating net new information that has a contextual understanding of itself. These models predict the next word in a sequence based on the previous words in that sequence. It’s important not to treat them as a source of authority, an oracle or anything that has a mind behind it.”
What can generative AI do for enterprise IT?
At the networking layer, large language models can perform functions like generating network configurations, writing scripts for IT automation tools and creating networking maps, says Shamus McGillicuddy, vice-president of research at Enterprise Management Associates.
“It’s very good for inspiration, imagination, anti-procrastination. One can use it to get started with a task or project. Ask it to give you something, like a piece of content or code. Then one can use his or her knowledge and skills to turn it into something good, whether it’s a policy paper or a network configuration file,” McGillicuddy says.
In software development, generative AI can spit out code snippets and has the ability to de-bug code. Large language models use the term “token” the way IT pros talk about bytes. With ChatGPT, one token represents four characters, or roughly three-fourths of a word. This is important because each ChatGPT query/response has a limit of around 4,000 tokens, and the query wording itself counts toward that limit. So, generative AI systems can write pieces of code in a variety of programming languages, but don’t ask them to come up with new versions of an operating system, because when it hits that limit, it stops and resets.
At the strategic level, as IT leaders become more familiar and comfortable with generative AI, they will be able to roll it out across the enterprise to make employees more productive, streamline business processes, improve customer service, and drive digital transformation.
Bedi says generative AI’s ability to take large pieces of disparate, complex information and summarize them for human consumption has applications for ITOps, analysis of security and event logs, customer support, call center, help desk, finance, HR, sales, and marketing. “Everybody is awash in tons of content; generative AI has the ability to distill it into something useful and consumable. It can speed up every operation in the company,” he adds.
Generative AI pitfalls.
If this all this sounds too good to be true, and that’s because it probably is—at least for now. A McKinsey report cautions, “The outputs generative AI models produce may often sound extremely convincing. But sometimes the information they generate is just plain wrong. Worse, sometimes it’s biased (because it’s built on the gender, racial, and myriad other biases of the internet and society more generally) and can be manipulated to enable unethical or criminal activity.”
Forrester’s Curran uses the term “coherent nonsense” to describe this phenomenon. But the term gaining the most traction in the generative AI ecosystem is “hallucination.”
Futurist Bernard Marr says, “Hallucination in AI refers to the generation of outputs that may sound plausible but are either factually incorrect or unrelated to the given context. These outputs often emerge from the AI model’s inherent biases, lack of real-world understanding, or training data limitations. In other words, the AI system ‘hallucinates’ information that it has not been explicitly trained on, leading to unreliable or misleading responses.”
This means that enterprise IT should not put generative AI software code or networking scripts into production without a person double-checking it first—an approach dubbed “human in the loop.” And organizations should have systems in place to catch instances in which the chatbot might interact with customers in a way that might be considered argumentative, offensive, or inappropriate.
The explosion of interest in ChatGPT has also triggered worries about data privacy and shadow generative AI, since it must be assumed that employees at all levels are asking ChatGPT questions.
“I am very concerned about what data people put into ChatGPT when they are putting queries into it,” says McGillicuddy. “I’m concerned about how that data is used and stored and what rights Open AI asserts to it.”
Keatron Evans, principal cybersecurity advisor at the InfoSec Institute, cautions, “Don’t use any protected data or personal information when utilizing or experimenting with AI. For instance, say you have a confidential sales report and want to generate a summary using AI. You upload the report, but now the data that you entered is stored on ChatGPT’s servers, and it will use that data to answer queries from other people, possibly exposing your company’s confidential information.”
He adds that hackers could exploit ChatGPT code vulnerabilities to steal user information or find a way to steal that data directly from the app itself. “Regardless, uploading sensitive data or information could violate privacy laws, which would result in your company possibly facing large fines,” Evans points out.
Another, more fuzzy issue concerns ownership of intellectual property. Let’s say an employee uploads proprietary software code to ChatGPT, asking it to de-bug the code or add a piece of functionality. That code goes into the ChatGPT database. What happens when someone else at a different company queries ChatGPT and the output includes chunks of that original code?
Samsung recently banned the use of ChatGTP by employees after an engineer “accidentally leaked internal source code by uploading it to ChatGPT,” according to an internal memo.
How should organizations acquire generative AI.
The vendor community is racing to provide generative AI for enterprises. The usual suspects are leading the way (Google, AWS, Microsoft, IBM) since they have the resources to develop these large language models. But just about every vendor is figuring out a way to embed generative AI into its platforms.
IDC analyst Nancy Gohring says, “Vendors in ITSM and ITOps are already applying generative AI to a variety of use cases, predominantly with the aim of improving tool usability, speeding response times, and expanding use cases. While ensuring human oversight is critical, particularly given the immaturity of the technology, enterprises should seriously consider embracing new offerings as a way to improve efficiencies.”
For IT leaders, a sensible approach would be to work with existing platform partners to determine how the vendor roadmap aligns with their enterprises’ style of technology acquisition. Does the enterprise have the willingness and the readiness (either from a skills, funding, or data-processing infrastructure perspective) to spin up its own generative AI capabilities? This approach might deliver competitive advantage, but it also takes time and effort.
Or would it make more sense to leverage existing technology providers who are embedding generative AI into their platforms? For example, Salesforce has released Einstein GPT which brings generative AI capabilities to the Salesforce CRM platform as well as to the Slack app.
Of course, similar to the way organizations have adopted hybrid-cloud architectures, it’s likely that enterprises will adopt a mixed model that encompasses both cloud and on-prem deployments. One option would be to build new generative AI apps in the AWS cloud, using the AWS infrastructure, large language models, and toolsets. Another would be to build custom generative AI functionality on top of vendor CRM or ERP platforms.
What some of the leading vendors are offering.
Here are some examples of how key vendors are ramping up their generative AI capabilities:
Microsoft
Microsoft, the major investor in OpenAI and its technology partner, is embedding ChatGPT technology throughout its portfolio. Microsoft has introduced Microsoft 365 Copilot, which integrates generative AI into Office productivity apps like Word, Excel, Outlook, and Teams. A feature called Business Chat combines a user’s calendar, emails, chats, documents, contacts, etc., into one knowledge base that can be queried in natural language. Microsoft has announced Dynamics 365 Copilot, which brings generative AI to CRM and ERP. And Microsoft has embedded ChatGPT functionality into its Bing search engine.
In a related development, OpenAI has partnered with GitHub to offer a commercial product called GitHub Copilot, a code-writing chatbot that can speak more than a dozen programming languages.
Google has announced Duet AI for Google Workspace, which embeds its Generative AI (Google’s large language model is called PaLM) into the Google productivity suite (Gmail, Google Docs, Sheets, Slides, and Meet). Google is putting Generative AI functionality into its Chrome browser. It has a platform called Vertex AI that enables enterprises and SaaS vendors to build their own applications; a service to help enterprises build AI-powered chat and search applications based on Google’s foundation models; and an answer to GitHub’s Copilot called Duet, designed to help developers write code.
Cisco
Cisco has built its own generative AI and recently announced plans to buy AI startup Armorblox. Cisco says it will embed Generative AI capabilities across its entire portfolio, starting with its Security Cloud service and Webex collaboration tool.