Technology

Generative AI Assistants in Enterprise Context Part II

Reading time: 9 min | May 13, 2024

Welcome back to our series on generative AI assistants and their role in enterprise contexts. In our previous article, we discussed the potential of AI assistants such as ChatGPT, Claude, and Gemini, highlighting their capabilities, architecture, and processing workflows. We also clarified terminology and emphasized the need for advanced configuration and customization when implementing these assistants in enterprise companies. In this second part, we will delve into the various generative AI assistants available on the market, examining their benefits and drawbacks.

The Key to Effective Deployment is understanding the Solution Layers

A generative AI assistant solution consists of two primary components: the Large Language Model (LLM) and the AI assistant application. The LLM is responsible for analyzing and responding to user queries based on the datasets it was trained on. The AI assistant application serves as a wrapper around the LLM, acting as the interface through which users interact with the LLM.

Customization options for these solutions are available at both the LLM and AI assistant application levels. For the LLM, you can either use it as is, trained on a vast array of general datasets, or further specialize it for your specific domain use-cases. This can be achieved through fine-tuning the model on your own curated dataset or employing prompt-engineering techniques to optimize the LLM's output based on its existing general knowledge. A hybrid approach combining these techniques is also possible.

At the AI assistant application level, there are multiple ways to construct an application that orchestrates how information is fed into the LLM and interacts with other components, such as connectors to ERP systems. This layer is crucial, especially in large-scale enterprise environments where the volume and complexity of queries demand robust, scalable solutions. So while the LLM is the foundation of generative AI assistants, the application layer is the part enterprises can and should influence the most, as it directly impacts cost efficiency and performance. The design of this layer must prioritize cost efficiency and performance to manage potential escalations in operational expenses effectively.

Overview of AI Assistant Application Architecture

Understanding AI Adoption

Many executives recognize that AI should not be treated as a separate business component but rather as the core fuel powering all components. However, in practice, AI is often initially adopted as a new, separate component before being integrated into the broader business ecosystem. This approach may work for a time, but as technology advances and use-cases expand, it can create technical debt for companies.

Based on current evidence, we can make two key assumptions:

  1. Large Language Models will continue to improve, becoming faster and more powerful.
  2. The future will see a shift from individual AI applications to complex compound AI systems.

With these assumptions in mind, it's essential to design and develop AI assistant applications with future complex scenarios in mind, as they will eventually become part of larger, interconnected AI systems.

Market Overview - Out of the Box Solutions

Let's examine the current options provided by some of the pioneers and major players in the generative AI market for building your own AI assistant on top of their models.

OpenAI

  • Models: GPT-4 Turbo and GPT-3.5 Turbo are OpenAI's flagship models, with GPT-4 offering multimodal capabilities such as processing text, images, or files. API-based access is available for these models.
  • AI Assistant Option: OpenAI's assistant offering is ChatGPT, which leverages GPT-4 or GPT-3.5 capabilities. ChatGPT offers various integration possibilities with third-party tools, web browsing capabilities, and memory features. For enterprise usage, ChatGPT for Enterprise provides enterprise-grade security and privacy, along with custom GPTs tailored to specific tasks.

Anthropic

  • Models: The Claude 3 series (Opus, Sonnet, Haiku) is designed for high-complexity tasks, emphasizing safety, trustworthiness, and reliability. These models excel at handling large context windows and demonstrate advanced accuracy over long documents with low hallucination rates.
  • AI Assistant Option: Anthropic offers the Claude AI assistant, leveraging the capabilities of the Claude 3 models. While individual subscriptions are available, the team plan's suitability for enterprises is unclear. The team plan includes admin tools, streamlined billing management, source verification features, and integration with other apps.

Google

  • Models: Gemini, including variants like Gemini Ultra, Gemini Pro, and Gemini Nano, is designed to be multimodal, processing and understanding text, images, and other types of data.
  • AI Assistant Option: Google Gemini is the AI assistant offered by Google, leveraging the power of their models. The Gemini Enterprise subscription plan allows connection to various Google apps and includes enhanced enterprise security features. A notable feature is the integration with Google Search, allowing AI-generated claims to be verified.

Microsoft

  • Models: Microsoft utilizes LLMs such as the GPT-4 model from OpenAI, although other specific models used in its offerings are not clear.
  • AI Assistant Option: Microsoft Copilot is available in two versions: one for individuals leveraging the GPT-4 model, and another for the Microsoft 365 suite, designed for businesses and large enterprises. The Microsoft 365 suite version offers enterprise-grade security, integration possibilities with other ERP data sources, and the Copilot Studio for customizing the assistant to specific use-cases.

Mistral AI

  • Models: Mistral AI offers multiple models, such as Mistral 7B, Mistral 8x7B, Mistral Large, Mistral Next, and Mistral Small, which can be consumed in various ways within AI assistant applications.
  • AI Assistant Option: Mistral's own AI assistant application, "LE CHAT," leverages the Large, Next, and Small model versions. Currently in beta testing, it will be available in individual and enterprise versions. Notable enterprise features include self-hosting and fine-grained moderation mechanisms.

Trade-offs and Considerations for Enterprises

After analyzing the different AI assistant application offerings from some of the most important players in the AI market, we can identify two key considerations. At the Large Language Model (LLM) layer, companies must first decide between the ease of deployment and the increased control needed. If they need a simple model deployment option, they can choose to consume models like GPT through the provider's own platform. However, this approach results in minimal control over the LLM and raises data privacy concerns. For most EU-enterprises, this option is likely not viable due to the need for increased control over the LLM and data access.

The second option for companies that require increased control is the self-deployment option. This approach allows them to deploy the models in their own cloud or on-premise infrastructure. While this option offers more control, it comes with potential disadvantages such as increased maintenance costs and the need for specialized software engineering teams.

At the application layer, enterprises must also make further trade-offs if they want to use the AI assistants provided by these companies out-of-the-box. Each of these assistants is suitable for personal productivity and use cases like synthesis, research, ideation, and writing. However, the number of users who will use the assistants is an important factor to consider. Depending on the business size, some companies might find offerings like Anthropic more appealing, while enterprises might prefer ChatGPT due to its scalability for large teams with many users.

There is also the use case of Gemini and Microsoft Co-pilot, which embed their assistants into their product offerings. This unified ecosystem experience can be attractive to organizations seeking seamless integration. However, if we ignore this use case, we are left with AI assistants that can be used for various use cases.

Apart from security and data privacy concerns, the real return on investment (ROI) will only become apparent if these solutions are integrated into the broader enterprise landscape, including ERP systems, third-party apps, and other data sources. While off-the-shelf AI products are attractive for their quick deployment capabilities and potential cost savings, their customization limitations can pose significant drawbacks. These limitations may lead to a dependency on the providers' terms and continuity, and there are potential risks with data privacy due to the necessity of processing data externally.

Another major issue is that most AI assistants from OpenAI, Anthropic, or Mistral lack output auditing capabilities. This lack of transparency can make it difficult for users to understand how the assistant generated the output, the steps it took, or the sources it leveraged. For tasks like financial report analysis, the inability to verify the output and sources might be a deal-breaker for many companies.

Cost optimization is another issue enterprises might encounter when leveraging such solutions. Since there is no one-size-fits-all LLM, companies may achieve better results and cost efficiency by using different LLMs for different tasks depending on their complexities and suitability. This level of customization is generally not possible with out-of-the-box solutions.

Working at the application layer

The alternative for companies is to develop their AI assistant apps from scratch or work with an AI assistant toolkit from a provider, which can be customized according to their needs and use cases. Custom-built AI solutions offer the highest level of customization, allowing companies to tailor features to their operational needs and maintain control over their data. However, this approach can come with higher development and maintenance costs, the need for substantial technical expertise, and the risks of project failure or integration issues.

In conclusion, whether a business should opt for an off-the-shelf product or invest in a custom solution depends on various factors. These factors include budget constraints, specific business and domain needs, available technical expertise, and desired speed of deployment. While ready-made solutions can provide quick and easy gains in productivity, they might fall short in delivering long-term value without significant customization. Custom AI solutions, while costly and complex to develop, offer tailored functionality that can significantly enhance operational efficiency and competitive advantage. In the third part of this blog series, we will delve into the development of custom generative AI assistants and provide best practices and recommendations for their implementation and adoption in enterprises.

Here’s more cool stuff

Read what's next.