Large Language Model Landscape
In the recent past I have been observing and describing current LLM-related technologies and trends. In this article I’m taking a step back to present an overview of the current Large Language Model (LLM) landscape.
The image above shows the ripples caused by the advent of LLMs which can be divided into six bands or zones. As these ripples extend, there are requirements and opportunities for products and services.
Some of these opportunities have been discovered, some are yet to be discovered. I would argue that the danger of being superseded as a product is greater in Zone 5 as apposed to Zone 6.
Zone 5 offers a bigger opportunity for differentiation, substantial built-in intellectual property and stellar UX enabling enterprises to leverage the power of LLMs.
Zone 1 — Available Large Language Models
Considering LLMs, in essence LLMs are language bound, however, multi-modal models or multi-modality have been introduced in terms of images, audio and more. This shift gave rise to a more generic term being used, namely Foundation Models.
Apart from increased modalities, there has been model diversification from the large commercial providers, offering multiple models which are more task specific. There has also been a slew of open-sourced models made available.
New prompting techniques have illustrated how models performance can be enhanced and how the market are moving towards a scenario where data discovery, data design, data development and data delivery can be leveraged to achieve this level of model-autonomy.
Zone 2 — General Use-Cases
With the advent of large language models, functionality was more segmented…models were trained for specific tasks. Sphere focussed on Knowledge Answering; something Meta called KI-NLP. Models like DialoGPT, GODEL and others focussed on dialog management, etc.
Recent developments in LLMs follows an approach where models incorporate these traits and astounding performance can be extracted using different prompting techniques.
The main implementations of LLMs are listed here, with text generation encompassing tasks like summarisation, rewriting, key-word extraction and more.
Text analysis is becoming increasingly important, and embeddings are vital for these type of implementations.
Speech recognition, also known as ASR is the process of converting audio speech into text. The accuracy of any ASR process can easily be measured via a method called Word Error Rate (WER). ASR opens up vast amounts of recorded language data for LLM training and use.
Two notable shifts in this zone are:
- Knowledge answering and Knowledge Intensive NLP (KI-NLP) approaches are superseded by RAG Prompt Engineering at inference.
- Dialog generation was spearheaded by developments like GODEL and DialoGPT. These have been superseded by specific implementations like ChatGPT, HuggingChat and Cohere Coral. Also by prompt engineering approaches where few-shot training is used with the dialog context presented in the prompt.
Zone 3 — Specific Implementations
A few specific-use models are listed in this zone. Implementations have been split between general, powerful LLMs, and LLM-based digital/personal assistants like ChatGPT, HuggingChat and Cohere Coral.
Zone 4 — Models
The most notable Large Language Model suppliers are listed here. Most of the LLMs have inbuilt knowledge and functionality including human language translation, capability of interpreting and writing code, dialog and contextual management via prompt engineering.
Zone 5 — Foundation Tooling
This sector considers tooling to harness the power of LLMs, including vector stores, playgrounds and prompt engineering tools. Hosting like HuggingFace enables no-code interaction via model cards and simple inference APIs.
Lastly, listed in this zone is the idea of data-centric tooling which focusses on repeatable, high value use of LLMs.
The market opportunity in this area is creating foundation tooling which will address a future need for data discovery, data design, data development and data delivery.
Zone 6 — End User UIs
Further out, there is a whole host of applications which focus on flow building, idea generation, content and writing assistants. These products focus on UX and adding varying degrees of value between LLMs and the user experience.
⭐️ Follow me on LinkedIn for updates on Large Language Models ⭐️
I’m currently the Chief Evangelist @ Kore AI. I explore & write about all things at the intersection of AI & language; ranging from LLMs, Chatbots, Voicebots, Development Frameworks, Data-Centric latent spaces & more.
Get an email whenever Cobus Greyling publishes.
Get an email whenever Cobus Greyling publishes. By signing up, you will create a Medium account if you don’t already…
Walking Down the Memory Maze: Beyond Context Limit through Interactive Reading
Large language models (LLMs) have advanced in large strides due to the effectiveness of the self-attention mechanism…
Does Submitting Long Context Solve All LLM Contextual Reference Challenges?
Large Language Models (LLMs) are known to hallucinate. Hallucination is when a LLM generates a highly succinct and…
RAG & LLM Context Size
In this article I consider the growing context of various Large Language Models (LLMs) to what extent it can be used…