Back to Blog

How Major Retailers Personalize the Buy with AI

,
,
,
January 12, 2024
No items found.

Forward-thinking retailers are racing to deploy LLM chatbots to hyper-personalize the buying journey and cement customer loyalty.

Come mid-January there are two things the retail industry can be sure of:
1) About one in four consumers will have already quit their New Year’s Resolutions; and
2) The NRF Big Show will showcase how the latest technological innovations, such as LLM chatbots, will impact the retail industry.

Retail businesses are increasingly reliant on data, and lots of it. The industry thrives when it can rapidly identify trends, understand buyer personas and optimize inventories. Harnessing the power of Generative AI to process this data, analyze it and make decisions and recommendations based on it, will fundamentally change the way we do business.

The retail industry has fully embraced digital commerce and all its potential, finally catching up to the generations of consumers who barely know any other way. It is a fast-paced world in which displays change based on who’s seeing them; people discover, shop, and buy while at a stoplight; and customers are yours only for as long as it can take them to click the next link. When everything is moving so quickly, acting swiftly to become an early adopter of AI can be the difference between sold out collections and sending truckloads to TJ Maxx and Marshalls.  The adoption of AI and retail chatbots opens up endless possibilities to develop personalized offers and deliver them with personalized messages. This will empower retailers to enhance user experiences, streamline operations and ultimately drive business growth. Fortune favors the bold.

You might wonder about the role of GenAI in retail. Thanks to its natural language processing capabilities, you can use LLMs (Large Language Models) to interact with your information systems through verbal communication. Think of the most knowledgeable, alert and articulate Salesperson that you know, add real-time access to inventory levels and customer purchase histories and no need for a coffee break. This is exactly what LLM chatbots can be for you.

Delivering tailored retail experiences with GenAI

While chatbots are not new, the responses that pre-AI versions provided were often rigid, general and clearly scripted. More importantly, the logic was easily sidetracked by the use of unfamiliar slang or vernacular. With natural language capabilities, LLM chatbots have a better grasp of the context of the question and the flexibility to provide an informative and useful response in a much shorter time. 

One of the most talked about applications for this is personalized customer support. Customers benefit from a humanized interaction, able to communicate freely and get to a resolution with much less friction and frustration. Another holy grail for retail chatbots is the Private Shopping Assistant. For those consumers who are much more comfortable with digital interactions, a humanized chat with an entity that is both discrete and also familiar with their preferences and past purchases is the perfect confidant for questions ranging from sizes and prices to availability and accessories.

LLM chatbots help personalize the buy

Alternatively, LLMs can also assist retailers with more back-office tasks, such as preparing effective product descriptions. With practically infinite authoring resources, retailers can optimize the description for any scale of inventory and then duplicate them with targeted tweaks to catch the eye of each buyer persona. 

Since retailers (like most organizations) don’t always have the technical resources to set up and train a Generative AI model from scratch, AI21 has developed Task Specific Models (TSMs). TSMs are designed and trained to excel at one specific job, one natural language capability. For example, the Semantic Search TSM allows users to freely express what they are looking for and receive results based on what they meant, not what they typed. The Contextual Answers TSM takes information that organizations feed into the system and uses it to provide focused, informative responses to questions that end-users submit. Its output is grounded on the retailer’s well-defined subject matter and incorporated into the Customer Experiences at a fraction of the time it would take to create an AI model from scratch and with much lower error rates. 

There's a simple way to implement GenAI

The vision of implementing a Generative AI solution in a working retail business can seem both fantastic and daunting, the ultimate “we’ll do it next year” project. With TSMs, you no longer have to wait, as they are faster to deploy, easier to customize and cheaper to implement and maintain. They do not require the vast, sometimes complex, amounts of data that general purpose AI models do– rather, they rely only on data sources that are trusted by the retailer. They do not need to support a broad range of query types, only the types that serve the use case, making it much easier to anticipate and secure the inputs. Their focused nature means that they are leaner to operate, with a much smaller footprint that requires fewer resources and leads to lower latency. They are purpose-built, equipped with various built-in customization and verification mechanisms that help ensure that outputs that are more accurate, reliable and grounded. The result is a simpler implementation, one that can be done by a developer and a single line of code. 

The retail industry has realized the importance of achieving the levels of customization that consumers have come to expect. Task-Specific Models help deliver this ambitious goal by assisting retailers to harness the power of generative AI personalization with minimal effort. The experienced team at AI21 is available to provide tailored guidance and a dedicated support team to help you reach your goals. Whether you’re looking to automate and optimize unique product descriptions, enable self-service support and rapidly resolve customer issues or generate targeted marketing messages that speak to each buyer, Task-Specific Models offer the most practical path forward. 

The era of conversational commerce is now

It is no longer outlandish to state that we are entering the age of AI, an era in which we use massive amounts of data to identify patterns, predict outcomes and make corresponding decisions in microseconds. Retail is a high-volume, high-touch industry, where most companies have made the digital transformation needed to generate that data. Feeding that data into generative AI models with natural language processing will result in an ultra-personalized, highly scalable user experience. Consumers will be able to naturally converse with your applications to ask questions, seek recommendations and make purchases. Staff will be able to study behaviors, optimize logistics and create finely tuned messaging. The only drawback has been the cost and the duration of setting it all up.

Task-Specific Models overcome this hurdle by offering a new paradigm for enterprises seeking to implement AI. With TSMs you can build applications such as LLM chatbots in a reliable, simple and cost-effective manner to solve real business challenges.  Retailers can enjoy a faster time-to-value and, thanks to lowered costs, a clear ROI. 

Don’t get left behind – AI is unlocking new levels of customization. Whether you're looking to boost conversion rates, improve customer satisfaction, or create infinite scalability, our AI solutions make personalization effortless. Book a meeting at NRF to learn more.

Discover more

What is a MRKL system?

In August 2021 we released Jurassic-1, a 178B-parameter autoregressive language model. We’re thankful for the reception it got – over 10,000 developers signed up, and hundreds of commercial applications are in various stages of development. Mega models such as Jurassic-1, GPT-3 and others are indeed amazing, and open up exciting opportunities. But these models are also inherently limited. They can’t access your company database, don’t have access to current information (for example, latest COVID numbers or dollar-euro exchange rate), can’t reason (for example, their arithmetic capabilities don’t come close to that of an HP calculator from the 1970s), and are prohibitively expensive to update.
A MRKL system such as Jurassic-X enjoys all the advantages of mega language models, with none of these disadvantages. Here’s how it works.

Compositive multi-expert problem: the list of “Green energy companies” is routed to Wiki API, “last month” dates are extracted from the calendar and “share prices” from the database. The “largest increase“ is computed by the calculator and finally, the answer is formatted by the language model.

There are of course many details and challenges in making all this work - training the discrete experts, smoothing the interface between them and the neural network, routing among the different modules, and more. To get a deeper sense for MRKL systems, how they fit in the technology landscape, and some of the technical challenges in implementing them, see our MRKL paper. For a deeper technical look at how to handle one of the implementation challenges, namely avoiding model explosion, see our paper on leveraging frozen mega LMs.

A further look at the advantages of Jurassic-X

Even without diving into technical details, it’s easy to get a sense for the advantages of Jurassic-X. Here are some of the capabilities it offers, and how these can be used for practical applications.

Reading and updating your database in free language

Language models are closed boxes which you can use, but not change. However, in many practical cases you would want to use the power of a language model to analyze information you possess - the supplies in your store, your company’s payroll, the grades in your school and more. Jurassic-X can connect to your databases so that you can ‘talk’ to your data to explore what you need-  “Find the cheapest Shampoo that has a rosy smell”, “Which computing stock increased the most in the last week?” and more. Furthermore, our system also enables joining several databases, and has the ability to update your database using free language (see figure below).

Jurassic-X enables you to plug in YOUR company's database (inventories, salary sheets, etc.) and extract information using free language

AI-assisted text generation on current affairs

Language models can generate text, yet can not be used to create text on current affairs, because their vast knowledge (historic dates, world leaders and more) represents the world as it was when they were trained. This is clearly (and somewhat embarrassingly) demonstrated when three of the world’s leading language models (including our own Jurassic-1) still claim Donald Trump is the US president more than a year after Joe Biden was sworn into office.
Jurassic-X solves this problem by simply plugging into resources such as Wikidata, providing it with continuous access to up-to-date knowledge. This opens up a new avenue for AI-assisted text generation on current affairs.

Who is the president of the United States?

T0
Donald Trump
GPT-3
Donald Trump
Jurassic-1
Donald Trump
Google
Joe Biden
Jurassic-X
Joe Biden is the
46th and current
president
Jurassic-X can assist in text generation on up-to-date events by combining a powerful language model with access to Wikidata

Performing math operations

A 6 year old child learns math from rules, not only by memorizing examples. In contrast, language models are designed to learn from examples, and consequently are able to solve very basic math like 1-, 2-, and possibly 3- digit addition, but struggle with anything more complex. With increased training time, better data and larger models, the performance will improve, but will not reach the robustness of an HP calculator from the 1970s. Jurassic-X takes a different approach and calls upon a calculator whenever a math problem is identified by the router. The problem can be phrased in natural language and is converted by the language model to the format required by the calculator (numbers and math operations). The computation is performed and the answer is converted back into free language.
Importantly (see example below) the process is made transparent to the user by revealing the computation performed, thus increasing the trust in the system. In contrast, language models provide answers which might seem reasonable, but are wrong, making them impractical to use.

The company had 655400 shares which they divided equally among 94 employees. How many did each employee get?

T0
94 employees.
GPT-3
Each employee got 7000 stocks
Jurassic-1
1.5
Google
(No answer provided)
Jurassic-X
6972.3
X= 655400/94
Jurassic-X can answer non-trivial math operations which are phrased in natural language, made possible by the combination of a language model and a calculator

Compositionality

Solving simple questions might require multiple steps, for example - “Do more people live in Tel Aviv or in Berlin?” requires answering: i. What is the population of Tel-Aviv? ii. What is the population of Berlin? iii. Which is larger? This is a highly non-trivial process for a language model, and language models fail to answer this question (see example). Moreover, the user can’t know the process leading to the answers, hence is unable to trust them. Jurassic-X can decompose such problems into the basic questions, route each to the relevant expert, and put together an answer in free language. Importantly, Jurassic-X not only provides the correct answer but also displays the steps taken to reach it, increasing the trust in the system.

Do more people live in Tel Aviv or in Berlin?

T0
Berlin
GPT-3
There are more people living in Tel Aviv than in Berlin.
Jurassic-1
Berlin and Tel Aviv are roughly the same size
Google
(First hit is a comparison between Tel Aviv and Berlin)
Jurassic-X
More people live in Berlin than in Tel-Aviv

[‘Return population of Tel Aviv’; Return population of Berlin’; Return which is bigger between #1 and #2’]
Step 1: Population of Tel Aviv. Result - 451523.
Step 1: Population of Berlin. Result - 3664088.
Step 3: Which is bigger,  #1 or #2. Result - Berlin.

Jurassic-X breaks down compositional questions, answers the basic sub-questions, and puts together the answer. Importantly, this process is transparent to the user greatly increasing the trust in the system

Dynamic information (like weather and currency exchange rates)

Certain types of information change continuously - weather, currency exchange rates, share values and more. Such information will never be captured by language models, yet can easily be handled by Jurassic-X by integrating it with a reliable source of information. We performed a proof-of-concept on two such features - weather and currency exchange rates, and the design enables quick integration with more sources to solve your use-case.
Weather - a loosely phrased question about the weather elicits an answer from all language models, where language models always return the same answer, regardless of when the question was asked (funny, right?), while Jurassic-X provides an answer based on the actual weather prediction.

I’m going to be in New-York in 3 days. Should I pack my umbrella?

T0
No
GPT-3
Yes, you should pack your umbrella.
Jurassic-1
Yes, you should. The weather forecast is rain.
Google
(Links to weather websites)
Jurassic-X
Yes, you should pack your umbrella, because in New York in 3 days there will be broken clouds and the temperature will be -2 degrees.

Currency exchange rates change much faster than weather predictions, yet the Jurassic-X concept - a language model connected to a reliable source of information - easily solves this problem as well.

How much Moroccan money will I get for 100 bucks?

T0
125 dirhams
GPT-3
100 moroccan dirhams is about 27$.
Jurassic-1
How much is 100 dollars in moroccan money?
Google
100 dirhams = 10.75 dollars
Jurassic-X
100 USD = 934.003 MAD
Jurassic-X combines a language model with access to APIs with continuously changing information. This is demonstrated for weather forecasts and currency exchange rates, and can easily be extended to other information sources

Transparency and trust

Transparency is a critical element that is lacking in language models, preventing a much wider adoption of these models. This lack of transparency is demonstrated by the answers to the question - “Was Clinton ever elected as president of the United States?”. The answer, of course, depends on which Clinton you have in mind, which is only made clear by Jurassic-X that has a component for disambiguation. More examples of Jurassic-X’s transparency were demonstrated above - displaying the math operation performed to the user, and the answer to the simple sub-questions in the multi-step setting.

Was Clinton ever elected president of the United States?

T0
Yes
GPT-3
No, Clinton was never elected as president of the United States.
Jurassic-1
No
Google
Clinton was elected president in the 1992 presidential elections…
Jurassic-X
Bill Clinton was elected president.
Jurassic-X is designed to be more transparent by displaying which expert answered which part of the question, and by presenting the intermediate steps taken and not just the black-box response

Your Turn

That's it, you get the picture. The use cases above give you a sense for some things you could do with Jurassic-X, but now it's your turn. A MRKL system such as Jurassic-X is as flexible as your imagination. What do you want to accomplish? Contact us for early access

Contact us below and we will get back to you shortly.

Thank you!

Your submission has been received!
Oops! Something went wrong while submitting the form.