(Large) language models
RAG for Optimizing Generative AI - Response generation from LLMs enhanced by information retrieval
Article REF: H6042 V1
(Large) language models
RAG for Optimizing Generative AI - Response generation from LLMs enhanced by information retrieval

Author : Patrice BELLOT

Publication date: October 10, 2025 | Lire en français

Logo Techniques de l'Ingenieur You do not have access to this resource.
Request your free trial access! Free trial

Already subscribed?

2. (Large) language models

The first probabilistic language models were created many years ago, and were inspired by work to model human and computer languages using Markov chains. Models have been used in automatic language processing and speech recognition since the 1980s-1990s, making it possible to estimate the probability of a word's appearance based on previous words. Using a short history (a single previous word for unigram models, two and three for bigram and trigram models), these models were unable to take into account dependencies beyond a few words, and had difficulty handling rare and new words.

Large Language Models (LLMs) emerged in the late 2010s. They have benefited both from theoretical advances in machine learning, with deep neural networks

You do not have access to this resource.
Logo Techniques de l'Ingenieur

Exclusive to subscribers. 97% yet to be discovered!

You do not have access to this resource. Click here to request your free trial access!

Already subscribed?


Article included in this offer

"Software technologies and System architectures"

( 227 articles )

Complete knowledge base

Updated and enriched with articles validated by our scientific committees

Services

A set of exclusive tools to complement the resources

View offer details