What is a large language model (LLM)?

A large language model is a type of AI trained on massive amounts of text to understand and generate human language. It's the reasoning engine behind modern AI voice agents, chatbots, and writing tools.

Written By Catherine Weir

Last updated About 2 hours ago

A large language model, usually shortened to LLM, is a type of AI that has been trained on massive amounts of human-written text to understand and generate language. It's the reasoning engine inside ChatGPT, Claude, Gemini, and every modern AI voice agent โ€” including ours.

When you ask an AI phone agent a question, the LLM is the part that decides what to say back. It takes the transcribed words the caller said, looks at the instructions and knowledge base the business provided, and produces a response in plain language.

What "large" means

"Large" refers to the number of parameters โ€” the internal values the model has learned during training. Modern production LLMs have tens of billions to trillions of parameters. More parameters generally means more capable reasoning, though also more expensive to run.

The largest and most capable LLMs include Claude (Anthropic), GPT-4 and GPT-5 family (OpenAI), and Gemini (Google). There are also open-source models like Llama (Meta) and Mistral that perform competitively for many tasks.

What LLMs are actually good at

  • Understanding what someone is asking, even when the phrasing is unusual or the words are ambiguous

  • Generating grammatical, context-aware responses in natural language

  • Following complex instructions ("you are a receptionist for a dental practice; always greet callers by name; never book appointments for Fridays")

  • Reasoning across long contexts โ€” remembering what was said earlier in the same conversation

  • Using tools โ€” calling APIs, looking up records, scheduling appointments โ€” as part of generating a response

What LLMs are not good at

  • Being consistently factual about recent events or specific customer data unless explicitly given the information to work from

  • Math without a calculator tool โ€” LLMs are not reliable at arithmetic

  • Avoiding plausible-but-wrong answers (a phenomenon called "hallucination") when the model is unsure

  • Remembering specific details about a customer between calls unless the information is loaded into the conversation context

Good AI voice agent platforms work around these limitations by grounding the LLM in your actual business data, giving it access to real tools (calendars, CRM records, knowledge bases), and escalating to a human when the LLM's confidence drops.

How LLMs work in voice AI

In a voice AI pipeline, the LLM sits in the middle. Speech-to-text feeds it the caller's words. The LLM consults its instructions and the business's knowledge base, decides what to do, and produces a response. That response goes to text-to-speech, which speaks it back to the caller. The whole loop runs in fractions of a second.

Related concepts

See it in action

The Receptionist Agent at 365agents uses state-of-the-art LLMs as the reasoning engine for every call. We continuously evaluate new model releases and upgrade our underlying models as better options ship โ€” without you needing to do anything. See our ISO 42001 AI management practices for how we govern those upgrades.