often begins the identical method. In a management assembly, somebody says: “Let’s use AI!” Heads nod, enthusiasm builds, and earlier than you realize it, the room lands on the default conclusion: “Certain — we’ll construct a chatbot.” That intuition is comprehensible. Giant language fashions are highly effective, ubiquitous, and interesting. They promise intuitive entry to common data and performance.
The staff walks away and begins constructing. Quickly, demo time comes round. A refined chat interface seems, accompanied by assured arguments about why this time, it will likely be completely different. At that time, nevertheless, it often hasn’t reached actual customers in actual conditions, and analysis is biased and optimistic. Somebody within the viewers inevitably comes up with a customized query, irritating the bot. The builders promise to repair “it”, however normally, the underlying concern is systemic.
As soon as the chatbot hits the bottom, preliminary optimism is commonly matched by consumer frustration. Right here, issues get a bit private as a result of over the previous weeks, I used to be compelled to spend a while speaking to completely different chatbots. I are inclined to delay interactions with service suppliers till the scenario turns into unsustainable, and a few these circumstances had piled up. Smiling chatbot widgets turned my final hope earlier than an everlasting hotline name, however:
- After logging in to my automotive insurer’s web site, I requested to elucidate an unannounced value enhance, solely to understand the chatbot had no entry to my pricing information. All it may provide was the hotline quantity. Ouch.
- After a flight was canceled on the final minute, I requested the airline’s chatbot for the explanation. It politely apologized that, because the departure time was already previously, it couldn’t assist me. It was open to debate all different matters, although.
- On a telco web site, I requested why my cellular plan had all of the sudden expired. The chatbot confidently replied that it couldn’t touch upon contractual issues and referred me to the FAQs. As anticipated, these have been lengthy however irrelevant.
These interactions didn’t convey me nearer to an answer and left me on the reverse finish of pleasure. The chatbots felt like overseas our bodies. Sitting there, they consumed actual property, latency, and a spotlight, however didn’t add worth.
Let’s skip the controversy on whether or not these are intentional darkish patterns. The very fact is, legacy methods because the above carry a heavy burden of entropy. They arrive with tons of distinctive information, data, and context. The second you attempt to combine them with a general-purpose LLM, you make two worlds conflict. The mannequin must ingest the context of your product so it could purpose meaningfully about your area. Correct context engineering requires talent and time for relentless analysis and iteration. And earlier than you even get to that time, your information must be prepared, however in most organizations, information is noisy, fragmented, or simply lacking.
On this publish, I’ll recap insights from my guide The Artwork of AI Product Improvement and my latest discuss on the Google Net AI Summit and share a extra natural, incremental strategy to integrating AI into current merchandise.
Utilizing smaller fashions for low-risk, incremental AI integration
“When implementing AI, I see extra organizations fail by beginning too large than beginning too small.” ( Andrew Ng).
AI integration wants time:
- Your technical staff wants to organize the information and be taught the accessible strategies and instruments.
- You could prototype and iterate to search out the candy spots of AI worth in your product and market.
- Customers must calibrate their belief when shifting to new probabilistic experiences.
To adapt to those studying curves, you shouldn’t rush to show AI — particularly open-ended chat performance — to your customers. AI introduces uncertainty and errors into the expertise, which most individuals don’t like.
One efficient method to tempo your AI journey within the brownfield context is through the use of small language fashions (SLMs), which usually vary from just a few hundred million to a couple billion parameters. They’ll combine flexibly along with your product’s current information and infrastructure, somewhat than including extra technological overhead.
How SLMs are educated
Most SLMs are derived from bigger fashions by data distillation. On this setup, a big mannequin acts because the instructor and a smaller one as the scholar. For instance, Google’s Gemini served because the instructor for Gemma 2 and Gemma 3 , whereas Meta’s Llama Behemoth educated its herd of smaller Llama 4 fashions. Simply as a human instructor condenses years of research into clear explanations and structured classes, the big mannequin distills its huge parameter area right into a smaller, denser illustration that the scholar can take in. The result’s a compact mannequin that retains a lot of the instructor’s competence however operates with far fewer parameters and dramatically decrease computational price.

Utilizing SLMs
One of many key benefits of SLMs is their deployment flexibility. Not like LLMs which might be largely used by exterior APIs, smaller fashions might be run regionally, both in your group’s infrastructure or instantly on the consumer’s gadget:
- Native deployment: You may host SLMs by yourself servers or inside your cloud atmosphere, preserving full management over information, latency, and compliance. This setup is good for enterprise purposes the place delicate data or regulatory constraints make third-party APIs impractical.
📈 Native deployment additionally affords you versatile fine-tuning alternatives as you accumulate extra information and wish to reply to rising consumer expectations.
- On-device deployment by way of the browser: Trendy browsers have built-in AI capabilities that you may depend on. As an example, Chrome integrates Gemini Nano by way of the built-in AI APIs, whereas Microsoft Edge consists of Phi-4 (see Immediate API documentation). Working fashions instantly within the browser allows low-latency, privacy-preserving use circumstances reminiscent of sensible textual content recommendations, kind autofill, or contextual assist.
If you need to be taught extra in regards to the technicalities of SLMs, listed here are a few helpful sources:
Let’s now transfer on and see what you’ll be able to construct with SLMs to supply consumer worth and make regular progress in your AI integration.
Product alternatives for SLMs
SLMs shine in targeted, well-defined duties the place the context and information are already recognized — the sorts of use circumstances that reside deep inside current merchandise. You may consider them as specialised, embedded intelligence somewhat than general-purpose assistants. Let’s stroll by the principle buckets of alternative they unlock within the brownfield, as illustrated within the following alternative tree.

1. Higher product analytics
Earlier than exposing AI options to customers, search for methods to enhance your product from the within. Most merchandise already generate a steady stream of unstructured textual content — help chats, assist requests, in-app suggestions. SLMs can analyze this information in actual time and floor insights that inform each product choices and speedy consumer expertise. Listed here are some examples:
- Tag and route help chats as they occur, directing technical points to the proper groups.
- Flag churn alerts throughout a session, prompting well timed interventions.
- Counsel related content material or actions based mostly on the consumer’s present context.
- Detect repeated friction factors whereas the consumer remains to be within the circulate, not weeks later in a retrospective.
These inside enablers maintain danger low whereas including worth and giving your staff time to be taught. They strengthen your information basis and put together you for extra seen, user-facing AI options down the street.
2. Take away friction
Subsequent, take a step again and audit UX debt that’s already there. Within the brownfield, most merchandise aren’t precisely a designer’s dream. They have been designed beneath the technical and architectural constraints of their time. With AI, we now have a possibility to raise a few of these constraints, lowering friction and creating sooner, extra intuitive experiences.
An excellent instance is the sensible filters on search-based web sites like Reserving.com. Historically, these pages use lengthy lists of checkboxes and classes that attempt to cowl each doable consumer choice. They’re cumbersome to design and use, and in the long run, many customers can’t discover the setting that issues to them.
Language-based filtering modifications this. As a substitute of navigating a fancy taxonomy, customers merely sort what they need (for instance “pet-friendly resorts close to the seaside”), and the mannequin interprets it right into a structured question behind the scenes.

Extra broadly, search for areas in your product the place customers want to use your inside logic — your classes, buildings, or terminology — and substitute that with pure language interplay. Every time customers can categorical intent instantly, you take away a layer of cognitive friction and make the product smarter and friendlier.
3. Increase
Along with your consumer expertise decluttered, it’s time to consider augmentation — including small, helpful AI capabilities to your product. As a substitute of reinventing the core expertise, take a look at what customers are already doing round your product — the facet duties, workarounds, or exterior instruments they depend on to achieve their aim. Can targeted AI fashions assist them do it sooner or smarter?
For instance, a journey app may combine a contextual journey notice generator that summarizes itinerary particulars or drafts messages for co-travelers. A productiveness instrument may embrace a gathering recap generator that summarizes discussions or motion objects from textual content notes, with out sending information to the cloud.
These options develop organically from actual consumer habits and lengthen your product’s context as a substitute of redefining it.
4. Personalize
Profitable personalization is the holy grail of AI. It flips the normal dynamic: as a substitute of asking customers to be taught and adapt to your product, your product now adapts to them like a well-fitting glove.
Whenever you begin, maintain ambition at bay — you don’t want a completely adaptive assistant. Slightly, introduce small, low-risk changes in what customers see, how data is phrased, or which choices seem first. On the content material stage, AI can adapt tone and elegance, like utilizing concise wording for specialists and extra explanatory phrasing for newcomers. On the expertise stage, it could create adaptive interfaces. As an example, a project-management instrument may floor essentially the most related actions (“create process,” “share replace,” “generate abstract”) based mostly on the consumer’s previous workflows.
⚠️ When personalization goes fallacious, it shortly erodes belief. Customers sense that they’ve traded private information for an expertise that doesn’t really feel higher. Thus, introduce personalization solely as soon as your information is able to help it.
Why “small” wins over time
Every profitable AI function — be it an analytics enchancment, a frictionless UX touchpoint, or a customized step in a bigger circulate — strengthens your information basis and builds your staff’s iteration muscle and AI literacy. It additionally lays the groundwork for bigger, extra complicated purposes later. When your “small” options work reliably, they change into reusable elements in greater workflows or modular agent methods (cf. Nvidia’s paper Small Language Fashions are the Way forward for Agentic AI).
To summarize:
✅ Begin small — favor gradual enchancment over disruption.
✅ Experiment quick — smaller fashions imply decrease price and sooner suggestions loops.
✅ Be cautious — begin internally; introduce user-facing AI when you’ve validated it.
✅ Construct your iteration muscle — regular, compound progress beats headline tasks.
Initially printed at https://jannalipenkova.substack.com.
















