Classes from constructing manufacturing AI techniques that no person talks about.
The dialog round AI brokers has moved quick. A 12 months in the past, everybody was optimizing RAG pipelines. Now the discourse facilities on context engineering, MCP/A2A protocols, agentic coding instruments that learn/handle complete codebases, and multi-agent orchestration patterns. The frameworks maintain advancing.
After 18 months constructing the AI Assistant at Cisco Buyer Expertise (CX), we’ve discovered that the challenges figuring out real-world success are hardly ever those getting consideration. Our system makes use of multi-agent design patterns over structured enterprise information (principally SQL, like most enterprises). The patterns that comply with emerged from making that system really helpful to the enterprise.
This put up isn’t in regards to the apparent. It’s about a number of the unglamorous patterns that decide whether or not your system will get used or deserted.
1. The Acronym Downside
Enterprise environments are dense with inside terminology. A single dialog would possibly embody ATR, MRR, and NPS, every carrying particular inside that means that differs from frequent utilization.
To a basis mannequin, ATR would possibly imply Common True Vary or Annual Taxable Income. To our enterprise customers, it means Accessible to Renew. The identical acronym can even imply utterly various things inside the firm, relying on the context:
Consumer: “Arrange a gathering with our CSM to debate the renewal technique”
AI: CSM → Buyer Success Supervisor (context: renewal)
Consumer: “Examine the CSM logs for that firewall situation”
AI: CSM → Cisco Safety Supervisor (context: firewall)
NPS may very well be Web Promoter Rating or Community Safety Options, each utterly legitimate relying on context. With out disambiguation, the mannequin guesses. It guesses confidently. It guesses incorrect.
The naive resolution is to increase acronyms in your immediate. However this creates two issues: first, it is advisable to know which acronyms want enlargement (and LLMs hallucinate expansions confidently). Second, enterprise acronyms are sometimes ambiguous even inside the similar group.
We preserve a curated company-wide assortment of over 8,000 acronyms with domain-specific definitions. Early within the workflow, earlier than queries attain our area brokers, we extract potential acronyms, seize surrounding context for disambiguation, and lookup the proper enlargement.
50% of all queries requested by CX customers to the AI Assistant include a number of acronyms and obtain disambiguation earlier than reaching our area brokers.
The important thing element: we inject definitions as context whereas preserving the consumer’s unique terminology. By the point area brokers execute, acronyms are already resolved.
2. The Clarification Paradox
Early in growth, we constructed what appeared like a accountable system: when a consumer’s question lacked adequate context, we requested for clarification. “Which buyer are you asking about?” “What time interval?” “Are you able to be extra particular?”
Customers didn’t prefer it, and a clarification query would typically get downvoted.
The issue wasn’t the questions themselves. It was the repetition. A consumer would ask about “buyer sentiment,” obtain a clarification request, present a buyer identify, after which get requested about time interval. Three interactions to reply one query.
Analysis on multi-turn conversations exhibits a 39% efficiency degradation in comparison with single-turn interactions. When fashions take a incorrect flip early, they hardly ever get well. Each clarification query is one other flip the place issues can derail.
The repair was counterintuitive: classify clarification requests as a final resort, not a primary intuition.
We carried out a priority system the place “proceed with affordable defaults” outranks “ask for extra data.” If a consumer gives any helpful qualifier (a buyer identify, a time interval, a area), assume “all” for lacking dimensions. Lacking time interval? Default to the following two fiscal quarters. Lacking buyer filter? Assume all clients inside the consumer’s entry scope.
That is the place clever reflection additionally helps tremendously: when an agent’s preliminary try returns restricted outcomes however an in depth different exists (say, a product identify matching a barely completely different variation), the system can robotically retry with the corrected enter somewhat than bouncing a clarification query again to the consumer. The objective is resolving ambiguity behind the scenes every time potential, and being clear to customers about what filters the brokers used.
Early variations requested for clarification on 30%+ of queries. After tuning the choice stream with clever reflection, that dropped under 10%.

Determine: Determination stream for clarification, with clever reflection
The important thing perception: customers would somewhat obtain a broader outcome set they will filter mentally than endure a clarification dialogue. The price of displaying barely extra information is decrease than the price of friction.
3. Guided Discovery Over Open-Ended Dialog
We added a function referred to as “Compass” that implies a logical subsequent query after every response. “Would you want me to interrupt down buyer sentiment by product line?”
Why not simply ask the LLM to counsel follow-ups? As a result of a basis mannequin that doesn’t perceive your enterprise will counsel queries your system can’t really deal with. It’ll hallucinate capabilities. It’ll suggest evaluation that sounds affordable however leads nowhere.
Compass grounds recommendations in precise system capabilities. Quite than producing open-ended recommendations (“Is there anything you’d wish to know?”), it proposes particular queries the system can undoubtedly fulfill, aligned to enterprise workflows the consumer cares about.
This serves two functions. First, it helps customers who don’t know what to ask subsequent. Enterprise information techniques are advanced; enterprise customers typically don’t know what information is on the market. Guided recommendations educate them the system’s capabilities by instance. Second, it retains conversations productive and on-rails.
Roughly 40% of multi-turn conversations inside the AI Assistant embody an affirmative follow-up, demonstrating how contextually related comply with up recommendations can enhance consumer retention, dialog continuity and information discovery.
We discovered this sample useful sufficient that we open-sourced a standalone implementation: langgraph-compass. The core perception is that follow-up era must be decoupled out of your major agent so it may be configured, constrained, and grounded independently.
4. Deterministic Safety in Probabilistic Methods
Function-based entry management can’t be delegated to an LLM.
The instinct is perhaps to inject the consumer’s permissions into the immediate: “This consumer has entry to accounts A, B, and C. Solely return information from these accounts.” This doesn’t work. The mannequin would possibly comply with the instruction. It may not. It’d comply with it for the primary question and overlook by the third. It may be jailbroken. It may be confused by adversarial enter. Immediate-based id isn’t id enforcement.
The danger is delicate however extreme: a consumer crafts a question that tips the mannequin into revealing information outdoors their scope, or the mannequin merely drifts from the entry guidelines mid-conversation. Compliance and audit necessities make this untenable. You can not clarify to an auditor that entry management “normally works.”
Our RBAC implementation is solely deterministic and utterly opaque to the LLM. Earlier than any question executes, we parse it and inject entry management predicates in code. The mannequin by no means sees these predicates being added; it by no means makes entry choices. It formulates queries; deterministic code enforces boundaries.
When entry filtering produces empty outcomes, we detect it and inform the consumer: “No information are seen together with your present entry permissions.” They know they’re seeing a filtered view, not a whole absence.
Liz Centoni, Cisco’s EVP of Buyer Expertise, has written about the broader framework for constructing belief in agentic AI, together with governance by design and RBAC as foundational rules. These aren’t afterthoughts. They’re stipulations.
5. Empty Outcomes Want Explanations
When a database question returns no rows, your first intuition is perhaps to inform the consumer “no information discovered.” That is virtually at all times the incorrect reply.
“No information discovered” is ambiguous. Does it imply the entity doesn’t exist? The entity exists however has no information for this time interval? The question was malformed? The consumer doesn’t have permission to see the info?
Every situation requires a unique response. The third is a bug. The fourth is a coverage that wants transparency (see part above).
System-enforced filters (RBAC): The information exists, however the consumer doesn’t have permission to see it. The correct response: “No information are seen together with your present entry permissions. Information matching your standards exist within the system.” That is transparency, not an error.
Consumer-applied filters: The consumer requested for one thing particular that doesn’t exist. “Present me upcoming subscription renewals for ACME Corp in Q3” returns empty as a result of there are not any renewals scheduled for that buyer in that interval. The correct response explains what was searched: “I couldn’t discover any subscriptions up for renewal for ACME Corp in Q3. This might imply there are not any energetic subscriptions, or the info hasn’t been loaded but.”
Question errors: The filter values don’t exist within the database in any respect. The consumer misspelled a buyer identify or used an invalid ID. The correct response suggests corrections.
We deal with this at a number of layers. When queries return empty, we analyze what filters eradicated information and whether or not filter values exist within the database. When entry management filtering produces zero outcomes, we verify whether or not outcomes would exist with out the filter. The synthesis layer is instructed to by no means say “the SQL question returned no outcomes.”
This transparency builds belief. Customers perceive the system’s boundaries somewhat than suspecting it’s damaged.
6. Personalization is Not Non-compulsory
Most enterprise AI is designed as a one-size-fits-all interface. However folks count on an “assistant” to adapt to their distinctive wants and help their approach of working. Pushing a inflexible system with out primitives for personalization causes friction. Customers attempt it, discover it doesn’t match their workflow, and abandon it.
We addressed this on a number of fronts.
Shortcuts enable customers to outline command aliases that increase into full prompts. As a substitute of typing out “Summarize renewal danger for ACME Corp, present a two paragraph abstract highlighting key danger elements which will affect chance of non-renewal of Meraki subscriptions”, a consumer can merely kind /danger ACME Corp. We took inspiration from agentic coding instruments like Claude Code that help slash instructions, however constructed it for enterprise customers to assist them get extra executed shortly. Energy customers create shortcuts for his or her weekly reporting queries. Managers create shortcuts for his or her workforce overview patterns. The identical underlying system serves completely different workflows with out modification.
Primarily based on manufacturing visitors, we’ve seen essentially the most energetic shortcut customers common 4+ makes use of per shortcut per day. Energy customers who create 5+ shortcuts generate 2-3x the question quantity of informal customers.
Scheduled prompts allow automated, asynchronous supply of knowledge. As a substitute of synchronous chat the place customers should bear in mind to ask, duties ship insights on a schedule: “Each Monday morning, ship me a abstract of at-risk renewals for my territory.” This shifts the assistant from reactive to proactive.
Lengthy-term reminiscence remembers utilization patterns and consumer behaviors throughout dialog threads. If a consumer at all times follows renewal danger queries with product adoption metrics, the system learns that sample and recommends it. The objective is making AI really feel actually private, prefer it is aware of the consumer and what they care about, somewhat than beginning contemporary each session.
We observe utilization patterns throughout all these options. Closely-used shortcuts point out workflows which might be price optimizing and generalizing throughout the consumer group.
7. Carrying Context from the UI
Most AI assistants deal with context as chat historical past. In dashboards with AI assistants, one of many challenges is context mismatch. Customers could ask a couple of particular view, chart or desk they’re viewing, however the assistant normally sees chat textual content and broad metadata or carry out queries which might be outdoors the scope the consumer switched from. The assistant doesn’t reliably know the precise reside view behind the query. As filters, aggregations, and consumer focus change, responses turn into disconnected from what the consumer really sees. For instance, a consumer could apply a filter for belongings which have reached end-of-support for a number of architectures or product varieties, however the assistant should still reply from a broader prior context.
We enabled an possibility through which UI context is specific and steady. Every AI flip is grounded within the precise view state of the chosen dashboard content material and even objects, not simply dialog historical past. This offers the assistant exact situational consciousness and retains solutions aligned with the consumer’s present display screen. Customers are made conscious that they’re inside their view context after they change to the assistant window,
For customers, the largest achieve is accuracy they will confirm shortly. Solutions are tied to the precise view they’re , so responses really feel related as an alternative of generic. It additionally reduces friction: fewer clarification loops, and smoother transitions when switching between dashboard views and objects. The assistant feels much less like a separate chat device and extra like an extension of the interface.
8. Constructing AI with AI
We develop these agentic techniques utilizing AI-assisted workflows. It’s about encoding a senior software program engineer’s data into machine-readable patterns that any new workforce member, human or AI, can comply with.
We preserve guidelines that outline code conventions, architectural patterns, and domain-specific necessities. These guidelines are at all times energetic throughout growth, making certain consistency no matter who writes the code. For advanced duties, we preserve command recordsdata that break multi-step operations into structured sequences. These are shared throughout the workforce, so a brand new developer can decide issues up shortly and contribute successfully from day one.
Options that beforehand required multi-week dash cycles now ship in days.
The important thing perception: the worth isn’t essentially in AI’s basic intelligence and what state-of-the-art mannequin you employ. It’s within the encoded constraints that channel that intelligence towards helpful outputs. A general-purpose mannequin with no context writes generic code. The identical mannequin with entry to undertaking conventions and instance patterns writes code that matches the codebase.
There’s a moat in constructing a undertaking as AI-native from the beginning. Groups that deal with AI help as infrastructure, that spend money on making their codebase legible to AI instruments, transfer sooner than groups that bolt AI on as an afterthought.
Conclusion
None of those patterns are technically refined. They’re apparent in hindsight. The problem isn’t understanding them; it’s prioritizing them over extra thrilling work.
It’s tempting to chase the newest protocol or orchestration framework. However customers don’t care about your structure. They care whether or not the system helps them do their job and is evolving shortly to inject effectivity into extra components of their workflow.
The hole between “technically spectacular demo” and “really great tool” is stuffed with many of those unglamorous patterns. The groups that construct lasting AI merchandise are those prepared to do the boring work effectively.
These patterns emerged from constructing a manufacturing AI Assistant at Cisco’s Buyer Expertise group. None of this may exist with out the workforce of architects, engineers and designers who argued about the best abstractions, debugged the sting circumstances, and stored pushing till the system really labored for actual customers.