The Structural Shift in AI Value Creation
The Akamai Digital Leadership Summit revealed that India's AI deployment challenges are creating a fundamentally different engineering discipline that prioritizes extreme cost optimization over model sophistication. This shift matters because it represents the first systematic blueprint for scaling AI in constrained environments that will influence global deployment economics.
India's infrastructure layer presents challenges fundamentally different from Western markets: not labor shortages or aging populations, but the need to deliver services at population scale for near-zero cost per transaction. Jigar Halani from NVIDIA's solution architecture team framed this starkly: "AI, for India, is not a replacement of humans. It is solving something at population scale that humans simply haven't been able to solve over a long period of time." This reality has forced Indian companies to develop specialized expertise that Western counterparts haven't needed to cultivate.
The summit's discussions revealed that the foundation model debate is largely settled in India. What remains is the harder question: how to actually run AI in production at India's scale, under India's cost constraints, without the infrastructure budgets that the problems seem to demand. This shift from model development to production engineering represents a critical inflection point in AI's evolution.
The Cost Optimization Imperative
Mohit Saxena, Co-Founder and CTO of InMobi and GlanceAI, provided the most revealing data point: when Glance first started generating AI images, the cost was $30 per image, which was unviable for India's price-sensitive market. Through systematic optimization, the team brought this down to $1.50, then to a few cents. "If you make it at three cents, the same product that you launch in the US, then India is sorted," Saxena explained.
The method behind this dramatic cost reduction reveals a structural advantage: "Almost 60% of queries are repetitions. You don't need to call the LLM for every one of them." Combined with batch processing and multiple specialized models, Glance reduced its effective model invocations substantially. This approach represents a fundamental rethinking of AI economics that Western companies operating in resource-abundant environments haven't needed to develop.
Kiran Kumar Katreddi, VP of Platform Engineering at Meesho, extended this cost conversation into even more constrained territory. With over 200 million users, many first-time internet users in Tier-III and Tier-IV towns, Meesho operates under specific engineering constraints: a 14MB app size, voice and image search in eight Indian languages, personalization that updates within a 500-millisecond session window, and AI-assisted address resolution for deliveries to locations described as "opposite the previous sarpanch's house."
The Production Engineering Breakthrough
Sanath Moguluri, VP of Voice AI Engineering at Reliance Jio, described the latency challenge of serving hundreds of millions of users across smartphones, televisions, and automotive systems. "On a telecom network, achieving 500 milliseconds is very challenging. In our experience, around one second of latency is good enough for people to converse with agents, when the use case is specific and domain-focused."
Jio's hybrid deployment model reveals a sophisticated understanding of production requirements: "Not everything needs to go to LLMs. We have hybrid deployments, from edge to cloud, and even within the cloud, there are smaller and larger models doing orchestration." This layered approach to AI deployment represents a maturity that many Western companies haven't yet achieved.
Sagar Gaonkar, CTO of Eloelo, described his team's approach to live content moderation, separating edge decisions from cloud decisions, with human review reserved for boundary cases. "A good 80% of the cases are very black and white. That last 20% is where you want the human in the loop." The end-to-end cycle runs in under 10 seconds, demonstrating how Indian companies are achieving production-grade performance under severe constraints.
The Security Implications of AI Expansion
Vijay Kolli, Akamai's Regional VP for Enterprise Security, shifted the focus to what happens when AI systems expand the attack surface. The number that got the room's attention: API growth on Akamai's network is no longer 100% annually. "It's literally 1,000% and more." AI agents accessing internal databases and inheriting permissions without judgment have changed the threat model in ways that legacy architectures were not designed for.
Mukesh Solanki, CISO at KreditBee, was frank about the attacker advantage: "Hackers will get much more sophisticated with generative AI tools, and will find better ways to poison data so that someone who isn't eligible for a loan ends up getting one." His company processes a million loans a month, making security non-negotiable.
Sujatha Iyer, Head of AI Security at Zoho Corp, made the case for deterministic models where explainability is non-negotiable: "If your monitoring solution is telling you there's an 80% chance your server is going to face an outage, it has to come with an explanation." Her closing line was unambiguous: "Security — imbibe it right from day one of software development. It's not an afterthought anymore."
The Sovereign AI Movement
Ganesh Gopalan, CEO of Gnani.ai, closed the formal sessions with a grounded take on sovereign AI. He said the commercial rationale is straightforward: enterprises want to retain ownership of the intelligence embedded in their systems. Gnani's response is to own every layer of its voice stack, ASR, TTS, turn-taking, denoising, and a small language model tuned for voice.
"Unless you develop that tech, you're going to struggle – firstly in terms of protecting your customers' data, secondly about superior experiences, and thirdly about cost," he explained. The structural problem with current voice pipelines, he argued, is that converting voice to text and back again loses emotional information that matters. Gnani is building a voice-to-voice model that preserves it.
On guardrails, his observation was pointed: "A couple of years back, we very proudly told customers that 55% of our prompting was guardrails. Today, if you say that to a customer, they will throw you out of the room. The benchmark now is that a minimum of 75 to 80% needs to be guardrails." This represents a significant maturation in AI deployment standards.
The Engineering Talent Transformation
Mohit Saxena pushed back against easy conclusions about AI's impact on engineering talent: "AI has reduced the bar of being an average engineer. But it has really raised the bar of a good engineer. The average is not good enough anymore." Today, roughly 70% of code at Glance is written by AI in the IDE, but the integration work, he said, still requires the best engineers.
This insight reveals a structural shift in what constitutes valuable engineering talent in the AI era. The ability to work with AI tools, optimize deployment, and integrate systems has become more valuable than raw coding ability. Indian companies, operating under severe constraints, are developing this talent pool faster than their Western counterparts.
Pranav Tiwari, Head of Engineering APAC at Postman, offered a wider lens on what agentic AI is doing to the connectivity layer. "What used to connect applications very deterministically is changing fundamentally. Connectivity is transforming from plumbing between two applications to something with inference in the middle, business logic blended in, and a series of conversations that eventually get a task done." A show of hands confirmed that a significant portion of the room already has agent-written code running in production.
The Infrastructure Layer Differentiation
India's infrastructure challenges are creating specialized solutions that may not transfer easily to Western markets. The Bharat ML stack, which Meesho open-sourced in 2025 after 18 months of development, handles 3-4 trillion inferences and 1 million queries per second on model inference alone. This platform was built specifically to handle peak Diwali sale volumes, when order volumes hit 3-4x normal, and commercial inference platforms kept breaking.
"Most of our innovation," Katreddi said, "exists because of the AI investments we've made over the last four or five years." This long-term investment in production AI infrastructure represents a competitive advantage that cannot be easily replicated.
Akamai's partnership with NVIDIA, deploying RTX Pro 6000 GPUs across a distributed network, represents another structural advantage. As Dr Robert Blumofe, Akamai's CTO, noted: "A lot of companies who have come to it late find themselves becoming LLM one-trick ponies. To get real value out of AI, you not only need to know how to use the LLM, but you need to know how to use other forms of deep learning and other forms of ML."
The Global Implications
The discussions at the summit suggested that the next phase of AI adoption will not be defined solely by advances in models, but by the engineering discipline required to run them reliably in the real world. In markets like India, where platforms must serve hundreds of millions of users while maintaining cost efficiency, that discipline may ultimately determine which AI systems succeed in production.
This represents a fundamental shift in competitive dynamics. Companies that have developed expertise in constrained environments now possess knowledge and capabilities that are increasingly valuable as AI deployment becomes more widespread globally. The cost optimization techniques, production engineering practices, and security frameworks developed in India represent exportable intellectual property.
As Sumant Narayanan, Akamai's Regional Sales Director for India and SAARC, noted in his welcome address: "Over the last 12 months or so, a lot of the conversation around AI has been mostly about foundation models, making them bigger and bigger. But now, the conversation has shifted towards how enterprises use these foundational models and actually deliver value to their customers." This shift from model development to value delivery represents the maturation of the AI industry.
Source: YourStory
Rate the Intelligence Signal
Intelligence FAQ
Indian companies optimize through query repetition analysis (60% of queries don't need LLM calls), batch processing, specialized smaller models, and hybrid edge-cloud deployments that Western companies operating in resource-abundant environments haven't needed to develop.
Forced innovation under severe cost and infrastructure constraints has created specialized expertise in production AI engineering that represents exportable intellectual property and gives Indian companies first-mover advantage in emerging markets.
AI reduces the bar for average engineers but raises it significantly for good engineers who must master system integration, cost optimization, and production deployment—skills that Indian companies are developing faster due to their constraints.
API growth exceeding 1,000% annually, AI agents inheriting permissions without judgment, and sophisticated AI-powered attacks that can poison data systems represent new threat models that legacy security architectures weren't designed to handle.
Enterprises want ownership of embedded intelligence, data protection, superior localized experiences, and cost control—all driving demand for region-specific AI stacks rather than one-size-fits-all global solutions.

