Connect with us

Artificial Intelligence

How Amul is using AI dairy farming to put 36 million farmers first

Published

on

AI dairy farming has found its most ambitious deployment yet – not in a Silicon Valley lab nor a European agri-tech campus, but in the villages of Gujarat, India, where 36 lakh (3.6 million) women milk producers are now being served by an AI assistant named Sarlaben.

Amul, the world’s largest dairy cooperative, has launched what it calls Amul AI: a platform built on five decades of cooperative data, designed to give every farmer in its network round-the-clock, personalised guidance in their own language.

Amul was launched just ahead of India’s AI Impact Summit 2026 and backed by the Ministry of Electronics and Information Technology (MeitY) with the EkStep Foundation. It is a test case for whether AI – the kind being debated in boardrooms and policy forums globally – can actually reach the last mile.

Meet Sarlaben: The AI dairy farming assistant

Sarlaben draws from one of India’s most comprehensive agricultural data repositories. It’s accessible via the Amul Farmer mobile app – already downloaded by over 10 lakh (one million) users on Android and iOS – as well as through voice calls for farmers using feature phones or landlines.

The system is integrated with Amul’s Automatic Milk Collection System (AMCS) and the Pashudhan application, allowing it to offer personalised, cattle-specific guidance.

What makes Amul AI substantially different from most agricultural chatbots is the scale of its training data. The platform was built on a digital backbone managing over 200 crore (two billion) milk procurement transactions annually, veterinary treatment records from more than 1,200 doctors covering nearly 3 crore (30 million) cattle, approximately 70 lakh (seven million) artificial inseminations conducted each year, ISRO satellite imagery for fodder production mapping, and a cattle census conducted every five years.

Every animal in the system carries a unique ID, with individual records of feed intake, disease history and milking status. “Amul AI is about taking dependable, verified information directly to the farmer – instantly and in a language they are comfortable with,” said Jayen Mehta, Managing Director of the Gujarat Cooperative Milk Marketing Federation (GCMMF), which markets the Amul brand.

He said how, by using decades of structured data and integrating it with their operational systems, the platform will help farmers make timely decisions that improve animal productivity and income.

India’s productivity paradox

India is the world’s largest producer of milk, generating 347.87 million tonnes in 2024-25 according to the Department of Animal Husbandry and Dairying – more than double the US’s 102.70 million tonnes. And yet despite leading in volume, India’s per-animal milk yield remains among the lowest globally.

The reasons are structural. India’s dairy sector is characterised by small herd sizes, low-quality feed, limited access to veterinary care in rural areas, and widespread lack of awareness about modern breeding and husbandry practices. Amul’s network spans more than 18,600 villages in Gujarat, where farmers supply over 350 lakh litres (35 million litres) of milk daily.

But information asymmetry has long been a bottleneck – a farmer facing a sick animal at midnight in a remote village has few places to turn; the gap Amul AI is designed to close.

Available initially in Gujarati – the primary language of the cooperative’s farmer base – the platform is built on the government’s Bhashini multilingual framework and could, in principle, be extended to 20 Indian languages, reaching Amul’s presence in 20,000 villages in 20 states.

The cooperative model

The technology story here is inseparable from the institutional one. Amul’s cooperative structure – built over five decades under the original White Revolution – created the data infrastructure that makes Amul AI possible.

Most private agri-tech startups are working backwards: collecting data first, building products second. Amul already had the data. What was needed was a way to make it actionable at the farmer level.

Experts tracking the dairy-tech space see this as significant. Sreeshankar Nair, Founder of Brainwired, a dairy-tech startup, identifies three specific challenges that Amul AI could meaningfully address: farmer awareness, access to quality veterinary guidance, and connectivity to grazing and feed resources.

“If AI can integrate local dialects of Indian languages, India can have White Revolution 2.0,” Nair said, pointing to the transformative potential of vernacular AI in a sector where not every farmer speaks the same dialect.

Saswata Narayan Biswas, Director of the Institute of Rural Management, Anand (IRMA) – the institution closely associated with Amul’s founding ethos – frames it as an AI embedded in a cooperative framework. It becomes “not a technology upgrade, but an instrument of inclusive rural transformation.”

For Biswas, the specific abilities Amul AI brings – predictive disease detection, oestrus tracking, optimised feed formulation, localised weather risk advisories – are abilities Amul had been building for years. AI accelerates and democratises them.

Scale and the test ahead

The launch has drawn backing from the highest levels of government. Gujarat Chief Minister Bhupendra Patel launched the platform and confirmed it will be showcased at the AI Impact Summit 2026. The cooperative has acknowledged MeitY and the EkStep Foundation – an open digital infrastructure nonprofit – as partners in building the AI layer.

Farmers not affiliated with Amul can also access general dairying and animal husbandry information through the app. At its current scale, Amul AI already covers more cattle – nearly 3 crore (30 million) – than most national veterinary databases anywhere in the world.

The harder question, as with most AI deployments at a population scale, is whether the tool will serve those who need it most. The farmers most likely to benefit first – those already comfortable with smartphones, already plugged into Amul’s digital system – may not be the ones with the greatest information deficit.

The rollout of Bhashini-enabled dialect support, the adoption rate among feature-phone users relying on voice calls, and whether AI-driven advisories translate into measurable yield improvements will be the metrics that determine whether this is genuinely White Revolution 2.0.

Amul has built an AI system grounded in half a century of real cooperative transactions, real animals, and real farmers. Such an infrastructure is, arguably, the most credible foundation for AI dairy farming at scale. Whether it fulfils its promise will depend on execution – and on whether Sarlaben’s voice can reach in the last few miles; those that have always been the hardest to cross.

See also: Hitachi bets on industrial expertise to win the physical AI race

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and co-located with other leading technology events. Click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Artificial Intelligence

Mastercard’s AI payment demo points to agent-led commerce

Published

on

A recent demonstration from Mastercard suggests that payment systems may be heading toward a future where software agents, not people, complete purchases. During the India AI Impact Summit 2026, Mastercard showed what it described as its first fully authenticated “agentic commerce” transaction.

In the demo, as reported by Times of India, an AI agent searched for a product, assessed the website, and completed the purchase using stored payment credentials, without the user opening an app or entering card details. The company said the transaction took place inside a secure payment framework designed to verify both the user and the AI acting on their behalf.

The demonstration was controlled, not a public rollout. Mastercard executives told reporters that broader deployment would depend on regulatory approval and ecosystem readiness. Still, the test highlights a change that many enterprises may need to prepare for: the possibility that customers – or corporate systems – will increasingly rely on AI agents to initiate and complete transactions.

Assisted checkout to delegated spending

Digital payments have usually focused on reducing friction for human users through tokenisation, saved credentials, and one-click checkout. Agentic commerce goes further. Instead of helping a user complete a purchase, the system allows software to handle the process from start to finish once permission rules are in place.

The model relies on several building blocks already used in modern payments: identity verification, tokenised card data, and risk monitoring. What changes is who performs the action. If AI agents can act in defined limits, like spending caps or merchant restrictions, checkout may change from a user interaction to a background workflow.

For enterprises, the issue is if software can spend money automatically, procurement rules, approval chains, and audit trails need to account for machine decisions, not human ones. Finance teams may need clearer policies on when an AI agent can commit funds, how liability is assigned if something goes wrong, and how fraud detection should treat automated transactions.

Payment networks position for machine customers

Mastercard is not alone in exploring this direction. Across the payments sector, providers are testing ways to embed transactions into AI-driven tools and digital assistants. The goal is to ensure that when autonomous software begins purchasing goods or services, payment networks remain part of the trust and verification layer.

In public statements tied to the summit demo, Mastercard framed the effort as building infrastructure that allows AI agents to transact safely on behalf of users. That framing points to a broader industry race: not to build smarter AI shopping tools, but to control the authentication systems that make those tools safe enough for financial use.

For banks and fintech firms, the change could affect how customer identity is managed. Traditional authentication often assumes a person is present, entering a password or approving a prompt. Agentic commerce assumes the opposite: the user may not be involved at the moment of purchase. That means identity systems must verify both the account owner’s prior consent and the agent’s authority at the time of transaction.

Merchants may need API-ready storefronts

If AI agents begin acting as buyers, merchant systems may also need to adapt. Online stores built mainly for human browsing may struggle if automated agents become a meaningful share of customers.

To support machine-driven purchases, product catalogues, pricing data, and checkout processes may need to be accessible through structured APIs not only visual web pages. Inventory accuracy, transparent pricing, and clear return policies become more important when decisions are made by software trained to compare options instantly.

This could also influence competition. If agents optimise for price and delivery speed, merchants with inconsistent data or hidden fees may be filtered out before a human even sees them.

Security risks move, not disappear

While agentic commerce promises convenience, it also introduces new risks. A compromised AI assistant with payment authority could execute purchases at scale before detection. Fraud models that look for unusual user behaviour may need updating to distinguish between legitimate automated spending and malicious activity.

Regulators are likely to take a cautious approach. Mastercard’s own comments that the system still awaits approvals suggest that compliance frameworks for AI-initiated payments are still taking shape.

In enterprises deploying AI internally, similar concerns apply. Automated purchasing agents integrated into enterprise resource planning systems could streamline routine procurement, but they also expand the attack surface. Access controls and spending thresholds will matter more when software can execute financial actions without real-time human confirmation.

Where commerce may head

Mastercard’s demonstration does not mean agent-led payments will reach consumers immediately. Yet it offers a glimpse of how commerce may change as AI systems move from advisory roles into operational ones.

If the model matures, the most visible change may be that checkout disappears as a distinct step. Instead of visiting a site and paying, users or companies may set rules, and their software will handle the rest.

For enterprises, the important takeaway is less about Mastercard’s AI technology and more about the direction of travel. As AI agents gain the authority to act, payment systems, identity frameworks, and digital storefronts may need to treat software not as a tool, but as a participant in the transaction.

(Photo by Cova Software)

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and co-located with other leading technology events. Click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.

Continue Reading

Artificial Intelligence

Hitachi bets on industrial expertise to win the physical AI race

Published

on

Physical AI–the branch of artificial intelligence that controls robots and industrial machinery in the real world–has a hierarchy problem. At the top, OpenAI and Google are scaling multimodal foundation models. In the middle, Nvidia is building the platforms and tools for physical AI development. 

And then there is a third camp: industrial manufacturers like Hitachi and Germany’s Siemens, which are making the quieter but arguably more grounded argument that you cannot train machines to navigate the physical world without first understanding it.

That argument is now moving from boardroom strategy to factory floor deployment, as Hitachi revealed in a recent interview with Nikkei Asia.

Why Physical AI needs more than a good model

Kosuke Yanai, deputy director of Hitachi’s Centre for Technology Innovation-Artificial Intelligence, is direct about what separates viable physical AI from the theoretical kind. “Physical AI cannot be implemented in society without a systematic understanding that begins with foundational knowledge of physics and industrial equipment,” he told Nikkei.

Hitachi’s pitch is that it already holds much of that foundational knowledge–accumulated over decades of building railways, power infrastructure, and industrial control systems. The company has thermal fluid simulation technology that models the behaviour of gases and liquids, and signal-processing tools for monitoring equipment condition — what Yanai describes as the engineering foundation underpinning Hitachi’s ‘extensive knowledge of product design and control logic construction.’

From concept to deployment: Daikin and JR East

While Hitachi’s overarching physical AI architecture–the Integrated World Infrastructure Model (IWIM), which it describes as a mixture-of-experts system integrating multiple specialised models, simulators, and data sets–remains in the concept verification stage, two real-world deployments signal that the underlying approach is already producing results.

In collaboration with Daikin Industries, Hitachi has deployed an AI system that diagnoses malfunctions in commercial air-conditioner manufacturing equipment. The system, trained on equipment maintenance records, procedure manuals, and design drawings, can now identify which component is likely failing when an anomaly is detected–the kind of operational intuition that previously existed only in the heads of experienced engineers.

With East Japan Railway (JR East), Hitachi has built an AI that identifies the root cause of malfunctions in the control devices running the Tokyo metropolitan area’s railway traffic management system, and then assists operators in formulating a response plan. In a network where delays ripple across millions of daily journeys, the ability to accelerate fault diagnosis carries real operational weight.

The R&D pipeline: Cutting development time

Hitachi’s physical AI push is also showing up in its research output. In December 2025, the company published findings from two projects presented at ASE 2025, a top-tier software engineering conference, that address a persistent bottleneck in industrial AI: the time and effort required to write and adapt control software.

In the automotive sector, Hitachi and its subsidiary Astemo developed a system that uses retrieval-augmented generation to automatically produce integration test scripts for vehicle electronic control units (ECUs)–pulling from hardware-specific API information and frontline engineering knowledge. In a pilot involving multi-core ECU testing, the technology reduced integration testing man-hours by 43% compared to manual execution.

In logistics, the company developed variability management technology that modularises robot control software into reusable components structured around a robot operating system (ROS). By mapping out the environmental variables and operational requirements of different warehouse settings in advance, the system lets operators adapt robotic picking-and-placing workflows to new products or layouts without rewriting software from scratch.

Safety as a structural requirement, not an afterthought

One thread that runs through all of Hitachi’s physical AI work is its emphasis on safety guardrails–not as a compliance checkbox, but as an engineering constraint baked into system design. Yanai told Nikkei that the company is integrating its control and reliability technology from social infrastructure development to prevent AI outputs from deviating from human-approved operating parameters. 

This includes input validation to screen out data that models should not be trained on, output verification to ensure machine actions do not endanger people or property, and real-time monitoring of the AI model itself for operational anomalies.

It is a meaningful distinction. Physical AI systems fail in the real world, not in a sandbox. The stakes for an AI controlling railway signalling or factory robotics are categorically different from those governing a chatbot.

Infrastructure to match the ambition

On the infrastructure side, Hitachi Vantara–the group’s data and digital infrastructure arm–is positioning itself as an early adopter of NVIDIA’s RTX PRO Servers, built on the RTX PRO 6000 Blackwell Server Edition GPU, designed to accelerate agentic and physical AI workloads. The hardware is being paired with Hitachi’s iQ platform and used to build digital twins–virtual replicas of physical systems–that can simulate everything from grid fluctuations to robotic motion at scale.

The IWIM concept, meanwhile, is designed to connect Nvidia’s open-source Cosmos physical AI development platform with specialised Japanese-language LLMs and visual language models via the model context protocol (MCP)–essentially a framework to stitch together the models, simulation tools, and industrial datasets that physical AI systems require.

The broader race in physical AI is far from settled. But Hitachi’s position–that domain expertise and operational data are as important as model architecture–is increasingly hard to dismiss, particularly as deployments with partners like Daikin and JR East begin to demonstrate what that expertise is actually worth in practice.

Sources: Nikkei Asia (Feb 21, 2026); Hitachi R&D (Dec 24, 2025); Hitachi Vantara Blog (Aug 27, 2025)

See also:Alibaba enters physical AI race with open-source robot model RynnBrain

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.

Continue Reading

Artificial Intelligence

Exploring AI in the APAC retail sector

Published

on

AI in the APAC retail sector is transitioning from analytics and pilots into workflows and daily operations.

Dense urban stores, high labour churn, and competitive quick-commerce ecosystems are driving the uptake. A Q4 2025 survey by GlobalData found that 45 percent of consumers in Asia and Australasia are very or quite likely to purchase a product based on AI recommendations or endorsements.

Jaya Dandey, Consumer Analyst at GlobalData, said: “Whether shoppers realise it or not, machine-learning systems have long been deciding when to encourage consumers to make purchases, which products they can see, and what discounts they can avail.

“Now, agentic systems can also complete shopping-related tasks end-to-end.” 

Computer vision and store automation

Enterprises evaluating computer vision and machine learning can observe early implementations in the region.

Lawson, for example, introduced AI-enabled ‘Lawson Go’ stores in Japan during 2022. The retailer collaborated with technology provider CloudPick in 2025 to integrate AI, machine learning, and computer vision. This integration eliminates check-out lines and cashiers to enhance the customer experience.

In South Korea, retail AI company Fainders.AI launched a compact and cashier-less MicroStore inside a gym in 2024. This deployment improved the accessibility of autonomous retail across different businesses.

AI also aids the forecasting and automation of retail replenishment—a capability that applies well to the APAC market, where store footprints are small and replenishment frequency is high.

Japanese food retail chain Coop Sapporo uses a camera-based AI system named Sora-cam, developed by Soracom. The system helps the chain avoid overstocking and reduce unsold merchandise on store shelves. Coop Sapporo employs an analytics team to evaluate the generated images. The team determines the optimal shelf display ratio. The Sora-cam system also alerts staff members to apply discount labels on food items close to expiry to prevent wastage.

AI models track waste and markdown timing while improving promotion efficiency. In Southeast Asian (SEA) markets characterised by high price sensitivity, minor improvements in promotion efficiency increase profit margins.

AI-driven labour optimisation measures include scheduling, task priority lists, and workload balancing. These measures assist retailers in Japan and South Korea, which face structural labour shortages. They also provide efficiency benefits in high-growth SEA markets.

Agentic AI systems in retail are improving APAC consumer interaction

“In food retail, agentic AI is best understood as an AI ‘operator’ that can understand a goal, plan steps, stay within budget or allergen constraints, execute actions across systems, ask clarifying questions, and learn preferences over time,” says Dandey. 

Customers can bypass individual item searches by outlining their overall intent. A customer, for example, might request an AI agent to “Plan five dinners for a family of four, mostly Asian recipes, no shellfish, under 45 minutes.” The agent then generates recipes, builds a shopping cart, sizes quantities, and adds missing staples to the cart.

This retail agentic AI capability aligns with regional behaviours, as many APAC households cook frequently and shop fresh. AI agents that recognise local cuisines – such as Korean banchan, Japanese bentos, and Indian spice bases – fit regional habits better than generic Western meal plans.

“In many APAC markets, shopping is already deeply integrated with digital wallets, messaging apps, ride-hailing, and delivery ecosystems, making it easier for agentic AI to plug into daily routines,” explains Dandey.

“Nevertheless, some key challenges need to be overcome; ensuring private data sharing consent, minimising hallucinations in terms of allergens and ingredients, and implementing proper localisation of the system with language nuance.”

See also: DBS pilots system that lets AI agents make payments for customers

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security & Cloud Expo. Click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.

Continue Reading

Trending