Agentic AI is being adopted across the software development lifecycle, but its success hinges on rigorous data integration and human-in-the-loop governance.
Operationalizing autonomous agents in software development
In a recent Infosys Knowledge Institute podcast, Prasad Banala, director of software engineering at a major U.S. retailer, described production-scale workflows in which autonomous agents manage requirement validation, test-case generation and analysis, issue triage and resolution, and governance enforcement through embedded review checkpoints. This approach moves beyond code completion, positioning agents as active participants in planning, quality assurance and incident response.
Early adopters report that autonomous agents have compressed validation cycles by roughly 30 percent and cut manual triage time by an estimated 40 percent, according to vendor briefings and retailer disclosures at industry conferences. Such figures align with Salesforce’s report that retailers investing in AI contributed to $229 billion in holiday sales in 2024, though those outcomes fuse traditional ML with emerging agentic workflows. What stands out is how teams are reshuffling tasks: software engineers are evolving into data stewards responsible for crafting acceptance-criteria schemas and maintaining the plumbing that feeds agents, while product managers and legal teams assume oversight roles in policy authoring and exception adjudication.
This structural shift underscores a core insight: agentic autonomy without end-to-end governance can accelerate decision-making, but only integrated data and human checkpoints preserve the safety and compliance of automated actions.
Data integration as the linchpin of autonomy
Industry surveys from Gartner and Deloitte consistently flag fragmented data and legacy APIs as the leading impediments to scaling agentic AI. Gartner’s 2026 CIO Agenda forecast indicates that roughly 48 percent of retailers plan agentic AI pilots within the next 12 months, yet only 39 percent currently operate unified data strategies. Deloitte’s 2025 AI adoption survey shows that 48 percent of respondents cite data searchability and 47 percent cite data reusability as critical blockers.
In practice, engineering teams are experimenting with multiple integration architectures: lightweight API gateways that expose sanitized event streams; middleware layers that map legacy ERP and CRM schemas into normalized object graphs; and knowledge-graph pilots that enable agents to traverse domain ontologies for richer context. These approaches aim to avoid expensive rip-and-replace projects while providing agents with consistent, real-time access to requirements, test outcomes, user feedback and incident logs. However, without cohesive data catalogs and metadata governance, agents can produce partial or outdated recommendations, triggering false positives or policy violations.

Engineering leaders are reporting multi-month timelines to achieve initial data harmonization—time during which automated pipelines can drift and require human intervention. These integration efforts reflect an emerging consensus that robust data plumbing is a non-negotiable prerequisite for reliable agentic workflows.
Human-in-the-loop governance and shifting decision authority
While autonomous agents promise throughput gains, governance models centered on human oversight are emerging as the fulcrum that prevents drift and error. In Banala’s production framework, agent proposals that exceed policy thresholds—such as changes to pricing modules or database migrations—are automatically flagged and routed to a multidisciplinary review board. Caroline Reppert, director at the National Retail Federation, has noted that NRF working groups are convening in early 2026 to define shared governance taxonomies and best practices for human escalation paths.
Such governance patterns reshape internal power structures. Business, legal and security stakeholders acquire formal veto rights over automated actions. Software engineers transition from sole code authors to curators of policy engines and escalation criteria. Organizationally, this creates new accountability lines: policy exceptions must be justified in real-time dashboards, and every override generates an audit-log entry that ties back to individual reviewers.
Kognitos’ analysis cautions that absence of these guardrails risks “brittleness, maintenance complexity and adaptation failures,” especially when agents encounter novel scenarios. The human-in-the-loop checkpoints, therefore, do more than mitigate risk—they redefine the locus of authority, ensuring that human values and strategic considerations remain embedded at every stage of automation.
Failure modes and industry outlook
Industry estimates suggest that more than 40 percent of agentic AI projects risk failure by 2027 due to legacy API bottlenecks and ETL data friction, according to Gartner and other analysts. These failure modes are not mere budgetary overruns: they entail loss of stakeholder trust, brand integrity risks when erroneous actions reach production, and burnout among engineering teams forced to patch fragmented pipelines.

McKinsey’s recent whitepaper identifies six technical domains—core platforms for dynamic pricing, in-store point-of-sale integration, inventory forecasting, retail media monetization, customer engagement engines and supply-chain optimization—where hybrid innovate-and-renovate strategies can mitigate bypass risks. Rather than delegating entire systems to agents, teams are carving out narrow use cases—test-case creation in isolated environments, code linting against security policies, or automated dependency upgrades—while preserving manual controls on critical paths.
These curbs reveal a deeper human impact: enterprise leaders must negotiate the tension between speed and stewardship. When governance frameworks lag behind agentic capabilities, policy exceptions proliferate, decision authority becomes opaque and the human sense of agency erodes, undermining institutional confidence in AI-driven processes.
Engineering practices emerging from early pilots
Diagnostic review of pilot programs shows common patterns in reconciling autonomy with control. Engineering teams are:
- Instrumenting agentic workflows with quantitative KPIs—such as test-coverage delta, agent confidence percentiles and policy-violation counts—to transform governance from checklist exercises into real-time operational metrics.
- Deploying middleware for data normalization—mapping legacy systems into canonical event streams—to grant agents consistent context without full platform replacements.
- Embedding scaled audit logs that record every agent decision, human override and policy exception, enabling compliance teams to reconstruct decision timelines and attribute accountability.
- Establishing cross-functional AI councils—comprising engineering, security, legal and operations representatives—that meet in sprint cycles to calibrate escalation thresholds and review audit-log trends.
- Pairing agentic pipelines with periodic human-driven “health checks,” where teams validate data freshness, governance rule coverage and drift in agent performance against business KPIs.
These practices underscore a diagnostic truth: agentic AI will not supplant human judgment but will shift it upstream, requiring new skills in policy definition, data stewardship and cross-functional collaboration. The real test lies not in writing autonomous agents but in weaving them into the social fabric of engineering organizations.
Next indicators to monitor
- Updated Gartner metrics on pilot-to-scale conversion rates and revised forecasts of project failure modes through 2027.
- Publications by the National Retail Federation and industry working groups detailing shared governance templates and human-in-the-loop taxonomies, expected in Q1 2026.
- Vendor disclosures and case studies quantifying MTTR improvements, test-coverage expansion and revenue impact linked specifically to agentic SDLC deployments.
- Emerging standards for audit-log interoperability, policy exception taxonomies and data-catalog governance that could coalesce into de facto industry norms.
- Academic studies or neutral assessments measuring the human factors—role satisfaction, decision authority clarity and governance overhead—introduced by agentic AI initiatives.
As agentic AI moves from pilot to production, its fate will rest not on the novelty of autonomous agents but on the resilience of data integration and the robustness of human governance frameworks that safeguard agency, accountability and institutional trust.



