Open-source technology is revolutionizing the way AI agents are developed and deployed. AWS is at the forefront of this movement, championing the Model Context Protocol (MCP) as a cornerstone for agentic AI. With MCP’s transition to the Linux Foundation, its path is now firmly in the hands of a global community, ensuring continuous innovation free from single-vendor dominance.
How AWS Is Shaping MCP’s Evolution
AWS has been instrumental in integrating MCP into its AI offerings. The company actively supports open-source MCP servers, enabling customers to deploy agentic solutions seamlessly via platforms like Amazon Bedrock AgentCore.
AWS’s dual commitment, to advancing the MCP specification and enhancing its implementations, positions MCP as a vital connector between AI agents and the tools, data, and services they rely on.
- Open governance through the Linux Foundation guarantees neutrality and fosters collaboration.
- AWS maintains multiple MCP server implementations, promoting integration across its AI suite.
- Customer use cases span from sophisticated code assistants to custom, serverless agent deployments.
Innovating with Tasks and Elicitations
MCP Tasks: Enabling Reliable Asynchronous Operations
Agentic AI often requires managing long-running operations that can be disrupted by connectivity issues. AWS addressed this by introducing MCP Tasks, which allow agents to initiate operations asynchronously. Using task handles, clients can check status or retrieve results even after interruptions, echoing the reliability found in AWS services like SQS and Step Functions.
- Boosts reliability by supporting operation continuity across disruptions.
- Builds a foundation for advanced features like webhooks and multi-step workflows.
- Facilitates seamless integration with existing AWS asynchronous processes.
Elicitations: Smarter Human-in-the-Loop Workflows
Elicitations take agentic AI a step further by allowing agents to request missing user information only when needed. This streamlines user interaction and creates more natural, efficient workflows, especially for chatbots and automation tools.
- Reduces up-front complexity by prompting for required data only when necessary.
- Enables more contextual, conversational agent interactions.
- Widely adopted in leading MCP implementations, enhancing agentic applications.
The Growing MCP Ecosystem Within AWS
New MCP-powered features are rapidly expanding across AWS. The AWS Knowledge MCP server now makes documentation and over 15,000 APIs accessible to AI assistants.
Meanwhile, the AWS MCP Server (currently in preview) empowers agents to automate complex workflows using standardized procedures, unlocking new areas of productivity.
- Deep MCP support is found in developer tools like Kiro and Amazon Q Developer.
- Amazon Bedrock AgentCore delivers managed, serverless hosting for MCP servers.
- Customers can build AI solutions tailored to their unique data and business processes.
Collaborative Progress and Future Outlook
MCP’s ongoing evolution under the Linux Foundation signals a new era of open, collaborative AI development. AWS’s expanding contributions and commitment to open standards invite partners and developers to join in shaping the future of agentic AI. The result: a robust ecosystem, open governance, and innovative workflows that redefine what’s possible for intelligent agents in the cloud.
Conclusion
By investing in MCP and championing open-source stewardship, AWS sets a high bar for collaboration and advancement in AI. As MCP matures, organizations can expect richer workflows, seamless AWS integration, and a dynamic, community-driven ecosystem.
Source: AWS Open Source Blog

AWS Champions Open Innovation for Agentic AI: The Future of MCP