The OpenSearch project has experienced remarkable momentum since joining the Linux Foundation, with active contributors jumping 46% in the past six months and enterprise adoption accelerating across AI and observability use cases.
Mukul Karnik, Director, AWS OpenSearch, revealed these growth metrics during a recent interview, highlighting how the open source search and analytics platform has evolved far beyond its origins to become a critical infrastructure component for modern AI applications.
📹 Going on record for 2026? We're recording the TFiR Prediction Series through mid-February. If you have a bold take on where AI Infrastructure, Cloud Native, or Enterprise IT is heading—we want to hear it. [Reserve your slot
From Foundation Move to Market Leadership
Nearly a year after OpenSearch’s contribution to the Linux Foundation, the project has seen “significant acceleration in momentum,” according to Karnik. The 46% increase in active contributors represents more than just numbers – it reflects what Karnik calls “meaningful contributions” from developers with deep understanding of the codebase.
“You can have people come in and make small tweaks and improvements—which is all good and necessary—but you also need people who have a deep understanding of the codebase and can make meaningful contributions,” Karnik explained.
This quality of engagement is evident in recent enterprise implementations. Uber recently published details about leveraging OpenSearch to power their Uber Eats search platform, making substantial contributions back to the project while reimagining their search infrastructure.
Enterprise Scale and Real-World Impact
The platform’s enterprise credibility is perhaps best demonstrated by SAP’s implementation, which runs “thousands of clusters of OpenSearch” to power logging and observability for their customer base. Adobe has integrated OpenSearch into Acrobat’s AI-powered PDF summarization features, while numerous organizations are adopting it as their vector database for GenAI applications.
“OpenSearch is both a leading search engine and a vector database project,” Karnik noted, positioning it at the intersection of traditional search and emerging AI workloads. This dual capability has proven crucial as organizations build retrieval-augmented generation (RAG) systems and other AI applications requiring sophisticated search functionality.
Major Releases Drive Innovation
The 3.0 release marked what Karnik described as “a leap forward moment” – the biggest launch since the foundation move. Key improvements included enhanced performance, expanded agentic AI capabilities, and strengthened observability features. The subsequent 3.1 release continued this trajectory of rapid innovation.
“We introduced agentic AI capabilities as well as observability capabilities,” Karnik said, emphasizing the platform’s evolution beyond traditional search into areas like agent-based systems and infrastructure monitoring. “So we’re innovating across many different areas.”
AI Integration: Both Platform and Tool
OpenSearch’s AI strategy operates on two levels. First, it serves as infrastructure for AI applications, providing the search and vector database capabilities that power modern AI systems. Second, it increasingly incorporates AI to enhance its own functionality.
“Search fundamentally is changing” in the AI era, Karnik observed. “If you think of OpenSearch as the next generation search platform,” it’s enabling users to ask ambiguous questions and receive insightful answers through the combination of large language models and OpenSearch’s analytical capabilities.
The platform now supports the Model Context Protocol, enabling more sophisticated agentic workflows. This positions OpenSearch not just as a data store, but as an active participant in AI-driven decision making.
Observability and DevOps Transformation
Beyond AI applications, OpenSearch continues strengthening its position in observability and log analytics. The platform’s analytical capabilities, combined with GenAI features, are helping DevOps teams accelerate root cause analysis and issue resolution.
“We are now able to leverage OpenSearch’s analytical capabilities combined with the GenAI capabilities to get to the root cause very quickly,” Karnik explained. This can significantly accelerate debugging and reduce mean time to resolution for production issues.
Looking Ahead: Agentic Memory and Beyond
Future development will focus on agentic memory capabilities – a critical component for AI agents that need to maintain context across interactions. “Agents, in addition to using tools, require memory to maintain context,” Karnik noted. “We are exploring like OpenSearch could be a powerful platform to build that kind of capability, given it’s a search and a vector database.”
This roadmap reflects OpenSearch’s positioning at the forefront of enterprise AI infrastructure, where traditional boundaries between search, analytics, and AI are increasingly blurred.
Community Growth and Sustainability
With over 90 partners and a growing community of core maintainers, OpenSearch is building the sustainable foundation necessary for long-term success. The project’s Technical Steering Committee is driving architectural modernization toward a more cloud-native design – changes that require deep technical expertise and community commitment.
As organizations increasingly adopt AI-first architectures, OpenSearch’s dual role as both search platform and AI infrastructure positions it as essential enterprise technology. The 46% contributor growth suggests the broader tech community recognizes this potential, contributing not just code but strategic direction for the platform’s future.





