Image source: The Motley Fool.
Thursday, Feb. 26, 2026 at 5 p.m. ET
Need a quote from a Motley Fool analyst? Email pr@fool.com
Management confirmed that CRPO surpassed $1 billion for the first time, providing heightened forward revenue visibility. Elastic N.V.'s AI offerings now account for over a quarter of its $100,000+ ACV customer segment, illustrating rapid AI adoption and expansion momentum. Product innovation continued with new AI-centric capabilities—agent builder, technical preview of Elastic Workflows, and integration of high-performing GINA AI models—designed to both attract new enterprise workloads and deepen hybrid flexibility. Integration with major partners, including Dell and NVIDIA, advanced the company's technical moat and streamlined access to GPU-powered AI. Shareholder returns accelerated, with over 60% of the existing repurchase authorization executed by quarter-end. Guidance for both the fiscal fourth quarter and fiscal 2026 was raised, reflecting pipeline strength and sustained sales execution.
Operator: Good afternoon, and welcome to the Elastic N.V. Third Quarter Fiscal 2026 Earnings Results Conference Call. All participants will be in listen-only mode. Should you need assistance, please signal a conference specialist by pressing the star key followed by zero. After today's presentation, there will be an opportunity to ask questions. Please note this event is being recorded. I would now like to turn the conference over to Eric Prengel, Global Vice President of Finance. Please go ahead.
Eric Prengel: Thank you, and good afternoon, and thank you for joining us on today's conference call to discuss Elastic N.V.'s third quarter fiscal 2026 financial results. On the call, we have Ashutosh Kulkarni, Chief Executive Officer, and Navam Welihinda, Chief Financial Officer. Following their prepared remarks, we will take questions. Our press release was issued today after the close of market and is posted on our website. Slides which are supplemental to the call can also be found on the Elastic N.V. Investor Relations website at ir.elastic.co. Our discussion will include forward-looking statements, which may include predictions, estimates, our expectations regarding demand for our products and solutions, and our future revenue and other information.
These forward-looking statements are based on factors currently known to us, speak only as of the date of this call, and are subject to risks and uncertainties that could cause actual results to differ materially. We disclaim any obligation to update or revise these forward-looking statements unless required by law. Please refer to the risks and uncertainties included in the press release that we issued earlier today, included in the slides posted on the Investor Relations website, and those more fully described in our filings with the Securities and Exchange Commission. We will also discuss certain non-GAAP financial measures. Disclosures regarding non-GAAP measures, including reconciliations with the most comparable GAAP measures, can be found in the press release and slides.
Unless specifically noted otherwise, all results and comparisons are on a fiscal year-over-year basis. The webcast replay of this call will be available on our company website under the Investor Relations link. Our fourth quarter fiscal 2026 quiet period begins at the close of business on Thursday, 04/16/2026. We will be participating in the Morgan Stanley Technology, Media and Telecom Conference on March 2. With that, I will turn it over to Ashutosh Kulkarni.
Ashutosh Kulkarni: Thank you, Eric, and good afternoon, everyone. Thank you for joining today's call. Elastic N.V. delivered yet another outstanding quarter, beating the high end of guidance across all key metrics, and showcasing the power of the Elastic N.V. platform in our business model. Sustained platform demand, strong sales execution, and our relentless focus on customers drove Q3 momentum. As LLMs rapidly evolve their capabilities around inference and reasoning, it is becoming increasingly clear that context is the most important ingredient in making these models useful within an enterprise. We continued to see enterprises choose Elastic N.V. with that backdrop in Q3 to power context for their most critical AI needs.
Translating the success to our performance, we achieved 18% total revenue growth and an 18.6% non-GAAP operating margin. Sales-led subscription revenue accelerated to 21% alongside our growing cohort of $100,000 ACV customers, which now exceeds 1,660. Q3 marked our sixth consecutive quarter of strong field execution, driving solid customer commitments and supporting healthy CRPO growth. That execution is also translating into a strong pipeline as we head into Q4. The lifeblood of organizations is the proprietary data that they create, manage, and analyze every day to drive business decisions and operations. This data is massive, often many petabytes in scale, and simply cannot be moved for cost and security reasons outside of the organization's control.
For businesses to use agentic AI, the LLM needs to come to the data. This is where Elastic N.V. comes in, with our ability to help organizations store and manage all of their data in very cost-effective ways and by providing accurate, real-time context to AI by searching through all of this organizational data in real time. Furthermore, Elastic N.V. is capable of doing this consistently across cloud and self-managed environments. This hybrid flexibility allows sensitive data and workloads to remain in their preferred environments, eliminating the need for costly replatforming. This unique flexibility is why we continue to displace legacy vendors and niche cloud-native players alike. And the results are clear.
The number of commitments for over $1 million in annual commitment value signed this quarter grew over 30% compared to the same period last year, driven by new logos and customer expansion. Consolidation and AI are powerful tailwinds. As organizations manage exploding data volumes, they are turning to Elastic N.V. to drive both innovation and efficiency for their search, observability, and security needs. For example, we signed a seven-figure new logo deal with a Fortune 100 insurance institution for Elastic Security. Seeking to modernize their security operations, the company initiated a competitive process to replace a legacy SIEM solution that was plagued by slow query speeds, inefficient data retention, and rigid SOC workflows.
By leveraging features like LogsDB and searchable snapshots, they are consolidating data into a single cyber data lake with integrated AI-powered SIEM workflows, all powered by Elastic N.V. Its capabilities, including AI assistant, attack discovery, and AI-driven orchestration, enable their analysts to achieve markedly faster cybersecurity detection and remediation outcomes while meeting strict regulatory requirements. Another large deal from the quarter, a global leader in data resiliency software chose Elastic Observability to power the monitoring layer for its new cloud offering. As they migrate their vast user base to the cloud, they are leveraging our full observability suite, including AI assistant and LogsDB, to transform from reactive troubleshooting to intelligent semantic-aware analysis.
By integrating OpenTelemetry and our vector search capabilities, the customer is now able to proactively detect anomalies and remediate issues using natural language queries, significantly reducing mean time to resolution. They chose Elastic N.V. over incumbents due to our deep integration flexibility, superior handling of unstructured data, and the ability to provide a single source of truth across the organization. Crucially, as companies navigate their cloud migrations, they require a platform that does not force them to choose between their existing data centers and the cloud. Our asymmetric advantage in supporting modern cloud and hybrid environments drove a significant win with a global financial group.
During the quarter, we closed a seven-figure expansion deal for Elasticsearch which serves as the core of their online banking application for tens of millions of users. They needed a central data repository capable of supporting both cloud and self-managed architectures, allowing them to run mission-critical workloads in their preferred environment without compromising performance. Elastic N.V. succeeded where their existing MongoDB implementation failed to provide the scalable retriever and precision necessary to move beyond simple search into production-grade context engineering. Moving forward, they are integrating semantic search and advanced AI features to further personalize the user experience through faster, more accurate retrieval. Central to these enterprise engagements is the rise of agentic AI.
Customers are moving from passive Q&A to active agents that drive workflows. Precise action requires precise data. The conversation has shifted from which model to use to how to feed it the most accurate context. Enterprises realize that to unlock the value of AI, they must bridge the gap between their LLMs and their proprietary unstructured and structured data. Elastic N.V. makes this AI work. We are the engine that allows enterprises to build production-grade AI systems that are actually worthy of their business. While others offer simple vector databases, we know that vectors alone are not enough.
We deliver the full retrieval toolkit from hybrid search to advanced re-ranking, ensuring that agents have the relevant context they need to take precise actions. This ability to bridge enterprise data to the LLM with our platform is directly translating into expanded AI adoption. New customer commitments with AI continued to grow in Q3, and we now have over 2,700 customers on Elastic Cloud using us as a vector database, with additional customers using us for broader AI capabilities including agent builder and attack discovery, bringing our total count of AI customers to over 3,000. We now have over 470 customers with an ACV of $100,000 or greater using us for AI.
This includes more than 410 using us as a vector database. Cumulatively, AI use cases have now penetrated over a quarter of our $100,000 ACV customer cohort. We are seeing sustained demand from the largest companies in the world alongside interest from the new wave of AI-native companies. During the quarter, we closed multiple new logo and expansion deals with AI-first innovators, validating that our platform is the standard for both established enterprises and disruptors. A leading AI recruiting platform used by large enterprises and startups alike chose Elastic N.V.'s vector database to power their core customer-facing software because our search performance at scale was better than competitors.
An AI-enabled driver and fleet safety company expanded their use of Elasticsearch in Q3 as they scale into new global regions. Elastic N.V. provides the real-time retrieval necessary to power their platform, ensuring they can manage increasing data volumes without sacrificing performance. And a leading AI-native cybersecurity company focused on AI-automated penetration testing has integrated our SIEM solution into their product. Elastic N.V. centralizes all of their logs without complication, allowing them to effortlessly scale through their massive growth trajectory. At the heart of these wins is the performance of our Search AI platform. Beyond just adding features, we are aggressively optimizing our engine, focusing our development efforts on delivering market-leading relevance, speed, and efficiency.
In the last 18 months, we have driven two orders of magnitude less RAM required for vector search through innovations like better binary quantization (BBQ), disk BBQ, and our ACON filtering algorithm, among other things. This investment makes Elasticsearch vector search up to eight times faster than OpenSearch. Our superior performance led to a seven-figure deal with a global heavy equipment manufacturer. The customer continues to migrate mission-critical workloads over to Elastic Cloud from OpenSearch to improve scalability and performance. They are relying on our platform to power their high-speed search for telemetry data collected via the Starlink network. By leveraging LogsDB, they have achieved a significant reduction in cloud costs while managing seven years of historical customer data.
Our focus on performance extends to our partnership with NVIDIA as well, where together we help enterprises deploy AI applications faster without draining IT infrastructure. We recently announced the technical preview of our Elasticsearch GPU plugin for a GPU-accelerated vector database, which allows for 12 times faster indexing. Additionally, the Dell AI data platform, now with NVIDIA and Elastic N.V., delivers a tightly integrated AI stack that streamlines the ability to build, deploy, and scale AI. By making Elasticsearch a core component of the Dell and NVIDIA AI factories, we are meeting the critical demand for building AI on customer-controlled infrastructure. As we deepen these technical advantages, we strengthen our technical moat while removing friction from scaling AI.
This quarter, we reached several product milestones designed to simplify the path from data to action for our customers. We are providing an end-to-end framework for building the next generation of intelligent applications. First, we officially launched the general availability of agent builder. Agent builder allows developers to build secure, context-driven AI agents in minutes. Unlike consumer apps that surf the web, our focus is on internal business applications using company data. We piloted agent builder with a Global 100 financial group to investigate and troubleshoot its production infrastructure, demonstrating an order of magnitude improvement in performance for complex issues and democratizing the specialized expertise necessary for rapid troubleshooting.
An international entertainment and media company created a chat interface for customer interactions. They found the agent builder results to be significantly more reliable and accurate than the other LLM-centric approaches they had tried. Building an agent is only half the battle. The other half is ensuring that agent has the most relevant information at its fingertips. This quarter, we expanded our Elastic Inference service to include GINA AI's multilingual re-ranking models. GINA AI delivers a best-in-class model for search accuracy with GINA v3, currently the number one re-ranker in its model size category on the MTEB English retrieval benchmark, a gold standard for search and RAG relevance.
GINA AI's v5 nano and v5 small models continue to outpace peers as well, scoring high in retrieval, re-ranking, and other tasks. By making these models available natively, we are allowing our customers to tune their AI applications for maximum precision and recall. Re-ranking is the critical next step in a context engineering pipeline that ensures the most relevant data is presented to the LLM. GINA's state-of-the-art models deliver superior performance across over 80 languages. While AI provides the reasoning, enterprises still require the reliability of rule-based automation for critical business tasks. This is why we introduced Elastic Workflows in technical preview.
Workflows add automation capability directly into our platform, allowing agents to orchestrate actions across internal and external systems like Slack or ServiceNow. It moves Elastic N.V. from being a search box to a complete system of action. Finally, we are delivering on our promise of hybrid flexibility with Cloud Connect for self-managed customers. We recognize that many of our largest customers, particularly in financial services and government, maintain data on premises for regulatory or sovereignty reasons. However, procuring and managing GPU hardware for AI is a massive hurdle for these teams. Cloud Connect allows customers to keep their data local while securely bursting to Elastic Cloud to leverage NVIDIA GPUs for high-performance inference.
This ability to bridge modern AI capabilities with rigorous enterprise requirements is exactly why we are winning large-scale displacements against legacy providers. As organizations prioritize both innovation and operational efficiency, they are moving away from fragmented legacy tools in favor of Elastic N.V.'s unified Search AI platform. The results of this quarter—accelerating growth, large deal momentum, and major competitive displacements—confirm that our strategy is resonating and that we are winning the race to become the essential infrastructure for the next generation of AI-powered businesses. I want to thank our customers for their partnership, our shareholders for their trust, and most importantly, our employees for their tireless spirit of innovation.
With that, I will turn the call over to Navam to review our financial results in more detail.
Navam Welihinda: Thank you, Ash. Good afternoon, everyone. We delivered yet another outstanding quarter. We outperformed the high end of revenue and profitability guidance ranges, driven by another quarter of consistent execution, strong consumption, and strong customer commitments across search, security, and observability. The momentum in our performance throughout this fiscal year is a testament to our team's ability to deliver rapid innovation and sales execution consistently quarter over quarter. The ongoing market demand we see is translating to total revenue growth, sales-led subscription revenue growth, and healthy increases in pipeline generation to support our future growth. These factors together underscore our increasingly strategic value as a critical data platform in the age of AI.
Our total revenue in the third quarter was $450 million, representing growth of approximately 18% as reported and 16% on a constant currency basis. Sales-led subscription revenue in the third quarter was $376 million, growing 21% as reported and 19% on a constant currency basis. We saw commitment contribution from both our self-managed and cloud offerings, and aggregate consumption trends in the third quarter remained strong. Our current remaining performance obligations, or CRPO, which is a portion of RPO that we expect to recognize as revenue within the next 12 months, crossed the $1 billion mark for the first time in Q3.
CRPO accelerated to approximately $1.06 billion, growing 19% as reported and 15% on a constant currency basis, providing a clear view into the revenue we will recognize in the next 12 months, giving us visibility and confidence in our business. As Ash mentioned, we saw deal momentum continue in Q3. This quarter's strength was balanced across all geographies, and we continue to see customers make multiyear commitments this quarter, which serves as a clear indicator of how our customers view the Elastic N.V. platform as a critical foundational element in their long-term data architectures. The positive momentum was reflected in our RPO, which saw strong growth of 22% in the quarter as reported and 18% on a constant currency basis.
Our deal momentum is also evident in the growth of the count of customers with over $100,000 in annual contract value. We ended the third quarter with over 1,660 customers with ACV of more than $100,000, growing 14%. Quarter over quarter, we added approximately 60 net new $100,000 ACV customers. We saw strong field execution and healthy growth across our solutions, where Search continues to see ongoing momentum from AI. This demand is benefiting both cloud and self-managed, where both form factors are relevant for AI use cases. We continue to see customers taking a self-managed license and deploying Elastic N.V. into their own modern cloud and hybrid environments.
The demand reflects customers' preference for Elastic N.V., which uniquely provides the necessary control and cost efficiency for AI initiatives. AI also continues to be a powerful catalyst for customer expansion. Twenty-eight percent of our greater than $100,000 cohort now utilizes Elastic N.V. for AI, which includes incremental AI capabilities like attack discovery and agent builder. Today, we are still in the early stages of expansion, and we see considerable opportunity for ongoing upside for both new and existing customers to accelerate their AI adoption in the years ahead, particularly as they scale into and within our $100,000 ACV cohort. Now turning to third quarter margins and profitability. I will discuss all measures on a non-GAAP basis.
Our commitment to balancing growth with disciplined spending translated into robust operating leverage and strong bottom-line results. We continue to focus on costs and efficiency in our business. We recorded subscription gross margins of 82% and total gross margins of 78%, delivering an operating margin of 18.6%. The outperformance on Q3 operating margin was the result of our strong revenue performance, the sustained leverage in our model, as well as some Q3 expenses moving into Q4. Due to this outperformance, we now expect to see our full-year margins come in slightly ahead of what we previously anticipated, with updated FY 2026 operating margin guidance now at 16.3%.
Regarding cash flow, adjusted free cash flow was approximately $54 million in Q3, representing a margin of approximately 12%. Our cash flows are expected to fluctuate on a quarterly basis based on the timing of bookings and collections related to the enterprise booking seasonality, so we continue to manage cash flow on a full-year basis. For fiscal 2026, we do not see any change in our full-year outlook, where we continue to expect to sustain the level of adjusted free cash flow margins that we achieved in fiscal 2025. We have made significant progress on the $500 million share repurchase program that we announced in October.
During the third quarter, we returned approximately $186 million to shareholders, representing purchases of approximately 2,400,000 shares. Cumulatively, we have repurchased 3,800,000 shares. I mentioned at our Financial Analyst Day in October that we expect to use more than 50% of the $500 million authorized amount in fiscal 2026. We have already exceeded this goal. As of the end of Q3, we have completed 60% of our repurchase program, and we are continuing our repurchase program here in Q4. Let's move to our outlook for the fourth quarter and the remainder of fiscal 2026.
For Q4 2026, we expect total revenue in the range of $445 million to $447 million, representing 15% growth at the midpoint, or 13% constant currency growth at the midpoint. We expect sales-led subscription revenue in the range of $371 million to $373 million, representing 18% growth at the midpoint, or 15% constant currency growth at the midpoint. We expect non-GAAP operating margins to be approximately 14.5%. We expect non-GAAP diluted earnings per share in the range of $0.55 to $0.57, using between 105,500,000 and 106,500,000 diluted weighted average ordinary shares outstanding. Based on our fourth quarter guidance, we are raising our full-year total revenue and sales-led subscription revenue targets as well.
We expect total revenue in the range of $1.734 billion to $1.736 billion, representing approximately 17% growth at the midpoint, or 15% constant currency growth at the midpoint. We expect sales-led subscription revenue in the range of $1.434 billion to $1.436 billion, representing 20% growth at the midpoint, or 18% constant currency growth at the midpoint. We expect non-GAAP operating margin for full fiscal 2026 to be approximately 16.3%. We expect non-GAAP diluted earnings per share in the range of $2.50 to $2.54, using between 107,000,000 and 108,000,000 diluted weighted average ordinary shares outstanding. A few other financial modeling points to keep in mind. The diluted weighted average shares outstanding reflect only share buybacks completed as of 01/31/2026.
As you consider comparing sequential quarters, keep in mind that Q4 has three fewer days than we had in each of the first three quarters of the year, which creates a sequential headwind to revenue, which we have accounted for in our guidance. Also, as is typical with prior Q4 periods, we expect to see seasonally higher expenses related to the timing of employee benefit costs. These expenses were already part of the guidance that we had initially laid out for the year. As in past years, we finalized our plans for the upcoming fiscal year during the fourth quarter, and we will provide our initial FY 2027 guide during our earnings call in May.
In summary, Q3 was another very strong quarter at Elastic N.V. Consistent sales execution throughout FY 2026 continues to drive our sales-led subscription revenue growth expectations higher for the year, validating the durability of this business motion. As I said last quarter, while quarterly revenue can naturally vary in a consumption model, our strong customer commitments drive strong annual growth. Fueled by a highly differentiated platform and the expanding value we deliver to our customers, we remain on track to achieve our medium-term targets for both sales-led subscription revenue growth and adjusted free cash flow. Looking forward, we are confident in our ability to continue to drive profitable growth.
We are the critical technology that accelerates data discovery, secures infrastructure, and maximizes application performance. With that, we will now open for questions.
Operator: We will now begin the question and answer session. If you are using a speakerphone, please pick up your handset before pressing the keys. Our first question today is from Sanjit Singh with Morgan Stanley. Please go ahead.
Sanjit Singh: Thank you for taking the question, and congrats on the stability that we are seeing across the business. Navam, I wanted to go back to some of the themes on the Investor Day a couple of months ago. There was a data point that you provided around the AI-native customers or the AI customers being a relatively small amount of the customers in fiscal year 2024, but driving an outsized degree of expansion—that sort of year one to year two expansion.
So the gist of this question is that as you know, we get to, like, 25% of penetration of your $100,000 customer cohort, is there an opportunity here for growth to not just be stable, but actually to accelerate on a more sustained basis as we hit those critical tipping points, if you will?
Navam Welihinda: Yeah, thanks for the question, Sanjit. The trends that we laid out during the Financial Analyst Day for the generative AI cohort, they remain the same. So we continue to perform well, and we are seeing stronger growth on those generative AI cohorts today as it was when we disclosed it to you during Financial Analyst Day. So we are seeing these tailwinds right now, and we are seeing more of our customers reach the $100,000 mark. Now remember that each of these customers in the $100,000 mark are also early in their journey, so there is this other dimension of additional penetration and maturation in their own AI journey, which will drive faster growth as well.
So we are seeing the tailwinds right now. We have seen tailwinds that average to 5%, but there are obviously some customers that have higher growth than that. And to answer your question, yeah, absolutely. There is a possibility—the art of the possible is there for us to actually accelerate beyond that 5% that we laid out during Financial Analyst Day. And the trends remain positive.
Ashutosh Kulkarni: And, Sanjit, just to add to that, that is exactly why we are so focused on the penetration of AI within our customer base. And as these customers—you know, right now, every quarter, you are seeing us increase the penetration—the penetration, you know, initially starts with them using us in some small way. And as that usage grows, as you rightly pointed out, that is going to add to the consumption. It is going to add to the overall revenue. And that is going to show in the continued strength and acceleration of the business.
Sanjit Singh: Understood. And, Ash, maybe a question for you. You made the point in your script about vector search and vector databases are not enough in terms of building a resilient and powerful AI application. And I think a lot of people would agree with that statement. So when we brand the company as a context engine, what are the core pieces that are mandatory to secure your status as the leading provider of context for AI applications?
Ashutosh Kulkarni: That is a great question. I think the most important thing to keep in mind is, you know, context is going to change from task to task. And so the data platform, the context engineering platform that you provide, needs to be able to do a whole bunch of things altogether in a very consistent way. The first is the ability to bring in any and all kinds of data. And as you know, you know, we have some unique capabilities in our ability to bring in, you know, not just structured information, but also unstructured, really, really messy information.
The second is to then take that data and convert it into vectors for vector search, which is a very powerful technique, especially in the AI world, for semantic search, but then also to be able to mix it with hybrid search techniques. That includes textual search and then being able to re-rank against multiple techniques to get the most accurate context. The GINA AI embedding models, the GINA AI re-ranker models, those are a key part of that overall platform infrastructure. On top of it, then you need something that will allow you to assemble agents using all of these capabilities, and that is what agent builder was all about.
As you know, it is a relatively new feature from us and a relatively new capability. We are seeing great traction and adoption within our customer base. On top of it, you need workflows because agents are not just about chat anymore. They are not just about conversations. They are about taking precise actions. And that is where the workflow functionality that we released becomes really important. And lastly, the ability to monitor all of this, and that is where our LLM observability functionality becomes key. So we believe that it is all of these capabilities, Sanjit, that taken together make the platform a very compelling platform for context engineering.
On top of that, we have also added our Elastic Inference service, so you do not need to bring your own LLM. We can help you proxy to any LLM of choice that you might want to use. We integrate with pretty much all of them.
Sanjit Singh: Appreciate the color, Raj. Thank you.
Operator: The next question is Rob Owens with Piper Sandler. Please go ahead.
Rob Owens: Great. Thank you very much for taking my question. I will apologize upfront for the flurry of questions in one here, but I will keep it to one but maybe three parts. I really want to focus on the outperformance in other subscription, and I understand you are going to meet customers where they want to buy. So I guess upfront, was some of that strength potentially push-outs that you saw in the prior quarter? Then if I look at your sales-led subscription forecast for Q4 and the fact it is down quarter over quarter, which you have not seen historically, it is usually a little bit up.
Is that really a function of the strength you saw here in the January quarter or something else to be read into that? And lastly, when we think about monetization of self-managed versus cloud customers and your ability to expand them over the coming years, can you maybe articulate the difference between the two, if there is much there? So again, apologize for the three questions, but hopefully they are brief answers.
Ashutosh Kulkarni: Yeah. Rob, let me start. This is Ash. Let me start and then pass it on to Navam. You know, in terms of our strength in self-managed, this is not just about push-outs or anything of that sort. We are continuing to see a lot of strength in our self-managed business. At the end of the day, what we are seeing now, especially with AI, is a lot of customers are applying AI on data that they consider to be extremely critical, extremely sensitive. This is not just with government customers. This is also in other regulated industries.
And for that reason, they are choosing, or they are preferring, to keep the data where it is within their control, within their environment. And that does not always mean in their own data centers. It might also mean within their own cloud VPCs. And we give them the flexibility to be able to do that. So these are modern workloads that continue to grow. As their usage of AI grows, we are going to continue to benefit from it, which is why we believe it is really important to not just look at cloud, but to look at the whole picture and take into account the strong growth that we are seeing even on self-managed.
And I will let Navam address the other questions.
Navam Welihinda: Yeah. Rob, I will address your quarterly sequential question. So, you know, overall, I will start with how the business is doing. We are continuing to execute very well on the sales-led motion. This is another quarter of good execution from the sales side, and we saw that play out in our CRPO and RPO numbers accelerating as well. If you are looking at commitments, we are seeing good commitment volume, and there is no deceleration on that. And on top of that, the pipeline is very healthy and growing each quarter. So overall, from a business perspective, very happy with where the quarter turned out and very positive about the future quarters as well.
So that leads us to the guide. When you think about the guide, we always guide with an appropriate amount of prudence on what we can achieve and outperform every quarter. So when you look at historical numbers versus actuals and guidance, you are comparing an actual number against a guidance number, and the guidance number has risk incorporated into that forward-looking projection. So I will first point to that.
And the second point I would make is that the fourth quarter has three fewer days, which translates to a $14 million to $15 million headwind—a 3% headwind for us on a revenue basis—because there are just fewer days of revenue to recognize, and all of that is incorporated in the guide. And if you look at last year's Q4 guides or Q4 guides in the past, there have been occasions where we have guided lower than the current quarter. So just keep that in mind. We continue to keep well on track with achieving our midterm targets, and we feel very positive about the strength of the business itself.
Rob Owens: Great. Thank you.
Operator: The next question is from Matt Hedberg with RBC. Please go ahead.
Matt Hedberg: Great. Thanks for taking my question, guys. Ash, I wanted to ask you about AI. And obviously, we have all seen pressure in the software market, and I appreciate your comments at the start of the call—that was really helpful to kind of get your perspective on AI. And it seems like there is a lot of great momentum from a customer perspective. I guess my question is, when we are looking at these frontier models, do you see them as future competition or more of a partnership opportunity?
Ashutosh Kulkarni: You know, really, we do not—in our opinion, AI does not displace us. It really depends on us, because if you think about these frontier models, they are amazing reasoning engines. The way I think about them is they are going to be the operating systems of tomorrow. But just as operating systems today also require data systems to feed appropriate data and context to these operating systems to actually, you know, build applications with, you are going to need the same thing going forward.
And our role in this whole ecosystem is to make sure that we can very quickly, in real time, across all of the petabytes of data that every organization holds, give the right context to these LLMs so they can do their job. And that is the reason why I believe that, you know, in the world of tomorrow, you are going to have agents talking to each other. You are going to have agents that you build with Elastic N.V. agent builder that are talking to Cloud Co-Work, that are talking to, you know, things that you build with OpenAI Frontier, and we already support the MCP protocols, the A2A protocols, that allow for that kind of communication.
So this is a world where we feel that, you know, the fact that we have this tremendous position—capabilities with our vector database, the capabilities with our entire context engineering platform—to become a critical part of the infrastructure going forward. And we are already partnering with hyperscalers and already integrate with all of these frontier-class models today.
Matt Hedberg: That is a great perspective. Thank you for that. And then maybe just one quick follow-up about Elastic N.V. internally using AI. How are you seeing some of the tangible benefits, and how might that impact headcount in the future?
Ashutosh Kulkarni: Yes. Look, we are all in on AI, not just in terms of what we are doing externally in terms of providing the platform that we are building, but also in terms of how we are using AI internally. Just to give you some context on this, a couple of years ago, we built out our first agent, our first support agent within the company, and that has been in production for a long time now. It is what our customers first hit when they have support questions.
And the amount of queries it is able to answer and the number of support tickets it is able to deflect has not only improved the overall performance, the overall experience for our customers when they come to us for support, but it has also significantly reduced the demand on headcount from our side. So in the last two years, even as our business has been growing—and as you can imagine, typically support workloads grow with the business—we have been able to manage that workload growth without adding any headcount to that support team. You know, in other parts of the business, whether it is in HR, whether it is in finance, in legal, we are heavily using AI tools.
Some of these are built on our stack. Some of them might be external products that we are leveraging. And even in engineering, we are finding tremendous value in using multiple different, you know, code generation tools that we use within the company. So, overall, we believe that this is going to definitely help us not just accelerate the pace of innovation—we are already seeing that now—but also the productivity and improve the overall efficiency of the business. And that is what is exciting about this. You know, we are able to help our customers with this, but we are also able to benefit from it ourselves.
Matt Hedberg: Got it. Thanks.
Operator: The next question is from Brian Essex with JPMorgan. Please go ahead.
Brian Lee Essex: Hi, good afternoon. Thank you for taking the question. I appreciate your question with regard to your vector database capabilities and context engineering platform. I guess as you look at the changing landscape and you look at different approaches, different ways to think about things, are there things to consider about the platform and its ability to adhere to some of those approaches? Like, for example, the page index approach to RAG. If they solve the cost and latency issues involved with that approach, are you well positioned to benefit from something like that and pivot with your approach?
Ashutosh Kulkarni: Yeah. Look, RAG—retrieval augmented generation itself—has progressed a lot since the last several years when it was first introduced as a concept. But fundamentally, you know, this comes down to finding the most appropriate context that is relevant for the LLM to do its job. Sometimes that requires you to understand, you know, specific data relationships that might exist. Sometimes it requires you to just search through all of your data. Sometimes it requires you to understand, you know, specific things like preferences and so on you might have captured in other data systems. And it is an amalgamation of all of this.
And as RAG continues to evolve, as these techniques become more and more sophisticated, we are actually on the leading front of capturing, you know, more than one single technique into our platform. We were one of the first to adopt hybrid search, and we were the first to talk about it. Since then, we have continued with that kind of momentum. So, absolutely, I feel very, very confident that we are going to be on the bleeding edge. You know, this is at the end of the day, what Elastic N.V. was born to do. We have always been in the business of relevance. Without relevance, you do not get good search.
Without relevance, you do not get good, accurate AI.
Brian Lee Essex: Great. That is super helpful. Maybe just one quick follow-up. Any traction from the recent CISA win that you had? Or any Fed agencies leveraging that for assumed referenceability? And are you seeing better activity on the back of that win?
Ashutosh Kulkarni: It has been a great success. Yeah. Thank you. It has been a great success for us already. I think we mentioned it in our press release as well, that, you know, SIEM as a service with CISA continues to grow. So I would expect that CISA win—and we saw additional agencies coming on board even in Q3—to be just the beginning of, you know, multiple agencies coming on to that service over the next several quarters. Fundamentally, CISA is considered to be, you know, the primary agency responsible for cybersecurity in the government in the United States. And that kind of endorsement is something that goes a long way. So it is a very exciting win.
Like I said, you know, we are going to benefit from it for many quarters and many years to come.
Brian Lee Essex: Great. Super helpful. Thank you very much, and congrats.
Navam Welihinda: Thank you.
Operator: The next question is from Brent Thill with Jefferies. Please go ahead.
Brent Thill: Just on the CRPO—15% constant currency this quarter, 15% last quarter—I guess, I mean, mid-teens growth, but I think everyone is asking, why are we not seeing a faster inflection? I know you have a true north of 20. It seems like the numbers support that you can accelerate to 20–20. But just curious kind of how you bridge to 20–20 and perhaps why maybe you are not seeing a little bit stronger AI tailwind in the near term?
Navam Welihinda: Yes. Thanks for the question, Brent. So I will start. CRPO has crossed over $1 billion. We are at 19% growth right now. RPO is at 22% growth. That is the best we have seen in two years, and we are very happy with the progress that we are making. If you just look at the absolute dollar additions that we added in the quarter, it is progressing very, very well.
So that is all pointing to the core things that are driving that CRPO growth, which are strong customer commitments, which now we have been talking about for a couple of quarters now, and it has been yet another quarter of very good sales execution leading to strong customer commitments. So the AI tailwinds we talked about during Financial Analyst Day, we are seeing them right now, and they are continuing to grow as we see more and more of the $100,000 customers adopt AI workloads from us. So we think that there is a good strong trajectory from this point ahead as we see more AI penetration among our $100,000 customer base.
Ashutosh Kulkarni: The other thing that I will say to this, Brent, is that if you look at the full-year guide for sales-led subscription revenue, you can see that the strength in our business continues. And, look, the midterm guide that we laid out is not the place where we end up. That is the place that, you know, we believe we can go beyond. If you remember, we talked about 20-plus. And, really, that is the way we see it. So as more and more customers adopt our AI functionality, given the fact that those cohorts tend to grow and expand faster, we feel very, very good about how we are tracking to that midterm.
And we feel very good about the fact that as that traction continues, we feel good about even exceeding what we have talked about in the past.
Operator: The next question is from Howard Ma with Guggenheim. Please go ahead.
Howard Ma: Hey, great. Thanks. I wanted to ask about cloud, and I guess this one is for Navam. You know, I want to throw out a caveat first, which is that I appreciate your deployment agnosticism and fewer days in Q4. When I look at cloud revenue in Q4 versus Q3 in FY 2022 and earlier, there was more of a sequential step-up than in FY 2023 through FY 2025, which were obviously impacted by, you know, industry-wide cloud optimizations. Also, Elastic N.V. had company-specific go-to-market issues.
Now that the go-to-market execution has improved significantly, and given the visibility that you now have into how large customers ramp consumption relative to the commits—and that includes some of the $10 million-plus TCV contracts that you signed last quarter—the question is, is there any reason why the sequential cloud growth in Q4 would not be more in line with the earlier years?
Navam Welihinda: So I will start off with what I always start off on, which is sales-led subscription revenue growth is the right metric for you to focus on measuring us as a barometer of the success of the company. And I talked about this during our prepared remarks as well. There are multiple examples, including this quarter, of AI workloads being sold as self-managed and deployed either in the customer's cloud or in their hybrid environments. So sales-led subscription grew a healthy 21% this year. If you look at just cloud—and the number there again is what is the sales-led cloud number—that grew 27% year over year this quarter.
So we are seeing very good traction on the metric that matters to us, which is sales-led subscription revenue. Also, the annual cloud number this quarter was very good as well at 27%. For the forward quarter, number one, you have three fewer days, so that is three fewer days to focus on. The forward quarter is a risk-adjusted number, so you cannot really compare an actual to a guidance number. But the point I would like to make is that we are seeing very strong commitments and very strong performance on sales-led.
Howard Ma: Okay. Thanks.
Operator: The next question is from Ryan MacWilliams with Wells Fargo. Please go ahead.
Dushan: Hi, team. This is Dushan on for Ryan McWilliams. I wanted to ask—it really seems that, on some of the work we have been doing, the number of agents and AI services in production have really increased over the past couple of months. And I want to hear from you what you are seeing within your customers. Are you seeing the types of AI use cases broaden out compared to what you were seeing maybe two quarters ago? How is that impacting usage and spend amongst those customers?
Ashutosh Kulkarni: Yes. We are seeing the usage broaden out in the sense that we are seeing more and more variety of use cases that involve AI. You know, eight quarters ago, the bulk of what we were seeing was only around vector databases, vector search, hybrid search, semantic search. It was mostly around the chat-style interface kind of work. Now we are seeing, you know, agentic workflows being put together not just around what you would typically think of as search-related workflows—around security workflows, around observability workflows. That was the reason why we gave the stat around our total count of customers using us for various AI use cases beyond just vector database.
And, you know, that includes things like agent builder. That includes things like attack discovery. In these kinds of scenarios, people are trying to automate their SOC workflows, their cybersecurity workflows for detection, for remediation; they are trying to do the same for SRE workflows around observability. So the variety of use cases is growing. And as that grows, you know, we see an opportunity not just in our core search business, but also in the work that we are doing in security and observability.
Dushan: Thanks, guys. Appreciate it.
Operator: The next question is from Miller Jump with Truist Securities. Please go ahead.
Miller Jump: Hey, great. Thank you for taking the question, and congrats on the sales momentum. Ash, you mentioned a MongoDB competitive win in the prepared remarks. We had not heard as much about this head-to-head between the two of you until fairly recently. So are you seeing MongoDB increasingly in bake-offs as customers look to build AI apps, or is that more of a one-off?
Ashutosh Kulkarni: No. This was a situation where the customer had started to use, you know, that technology for a basic search application. They had some issues scaling it. And as they were trying to build a more scalable solution, especially for hybrid search, they realized that they needed something that could perform, and that was the customer win that I talked about. You know, at the end of the day, where we tend to typically play is in the area of unstructured data. We do not tend to see them as much. But from time to time, you know, you do see these kinds of situations.
Miller Jump: Thanks. And if I could just ask a quick follow-up for Navam. As large deals are becoming more of a contributor in your go-to-market strategy moving up market, can you just remind us how you are handling those large deals in your guidance process? Any considerations around seasonality there? Thanks.
Navam Welihinda: Yeah. Seasonality-wise, I think it just follows the normal, typical enterprise seasonality pattern, where they end up being more tail-end weighted in Q3 and Q4. But, you know, we talked about large deals in the last quarter. They happen every quarter. It is just the volume of bookings are bigger towards the tail end of the year. In terms of how we handle it, I think that this is a natural byproduct of just being successful with our customers, particularly the larger customers within the G2K. So we welcome it.
When we look at our guidance and what we expect the full year to be, we naturally take a haircut on specific deals that could move from one quarter to another. So that is how we incorporate it into our guidance—risk-adjusted, not actually counting on everything going our way.
Miller Jump: Thanks very much.
Operator: The next question is from Koji Ikeda with Bank of America. Please go ahead.
George McGuin: Hi. This is George McGuin on for Koji. I appreciate you guys taking our questions today. Wanted to ask, just in the conversations that you guys have with customers and their strategy around adopting AI, how would you say that the tone and the conversations differ versus a year ago? And what kind of inning are they in today versus maybe a year ago in their adoption journey with Elastic N.V.? Thank you.
Ashutosh Kulkarni: The general tone is definitely one of greater enthusiasm for AI. I think there have been enough proof points now for AI helping in all kinds of use cases, whether it be around, you know, code development, whether it be around customer support, and legal e-discovery—like, lots and lots of use cases across all functions. And so we are seeing the conversations be less evangelism and more about helping them put together, you know, these kinds of sophisticated agentic applications. So there has definitely been maturity. In terms of the total number of, you know, these agents that people have within their organization, that number is still in the early days.
Like, if you think about the total number of business processes and workflows that can be automated by AI, I think you have to be realistic that we are still in the early days because, you know, AI just is a pretty powerful and transformative capability. And, you know, what you can do with these LLMs in terms of reasoning can be applied to many, many different functions and different work processes. So we believe that the opportunity is still, you know, very significant and still ahead of us.
Operator: The next question is from Mike Cikos with Needham. Please go ahead.
Matt Cletri: Hey, guys. This is Matt Cletri on for Michael Joseph Cikos over at Needham. Thanks for taking our questions. With all the advancements you are making to search with things like the GINA re-ranking models, are you able to charge customers more, or is the improved speed and accuracy more of an acquisition vehicle?
Ashutosh Kulkarni: So we do charge in terms of consumption. Right? So we have a consumption model, as you know. So pretty much everything that you do on our platform, it is metered, effectively based on compute, based on storage, and so on. And for anything that is LLM- or model-related, it is based on tokens. And all of our pricing is public on our pricing pages. But, yes, with these newer models, we are monetizing everything. And as the usage continues to grow, as customers do more and more on our platform, that is what drives revenue for us.
Matt Cletri: Got it. Very helpful. Thanks. And then, maybe just taking a different slice of the guidance question here. So you beat on the 3Q guide in constant currency, and then you raised the constant currency guide for sales-led subscription revenue, but you left constant currency unchanged for the full-year guide. And I can appreciate the three fewer days and the risk-adjusted guidance, but that would have been baked into the prior guide. Can you just help walk through the mechanics there of why that would not have increased?
Navam Welihinda: Yes. I mean, it is quite simple. The number that we care about is sales-led subscription revenue. We handily beat that number this quarter, and we raised more than we beat. That is a reaction to what we think is happening with the business and the positive momentum that we are seeing on the sales line. So overall, we are not thinking about it too much more than we feel good about the forward momentum of sales-led subscription revenue, and we beat the number and we are raising more than we beat.
Matt Cletri: Got it. Thank you.
Operator: The next question is from Eric Heath with KeyBanc Capital Markets. Please go ahead. Mr. Heath, your line is open on our end; perhaps it is muted on yours. Showing no further questions, this concludes our question and answer session. I would like to turn the conference back over to Ashutosh Kulkarni for any closing remarks.
Ashutosh Kulkarni: Thank you all for joining us today. We here at Elastic N.V. are very proud of our results and excited about the opportunity ahead. Thank you.
Operator: The conference has now concluded. Thank you for attending today's presentation. You may now disconnect.
Before you buy stock in Elastic, consider this:
The Motley Fool Stock Advisor analyst team just identified what they believe are the 10 best stocks for investors to buy now… and Elastic wasn’t one of them. The 10 stocks that made the cut could produce monster returns in the coming years.
Consider when Netflix made this list on December 17, 2004... if you invested $1,000 at the time of our recommendation, you’d have $445,995!* Or when Nvidia made this list on April 15, 2005... if you invested $1,000 at the time of our recommendation, you’d have $1,198,823!*
Now, it’s worth noting Stock Advisor’s total average return is 927% — a market-crushing outperformance compared to 194% for the S&P 500. Don't miss the latest top 10 list, available with Stock Advisor, and join an investing community built by individual investors for individual investors.
See the 10 stocks »
*Stock Advisor returns as of February 26, 2026.
This article is a transcript of this conference call produced for The Motley Fool. While we strive for our Foolish Best, there may be errors, omissions, or inaccuracies in this transcript. Parts of this article were created using Large Language Models (LLMs) based on The Motley Fool's insights and investing approach. It has been reviewed by our AI quality control systems. Since LLMs cannot (currently) own stocks, it has no positions in any of the stocks mentioned. As with all our articles, The Motley Fool does not assume any responsibility for your use of this content, and we strongly encourage you to do your own research, including listening to the call yourself and reading the company's SEC filings. Please see our Terms and Conditions for additional details, including our Obligatory Capitalized Disclaimers of Liability.
The Motley Fool recommends Elastic. The Motley Fool has a disclosure policy.