-

Elasticsearch Open Inference API Extends Support for Hugging Face Models with Semantic Text

Applications using Hugging Face embeddings on Elasticsearch now benefit from native chunking

SAN FRANCISCO--(BUSINESS WIRE)--Elastic (NYSE: ESTC), the Search AI Company, today announced the Elasticsearch Open Inference API now supports Hugging Face models with native chunking through the integration of the semantic_text field. Developers can now quickly ship generative AI (GenAI) applications without the burden of writing custom chunking logic, leveraging the Elasticsearch Open Inference API integration with Hugging Face Inference Endpoints.

“Combining Hugging Face’s embeddings with Elastic’s retrieval relevance tools helps users gain better insights and improve search functionality,” said Jeff Boudier, head of product at Hugging Face. “Hugging Face makes it easy for developers to build their own AI. With this integration, developers get a complete solution to leverage the best open models for semantic search, hosted on Hugging Face multi-cloud GPU infrastructure, to build semantic search experiences in Elasticsearch without worrying about storing or chunking embeddings.”

“Developers are at the heart of our business, and extending more of our GenAI and search primitives to Hugging Face developers deepens our collaboration,” said Matt Riley, global vice president & general manager of search at Elastic. “The integration of our new semantic_text field, simplifies the process of chunking and storing embeddings, so developers can focus on what matters most, building great applications.”

The integration of semantic_text support follows the addition of Hugging Face embeddings models to Elastic’s Open Inference API.

Read the Elastic blog for more information.

About Elastic

Elastic (NYSE: ESTC), the Search AI Company, enables everyone to find the answers they need in real-time using all their data, at scale. Elastic’s solutions for search, observability and security are built on the Elastic Search AI Platform, the development platform used by thousands of companies, including more than 50% of the Fortune 500. Learn more at elastic.co.

Elastic and associated marks are trademarks or registered trademarks of Elastic N.V. and its subsidiaries. All other company and product names may be trademarks of their respective owners.

Contacts

Elastic PR
PR-team@elastic.co

Elastic N.V.

NYSE:ESTC

Release Versions

Contacts

Elastic PR
PR-team@elastic.co

More News From Elastic N.V.

Elastic Delivers Performance Gains for Users Running Elasticsearch on Google Axion Processors

SAN FRANCISCO--(BUSINESS WIRE)--Elastic (NYSE: ESTC), the Search AI Company, announced Elasticsearch runs with up to 40% higher indexing throughput on C4A VMs, powered by Google Axion, Google’s first custom Arm-based CPU, compared to previous generations of VMs on Google Cloud. Elastic used a macro benchmarking framework for Rally with the elastic/logs track to determine the maximum indexing performance on Google Axion-powered VMs. C4A also powers Elastic Cloud Serverless. “Elastic is driving i...

Elasticsearch Now Available as a Native Grounding Engine on Google Cloud’s Vertex AI Platform

SAN FRANCISCO--(BUSINESS WIRE)--Elastic (NYSE: ESTC), the Search AI Company, announced that the Elasticsearch vector database is now a natively supported information retrieval engine for Google Cloud’s Vertex AI platform. “By integrating Elasticsearch as a native grounding engine in Google Cloud’s Vertex AI platform, our joint customers can seamlessly build RAG applications in a single workflow,” said Shay Banon, founder and chief technology officer at Elastic. “This solution enables faster dep...

Elastic Announces General Availability of LLM Observability for Google Cloud’s Vertex AI

SAN FRANCISCO--(BUSINESS WIRE)--Elastic (NYSE: ESTC), the Search AI Company, announced the general availability of the Elastic Google Cloud Vertex AI platform integration in Elastic Observability. This integration offers large language model (LLM) observability support for models hosted in Google Cloud’s Vertex AI platform, providing insights into costs, token usage, errors, prompts, responses and performance. Site Reliability Engineers (SRE) can now optimize resource usage, identify and resolv...
Back to Newsroom
  1. There was an issue with the authorization server. Please contact support if the issue persists.