About Semantic Encoder (SE)

In this article
Important

When used in a query pipeline, a Semantic Encoder (SE) model uses vector search to retrieve items from your index based on semantic similarity with the query. The model does this by first creating embeddings for the enterprise content that you specify for the model, and then referencing the embeddings at query time.

Each query passes through a query pipeline where pipeline rules and machine learning are applied to optimize relevance. However, in addition to the traditional lexical (keyword) search, the SE model adds vector-based search capabilities to retrieve the items from the index with high semantic similarity with the query. The most relevant search results are then sent to the RGA or CPR model for answer generation or passage retrieval (see RGA overview or CPR overview).

Traditional lexical search relies on matching keywords or phrases that appear in a query with the words and phrases in items. Due to the exact-match nature of lexical search, user queries that yield the best results tend to be concise and precise. The user needs to know what they’re searching for and use the proper keywords.

However, what if a user doesn’t know exactly what they’re looking for, or what if the query is more complex? With the emergence of generative AI, customer expectations are evolving and search interfaces must be able to understand and provide answers for more complex natural language queries. A complex query is one where the user provides context and asks a question.

For instance, if a user enters the query "What is Coveo Machine Learning and what are its benefits," lexical search results will include items that discuss "Coveo Machine Learning." The results may also contain items with high occurrences of the word "benefit," which may not be relevant to the query. Lexical search is fast, cost-efficient, and has a proven track record in many enterprises. However, lexical search doesn’t consider the meaning of words or the context. While lexical search isn’t suited to finding similarities in items based on meaning, vector-based search is designed for just that purpose.

When combined with Relevance Generative Answering (RGA) or Passage Retrieval (CPR), an SE model ensures that the RGA or CPR model uses the most relevant content to generate answers and retrieve passages, respectively.

Tip

By default, content retrieval is supported only for English content. However, Coveo offers beta support for content retrieval in languages other than English. Learn more about multilingual content retrieval and answer generation.

What’s next?