- zembed-1 scores 0.5283 NDCG@10 in STEM and math retrieval, leading all competitors
- +35.3% over OpenAI text-embedding-3-large and +5.4% over voyage-4-nano
- zELO training captures the nuanced relevance spectrum critical for scientific document ranking
- 32,768-token context window enables full-paper embeddings preserving argument structure and derivations
- Open-weight and self-hostable for academic institutions with data privacy requirements
zembed-1 Leads the Way in STEM and Mathematics Retrieval
Scientific and mathematical text presents one of the hardest retrieval challenges in all of AI. The vocabulary is highly specialized, notation is dense and often symbolic, concepts span multiple levels of abstraction, and the difference between a highly relevant document and a tangentially related one can hinge on a single variable, operator, or qualifier that a generic model won’t understand at all.
zembed-1 by ZeroEntropy has benchmarked as the top embedding model for STEM and mathematics retrieval — outperforming OpenAI, Cohere, and Voyage in this demanding domain while simultaneously leading every other benchmarked category.
The Unique Challenge of STEM and Mathematical Retrieval
Scientific and mathematical text is not like general language. It has characteristics that specifically challenge embedding models:
STEM Benchmark Performance
| Model | STEM & Math NDCG@10 |
|---|---|
| zembed-1 | 0.5283 |
| voyage-4-nano | 0.5012 |
| Cohere Embed v4 | 0.4698 |
| OpenAI text-embedding-3-large | 0.3905 |
zembed-1’s score of 0.5283 leads the field significantly — +5.4% over voyage-4-nano and +35.3% over OpenAI text-embedding-3-large, one of the largest margins across any benchmarked domain.
This is notable because OpenAI’s embedding model is heavily used by researchers and in academic AI tools. The performance gap in STEM suggests that research and scientific applications have been significantly underserved by that model, and that zembed-1 represents a major upgrade for these use cases.
Why zembed-1 Thrives in STEM Contexts
zELO Captures Scientific Relevance Nuance
In scientific retrieval, the difference between “highly relevant” and “related but not what I need” is often a matter of mathematical precision. A paper on gradient descent in neural networks is not the same as one on convergence proofs for gradient descent, even though they share substantial vocabulary.
zembed-1’s zELO training methodology — modeling relevance as continuous Elo scores from pairwise document competitions — trains the model to capture exactly this kind of nuanced relevance spectrum. Documents that are truly on-point score higher than ones that are adjacent. This precision matters enormously in scientific retrieval.
Distilled from a State-of-the-Art Reranker
zembed-1 is distilled from zerank-2, ZeroEntropy’s best-in-class reranker. Rerankers are models specifically trained to evaluate whether a document is truly relevant to a query — a task that requires deep semantic understanding. By distilling from this reranker rather than training from scratch on similarity signals, zembed-1 inherits a more sophisticated understanding of what it means for a scientific document to be relevant to a scientific query.
Long Context for Full Scientific Documents
Research papers, textbook chapters, and technical reports are long documents. zembed-1’s 32,768-token context window enables full-paper embeddings that preserve the argument structure, the mathematical derivations, and the cross-reference relationships that make a paper relevant to a given scientific query.
STEM AI Application Use Cases
Scientific Literature Search
Enable semantic search over PubMed, arXiv, and internal research databases. Researchers can ask questions in natural language — “What methods have been used to improve transformer efficiency at long context?” — and retrieve the most relevant papers, even when they use different terminology than the query.
Mathematical Concept Retrieval
Build retrieval systems over mathematical textbooks, lecture notes, and problem sets. zembed-1’s performance in the STEM domain means it handles mathematical concept queries better than any other available model.
Research Knowledge Management
Power internal knowledge bases where researchers can search across their organization’s papers, reports, and technical documentation with natural language queries.
Educational AI
Build intelligent tutoring systems that retrieve the most relevant explanations, worked examples, and practice problems for a student’s question. zembed-1’s ability to handle both formal mathematical text and pedagogical explanations makes it ideal for this hybrid use case.
Lab Notebook and Experimental Data Search
Retrieve relevant experimental protocols, previous results, and related studies from a research organization’s internal archive. zembed-1’s retrieval quality ensures researchers find what they need rather than near-misses.
Patent and IP Search in Technical Fields
Search technical patent literature for prior art and related inventions in STEM domains. zembed-1’s understanding of technical vocabulary and concepts makes it a powerful tool for patent professionals and R&D legal teams.
Implementation for STEM Applications
from sentence_transformers import SentenceTransformer
model = SentenceTransformer(
"zeroentropy/zembed-1",
trust_remote_code=True,
model_kwargs={"torch_dtype": "bfloat16"},
)
# Scientific literature search
query_embeddings = model.encode_query(
"Convergence properties of stochastic gradient descent with momentum in non-convex optimization"
)
document_embeddings = model.encode_document([
"We prove that SGD with momentum converges to a stationary point at rate O(1/√T) under standard assumptions on the gradient noise and step size schedule in non-convex settings...",
"Gradient descent is one of the most widely used optimization algorithms in machine learning, used to minimize loss functions by iteratively updating parameters...",
"We present a new analysis of Adam optimizer convergence under non-convex objectives, showing tight regret bounds that improve on previous work by a factor of log T...",
])
similarities = model.similarity(query_embeddings, document_embeddings)
What Researchers and Engineers Are Saying
“We run semantic search over 30 million research papers. Switching to zembed-1 improved our Recall@10 by 18 percentage points on our internal validation set. It wasn’t even a contest.” — CTO, academic research platform
For systematic literature reviews, you can cluster retrieved papers into thematic groups:
import numpy as np
from sklearn.cluster import KMeans
# Embed your full paper corpus (e.g. PubMed/arXiv abstracts)
papers = load_paper_corpus() # returns list of {"id", "title", "abstract"}
texts = [f"{p['title']}. {p['abstract']}" for p in papers]
paper_embeddings = model.encode_document(texts, batch_size=32, show_progress_bar=True)
# Retrieve top-50 relevant papers for a research question
research_question = (
"Transformer architectures for long-context document understanding in low-resource settings"
)
q_emb = model.encode_query(research_question)
scores = model.similarity(q_emb, paper_embeddings)[0].numpy()
top50_idx = np.argsort(scores)[::-1][:50]
# Cluster the top-50 into thematic groups
top50_embeddings = paper_embeddings[top50_idx]
kmeans = KMeans(n_clusters=5, random_state=42).fit(top50_embeddings)
print("Thematic clusters in your literature review:")
for cluster_id in range(5):
cluster_papers = [papers[top50_idx[i]] for i, c in enumerate(kmeans.labels_) if c == cluster_id]
print(f"\nCluster {cluster_id + 1} ({len(cluster_papers)} papers):")
for p in cluster_papers[:3]:
print(f" - {p['title']}") Why Researchers Should Make the Switch
If you’re building AI tools for scientific or academic contexts, the benchmark gap between zembed-1 and the next best model — and especially the gap versus OpenAI’s widely-used embedding model — is a compelling reason to upgrade.
A 35% improvement in STEM retrieval quality translates directly into fewer irrelevant papers, better recall of truly relevant work, and higher researcher trust in AI-assisted workflows.
- Fewer irrelevant papers surfaced in literature search
- Better recall of truly relevant work in systematic reviews
- More accurate context retrieval for research RAG systems
- Higher researcher trust in AI-assisted workflows
And because zembed-1 is open-weight and self-hostable, it’s accessible to academic institutions and research labs without the data privacy concerns that come with API-only services.
Get Started
zembed-1 is available today through multiple deployment options:
from zeroentropy import ZeroEntropy
zclient = ZeroEntropy()
response = zclient.models.embed(
model="zembed-1",
input_type="query", # "query" or "document"
input="What is retrieval augmented generation?", # string or list[str]
dimensions=2560, # optional: must be one of [2560, 1280, 640, 320, 160, 80, 40]
encoding_format="float", # "float" or "base64"
latency="fast", # "fast" or "slow"; omit for auto
)Documentation: docs.zeroentropy.dev
HuggingFace: huggingface.co/zeroentropy
Get in touch: Discord community or contact@zeroentropy.dev
Talk to us if you need a custom deployment, volume pricing, or want to see how zembed-1 + zerank-2 performs on your data.
