The Best Embedding Model for STEM and Mathematics in 2026

Apr 10, 2026 · GitHub Twitter Slack LinkedIn Discord
The Best Embedding Model for STEM and Mathematics in 2026
TL;DR
  • zembed-1 scores 0.5283 NDCG@10 in STEM and math retrieval, leading all competitors
  • +35.3% over OpenAI text-embedding-3-large and +5.4% over voyage-4-nano
  • zELO training captures the nuanced relevance spectrum critical for scientific document ranking
  • 32,768-token context window enables full-paper embeddings preserving argument structure and derivations
  • Open-weight and self-hostable for academic institutions with data privacy requirements

zembed-1 Leads the Way in STEM and Mathematics Retrieval

Scientific and mathematical text presents one of the hardest retrieval challenges in all of AI. The vocabulary is highly specialized, notation is dense and often symbolic, concepts span multiple levels of abstraction, and the difference between a highly relevant document and a tangentially related one can hinge on a single variable, operator, or qualifier that a generic model won’t understand at all.

zembed-1 by ZeroEntropy has benchmarked as the top embedding model for STEM and mathematics retrieval — outperforming OpenAI, Cohere, and Voyage in this demanding domain while simultaneously leading every other benchmarked category.

The Unique Challenge of STEM and Mathematical Retrieval

Scientific and mathematical text is not like general language. It has characteristics that specifically challenge embedding models:

STEM Benchmark Performance

ModelSTEM & Math NDCG@10
zembed-10.5283
voyage-4-nano0.5012
Cohere Embed v40.4698
OpenAI text-embedding-3-large0.3905

zembed-1’s score of 0.5283 leads the field significantly — +5.4% over voyage-4-nano and +35.3% over OpenAI text-embedding-3-large, one of the largest margins across any benchmarked domain.

This is notable because OpenAI’s embedding model is heavily used by researchers and in academic AI tools. The performance gap in STEM suggests that research and scientific applications have been significantly underserved by that model, and that zembed-1 represents a major upgrade for these use cases.

Why zembed-1 Thrives in STEM Contexts

zELO Captures Scientific Relevance Nuance

In scientific retrieval, the difference between “highly relevant” and “related but not what I need” is often a matter of mathematical precision. A paper on gradient descent in neural networks is not the same as one on convergence proofs for gradient descent, even though they share substantial vocabulary.

zembed-1’s zELO training methodology — modeling relevance as continuous Elo scores from pairwise document competitions — trains the model to capture exactly this kind of nuanced relevance spectrum. Documents that are truly on-point score higher than ones that are adjacent. This precision matters enormously in scientific retrieval.

Distilled from a State-of-the-Art Reranker

zembed-1 is distilled from zerank-2, ZeroEntropy’s best-in-class reranker. Rerankers are models specifically trained to evaluate whether a document is truly relevant to a query — a task that requires deep semantic understanding. By distilling from this reranker rather than training from scratch on similarity signals, zembed-1 inherits a more sophisticated understanding of what it means for a scientific document to be relevant to a scientific query.

Long Context for Full Scientific Documents

Research papers, textbook chapters, and technical reports are long documents. zembed-1’s 32,768-token context window enables full-paper embeddings that preserve the argument structure, the mathematical derivations, and the cross-reference relationships that make a paper relevant to a given scientific query.

STEM AI Application Use Cases

01

Scientific Literature Search

Enable semantic search over PubMed, arXiv, and internal research databases. Researchers can ask questions in natural language — “What methods have been used to improve transformer efficiency at long context?” — and retrieve the most relevant papers, even when they use different terminology than the query.

02

Mathematical Concept Retrieval

Build retrieval systems over mathematical textbooks, lecture notes, and problem sets. zembed-1’s performance in the STEM domain means it handles mathematical concept queries better than any other available model.

03

Research Knowledge Management

Power internal knowledge bases where researchers can search across their organization’s papers, reports, and technical documentation with natural language queries.

04

Educational AI

Build intelligent tutoring systems that retrieve the most relevant explanations, worked examples, and practice problems for a student’s question. zembed-1’s ability to handle both formal mathematical text and pedagogical explanations makes it ideal for this hybrid use case.

05

Lab Notebook and Experimental Data Search

Retrieve relevant experimental protocols, previous results, and related studies from a research organization’s internal archive. zembed-1’s retrieval quality ensures researchers find what they need rather than near-misses.

06

Patent and IP Search in Technical Fields

Search technical patent literature for prior art and related inventions in STEM domains. zembed-1’s understanding of technical vocabulary and concepts makes it a powerful tool for patent professionals and R&D legal teams.

Implementation for STEM Applications

from sentence_transformers import SentenceTransformer

model = SentenceTransformer(
    "zeroentropy/zembed-1",
    trust_remote_code=True,
    model_kwargs={"torch_dtype": "bfloat16"},
)

# Scientific literature search
query_embeddings = model.encode_query(
    "Convergence properties of stochastic gradient descent with momentum in non-convex optimization"
)

document_embeddings = model.encode_document([
    "We prove that SGD with momentum converges to a stationary point at rate O(1/√T) under standard assumptions on the gradient noise and step size schedule in non-convex settings...",
    "Gradient descent is one of the most widely used optimization algorithms in machine learning, used to minimize loss functions by iteratively updating parameters...",
    "We present a new analysis of Adam optimizer convergence under non-convex objectives, showing tight regret bounds that improve on previous work by a factor of log T...",
])

similarities = model.similarity(query_embeddings, document_embeddings)

What Researchers and Engineers Are Saying

“We run semantic search over 30 million research papers. Switching to zembed-1 improved our Recall@10 by 18 percentage points on our internal validation set. It wasn’t even a contest.” — CTO, academic research platform

For systematic literature reviews, you can cluster retrieved papers into thematic groups:

import numpy as np
from sklearn.cluster import KMeans

# Embed your full paper corpus (e.g. PubMed/arXiv abstracts)
papers = load_paper_corpus()  # returns list of {"id", "title", "abstract"}
texts = [f"{p['title']}. {p['abstract']}" for p in papers]
paper_embeddings = model.encode_document(texts, batch_size=32, show_progress_bar=True)

# Retrieve top-50 relevant papers for a research question
research_question = (
    "Transformer architectures for long-context document understanding in low-resource settings"
)
q_emb = model.encode_query(research_question)
scores = model.similarity(q_emb, paper_embeddings)[0].numpy()
top50_idx = np.argsort(scores)[::-1][:50]

# Cluster the top-50 into thematic groups
top50_embeddings = paper_embeddings[top50_idx]
kmeans = KMeans(n_clusters=5, random_state=42).fit(top50_embeddings)

print("Thematic clusters in your literature review:")
for cluster_id in range(5):
    cluster_papers = [papers[top50_idx[i]] for i, c in enumerate(kmeans.labels_) if c == cluster_id]
    print(f"\nCluster {cluster_id + 1} ({len(cluster_papers)} papers):")
    for p in cluster_papers[:3]:
        print(f"  - {p['title']}")

Why Researchers Should Make the Switch

If you’re building AI tools for scientific or academic contexts, the benchmark gap between zembed-1 and the next best model — and especially the gap versus OpenAI’s widely-used embedding model — is a compelling reason to upgrade.

A 35% improvement in STEM retrieval quality translates directly into fewer irrelevant papers, better recall of truly relevant work, and higher researcher trust in AI-assisted workflows.

What Better STEM Retrieval Means
  • Fewer irrelevant papers surfaced in literature search
  • Better recall of truly relevant work in systematic reviews
  • More accurate context retrieval for research RAG systems
  • Higher researcher trust in AI-assisted workflows

And because zembed-1 is open-weight and self-hostable, it’s accessible to academic institutions and research labs without the data privacy concerns that come with API-only services.

Get Started

zembed-1 is available today through multiple deployment options:

from zeroentropy import ZeroEntropy
zclient = ZeroEntropy()
response = zclient.models.embed(
model="zembed-1",
input_type="query", # "query" or "document"
input="What is retrieval augmented generation?", # string or list[str]
dimensions=2560, # optional: must be one of [2560, 1280, 640, 320, 160, 80, 40]
encoding_format="float", # "float" or "base64"
latency="fast", # "fast" or "slow"; omit for auto
)

Documentation: docs.zeroentropy.dev

HuggingFace: huggingface.co/zeroentropy

Get in touch: Discord community or contact@zeroentropy.dev

Talk to us if you need a custom deployment, volume pricing, or want to see how zembed-1 + zerank-2 performs on your data.

Related Blogs

Catch all the latest releases and updates from ZeroEntropy.

ZeroEntropy
The best AI teams retrieve with ZeroEntropy
Follow us on
GitHubTwitterSlackLinkedInDiscord