Veni AI

OpenAI Text Embedding 3 Large

Text Embedding 3 Large produces high-quality text embeddings for search, clustering, and recommendations.

Try Now
SCROLL
01

What is Text Embedding 3 Large?

Text Embedding 3 Large from OpenAI encodes text into dense vectors for retrieval and analytics. Use it for RAG pipelines, semantic search, recommendations, and topic detection across languages. Optimized for quality and latency so it scales to large corpora.

02

Technical Specifications

Context Window

8,191 token

Max Output

3072 boyutlu vektör

Training Cutoff

2024

Active

Active

03

Capabilities

High-quality text embeddings for search and clustering
Handles multi-language inputs
Optimized for semantic retrieval latency
04

Benchmark Scores

MTEB Average
64.6%
Dimension
3072
Max Input
8191
Accuracy vs Ada-002
+30%
Languages
100+
Cost Efficiency
95%
05

Pros & Cons

Pros

  • Strong retrieval quality
  • Fast inference and small vectors
  • Works across languages

Cons

  • Not a generative model
  • Needs good chunking to avoid drift
  • Quality depends on downstream index settings
06

Features

01

Semantic search

Encode queries and documents into the same vector space.

02

Multi-task

Use one embedding for search, recommendations, and clustering.

03

Scalable

Low latency and small vectors for large corpora.

07

Use Cases

01

RAG indexing

Embed knowledge bases for accurate retrieval-augmented generation.

02

Recommendations

Cluster similar items and surface relevant content.

03

Analytics

Detect topics, intent, and anomalies across text streams.

09

FAQ

10

Related Models