OpenAI Text Embedding 3 Large
Text Embedding 3 Large produces high-quality text embeddings for search, clustering, and recommendations.
Try Now
SCROLL
01
What is Text Embedding 3 Large?
Text Embedding 3 Large from OpenAI encodes text into dense vectors for retrieval and analytics. Use it for RAG pipelines, semantic search, recommendations, and topic detection across languages. Optimized for quality and latency so it scales to large corpora.
02
Technical Specifications
Context Window
8,191 token
Max Output
3072 boyutlu vektör
Training Cutoff
2024
Active
Active
03
Capabilities
High-quality text embeddings for search and clustering
Handles multi-language inputs
Optimized for semantic retrieval latency
04
Benchmark Scores
MTEB Average
64.6%Dimension
3072Max Input
8191Accuracy vs Ada-002
+30%Languages
100+Cost Efficiency
95%05
Pros & Cons
Pros
- Strong retrieval quality
- Fast inference and small vectors
- Works across languages
Cons
- Not a generative model
- Needs good chunking to avoid drift
- Quality depends on downstream index settings
06
Features
01
Semantic search
Encode queries and documents into the same vector space.
02
Multi-task
Use one embedding for search, recommendations, and clustering.
03
Scalable
Low latency and small vectors for large corpora.
07
Use Cases
01
RAG indexing
Embed knowledge bases for accurate retrieval-augmented generation.
02
Recommendations
Cluster similar items and surface relevant content.
03
Analytics
Detect topics, intent, and anomalies across text streams.
09
FAQ
10