Model name | model size |
---|---|
nomic-ai/nomic-embed-text-v1.5 (recommended) | 137M |
nomic-ai/nomic-embed-text-v1 | 137M |
WhereIsAI/UAE-Large-V1 | 335M |
thenlper/gte-large | 335M |
thenlper/gte-base | 109M |
BAAI/bge-base-en-v1.5 | 109M |
BAAI/bge-small-en-v1.5 | 33M |
mixedbread-ai/mxbai-embed-large-v1 | 335M |
sentence-transformers/all-MiniLM-L6-v2 | 23M |
sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2 | 118M |
/v1/embeddings
endpoint. This includes all the architectures supported for uploading custom models, such as Llama, Qwen, DeepSeek, Mistral, Mixtral, and many others.
Here are some examples of LLM-based embedding models that work with the embeddings API:
Model name |
---|
fireworks/llama4-scout-instruct-basic |
fireworks/glm-4p5 |
fireworks/gpt-oss-20b |
fireworks/kimi-k2-instruct |
fireworks/qwen3-30b-a3b |
fireworks/deepseek-r1 |
search_document: Spiderman was a particularly entertaining movie with...
and returns the following
/v1/embeddings
endpoint with your chosen model.
search_document:
prefix. Nomic models have been fine-tuned to take prefixes, so for user queries, you will need add thesearch_query:
prefix, and for documents, you need to prefix with search_document:
Here’s a quick example:
search_document:
embeddings.create()
request