sglang.0.4.8.post1/sglang/docs/supported_models/embedding_models.md

3.1 KiB
Raw Permalink Blame History

Embedding Models

SGLang provides robust support for embedding models by integrating efficient serving mechanisms with its flexible programming interface. This integration allows for streamlined handling of embedding tasks, facilitating faster and more accurate retrieval and semantic search operations. SGLang's architecture enables better resource utilization and reduced latency in embedding model deployment.

They are executed with `--is-embedding` and some may require `--trust-remote-code`

Example Launch Command

python3 -m sglang.launch_server \
  --model-path Alibaba-NLP/gme-Qwen2-VL-2B-Instruct \
  --is-embedding \
  --host 0.0.0.0 \
  --chat-template gme-qwen2-vl \
  --port 30000

Example Client Request

import requests

url = "http://127.0.0.1:30000"

text_input = "Represent this image in embedding space."
image_path = "https://huggingface.co/datasets/liuhaotian/llava-bench-in-the-wild/resolve/main/images/023.jpg"

payload = {
    "model": "gme-qwen2-vl",
    "input": [
        {
            "text": text_input
        },
        {
            "image": image_path
        }
    ],
}

response = requests.post(url + "/v1/embeddings", json=payload).json()

print("Embeddings:", [x.get("embedding") for x in response.get("data", [])])

Supported models

Model Family (Embedding) Example HuggingFace Identifier Chat Template Description
Llama/Mistral based (E5EmbeddingModel) intfloat/e5-mistral-7b-instruct N/A Mistral/Llama-based embedding model finetuned for highquality text embeddings (topranked on the MTEB benchmark).
GTE (QwenEmbeddingModel) Alibaba-NLP/gte-Qwen2-7B-instruct N/A Alibabas general text embedding model (7B), achieving stateoftheart multilingual performance in English and Chinese.
GME (MultimodalEmbedModel) Alibaba-NLP/gme-Qwen2-VL-2B-Instruct gme-qwen2-vl Multimodal embedding model (2B) based on Qwen2VL, encoding image + text into a unified vector space for crossmodal retrieval.
CLIP (CLIPEmbeddingModel) openai/clip-vit-large-patch14-336 N/A OpenAIs CLIP model (ViTL/14) for embedding images (and text) into a joint latent space; widely used for image similarity search.
BGE (BgeEmbeddingModel) BAAI/bge-large-en-v1.5 N/A Currently only support attention-backend triton and torch_native. BAAI's BGE embedding models optimized for retrieval and reranking tasks.