96 lines
3.4 KiB
ReStructuredText
96 lines
3.4 KiB
ReStructuredText
SGLang Documentation
|
|
====================
|
|
|
|
SGLang is a fast serving framework for large language models and vision language models.
|
|
It makes your interaction with models faster and more controllable by co-designing the backend runtime and frontend language.
|
|
The core features include:
|
|
|
|
- **Fast Backend Runtime**: Provides efficient serving with RadixAttention for prefix caching, zero-overhead CPU scheduler, prefill-decode disaggregation, speculative decoding, continuous batching, paged attention, tensor/pipeline/expert/data parallelism, structured outputs, chunked prefill, quantization (FP4/FP8/INT4/AWQ/GPTQ), and multi-lora batching.
|
|
- **Flexible Frontend Language**: Offers an intuitive interface for programming LLM applications, including chained generation calls, advanced prompting, control flow, multi-modal inputs, parallelism, and external interactions.
|
|
- **Extensive Model Support**: Supports a wide range of generative models (Llama, Qwen, DeepSeek, Kimi, GPT, Gemma, Mistral, etc.), embedding models (e5-mistral, gte, mcdse) and reward models (Skywork), with easy extensibility for integrating new models.
|
|
- **Active Community**: SGLang is open-source and backed by an active community with wide industry adoption.
|
|
|
|
.. toctree::
|
|
:maxdepth: 1
|
|
:caption: Get Started
|
|
|
|
get_started/install.md
|
|
|
|
.. toctree::
|
|
:maxdepth: 1
|
|
:caption: Basic Usage
|
|
|
|
basic_usage/send_request.ipynb
|
|
basic_usage/openai_api.rst
|
|
basic_usage/offline_engine_api.ipynb
|
|
basic_usage/native_api.ipynb
|
|
basic_usage/sampling_params.md
|
|
basic_usage/deepseek.md
|
|
basic_usage/gpt_oss.md
|
|
basic_usage/llama4.md
|
|
basic_usage/qwen3.md
|
|
|
|
.. toctree::
|
|
:maxdepth: 1
|
|
:caption: Advanced Features
|
|
|
|
advanced_features/server_arguments.md
|
|
advanced_features/hyperparameter_tuning.md
|
|
advanced_features/speculative_decoding.ipynb
|
|
advanced_features/structured_outputs.ipynb
|
|
advanced_features/structured_outputs_for_reasoning_models.ipynb
|
|
advanced_features/tool_parser.ipynb
|
|
advanced_features/separate_reasoning.ipynb
|
|
advanced_features/quantization.md
|
|
advanced_features/lora.ipynb
|
|
advanced_features/pd_disaggregation.md
|
|
advanced_features/vlm_query.ipynb
|
|
advanced_features/router.md
|
|
advanced_features/observability.md
|
|
advanced_features/attention_backend.md
|
|
|
|
.. toctree::
|
|
:maxdepth: 1
|
|
:caption: Supported Models
|
|
|
|
supported_models/generative_models.md
|
|
supported_models/multimodal_language_models.md
|
|
supported_models/embedding_models.md
|
|
supported_models/reward_models.md
|
|
supported_models/rerank_models.md
|
|
supported_models/support_new_models.md
|
|
supported_models/transformers_fallback.md
|
|
supported_models/modelscope.md
|
|
|
|
.. toctree::
|
|
:maxdepth: 1
|
|
:caption: Hardware Platforms
|
|
|
|
platforms/amd_gpu.md
|
|
platforms/blackwell_gpu.md
|
|
platforms/cpu_server.md
|
|
platforms/tpu.md
|
|
platforms/nvidia_jetson.md
|
|
platforms/ascend_npu.md
|
|
|
|
.. toctree::
|
|
:maxdepth: 1
|
|
:caption: Developer Guide
|
|
|
|
developer_guide/contribution_guide.md
|
|
developer_guide/development_guide_using_docker.md
|
|
developer_guide/benchmark_and_profiling.md
|
|
developer_guide/bench_serving.md
|
|
|
|
.. toctree::
|
|
:maxdepth: 1
|
|
:caption: References
|
|
|
|
references/faq.md
|
|
references/environment_variables.md
|
|
references/production_metrics.md
|
|
references/multi_node_deployment/multi_node_index.rst
|
|
references/custom_chat_template.md
|
|
references/frontend/frontend_index.rst
|
|
references/learn_more.md
|