chatai/sglang/python/sglang
hailin b6bb6585d8 first commit. 2025-05-14 22:47:16 +08:00
..
lang first commit. 2025-05-14 22:47:16 +08:00
srt first commit. 2025-05-14 22:47:16 +08:00
test first commit. 2025-05-14 22:47:16 +08:00
README.md first commit. 2025-05-14 22:47:16 +08:00
__init__.py first commit. 2025-05-14 22:47:16 +08:00
api.py first commit. 2025-05-14 22:47:16 +08:00
bench_offline_throughput.py first commit. 2025-05-14 22:47:16 +08:00
bench_one_batch.py first commit. 2025-05-14 22:47:16 +08:00
bench_one_batch_server.py first commit. 2025-05-14 22:47:16 +08:00
bench_serving.py first commit. 2025-05-14 22:47:16 +08:00
check_env.py first commit. 2025-05-14 22:47:16 +08:00
global_config.py first commit. 2025-05-14 22:47:16 +08:00
launch_server.py first commit. 2025-05-14 22:47:16 +08:00
llama3_eval.py first commit. 2025-05-14 22:47:16 +08:00
utils.py first commit. 2025-05-14 22:47:16 +08:00
version.py first commit. 2025-05-14 22:47:16 +08:00

README.md

Code Structures

  • lang: The frontend language.
  • srt: The backend engine for running local models. (SRT = SGLang Runtime).
  • test: The test utilities.
  • api.py: The public APIs.
  • bench_offline_throughput.py: Benchmark the throughput in the offline mode.
  • bench_one_batch.py: Benchmark the latency of running a single static batch without a server.
  • bench_one_batch_server.py: Benchmark the latency of running a single batch with a server.
  • bench_serving.py: Benchmark online serving with dynamic requests.
  • check_env.py: Check the environment variables and dependencies.
  • global_config.py: The global configs and constants.
  • launch_server.py: The entry point for launching the local server.
  • llama3_eval.py: Evaluation of Llama 3 using the Meta Llama dataset.
  • utils.py: Common utilities.
  • version.py: Version info.