-
Notifications
You must be signed in to change notification settings - Fork 3k
Description
1Panel Version
2.1.4
Please describe your needs or suggestions for improvements
Thank you to the 1Panel team for your excellent work on AI model deployment; the current support for Ollama and vLLM is fantastic.
As the open-source LLM ecosystem evolves, SGLang (https://github.com/sgl-project/sglang) is rapidly gaining traction as a next-generation, high-performance inference framework used by many developers and enterprises. Given 1Panel's commitment to providing comprehensive and cutting-edge productivity tools, I highly recommend adding built-in support for SGLang in the model deployment module.
Why is this needed?
While 1Panel already supports vLLM and Ollama, SGLang offers unique advantages that cover scenarios where existing solutions might not be optimal:
Exceptional Concurrency Performance (RadixAttention): SGLang's RadixAttention mechanism significantly maximizes KV Cache reuse when handling multi-turn conversations, long contexts, and complex Agent prompts. Its throughput is notably superior to vLLM in many of these scenarios.
Ultra-fast Structured Output: For AI Agent development, forced JSON structured output is a strict requirement. SGLang is deeply optimized for structured decoding, heavily outperforming other frameworks.
Native Multimodal Support: It provides out-of-the-box, high-performance support for mainstream Vision Large Language Models (VLMs).
Proposed Solution:
I suggest integrating SGLang similarly to how vLLM is currently handled—by adding an SGLang option in the engine dropdown under the Model Deployment settings.
The configuration form should ideally support the following core parameters (corresponding to SGLang's startup arguments):
Model Path (--model-path)
Service Port (--port)
Tensor Parallelism / GPU Count (--tp-size)
Context Length (--context-length)
API Key (--api-key)
On the backend, 1Panel could use SGLang's official Docker image (e.g., lmsysorg/sglang:latest) to generate the docker-compose file, spin up the service, and expose an OpenAI-compatible API.
Alternative Solutions:
Currently, users have to manually write a docker-compose.yml via 1Panel's "Container Orchestration" feature to deploy SGLang. While feasible, this method lacks the intuitive graphical management, log viewing, and convenient one-click start/stop capabilities provided by the dedicated Model Deployment module.
Additional Context:
SGLang Official Documentation: https://docs.sglang.ai/
SGLang Docker Deployment Guide: https://docs.sglang.ai/backend/docker.html
Please describe the solution you suggest
No response
Additional Information
No response