Skip to content

[Feature]Add support for SGLang in the "Model Deployment" module #12151

@maxbetas

Description

@maxbetas

1Panel Version

2.1.4

Please describe your needs or suggestions for improvements

Thank you to the 1Panel team for your excellent work on AI model deployment; the current support for Ollama and vLLM is fantastic.

As the open-source LLM ecosystem evolves, SGLang (https://github.com/sgl-project/sglang) is rapidly gaining traction as a next-generation, high-performance inference framework used by many developers and enterprises. Given 1Panel's commitment to providing comprehensive and cutting-edge productivity tools, I highly recommend adding built-in support for SGLang in the model deployment module.

Why is this needed?
While 1Panel already supports vLLM and Ollama, SGLang offers unique advantages that cover scenarios where existing solutions might not be optimal:

Exceptional Concurrency Performance (RadixAttention): SGLang's RadixAttention mechanism significantly maximizes KV Cache reuse when handling multi-turn conversations, long contexts, and complex Agent prompts. Its throughput is notably superior to vLLM in many of these scenarios.

Ultra-fast Structured Output: For AI Agent development, forced JSON structured output is a strict requirement. SGLang is deeply optimized for structured decoding, heavily outperforming other frameworks.

Native Multimodal Support: It provides out-of-the-box, high-performance support for mainstream Vision Large Language Models (VLMs).

Proposed Solution:
I suggest integrating SGLang similarly to how vLLM is currently handled—by adding an SGLang option in the engine dropdown under the Model Deployment settings.

The configuration form should ideally support the following core parameters (corresponding to SGLang's startup arguments):

Model Path (--model-path)

Service Port (--port)

Tensor Parallelism / GPU Count (--tp-size)

Context Length (--context-length)

API Key (--api-key)

On the backend, 1Panel could use SGLang's official Docker image (e.g., lmsysorg/sglang:latest) to generate the docker-compose file, spin up the service, and expose an OpenAI-compatible API.

Alternative Solutions:
Currently, users have to manually write a docker-compose.yml via 1Panel's "Container Orchestration" feature to deploy SGLang. While feasible, this method lacks the intuitive graphical management, log viewing, and convenient one-click start/stop capabilities provided by the dedicated Model Deployment module.

Additional Context:

SGLang Official Documentation: https://docs.sglang.ai/

SGLang Docker Deployment Guide: https://docs.sglang.ai/backend/docker.html

Please describe the solution you suggest

No response

Additional Information

No response

Metadata

Metadata

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions