Conversation
There was a problem hiding this comment.
Summary of Changes
Hello @hiworldwzj, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!
This pull request updates the installation documentation for the LightLLM service, specifically focusing on Docker deployment. It provides crucial information regarding shared memory requirements to ensure proper service startup and operation, and adjusts the recommended default shared memory allocation in the provided Docker commands.
Highlights
- Docker Installation Guidance: Enhanced the Docker installation instructions by adding comprehensive notes on shared memory (
--shm-size) requirements for the LightLLM service. - Shared Memory Recommendations: Provided specific shared memory recommendations for both text-only (2GB minimum, 16GB+ recommended) and multimodal (16GB+ recommended) LightLLM deployments within Docker.
- Default Shared Memory Adjustment: Increased the default
--shm-sizeparameter in the documented Docker run commands from1gto2gto better align with the service's needs. - Performance Tuning Tip: Included a tip to adjust the
--running_max_req_sizeparameter for users with limited shared memory, suggesting it can reduce memory footprint at the cost of concurrent requests.
Using Gemini Code Assist
The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.
Invoking Gemini
You can request assistance from Gemini at any point in your pull request via creating an issue comment (i.e. comment on the pull request page) using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands.
| Feature | Command | Description |
|---|---|---|
| Code Review | /gemini review |
Performs a code review for the current pull request in its current state. |
| Pull Request Summary | /gemini summary |
Provides a summary of the current pull request in its current state. |
| Comment | @gemini-code-assist | Responds in comments when explicitly tagged, both in issue comments and review comments. |
| Help | /gemini help |
Displays a list of available commands. |
Customization
To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.
Limitations & Feedback
Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments or fill out our survey to provide feedback.
You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.
Footnotes
-
Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution. ↩
| @@ -37,7 +43,7 @@ Lightllm 是一个纯python开发的推理框架,其中的算子使用triton | |||
| $ | |||
There was a problem hiding this comment.
为了保持一致性,建议在此处为手动构建的镜像添加关于共享内存要求的详细说明,就像为运行官方镜像所做的那样。这将避免选择从源代码构建的用户产生混淆。
$ # 运行服务, 注意现在的lightllm服务非常的依赖共享内存部分,在启动
$ # 前请确保你的docker设置中已经分配了足够的共享内存,否则可能导致
$ # 服务无法正常启动。
$ # 1.如果是纯文本服务,建议分配2GB以上的共享内存, 如果你的内存充足,建议分配16GB以上的共享内存.
$ # 2.如果是多模态服务,建议分配16GB以上的共享内存,具体可以根据实际情况进行调整.
$ # 如果你没有足够的共享内存,可以尝试在启动服务的时候调低 --running_max_req_size 参数,这会降低
$ # 服务的并发请求数量,但可以减少共享内存的占用。
| $ docker pull ghcr.io/modeltc/lightllm:main | ||
| $ | ||
| $ # Run | ||
| $ # Run,The current LightLLM service relies heavily on shared memory. |
| $ # | ||
| $ # 1. For text-only services, it is recommended to allocate more than 2GB of shared memory. | ||
| $ # If your system has sufficient RAM, allocating 16GB or more is recommended. | ||
| $ # 2.For multimodal services, it is recommended to allocate 16GB or more of shared memory. |
| $ docker build -t <image_name> . | ||
| $ | ||
| $ # Run | ||
| $ # Run, |
There was a problem hiding this comment.
本节缺少为官方镜像添加的关于共享内存的详细说明。$ # Run, 这一行似乎是不完整的编辑。为了确保一致性并向用户提供完整的信息,也请在此处添加完整的说明。
$ # Run: The current LightLLM service relies heavily on shared memory.
$ # Before starting, please make sure that you have allocated enough shared memory
$ # in your Docker settings; otherwise, the service may fail to start properly.
$ #
$ # 1. For text-only services, it is recommended to allocate more than 2GB of shared memory.
$ # If your system has sufficient RAM, allocating 16GB or more is recommended.
$ # 2. For multimodal services, it is recommended to allocate 16GB or more of shared memory.
$ # You can adjust this value according to your specific requirements.
$ #
$ # If you do not have enough shared memory available, you can try lowering
$ # the --running_max_req_size parameter when starting the service.
$ # This will reduce the number of concurrent requests, but also decrease shared memory usage.
No description provided.