-
Notifications
You must be signed in to change notification settings - Fork 406
Pull requests: InternLM/lmdeploy
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Call cuda empty_cache to prevent OOM when quantizing model
#2671
opened Oct 28, 2024 by
AllentDan
Loading…
feat: support dynamic ntk scaling rotary embedding in ascend graph mode
#2670
opened Oct 28, 2024 by
tangzhiyi11
•
Draft
support yarn in turbomind backend
enhancement
New feature or request
#2519
opened Sep 26, 2024 by
irexyc
Loading…
[Feature] Support vision module w8a8 inference
improvement
#2308
opened Aug 14, 2024 by
AllentDan
Loading…
better formatted table of 'lmdeploy list'
improvement
WIP
#2289
opened Aug 12, 2024 by
lvhan028
Loading…
[Feature] support qqq(w4a8) for lmdeploy
#2274
opened Aug 9, 2024 by
HandH1998
Loading…
6 tasks done
[Feature] Support XTuner Lite Llava
enhancement
New feature or request
#2191
opened Jul 31, 2024 by
pppppM
Loading…
[benchmark] optimize benchmark: counting tokenlizer tokens and error requests
#1607
opened May 17, 2024 by
NiuBlibing
Loading…
fix: update api_server_backend.py to adapt latest gradio
improvement
#1541
opened May 3, 2024 by
kv-chiu
Loading…
Visualize layer activations and weights to simplify the quantization process.
#607
opened Oct 24, 2023 by
HIT-cwh
Loading…
ProTip!
Find all pull requests that aren't related to any open issues with -linked:issue.