Comments (12)
but having ggml also can attract more contributors, also i'm curious whether inferLLM support cuda, doesn't see that on the page
from lmdeploy.
Indeed llama.cpp/GGML is famous (about 30k star), but the code is really hard to read.
InferLLM may be a better choice.
from lmdeploy.
wow what is it can you elaborate a bit more, to what i understand it's just that you need to implement the ggml itself into the triton backend and later on you can reuse the triton backend and have the benefit of ggml update after every iteration
from lmdeploy.
will not be re-using fastertransformer i guess but more about how the ggml itself can be integrated with tritonbackend
from lmdeploy.
will not be re-using fastertransformer i guess but more about how the ggml itself can be integrated with tritonbackend
Here is my opinion, when you integrate a inference backend, you have to make sure of service quality.
Let's take a assumption, after integration, if ggml
has a bug, lmdeploy has responsibility to locate it or fix it.
The greater cost of software is maintenance, so we need to consider the code complexity of ggml.
from lmdeploy.
About InferLLM https://github.com/MegEngine/InferLLM
from lmdeploy.
saw enable gpu on the cmake file still
from lmdeploy.
but having ggml also can attract more contributors, also i'm curious whether inferLLM support cuda, doesn't see that on the page
yes, attracting more contributors is a good reason to integrate ggml.
from lmdeploy.
but having ggml also can attract more contributors, also i'm curious whether inferLLM support cuda, doesn't see that on the page
InferLLM cuda part is WIP MegEngine/InferLLM#27
from lmdeploy.
InferLLM will have many language barriers for non-chinese speakers contributors also, which is a very large audience, for most of the open source implementation it depends on how many contributors you have.
from lmdeploy.
InferLLM will have many language barriers for non-chinese speakers contributors also, which is a very large audience, for most of the open source implementation it depends on how many contributors you have.
Already passed on to InferLLM team .. QvQ
from lmdeploy.
Close since no more activity over two weeks. Feel free to reopen it if it is still an issue
from lmdeploy.
Related Issues (20)
- [Bug] InternVL-Chat-V1-5量化报错 HOT 4
- [Bug] w8a8量化后的InternVL-Chat-V1-5模型lmdeploychat启动transformers报错 HOT 2
- [Feature] Guided Decoding HOT 1
- Batch infer seems no speed up HOT 8
- [Bug] WSL2环境下,0.4.2做InternVL量化时,磁盘写入速度极低 HOT 6
- 请问 TurboMind 支持cogvlm系列么? HOT 1
- [Bug] UnboundLocalError: local variable 'head_num' referenced before assignment HOT 4
- [feature] need rope_scaling_factor args in benchmark/profile_generation.py to enable dynamic NTK.
- [Feature] AWQ量化的校准数据集
- [Feature] ModuleNotFoundError: No module named 'timm' with Internal Vision model HOT 2
- [Feature] health endpoint HOT 2
- [Feature] model name should be settable or follow original full HF link name, not random new name HOT 6
- [Bug] lmdeploy lite auto_awq: RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! HOT 1
- [Bug] result of W4A16 quantized Qwen1.5-1.8B-Chat model not correct HOT 1
- Support for SWIFT finetuned models HOT 4
- 总是看到一个using default GEMM algo的WARNING,是否会因为使用了默认的GEMM而影响速度或者吞吐量? HOT 1
- [Bug] 下载代码执行internvl-v1.5量化,导入本地模型时报错 HOT 11
- [Feature] peft<=0.9.0 要求的版本要求太低,与较多环境要求peft>0.10冲突,能否修改
- [Feature] Support for LLaVA-NeXT HOT 1
- [Docs] How are multiple images handled? HOT 5
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from lmdeploy.