- Notifications
You must be signed in to change notification settings - Fork0
JohnClaw/chatllm.vb
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
VB.NET api wrapper for llm-inference chatllm.cpp
All credits go to original repo:https://github.com/foldl/chatllm.cpp and Llama 405b (https://cloud.sambanova.ai/) which made 99% of work. I only guided it with prompts.
You can compile exe with this command line: dotnet build ChatLLM.vbproj
Then launch executable like this: main.exe -m qwen2.5-1.5b.bin
Links for quantatized models:
QWen-2.5 1.5B -https://modelscope.cn/api/v1/models/judd2024/chatllm_quantized_qwen2.5/repo?Revision=master&FilePath=qwen2.5-1.5b.bin
If you need more quantatized models use this python model downloader:https://github.com/foldl/chatllm.cpp/blob/master/scripts/model_downloader.py
You can convert custom safetensors model to inner chatllm.cpp format by using this script:https://github.com/foldl/chatllm.cpp/blob/master/convert.py
Converting tutorial:https://github.com/foldl/chatllm.cpp?tab=readme-ov-file#quantize-model
List of supported llm architecture types suitable for conversion:https://github.com/foldl/chatllm.cpp/blob/master/docs/models.md
About
VB.NET api wrapper for llm-inference chatllm.cpp