Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up

VB.NET api wrapper for llm-inference chatllm.cpp

NotificationsYou must be signed in to change notification settings

JohnClaw/chatllm.vb

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 

Repository files navigation

VB.NET api wrapper for llm-inference chatllm.cpp

All credits go to original repo:https://github.com/foldl/chatllm.cpp and Llama 405b (https://cloud.sambanova.ai/) which made 99% of work. I only guided it with prompts.

You can compile exe with this command line: dotnet build ChatLLM.vbproj

Then launch executable like this: main.exe -m qwen2.5-1.5b.bin

Links for quantatized models:

QWen-2.5 1.5B -https://modelscope.cn/api/v1/models/judd2024/chatllm_quantized_qwen2.5/repo?Revision=master&FilePath=qwen2.5-1.5b.bin

Gemma-2 2B -https://modelscope.cn/api/v1/models/judd2024/chatllm_quantized_gemma2_2b/repo?Revision=master&FilePath=gemma2-2b.bin

If you need more quantatized models use this python model downloader:https://github.com/foldl/chatllm.cpp/blob/master/scripts/model_downloader.py

You can convert custom safetensors model to inner chatllm.cpp format by using this script:https://github.com/foldl/chatllm.cpp/blob/master/convert.py

Converting tutorial:https://github.com/foldl/chatllm.cpp?tab=readme-ov-file#quantize-model

List of supported llm architecture types suitable for conversion:https://github.com/foldl/chatllm.cpp/blob/master/docs/models.md


[8]ページ先頭

©2009-2025 Movatter.jp