You signed in with another tab or window.Reload to refresh your session.You signed out in another tab or window.Reload to refresh your session.You switched accounts on another tab or window.Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+7-2Lines changed: 7 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -6,7 +6,7 @@ Llama Coder is a better and self-hosted Github Copilot replacement for VS Studio
6
6
7
7
##Features
8
8
* 🚀 As good as Copilot
9
-
* ⚡️ Fast. Works well on consumer GPUs. RTX 4090 is recommended for best performance.
9
+
* ⚡️ Fast. Works well on consumer GPUs.Apple Silicon orRTX 4090 is recommended for best performance.
10
10
* 🔐 No telemetry or tracking
11
11
* 🔬 Works with any language coding or human one.
12
12
@@ -27,10 +27,11 @@ Install [Ollama](https://ollama.ai) on dedicated machine and configure endpoint
27
27
28
28
##Models
29
29
30
-
Currently Llama Coder supports only Codellama. Model is quantized in different ways, but our tests shows that`q4` is an optimal way to run network. When selecting model the bigger the model is, it performs better. Always pick the model with the biggest size and the biggest possible quantization for your machine. Default one is`codellama:7b-code-q4_K_M` and should work everywhere,`codellama:34b-code-q4_K_M` is the best possible one.
30
+
Currently Llama Coder supports only Codellama. Model is quantized in different ways, but our tests shows that`q4` is an optimal way to run network. When selecting model the bigger the model is, it performs better. Always pick the model with the biggest size and the biggest possible quantization for your machine. Default one is`stable-code:3b-code-q4_0` and should work everywhere and outperforms most other models.
31
31
32
32
| Name| RAM/VRAM| Notes|
33
33
|---------------------------|----------|-------|
34
+
| stable-code:3b-code-q4_0| 3GB||
34
35
| codellama:7b-code-q4_K_M| 5GB||
35
36
| codellama:7b-code-q6_K| 6GB| m|
36
37
| codellama:7b-code-fp16| 14GB| g|
@@ -48,6 +49,10 @@ Most of the problems could be seen in output of a plugin in VS Code extension ou
48
49
49
50
##Changelog
50
51
52
+
##[0.0.11]
53
+
- Added Stable Code model
54
+
- Pause download only for specific model instead of all models
55
+
51
56
##[0.0.10]
52
57
- Adding ability to pick a custom model
53
58
- Asking user if they want to download model if it is not available