|
2 | 2 | "name": "llama-coder", |
3 | 3 | "displayName": "Llama Coder", |
4 | 4 | "description": "Better and self-hosted Github Copilot replacement", |
5 | | - "version": "0.0.6", |
| 5 | + "version": "0.0.7", |
6 | 6 | "icon": "icon.png", |
7 | 7 | "publisher": "ex3ndr", |
8 | 8 | "repository": { |
|
70 | 70 | "codellama:34b-code-q4_K_S", |
71 | 71 | "codellama:34b-code-q4_K_M", |
72 | 72 | "codellama:34b-code-q6_K", |
73 | | - "deepseek-coder:6.7b-instruct-q4_K_S", |
74 | | - "deepseek-coder:6.7b-instruct-q4_K_M", |
75 | | - "deepseek-coder:6.7b-instruct-q8_0", |
76 | | - "deepseek-coder:6.7b-instruct-fp16", |
77 | | - "deepseek-coder:33b-instruct-q4_K_S", |
78 | | - "deepseek-coder:33b-instruct-q4_K_M", |
79 | | - "deepseek-coder:33b-instruct-q8_0", |
80 | | - "deepseek-coder:33b-instruct-fp16" |
| 73 | + "deepseek-coder:1.3b-base-q4_0", |
| 74 | + "deepseek-coder:1.3b-base-q4_1", |
| 75 | + "deepseek-coder:1.3b-base-q8_0", |
| 76 | + "deepseek-coder:6.7b-base-q4_K_S", |
| 77 | + "deepseek-coder:6.7b-base-q8_0", |
| 78 | + "deepseek-coder:6.7b-base-fp16", |
| 79 | + "deepseek-coder:33b-base-q4_K_S", |
| 80 | + "deepseek-coder:33b-base-q4_K_M", |
| 81 | + "deepseek-coder:33b-base-fp16" |
81 | 82 | ], |
82 | 83 | "default": "codellama:7b-code-q4_K_M", |
83 | 84 | "description": "Inference model to use" |
|
0 commit comments