1 Star 0 Fork 0

littleketty/llama-coder

加入 Gitee
与超过 1200万 开发者一起发现、参与优秀开源项目,私有仓库也完全免费 :)
免费加入
克隆/下载
贡献代码
同步代码
取消
提示: 由于 Git 不支持空文件夾,创建文件夹后会生成空的 .keep 文件
Loading...
README
MIT

Llama Coder

Llama Coder is a better and self-hosted Github Copilot replacement for VS Code. Llama Coder uses Ollama and codellama to provide autocomplete that runs on your hardware. Works best with Mac M1/M2/M3 or with RTX 4090.

VS Code Plugin

Features

  • 🚀 As good as Copilot
  • ⚡️ Fast. Works well on consumer GPUs. Apple Silicon or RTX 4090 is recommended for best performance.
  • 🔐 No telemetry or tracking
  • 🔬 Works with any language coding or human one.

Recommended hardware

Minimum required RAM: 16GB is a minimum, more is better since even smallest model takes 5GB of RAM. The best way: dedicated machine with RTX 4090. Install Ollama on this machine and configure endpoint in extension settings to offload to this machine. Second best way: run on MacBook M1/M2/M3 with enough RAM (more == better, but 10gb extra would be enough). For windows notebooks: it runs good with decent GPU, but dedicated machine with a good GPU is recommended. Perfect if you have a dedicated gaming PC.

Local Installation

Install Ollama on local machine and then launch the extension in VSCode, everything should work as it is.

Remote Installation

Install Ollama on dedicated machine and configure endpoint to it in extension settings. Ollama usually uses port 11434 and binds to 127.0.0.1, to change it you should set OLLAMA_HOST to 0.0.0.0.

Models

Currently Llama Coder supports only Codellama. Model is quantized in different ways, but our tests shows that q4 is an optimal way to run network. When selecting model the bigger the model is, it performs better. Always pick the model with the biggest size and the biggest possible quantization for your machine. Default one is stable-code:3b-code-q4_0 and should work everywhere and outperforms most other models.

Name RAM/VRAM Notes
stable-code:3b-code-q4_0 3GB
codellama:7b-code-q4_K_M 5GB
codellama:7b-code-q6_K 6GB m
codellama:7b-code-fp16 14GB g
codellama:13b-code-q4_K_M 10GB
codellama:13b-code-q6_K 14GB m
codellama:34b-code-q4_K_M 24GB
codellama:34b-code-q6_K 32GB m
  • m - slow on MacOS
  • g - slow on older NVidia cards (pre 30xx)

Troubleshooting

Most of the problems could be seen in output of a plugin in VS Code extension output.

Changelog

[0.0.14]

  • Ability to pause completition (by @bkyle)
  • Bearer token support for remote inference (by @Sinan-Karakaya)

[0.0.13]

  • Fix remote files support

[0.0.12]

  • Remote support
  • Fix codellama prompt preparation
  • Add trigger delay
  • Add jupyter notebooks support

[0.0.11]

  • Added Stable Code model
  • Pause download only for specific model instead of all models

[0.0.10]

  • Adding ability to pick a custom model
  • Asking user if they want to download model if it is not available

[0.0.9]

  • Adding deepseek 1b model and making it default

[0.0.8]

  • Improved DeepSeek support and language detection

[0.0.7]

  • Added DeepSeek support
  • Ability to change temperature and top p
  • Fixed some bugs

[0.0.6]

  • Fix ollama links
  • Added more models

[0.0.4]

  • Initial release of Llama Coder
MIT License Copyright (c) 2023 Open Source Contributors Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.

简介

暂无描述 展开 收起
README
MIT
取消

发行版

暂无发行版

贡献者

全部

近期动态

不能加载更多了
马建仓 AI 助手
尝试更多
代码解读
代码找茬
代码优化
1
https://gitee.com/littleketty/llama-coder.git
git@gitee.com:littleketty/llama-coder.git
littleketty
llama-coder
llama-coder
main

搜索帮助