Falcon-7b fine-tuned on the CodeAlpaca 20k directions dataset by utilizing the tactic QLoRA with PEFT library.
Falcoder 7B Full Mannequin – https://huggingface.co/mrm8488/falcoder-7b
Falcoder Adapter – https://huggingface.co/mrm8488/falcon-7b-ft-codeAlpaca_20k-v2
Falcoder Colab – https://colab.research.google.com/drive/1F5rU85bg45YWQyLnMmBMkX21rm1KC6VZ?usp=sharing
(Colab and Thumbnail Mascot Credit score: Manuel Romero https://twitter.com/mrm8488 )
❤️ If you wish to assist the channel ❤️
Help right here:
Patreon – https://www.patreon.com/1littlecoder/
Ko-Fi – https://ko-fi.com/1littlecoder