lynx
lynx@sh.itjust.works
Joined
0 posts • 4 comments
If you want in line completions, you need a model that is trained on “fill in the middle” tasks. On their Huggingface page they even say that this is not supported and needs fine tuning:
We do not recommend using base language models for conversations. Instead, you can apply post-training, e.g., SFT, RLHF, continued pretraining, etc., or fill in the middle tasks on this model.
A model that can do it is:
- starcoder2
- codegemma
- codellama
Another option is to just use the qwen model, but instead of only adding a few lines let it rewrite the entire function each time.
I like that every bar has a different scaling.