There are lots of general-purpose models to use locally, and also coding-specific models.
But are there models specialized in one programming language? My thought was that a model that only needs to handle one language (e.g. Python) could be faster, or be better for a given size.
E.g If I need to code in Rust and is limited to an 8B model to run locally, I was hoping to get better results with a model that is narrower. I don’t need it to be able to help with Java.
This approach would of course require switching models, but that’s no problem for me.
I finally found a Python-specific model. Searching is tricky when almost everything around local llms involve Python. https://huggingface.co/codellama/CodeLlama-13b-Python-hf
I’m pretty sure this would not produce better output. Ai has a « thought » of his own.limiting is approach to be pure python won’t offer you the link it can see with a Java approach. With a more diverse knowledge, it can generate a more unique output. Not forcing it to be readable by human may give a more efficient solution. Kinda like compiled code.
I’d say it depends. Some patterns will be different. Like trying to create zig code when the LLM is trained mostly on javascript and python.
I’ve seen that some exist on huggingface but Idk if they are any better. I would assume if u finetune the model on specific libraries/ur specific codebase then it would be more accurate. If u can run the model locally u can finetune it. Give it a go.