Ask HN: Why don't programming language foundations offer "smol" models?
No synthesized answer yet. Check the discussion below.
Because if you want to prompt it in English, it has to be good at English as well. And it gets good at English by reading extreme quantities of it. Which incidentally is written on a wide variety of topics.
I also wonder this.
My suspicion — based on what I experienced with local image generating models, but otherwise poorly educated — is that they need all of the other stuff besides programming languages just to understand what your plain English prompt means in the first place, and they need to be quite bulky models to have any kind of coherency over token horizons longer than one single function.
Of interest: Apple does ship a coding LLM in Xcode that's (IIRC) 2 GB and it really just does feel like fancy Swift-only autocomplete.