• MudMan@fedia.io
    link
    fedilink
    arrow-up
    1
    ·
    2 days ago

    I don’t know, some of these guys have acccess to a LOT of code, and even more debate about what those good codebases entail.

    I think the other issue is more relevant. Even 128K tokens is not enough for something really big, and the memory and processing costs for that do skyrocket. People are trying to work around it with draft models and summarization models, so they try to pick out the relevant parts of a codebase in one pass and then base their code generation just on that, and… I don’t think that’s going to work reliably at scale. The more chances you give a language model to lose their goddamn mind and start making crap up unsupervised the more work it’s going to be to take what they spit out and shape it into something reasonable.