• SGforce@lemmy.ca
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    1
    ·
    edit-2
    11 days ago

    Local models are not capable of coding yet, despite what benchmarks say. Even if they get what you’re trying to do they spew out so many syntax errors and tool calling problems that it’s a complete waste of time. But if you’re using an API then I don’t see why not one editor over another. They’ll be different in implementation but generally pull off the same things

    • brucethemoose@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      ·
      edit-2
      11 days ago

      Local models are not capable of coding yet, despite what benchmarks say. Even if they get what you’re trying to do they spew out so many syntax errors and tool calling problems that it’s a complete waste of time.

      I disagree with this. Qwen Coder 32B and on have been fantastic for niches with the right settings.

      If you apply a grammar template and/or start/fill in their response, drop the temperature a ton, and keep the actual outputs short, it’s like night and day vs ‘regular’ chatbot usage.

      TBH one of the biggest problems with LLM is that they’re treated as chatbot genies with all sorts of performance-degrading workarounds, not tools to fill in little bits of text (which is what language models were originally concieved for).

    • hendrik@palaver.p3x.de
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      edit-2
      11 days ago

      Alright. I mean I haven’t used local models for coding. This was ChatGPT, AIstudio and Grok I tried. I can’t try Claude, since they want my phone number and I’m not going to provide that to them. I feel DeepSeek and a few other local models should be able to get somewhere in the realm of commercial services, though. At least judging by the coding benchmarks, we have some open-weight competition there.