• ☆ Yσɠƚԋσʂ ☆@lemmy.mlOP
    link
    fedilink
    arrow-up
    6
    ·
    24 days ago

    I’ve found coding assistance to be pretty lacklustre myself as well. That said, one area where language models might actually be good is emulating a type system for dynamic languages. Given how good these things are at figuring out general shape of the code, I suspect they could fairly accurately tell you argument and return types for functions. And you could probably get away with a pretty small model if it only targets a specific language.

    • Kache@lemm.ee
      link
      fedilink
      arrow-up
      6
      arrow-down
      1
      ·
      24 days ago

      Dedicated incremental static type checkers for dynamic languages already exist. In particular, Pyright for Python is fantastic and in many ways surpasses the type systems of classic typed languages

      • ☆ Yσɠƚԋσʂ ☆@lemmy.mlOP
        link
        fedilink
        arrow-up
        1
        ·
        24 days ago

        I’m not too familiar with tooling for Python, but my experience is that you get fairly limited support in dynamic languages unless you start adding hints. Ultimately, a static type checker can’t resolve information that’s not there.

  • 1984@lemmy.today
    link
    fedilink
    arrow-up
    7
    arrow-down
    2
    ·
    24 days ago

    For devops, it’s amazing. We use many tools that we are not experts in, and it’s incredible to get ready to use code examples how to configure them for various scenarios.

    I save many hours every week using open Ai latest models.

    • mihor@lemmy.ml
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      24 days ago

      I used it for GUI code as well, I hate frontend work so that saves me quite some time getting all the forms together with a few prompts.

  • loathsome dongeater@lemmygrad.ml
    link
    fedilink
    English
    arrow-up
    5
    ·
    24 days ago

    How do the SaaS AI code assistants work? I am guessing they have to send the entire file or the codebase to their datacenter. Won’t this be a problem for corporations who want to protect their codebase?

    • onoki@reddthat.com
      link
      fedilink
      arrow-up
      4
      ·
      edit-2
      24 days ago

      The corporations have their own contracts with e.g. Microsoft/OpenAI. The data will likely be sent the same way as with the public tools, but the providers promise not to use it for other purposes.

    • ☆ Yσɠƚԋσʂ ☆@lemmy.mlOP
      link
      fedilink
      arrow-up
      2
      ·
      24 days ago

      I think it’s definitely an issue that the code gets sent over. While, as onoki points out, the providers promise to keep the data private this still opens up a problem that their infrastructure could be compromised.