• @[email protected]
    link
    fedilink
    English
    591 month ago

    Until we either solve the problem of LLMs providing false information or the problem of people being too lazy to fact check their work, this is probably the correct course of action.

    • @[email protected]
      link
      fedilink
      English
      271 month ago

      I can’t imagine using any LLM for anything factual. It’s useful for generating boilerplate and that’s basically it. Any time I try to get it to find errors in what I’ve written (either communication or code) it’s basically worthless.

      • @[email protected]
        link
        fedilink
        English
        51 month ago

        My little brother was using gpt for homework and he asked it the probability of extra Sunday in a leap year(52 weeks 2 days) and it said 3/8. One of the possible outcomes it listed was fkng Sunday, Sunday. I asked how two sundays can come consecutively and it made up a whole bunch of bs. The answer is so simple 2/7. The sources it listed also had the correct answer.

        • @[email protected]
          link
          fedilink
          English
          41 month ago

          All it does it create answers that sound like they might be correct. It has no working cognition. People that ask questions like that expect a conversation about probability and days in a year. All it does is combine the two, it can’t think about it.

      • @[email protected]
        link
        fedilink
        English
        31 month ago

        Really? It spotted a missing push_back like 600 lines deep for me a few days ago. I’ve also had good success at getting it to spot missing semicolons that C++ compilers can’t because C++ is a stupid language.

        • BrikoXM
          link
          fedilink
          English
          51 month ago

          You can thank all open source developers for that by supporting them.

            • BrikoXM
              link
              fedilink
              English
              61 month ago

              All LLMs are trained on open source code without any acknowledgment or compliance with the licenses. So their hard work is responsible for you being able to take advantage of it now. You can say thank you by supporting them.

              • @[email protected]
                link
                fedilink
                English
                21 month ago

                Ah yes, I am aware. Gotta love open source :)

                Were you under the impression that I said anything to the contrary?

                • BrikoXM
                  link
                  fedilink
                  English
                  31 month ago

                  No, just taking any opportunity to spread the word and support open source.

      • WIZARD POPE💫
        link
        fedilink
        English
        11 month ago

        I find it useful for quickly reformating smaller sample sizes of tables and similar for my reports. It’s often far simpler and quicker to just drop that in there and say what to dp than to program a short python script

        • @[email protected]
          link
          fedilink
          English
          21 month ago

          It’s probably just the novelty wearing off. People expected very little from it initially, then it got hyped up. This raised expectations. Combining the raised expectations with the memory of it exceeding expectations will let you see all the flaws.

    • @[email protected]
      link
      fedilink
      English
      -31 month ago

      Imo the human laziness is the issue. Every thread where a lot of people chime in about ai, so many talking about how it’s useless because it’s wrong sometimes. It’s basically like people who use Wikipedia but can’t be bothered to cross reference… Except lazier. They literally expect a machine to be flawless because it seems confident or something?

      • @[email protected]
        link
        fedilink
        English
        121 month ago

        I think you’re missing the point. I don’t like copilot/chat gpt for important stuff because if I have to double check their solutions I barely gained any time. Especially since it’s correct more often than not because it will make me complacent over enough time (the professors who were patient enough to actually explain why we shouldn’t be using Wikipedia as a primary source also used the same point which I thought made a lot of sense).

        • @[email protected]
          link
          fedilink
          English
          11 month ago

          You’re going to need to fact check any code you get online anyways, why not have it hyper specific to your current use case? If you’re a good developer, review does not take nearly as long as manual implementation

          • @[email protected]
            link
            fedilink
            English
            6
            edit-2
            1 month ago

            I very rarely grab code online because I work in videogames and it’s very hard to find good code for the things I struggle with since all the publicly available stuff is for hobbyists and thus usually very basic/unoptimized as hell

            Most of the time the stuff I can’t figure out myself isn’t even mentioned anywhere on hobbyist forums because it’s not needed for these applications (for a recent example: assets management. For hobby projects you can usually get away with hard references to all of your assets, so it’s not even a thing)

            • @[email protected]
              link
              fedilink
              English
              3
              edit-2
              1 month ago

              If what you want is difficult to find publicly, then that also means an LLM is going to be weak in that area as well

              What you want is a “general AI” LLM, something capable of stringing together a solution based on past somewhat related solutions. We’re not here yet, so basically you’re asking it to do something beyond what it is capable of and it’s trying its best anyways

              Alternatively, you could try fine tuning your own LLM, if you have access to some sort of large repository with non-public solutions or something

            • @[email protected]
              link
              fedilink
              English
              11 month ago

              So you’re rewriting the wheel every time? I also have worked in games and we definitely utilized public resources whenever possible to save time/money. Asset management in particular has a lot of resources unless you’re talking about truly huge scale things like MMO scale streaming stuff.

  • @[email protected]
    link
    fedilink
    English
    251 month ago

    I’m lead 365 admin for a major corporation and have been working with MS to identify if Copilot would be beneficial and secure for my org. Some major takeaways from my recent meetings with them:

    There’s two parts to Copilot. 1. Copilot 2. Copilot for 365.

    The first is basically Chat GPT. It reaches out to the web to get info and essentially works as a search engine.

    The 2nd part is internal only. It can do things like summarize meetings, compare documents, and search your emails. It abides by the same security, compliance, encryption, and DLP policies as the rest of your tenant.

    You can open up access to one or both.

    Government tenants are a unique case. There’s a specific 365 license for government entities, and their offerings are different from other organizations. This news article isn’t surprising - all new 365 offerings take a while before they’re available to government licenses. It will eventually be available.

    • @[email protected]
      link
      fedilink
      English
      10
      edit-2
      1 month ago

      Few questions about that, unless they’re literally taking their model and putting it into your own box using it’s own compute power, I don’t see how that’s possible. They can call it “your” copilot all they want but if they’re reading your data and prompts and computing that on their own box then they’re using your data, right?

      • @[email protected]
        link
        fedilink
        English
        31 month ago

        Major organizations use encryption where they hold the keys so Microsoft is unable to read their data. They can have thousands of servers running on Microsoft’s Azure stack and yet Microsoft is unable to read the data that is being processed.

        • @[email protected]
          link
          fedilink
          English
          41 month ago

          If all auditors are uncorrupted, highly competent and have full overview. Boeing was able to corrupt it’s government auditors to save some money on redundant sensors. With Microsoft pushing big on gathering and selling data I wouldn’t trust a byte that passes their server.

          • @[email protected]
            link
            fedilink
            English
            01 month ago

            Microsoft has to compete with other cloud providers on security. Unlike Boeing who has no domestic competition. Any of Google, Amazon, or Oracle would love to find out that Microsoft is decrypting user data to sell to partners because they would be screaming to the high heavens that O365/Azure is insecure and enterprises must switch to their solutions. SaaS/IaaS subscriptions are much more profitable than selling user data, there is a near 0 chance that Microsoft is improperly handling enterprise data (on purpose)

      • @[email protected]
        link
        fedilink
        English
        21 month ago

        I’m not an admin, but I do provision ms cloud licensing and have run across this question more than a few times. At the enterprise level, I’m told the copilot data is “walled off” and secure, and not harvested by MS. I have nothing to back that up, but that’s what I’m told. I’m certain if it weren’t true, I would have heard about it by now.