• Flying Squid@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    ·
    7 months ago

    If your code doesn’t work because you rely entirely on an AI to do it, you don’t have a business you can run unless you want to go back to paper and pencil.

    • Jeena@jemmy.jeena.net
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      2
      ·
      7 months ago

      If your code doesn’t work because you rely on humans understanding it, you don’t have a business you can run. We already are there where humans have no idea why the computer does this or that decision because it’s so complex especially with all the machine learning and complex training data, etc. let’s not pretend it will get less complex with time.

      • Flying Squid@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        7 months ago

        So your argument is that people will rely on AI entirely without making any redundancies, unlike now where they have more than one human so they can check for these issues because humans make coding errors?

        • enkers@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          1
          ·
          edit-2
          7 months ago

          I kinda agree with them. Currently coding already is an abstraction. The average developer has very little idea what machine code their compiler actually produces, and for the most part they don’t need to care about this. Feeding an AI a specification is just a higher level of abstraction.

          For now, we’ll need people to check that AI produces code that does what we expect, but I believe at some point we’ll mostly take it for granted that they just do.

        • Jeena@jemmy.jeena.net
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          7 months ago

          My argument is that already today no human is able to and checks it when it comes to decision making models like for example if the car should go left or right around a obstacle. And over time we will have less straight forward classical programming doing decisions and more and more models doing decisions with hundreds or thousands of sensor inputs.

            • Jeena@jemmy.jeena.net
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              1
              ·
              7 months ago

              That means that it right now can not be error checked and it will be even more difficult in the future.

              • Flying Squid@lemmy.world
                link
                fedilink
                English
                arrow-up
                2
                ·
                7 months ago

                So you’re saying no code would be worth error checking by a human at all? There is no level of simpler code that an AI could get wrong and would need someone to fix it?

                • Jeena@jemmy.jeena.net
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  arrow-down
                  1
                  ·
                  7 months ago

                  My assumption is that it would even be the other way around, AI checking all human code, especially writing all the tests. So I guess AI would also write tests for it’s own code too.

                  When it comes to humans, I think they would probably be changing the prompt they give to AI instead of changing the code at the end. You can already see how it’s done with generation of pictures, while theoretically you could take a not so perfect AI picture and use Photoshop to fix it, but most of the time people change the prompt instead and regenerate the picture.

                  • Flying Squid@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    3
                    ·
                    7 months ago

                    The generation of pictures is full of fuckups like giving people extra legs, so I’m not sure that’s a very good example.

              • Jeena@jemmy.jeena.net
                link
                fedilink
                English
                arrow-up
                2
                arrow-down
                1
                ·
                7 months ago

                Or let me rephrase it with the context of the original assumption that if people don’t check the code which AI wrote the company will lose customers because the quality is bad.

                Right now there are tons of models out there which no human can understand why they decide this or that, still they bring so much more value that they get shipped even though they make some mistakes. If a company would try to only ship code checked by humans they would not be able to ship the products and would lose their customers to a company which does not check it but does ship.

          • lemmyvore@feddit.nl
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            7 months ago

            Except we already have fields (like pharma manufacturing) that have to deal with hundreds or thousands of inputs and variables, are automated, and we still manage to fully understand the stack as well as fully check everything.

            Hint: when someone tells you they “can’t” check or understand what their software is doing, it’s a scam.

            Normally they should be told to go back and figure it out before being allowed to ship any product. If you tried this in any other industry it would be laughable. Even in software it’s outrageous, imagine getting accounting software or even a simple file backup tool that doesn’t work some of the time and nobody can tell you how it works. Yet these companies get a pass putting cars like this on the road.