Buried in the story was a deceptively simple question: does your AI agent count as an employee?

At a recent conference, Microsoft executive Rajesh Jha floated a provocative idea. In a future where companies deploy fleets of AI agents, those agents may need their own identities — logins, inboxes, and even seats inside software systems. If so, AI wouldn’t shrink software revenue. It could expand it.

  • edgemaster72@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    MicroSlop: We have this AI for you to use so you can reduce workforce and associated costs

    Also Sloppy: j/k, fuck you pay me

  • SpatchyIsOnline@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    So the “amazing tool of the future” that’s “going to make software developers obsolete” is also going to need to buy software licenses?

    Which one is it Microslop?

  • favoredponcho@lemmy.zip
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    I think what they’re missing is that it becomes trivial to build software. If there is a license fee, someone will just have AI generate a version of that software that does not require a license. Software companies have no moat anymore.

    • KairuByte@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      I think you vastly overestimate the capabilities of these things, and vastly under estimate the complexity of a lot of different software.

      • smeenz@lemmy.nz
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 month ago

        Exactly… it can take decades to create the level of feature debt we see in software these days.

      • favoredponcho@lemmy.zip
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        1 month ago

        Building software for humans is over. I say this and I build software for a living. I don’t write code anymore. No one I work with writes code anymore.

        Everyone is at different stages of acceptance with this, so I understand people having an attitude about this. It doesn’t change anything.

        • fodor@lemmy.zip
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          1 month ago

          Private companies want the AI cash to ride the bubble. So they’ll use AI, or say they are, just to get the investment money. Doesn’t mean it’s good, true, or worthwhile, or efficient.

          The real test is what the open source community does. And right now, they aren’t doing what you’re doing. We’ll see what the future brings, but I don’t trust your gut any more than I trust my own.

            • KairuByte@lemmy.dbzer0.com
              link
              fedilink
              English
              arrow-up
              0
              ·
              1 month ago

              Cool, so until that point in time, my point still stands. You can’t just hand waive and say “it’ll happen eventually” and be expected to be taken seriously.

              • favoredponcho@lemmy.zip
                link
                fedilink
                English
                arrow-up
                0
                ·
                1 month ago

                You picked one arbitrary example and hold it up as proof that no one can build complex apps with AI? You know there is more than one example of a complex app. Apple has reported an 84% increase in App Store submissions. That’s pretty much all AI driven.

                • KairuByte@lemmy.dbzer0.com
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  1 month ago

                  You claimed that someone could just whip up a license free version of any tool to bypass the cost of a license fee. Was my choice of photoshop arbitrary? Absolutely, you didn’t give any sort of qualifier as to what counts as “any tool”.

                  You can’t both claim that anyone will just use AI to build any tool, and then complain that my choice of tool is arbitrary so should be discounted.

                  App Store submissions isn’t a good metric of complex applications. “The fart app” is an app that any AI tool could make, that anyone could then submit, but is in no way complex. Vibe coded apps have taken off, but what was the last long term (even 1 year) successful vibe coded app? Because the vast majority of the news I hear about vibe coded apps is how they had a major security breach.

                • fodor@lemmy.zip
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  1 month ago

                  Right right. There’s an AI bubble, and there are AI scams. Of course people will ride the bubble, and scammers will always be with us. Doesn’t mean any of that work is quality, or that it will edge out the other work.

  • utopiah@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    That’s the beauty of totally arbitrary restrictions, you can change them as you want.

    Pay by seat? Pay by client? Pay by byte of data stored? Pay by backup location?

    … pay by moonphase? Pay by AI personality? Pay by virtual AI seat?

    Such BS but why wouldn’t Microslop extend its business model. It worked well so far. It’s not about software, or datacenter, or AI, it’s just about entrenchment.

    • tehfishman@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      It’s also a billing strategy that only works in a monopoly situation. If there was healthy competition and no vendor lock-in for the office suite of tools, Microsoft wouldn’t be able to even float this as an idea.

  • DarkSurferZA@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    MMM, interesting. Would the AI companies then need to buy a license for all the information they stole to train their AI? Or would they need to buy a license everytime someone uses micro-slop AI to ask it a question about something that has been trademarked?

    Or does licencing only apply to their software

  • CatAssTrophy@safest.space
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    This gets close to an idea I heard long ago that I think has some merit.

    Hire an employee? You must not only pay them, but cover taxes to have them there. Buy a robot to replace them? It’s a business expense, no taxes!

    Okay, pay taxes for your robot usage. Use that money to fund UBI, social programs and/or retraining people for other jobs.

    • muusemuuse@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      Then they’ll just make one robot do multiple things. Suddenly the big company only has one taxable employee.

      • CatAssTrophy@safest.space
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        1 month ago

        Depends. If the tax is based on jobs replaced, not the abstractly defined number of robots that exist, it would have an impact. Also, monolithic solutions tend to be inherently less efficient than similarly developed defined ones, so limiting the robot models for a tax benefit would have another limit on their efficiency.

        It’s an issue that could be accounted for, if there were sufficient political will. If taxes from automation were committed to public good, there would likely be pretty widespread acceptance.

  • pdxfed@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    1 month ago

    The natural extension of a non-open internet ala Reddit and charging developers for API pulls.

  • lowspeedchase@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    Reads: Our flagship operating system and services have gotten to the point of such terrible shite for humans that we need to pivot to a less discerning customer base.

  • db2@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    A house of cards built on top of ten other houses of cards. What could possibly go wrong.

  • deliriousdreams@fedia.io
    link
    fedilink
    arrow-up
    0
    ·
    1 month ago

    If the AI Agent counts as an employee then the company “employing” it is liable for what it does.

    My guess is the argument will be that “it’s a tool”, not an employee, and therefore they take no responsibility. Though I’m sure that argument is not going to fly for very long. If your air hammer harms someone because the person operating it wasn’t using it correctly, you’re still liable.

    • gokayburuc@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      1 month ago

      Chain fraud activities are being carried out in chain systems like n8n, where AI agents are used together. It didn’t take them long to create systems that generate deepfake voices to sound like real people, directing users to buy a product or deposit money into an account. Many videos on this topic have surfaced in Türkiye, particularly on YouTube. If the users and system creators are to be penalized, then of course, information logs regarding these agents can be used.

      However, if this is being done to keep some agents out of the system using user license fees, it will completely backfire.

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      0
      ·
      1 month ago

      I don’t see how this distinction affects the question of responsibility at all. If anything, “it’s an employee” gives the company more room for deniability.

    • village604@adultswim.fan
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      1 month ago

      What? Companies aren’t liable if the user doesn’t follow the instructions or warnings and hurts themselves.

      DeWalt isn’t liable because I was using their mini chainsaw while holding a branch with my bare hand and the saw bounced and cut me. I’m liable for being stupid.

      • deliriousdreams@fedia.io
        link
        fedilink
        arrow-up
        0
        ·
        1 month ago

        I don’t think you understand the context of the situation I was proposing. I am not supposing that DeWalt would be liable. But let’s say we work in a shop together and I’m using an air hammer to I dunno. Punch rivets. If I as an employee of that shop use the air hammer and something involving the air hammer happens to my coworker or a customer or whatever, it is extremely likely that the company I work for would be on the hook. Could they try to penalize me personally? Yes. Could the person who was injured sue me personally? Certainly. Would the company be off the hook if the air hammer malfunctioned causing injury? Maybe - And at that point I would expect the manufacturer to be liable. But my comment never mentioned the manufacturer.

        The context was companies using AI as a tool not companies manufacturing AI.

  • WesternInfidels@feddit.online
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 month ago

    This is going to wind up granting AI agents a piecemeal, half-assed, legal-fiction version of “personhood,” like corporations have. The AIs will wind up with freedoms like: They can spend all the money they want, that’s “free speech.”

    And the fleshy unfortunates among us still won’t have a right to a living wage, to medical care, etc.