• Snot Flickerman
    link
    fedilink
    English
    20
    edit-2
    2 months ago

    This comes up again and again:

    AI only does a half-ass job, so you need a real human stepping in to “fix it in post.”

    We’re seeing corporations throwing money hand over fist at AI because corporations want it to replace workers.

    We’re burning an extra planets worth of energy for something that still needs human intervention to be usable.

    Maybe, just maybe, we could just pay humans a living fucking wage to do the same work to begin with instead of constantly trying to find more and more ways to just not pay people at all.

    Like, shocker, if you have a fully staffed customer service department, you actually will solve problems for people faster than an AI staffed customer service department.

    The corpos don’t care, they’re not actually interested in solving our problems.

    • just another devA
      link
      fedilink
      English
      32 months ago

      That’s a a bit too absolute way to look at it.

      From their point of view the goal isn’t to abolish human involvement, but to minimise the cost. So if they can do the job at the same quality with a quarter of the personnel through AI assistance for less cost, obviously they’re gonna do that.

      At the same time, just because humans having crappy jobs is the current way we solve the problem of people getting money, doesn’t mean we should keep on doing that. Basic income would be a much nicer solution for that, for example. Try to think a bit less conservatively.

      • Snot Flickerman
        link
        fedilink
        English
        62 months ago

        but to minimise the cost

        What about the cost to the environment? That cost is just a negative externality to them and you, apparently. Yet I’m the one accused of thinking “conservatively.”

        Burning ten times as many fossil fuels to “minimise the costs” is literally fucking stupid and short-sighted.

        • @[email protected]
          link
          fedilink
          42 months ago

          In the end it’s about money. If one had to pay for environmentally damages (e.g. a new tax on $energyUnit, $resourceUnit,…) and you’d not only pay for the resources + some markup for the producing company and just external externalize the “worth” of the damages (read: the taxpayer,…), then it’s cheaper to use these services instead of humans.

  • @[email protected]
    link
    fedilink
    132 months ago

    This was pretty clear when observing the output of tldrbot. It would just randomly select paragraphs, ignoring surrounding context, and call it a summary.

    • NaN
      link
      fedilink
      English
      32 months ago

      The bot demonstrated very well what this article is about. I don’t know the internals, but I also can’t image the bot was using the best and most expensive ways of doing analysis.

      It was pretty bad at “getting the point” even when it was obvious, a better system should be able to do so. Sometimes the point is more difficult to discern and there has to be some judgement, you can see this in comments sometimes where people discuss what “the point” was and not just the data. I imagine an AI would have some difficulty determining what is worth summarizing in these situations especially.

  • @[email protected]
    link
    fedilink
    English
    122 months ago

    Anecdotally, this was my experience as a student when I tried to use AI to summarize and outline textbook content. The result says almost always incomplete such that I’d have to have already read the chapter to include what the model missed.

    • just another devA
      link
      fedilink
      English
      42 months ago

      I’m not sure how long ago that was, but LLM context sizes have grown exponentially in the past year, from 4k tokens to over a hundred k. That doesn’t necessarily affect the quality of the output, although you can’t expect it to summarize what it can’t hold on memory.

  • @[email protected]
    link
    fedilink
    12 months ago

    Maybe we just need a different type of NLP to work with summarization. I have noticed before LLMs are unlikely to escape their ‘base’ knowledge.

    • ☆ Yσɠƚԋσʂ ☆OP
      link
      fedilink
      22 months ago

      That’s very likely the case. LLMs sucking up all the air in ML research right now, and we shouldn’t be using them as a hammer for every problem.