• David GerardOPM
    link
    fedilink
    English
    245 days ago

    ahahaha holy shit. I knew METR smelled a bit like AI doomsday cultists and took money from OpenPhil, but those “open source” projects and engineers? One of them was LessWrong.

    Here’s a LW site dev whining about the study, he was in it and i think he thinks it was unfair to AI

    I think if people are citing in another 3 months time, they’ll be making a mistake

    dude $NEXT_VERSION will be so cool

    so anyway, this study has gone mainstream! It was on CNBC! I urge you not to watch that unless you have a yearning need to know what the normies are hearing about this shit. In summary, they are hearing that AI coding isn’t all that actually and may not do what the captains of industry want.

    around 2:30 the two talking heads ran out of information and just started incorrecting each other on the fabulous AI future, like the worst work lunchroom debate ever but it’s about AI becoming superhuman

    the key takeaway for the non techie businessmen and investors who take CNBC seriously ever: the bubble starts not going so great

    • @[email protected]
      link
      fedilink
      English
      10
      edit-2
      5 days ago

      Yeah, METR was the group that made the infamous AI IS DOUBLING EVERY 4-7 MONTHS GRAPH where the measurement was 50% success at SWE tasks based on the time it took a human to complete it. Extremely arbitrary success rate, very suspicious imo. They are fanatics trying to pinpoint when the robo god recursive self improvement loop starts.

    • @[email protected]
      link
      fedilink
      English
      64 days ago

      I think if people are citing in another 3 months time, they’ll be making a mistake

      In 3 months they’ll think they’re 40% faster while being 38% slower. And sometime in 2026 they will be exactly 100% slower - the moment referred to as “technological singularity”.

    • @[email protected]
      link
      fedilink
      English
      64 days ago

      Here’s a LW site dev whining about the study, he was in it and i think he thinks it was unfair to AI

      There a complete lack of introspection. It seems like the obvious conclusion to draw from a study showing people’s subjective estimates of their productivity with LLMs were the exact opposite of right would inspire him to question his subjectively felt intuitions and experience but instead he doubles down and insists the study must be wrong and surely with the latest model and best use of it it would be a big improvement.