This is a nice post, but it has such an annoying sentence right in the intro:

At the time I saw the press coverage, I didn’t bother to click on the actual preprint and read the work. The results seemed unsurprising: when researchers were given access to AI tools, they became more productive. That sounds reasonable and expected.

What? What about it sounds reasonable? What about it sounds expected given all we know about AI??

I see this all the time. Why do otherwise skeptical voices always have the need to put in a weakening statement like this. “For sure, there are some legitimate uses of AI” or “Of course, I’m not claiming AI is useless” like why are you not claiming that. You probably should be claiming that. All of this garbage is useless until proven otherwise! “AI does not increase productivity” is the null hypothesis! It’s the only correct skeptical position! Why do you seem to need to extend benefit of the doubt here, like seriously, I cannot explain this in any way.

  • sturger
    link
    fedilink
    English
    52 months ago

    Of course, AI makes people more productive! How do I know? I asked my AI, “Do you make people more productive?” and it said, “Yes.” Just look at all the time it just saved me!

  • ________
    link
    fedilink
    English
    202 months ago

    Much like blockchain the FOMO is so strong people are afraid to say it’s bad even when there is nonstop evidence rolling in. With all the data they still are too cowardly to say anything critical.

    • @[email protected]
      link
      fedilink
      English
      92 months ago

      I feel the C-SUITE executives are pushing the AI way harder than they ever pushed crypto though, since they never understood the tech beyond a speculative asset, but the idea of replacing work-hours by AI-automation has been sold HARD to them.

  • @[email protected]
    link
    fedilink
    English
    142 months ago

    This was bizarre to me, as very few companies do massive amounts of materials research and which also is split fairly evenly across the spectrum of materials, in disparate domains such as biomaterials and metal alloys. I did some “deep research” to confirm this hypothesis (thank you ChatGPT and Gemini)

    “I know it’s not actually research, but I did it anyway.”

  • @[email protected]
    link
    fedilink
    English
    82 months ago

    “For sure, there are some legitimate uses of AI” or “Of course, I’m not claiming AI is useless” like why are you not claiming that.

    Yes, thank you!! I’m frustrated by that as well. Another one I have seen way too often is “Of course, AI is not like cryptocurrency, because it has some real benefits [blah blah blah]”… uhm… no?

    As for the “study”, due to Brandolini’s law this will continue to be a problem. I wonder whether research about “AI productivity gains” will eventually become like studies about the efficacy of pseudo-medicine, i.e. the proponents will just make baseless claims that an effect were present, and that science is just not advanced enough yet to detect or explain it.

  • @[email protected]
    link
    fedilink
    English
    142 months ago

    Reads “Does AI make researchers more productive? What? Why would it?”

    Thinks “When does statistically likely text without relation to truth make researchers more productive? Well, when they are faking research”

    Gets to article. Article is about faking research about AI making researchers more productive.

  • David GerardM
    link
    fedilink
    English
    82 months ago

    If you’re after streams-crossing - this guy is a rationalist who does Manifold Markets

      • David GerardM
        link
        fedilink
        English
        32 months ago

        I mean this post seems largely correct and reasonable, but ehh be a little cautious

  • @[email protected]
    link
    fedilink
    English
    42 months ago

    gpt-like AI is useful for what I’m doing (and others in similar boats), because I’m doing lore and world building for a fictional setting that almost nobody but me knows about. If the lying machine lies, that’s ok, because I can just choose to use it or not.

    Actual research on real world subjects should not use gpt-like AI. They’re trying to discover about an unknown whatever, and the lying machine is of course going to fill in gaps with plausible sounding bullshit.

    Anyone who is both 1) paying attention, and 2) isn’t pushing an agenda, already knows this.

  • @[email protected]
    link
    fedilink
    English
    14
    edit-2
    2 months ago

    Because we are witnessing the birth of a religion, it just happens to be a lame, very cult-like one that is friends with everyone in power.