• @[email protected]
    link
    fedilink
    2392 years ago

    What a silly article. 700,000 per day is ~256 million a year. Thats peanuts compared to the 10 billion they got from MS. With no new funding they could run for about a decade & this is one of the most promising new technologies in years. MS would never let the company fail due to lack of funding, its basically MS’s LLM play at this point.

    • R0cket_M00se
      link
      fedilink
      English
      42 years ago

      Almost every company uses either Google or Microsoft Office products and we already know that they’re working on an AI offering/solution for O365 integration, they can see the writing on the wall here and are going to profit massively as they include it in their E5 license structure or invent a new one that includes AI. Then they’ll recoup that investment in months.

    • @[email protected]
      link
      fedilink
      English
      172 years ago

      While title is click bite, they do say right at the beginning:

      *Right now, it is pulling through only because of Microsoft’s $10 billion funding *

      Pretty hard to miss, and than they go to explain their point, which might be wrong, but still stands. 700k i only one model, there are others and making new ones and running the company. It is easy over 1B a year without making profit. Still not significant since people will pour money into it even after those 10B.

    • @[email protected]
      link
      fedilink
      English
      422 years ago

      Openai biggest spending is infrastructure, Whis is rented from… Microsoft. Even if the company fold, they will have given back to Microsoft most of the money invested

      • @[email protected]
        link
        fedilink
        English
        252 years ago

        MS is basically getting a ton of equity in exchange for cloud credits. That’s a ridiculously good deal for MS.

    • P03 Locke
      link
      fedilink
      English
      1122 years ago

      When you get articles like this, the first thing you should ask is “Who the fuck is Firstpost?”

      • Altima NEO
        link
        fedilink
        English
        352 years ago

        Yeah where the hell do these posters find these articles anyway? It’s always from blogs that repost stuff from somewhere else

    • lemmyvore
      link
      fedilink
      English
      122 years ago

      I mean, you’re correct in the sense Microsoft basically owns their ass at this point, and that Microsoft doesn’t care if they make a loss because it’s sitting on a mountain of cash. So one way or another Microsoft is getting something cool out of it. But at the same time it’s still true that OpenAI’s business plan was unsustainable hyped hogwash.

      • @[email protected]
        link
        fedilink
        English
        32 years ago

        Also, their biggest expenses are cloud expenses, and they use the MS cloud, so that basically means that Microsoft is getting a ton of equity in a hot startup in exchange for cloud credits which is a ridiculously good deal for MS. Zero chance MS would let them fail.

      • @[email protected]
        link
        fedilink
        English
        182 years ago

        Their business plan got Microsoft to drop 10 billion dollars on them.

        None of my shitty plans have pulled that off.

        • lemmyvore
          link
          fedilink
          English
          12 years ago

          If they got any of that into their own pockets kudos to them.

          Mainly they used it to pay for the tech and research and it’s all reverting back to Microsoft eventually. Going bankrupt is not quite the same as being acquired.

  • Tony Bark
    link
    fedilink
    English
    32 years ago

    The thing about all GPT models is that they’re based on the frequency of the word to determine its usage. Which means the only way to get good results is if it’s running on cutting edge equipment designed specifically for that job, while being almost a TB in size. Meanwhile, Diffusion models are only GB and run on the GPU but still produce masterpieces because they already know what that word is associated with.

  • @[email protected]
    link
    fedilink
    English
    32 years ago

    Would help if they would offer more payment options than just credit card, which is not really popular in many countries

  • @[email protected]
    link
    fedilink
    English
    982 years ago

    That would explain why ChatGPT started regurgitating cookie-cutter garbage responses more often than usual a few months after launch. It really started feeling more like a chatbot lately, it almost felt talking to a human 6 months ago.

    • @[email protected]
      link
      fedilink
      English
      632 years ago

      I don’t think it does. I doubt it is purely a cost issue. Microsoft is going to throw billions at OpenAI, no problem.

      What has happened, based on the info we get from the company, is that they keep tweaking their algorithms in response to how people use them. ChatGPT was amazing at first. But it would also easily tell you how to murder someone and get away with it, create a plausible sounding weapon of mass destruction, coerce you into weird relationships, and basically anything else it wasn’t supposed to do.

      I’ve noticed it has become worse at rubber ducking non-trivial coding prompts. I’ve noticed that my juniors have a hell of a time functioning without access to it, and they’d rather ask questions of seniors rather than try to find information our solutions themselves, replacing chatbots with Sr devs essentially.

      A good tool for getting people on ramped if they’ve never coded before, and maybe for rubber ducking in my experience. But far too volatile for consistent work. Especially with a Blackbox of a company constantly hampering its outputs.

      • @[email protected]
        link
        fedilink
        English
        682 years ago

        As a Sr. Dev, I’m always floored by stories of people trying to integrate chatGPT into their development workflow.

        It’s not a truth machine. It has no conception of correctness. It’s designed to make responses that look correct.

        Would you hire a dev with no comprehension of the task, who can not reliably communicate what their code does, can not be tasked with finding and fixing their own bugs, is incapable of having accountibility, can not be reliably coached, is often wrong and refuses to accept or admit it, can not comprehend PR feedback, and who requires significantly greater scrutiny of their work because it is by explicit design created to look correct?

        ChatGPT is by pretty much every metric the exact opposite of what I want from a dev in an enterprise development setting.

        • JackbyDev
          link
          fedilink
          English
          352 years ago

          Search engines aren’t truth machines either. StackOverflow reputation is not a truth machine either. These are all tools to use. Blind trust in any of them is incorrect. I get your point, I really do, but it’s just as foolish as believing everyone using StackOverflow just copies and pastes the top rated answer into their code and commits it without testing then calls it a day. Part of mentoring junior devs is enabling them to be good problem solvers, not just solving their problems. Showing them how to properly use these tools and how to validate things is what you should be doing, not just giving them a solution.

          • @[email protected]
            link
            fedilink
            English
            82 years ago

            I agree with everything you just said, but i think that without greater context it’s maybe still unclear to some why I still place chatGPT in a league of it’s own.

            I guess I’m maybe some kind of relic from a bygone era, because tbh I just can’t relate to the “I copied and pasted this from stack overflow and it just worked” memes. Maybe I underestimate how many people in the industry are that fundamentally different from how we work.

            Google is not for obtaining code snippets. It’s for finding docs, for troubleshooting error messages, etc.

            If you have like… Design or patterning questions, bring that to the team. We’ll run through it together with the benefits of having the contextual knowledge of our problem domain, internal code references, and our deployment architecture. We’ll all come out of the conversation smarter, and we’re less likely to end up needing to make avoidable pivots later on.

            The additional time required to validate a chatGPT generated piece of code could have instead been spent invested in the dev to just do it right and to properly fit within our context the first time, and the dev will be smarter for it and that investment in the dev will pay out every moment forward.

            • JackbyDev
              link
              fedilink
              English
              22 years ago

              I guess I see your point. I haven’t asked ChatGPT to generate code and tried to use it except for once ages ago but even then I didn’t really check it and it was a niche piece of software without many examples online.

        • @[email protected]
          link
          fedilink
          English
          132 years ago

          Don’t underestimate C levels who read a Bloomberg article about AI to try and run their entire company off of it…then wonder why everything is on fire.

        • @[email protected]
          link
          fedilink
          English
          62 years ago

          Honestly once ChatGPT started giving answers that consistently don’t work I just started googling stuff again because it was quicker and easier than getting the AI to regurgitate stack overflow answers.

        • @[email protected]
          link
          fedilink
          English
          62 years ago

          Would you hire a dev with no comprehension of the task, who can not reliably communicate what their code does, can not be tasked with finding and fixing their own bugs, is incapable of having accountibility, can not be reliably coached, is often wrong and refuses to accept or admit it, can not comprehend PR feedback, and who requires significantly greater scrutiny of their work because it is by explicit design created to look correct?

          Not me, but my boss would… wait a minute…

      • @[email protected]
        link
        fedilink
        English
        14
        edit-2
        2 years ago

        Copilot is pretty amazing for day to day coding, although I wonder if a junior dev might get led astray with some of its bad ideas, or too dependent on it in general.

        Edit: shit, maybe I’m too dependent on it.

        • @[email protected]
          link
          fedilink
          English
          62 years ago

          I’m also having a good time with copilot

          Considering asking my company to pay for the subscription as I can justify that it’s worth it.

          Yes many times it is wrong but even if it it’s only 80% correct at least I get a suggestion on how to solve an issue. Many times it suggest a function and the code snippet has something missing but I can easily fix it or improve it. Without I would probably not know about that function at all.

          I also want to start using it for documentation and unit tests. I think there it’s where it will really be useful.

          Btw if you aren’t in the chat beta I really recommend it

          • @[email protected]
            link
            fedilink
            English
            32 years ago

            Just started using it for documentation, really impressed so far. Produced better docstrings for my functions than I ever do in a fraction of the time. So far all valid, thorough and on point. I’m looking forward to asking it to help write unit tests.

            • @[email protected]
              link
              fedilink
              English
              32 years ago

              it honestly seems better suited for those tasks because it really doesn’t need to know anything that you’d have to tell it otherwise.

              The code is already there, so it can get literally all the info that it needs, and it is quite good at grasping what the function does, even if sometimes it lacks the context of the why. But that’s not relevant for unit tests, and for documentation that’s where the user comes in. It’s also why it’s called copilot, you still make the decisions.

    • @[email protected]
      link
      fedilink
      English
      52 years ago

      I am unsure about the free version, but I really am very surprised by how good the paid version with the code interpreter has gotten in the last 4-6weeks. Feels like I have a c# syntax guru on 24/7 access. Used to make lots of mistakes a couple months ago, but rarely does now and if it does it almost always fixes in in the next code edit. It has saved my untold hours.

    • @[email protected]
      link
      fedilink
      English
      18
      edit-2
      2 years ago

      But what did they expect would happen, that more people would subscribe to pro? In the beginning I thought they just wanted to survey-farm usage to figure out what the most popular use cases were and then sell that information or repackage use-cases as an individual added-value service.

  • LemmyLefty
    link
    fedilink
    292 years ago

    Does it feel like these “game changing” techs have lives that are accelerating? Like there’s the dot com bubble of a decade or so, the NFT craze that lasted a few years, and now AI that’s not been a year.

    The Internet is concentrating and getting worse because of it, inundated with ads and bots and bots who make ads and ads for bots, and being existentially threatened by Google’s DRM scheme. NFTs have become a joke, and the vast majority of crypto is not far behind. How long can we play with this new toy? Its lead paint is already peeling.

  • @[email protected]
    link
    fedilink
    English
    822 years ago

    I mean apart from the fact it’s not sourced or whatever, it’s standard practice for these tech companies to run a massive loss for years while basically giving their product away for free (which is why you can use openAI with minimal if any costs, even at scale).

    Once everyone’s using your product over competitors who couldn’t afford to outlast your own venture capitalists, you can turn the price up and rake in cash since you’re the biggest player in the market.

    It’s just Uber’s business model.

    • @[email protected]
      link
      fedilink
      English
      272 years ago

      The difference is that the VC bubble has mostly ended. There isn’t “free money” to keep throwing at a problem post-pan. That’s why there’s an increased focus on Uber (and others) making a profit.

      • @[email protected]
        link
        fedilink
        English
        162 years ago

        This is what caused spez at Reddit and Musk at Twitter to go into desperation mode and start flipping tables over. Their investors are starting to want results now, not sometime in the distant future.

      • @[email protected]
        link
        fedilink
        English
        42 years ago

        I don’t know anything about anything, but part of me suspects that lots of good funding is still out there, it’s just being used more quietly and more scrupulously, & not being thrown at the first microdosing tech wanker with a great elevator pitch on how they’re going to make “the Tesla of dental floss”.

      • FlumPHP
        link
        fedilink
        English
        222 years ago

        In this case, Microsoft owns 49% of OpenAI, so they’re the ones subsidizing it. They can also offer at-cost hosting and in-roads into enterprise sales. Probably a better deal at this point than VC cash.

    • @[email protected]
      link
      fedilink
      English
      122 years ago

      Speaking of Uber, I believe it turned a profit the first time this year. That is, it never made any profit since its creation in whenever it was created.

      • ineedaunion
        link
        fedilink
        English
        162 years ago

        All it’s every done is rob from it’s employees so it can give money to stockholders. Just like every corporation.

  • @[email protected]
    link
    fedilink
    English
    22 years ago

    Whatever, it was always some fly by night operation anyway. It’s a cool toy but all this revolutionary crap talk was just like when NFT’s showed up.

    • @[email protected]
      link
      fedilink
      English
      32 years ago

      well I mean, chatGPT actually does have some real world use. personally, I find chatGPT more helpful than Stack Overflow when it comes to finding problems with my code

    • pachrist
      link
      fedilink
      English
      272 years ago

      ChatGPT has the potential to make Bing relevant and unseat Google. No way Microsoft pulls funding. Sure, they might screw it up, but they’ll absolutely keep throwing cash at it.

      • @[email protected]
        link
        fedilink
        English
        4
        edit-2
        2 years ago

        They seems to be killing Cortana… So I expect a new assistant at least based partially on this tbh.

  • @[email protected]
    link
    fedilink
    English
    62 years ago

    It’s definitely become a part of a lot of people’s workflows. I don’t think OpenAI can die. But the need of the hour is to find a way to improve efficiency multifold. This will make it cheaper, more powerful and more accessible

    • kitonthenet
      link
      fedilink
      52 years ago

      At a $250mm/yr burn rate and a revenue of… a lot less than that, they can die pretty quickly

      • @[email protected]
        link
        fedilink
        English
        22 years ago

        Agreed. But it will be a significant loss for a big chunk of people since other LLMs aren’t nearly as good as GPT-4.

    • Saganastic
      link
      fedilink
      52 years ago

      I think they’re just trying to get people hooked, and then they’ll start charging for it. It even says at the bottom of the page when you’re in a chat:

      Free Research Preview. ChatGPT may produce inaccurate information about people, places, or facts. ChatGPT August 3 Version

      • kitonthenet
        link
        fedilink
        3
        edit-2
        2 years ago

        I don’t think it’s at all clear that that’s a viable business strategy in a market where that kind of sleight of hand is as well known as it is right now

        • Saganastic
          link
          fedilink
          2
          edit-2
          2 years ago

          Except for the fact that they’ve said for the entire existence of chatgpt that it’s a free research preview.

  • @[email protected]
    link
    fedilink
    English
    9
    edit-2
    2 years ago

    Well, I was happily paying them to lewd up the chatbots, but then they emailed me telling me to stop. I guess they don’t want my money.

    • @[email protected]
      link
      fedilink
      English
      4
      edit-2
      2 years ago

      I’ve heard it’s not as good, but I think NovelAI works pretty well, and explicitly allows that.

      • @[email protected]
        link
        fedilink
        English
        22 years ago

        I actually started my journey into Lewd AI stuff with NovelAI. I stopped using it after awhile because I like chatbot rp specifically, not just something that will finish a story for me. Using Silly Tavern to try and emulate the NovelAI models into acting like chat bots just shows how not good they are at that.

        • @[email protected]
          link
          fedilink
          English
          22 years ago

          Makes sense. In that case I guess your next best option is probably to buy or rent hardware to run the local models that are suitable for chat rp.

          • @[email protected]
            link
            fedilink
            English
            22 years ago

            I have definitely been considering it. My current hardware gives me about an 80 second delay when I run an llm locally.

            • @[email protected]
              link
              fedilink
              English
              2
              edit-2
              2 years ago

              Same, at least for anything but the tiny ones that will fit in my limited vram. Hoping a gpu that’s actually good for LLM will come out in the next few years that’s not 15k and made for servers.

    • @[email protected]
      link
      fedilink
      English
      182 years ago

      No sources and even given their numbers they could continue running chatgpt for another 30 years. I doubt they’re anywhere near a net profit but they’re far from bankruptcy.

    • @[email protected]
      link
      fedilink
      English
      22 years ago

      The flow of the writing style felt kinda off, like someone was speaking really fast spewing random trivia and leaving

    • subversive_dev
      link
      fedilink
      English
      112 years ago

      Right!? I believe it has the hallmark repetitive blandness indicating AI wrote it (because oroboros)

  • pitninja
    link
    fedilink
    English
    72 years ago

    If ChatGPT only costs $700k to run per day and they have a $10b war-chest, assuming there were no other overhead/development costs, OpenAI could run ChatGPT for 39 years. I’m not saying the premise of the article is flawed, but seeing as those are the only 2 relevant data points that they presented in this (honestly poorly written) article, I’m more than a little dubious.

    But, as a thought experiment, let’s say there’s some truth to the claim that they’re burning through their stack of money in just one year. If things get too dire, Microsoft will just buy 51% or more of OpenAI (they’re going to be at 49% anyway after the $10b deal), take controlling interest, and figure out a way to make it profitable.

    What’s most likely going to happen is OpenAI is going to continue finding ways to cut costs like caching common query responses for free users (and possibly even entire conversations, assuming they get some common follow-up responses). They’ll likely iterate on their infrastructure and cut costs for running new queries. Then they’ll charge enough for their APIs to start making a lot of money. Needless to say, I do not see OpenAI going bankrupt next year. I think they’re going to be profitable within 5-10 years. Microsoft is not dumb and they will not let OpenAI fail.

    • @[email protected]
      link
      fedilink
      English
      42 years ago

      Details of the 10b aren’t public, but we know it’s a multi year deal, so it’s possible that OpenAI doesn’t actually have the full amount in cash now, and they could go bankrupt before they unlock the full amount. In the event of a bankruptcy, Microsoft could be in a position to acquire their assets for themselves on the cheap.

  • danielbln
    link
    fedilink
    English
    41
    edit-2
    2 years ago

    This article has been flagged on HN for being clickbait garbage.

    • @[email protected]
      link
      fedilink
      English
      14
      edit-2
      2 years ago

      It is clearly no sense. But it satisfies the irrational needs of the masses to hate on AI.

      Tbf I have no idea why. Why do people hate a extremely clever family of mathematical methods, which highlights the brilliance of human minds. But here we are. Casually shitting on one of the highest peak humanity has ever reached

      • @[email protected]
        link
        fedilink
        English
        12 years ago

        People are scared because it will make consolidation of power much easier, and make many of the comfyer jobs irrelevant. You can’t strike for better wages when your employer is already trying to get rid of you.

        The idealist solution is UBI but that will never work in a country where corporations have a stranglehold on the means of production.

        Hunger shouldn’t be a problem in a world where we produce more food with less labor than anytime in history, but it still is, because everything must have a monetary value, and not everyone can pay enough to be worth feeding.

        • @[email protected]
          link
          fedilink
          English
          12 years ago

          I agree with this. People should fight to democratize AI, public model, public data, public fair research. And should fight misuse of it from business schools’ type of guys.

      • @[email protected]
        link
        fedilink
        English
        12 years ago

        I probably sound like I hate it, but I’m just giving my annual “this new tech isn’t the miracle it’s being sold as” warning, before I go back to charging folks good money to clean up the mess they made going “all in” on the last one.

      • FaceDeer
        link
        fedilink
        62 years ago

        It seems to be a common thing. I gave up on /r/futurology and /r/technology over on Reddit long ago because it was filled with an endless stream of links to cool new things with comment sections filled with nothing but negativity about those cool new things. Even /r/singularity is drifting that way. And so it is here on the Fediverse too, the various “technology” communities are attracting a similar userbase.

        Sure, not everything pans out. But that’s no excuse for making all of these communities into reflections of /r/nothingeverhappens. Technology does change, sometimes in revolutionary ways. It’d be nice if there was a community that was more upbeat about that.

  • @[email protected]
    link
    fedilink
    English
    592 years ago

    If AI was so great, it would find a solution to operate at fraction of the cost it does now

    • @[email protected]
      link
      fedilink
      English
      712 years ago

      Wait, has anybody bothered to ask AI how to fix itself? How much Avocado testing does it do? Can AI pull itself up by its own boot partition, or does it expect the administrator to just give it everything?

      • @[email protected]
        link
        fedilink
        English
        52 years ago

        I asked Bard how ChatGPT can fix itself, here’s what it said:

        ChatGPT can fix itself in a number of ways, including:

        Retraining on a larger dataset. ChatGPT is trained on a massive dataset of text and code, but it can always be retrained on a larger dataset. This would help ChatGPT to learn new things and improve its performance.

        Using a different algorithm. ChatGPT is currently trained using the GPT-3 algorithm, but there are other algorithms that could be used. For example, ChatGPT could be trained using the Megatron-Turing NLG algorithm, which is one of the most powerful language models in the world.

        Using a combination of algorithms. ChatGPT could also be trained using a combination of algorithms. This would allow ChatGPT to benefit from the strengths of different algorithms.

        Using a feedback loop. ChatGPT could be used to generate text, and then this text could be used to train ChatGPT. This would allow ChatGPT to learn from its own mistakes and improve its performance over time.

        Using human feedback. ChatGPT could also be used to generate text, and then this text could be reviewed by humans. The human feedback could then be used to improve ChatGPT’s performance.

        • @[email protected]
          link
          fedilink
          English
          102 years ago

          “Using a feedback loop. ChatGPT could be used to generate text, and then this text could be used to train ChatGPT. This would allow ChatGPT to learn from its own mistakes and improve its performance over time.”

          So basically create its own Fox News and see how that goes.

          • FaceDeer
            link
            fedilink
            32 years ago

            The full suggestion includes “This would allow ChatGPT to learn from its own mistakes”, which implies that the text it generated would be evaluated and curated before being sent back into it for training. That, as well as including non-AI-generated text along with the AI generated stuff, should stop model collapse.

            Model collapse is basically inbreeding, with similar causes and similar solutions. A little inbreeding is not inherently bad, indeed it’s used frequently when you’re trying to breed an organism to have specific desirable characteristics.

        • FaceDeer
          link
          fedilink
          42 years ago

          If having an AI tell researchers that they should base its next iteration off of Megatron isn’t the plot of a Michael Bay Transformers movie already, it should have been.

      • @[email protected]
        link
        fedilink
        English
        142 years ago

        Really says something that none of your responses yet seem to have caught that this was a joke.

        • FaceDeer
          link
          fedilink
          32 years ago

          OP might have been intending it as a joke, but self-improvement is a very real subject of AI research so if that’s the case he accidentally said something about a serious topic.

          • @[email protected]
            link
            fedilink
            English
            22 years ago

            It’s an essential part of the idea of the technological singularity. An AI iterates itself and the systems it runs on, becoming more efficient, powerful, and effective at a rate that makes all of human progress up to that point look like nothing.

            • @[email protected]
              link
              fedilink
              English
              32 years ago

              While I’m inclined to believe the singularity is achievable, it’s important to remember that there’s no evidence today that it will ever be reached.

              Our hope for it, and the good than can come with it, can’t pull it into the realm of things we will see in our lifetimes. It could emerge soon, but it’s at least as likely to stay science fiction for another millennia.

              • @[email protected]
                link
                fedilink
                English
                22 years ago

                Yeah, when chat gpt 4 first came out, I thought we might be close. But as it’s capabilities and limitations became more clear, it doesn’t look like we’re close at all. I mean, it’s hard to say for sure since an LLM will just make up a part of an AI and maybe the other pieces are farther along but just not getting as much attention because there’s value in not making those things public.

                But as someone who works in one of the fields that would be involved in the technological singularity, no one really knows good ways to apply AI to the work we do and the best initiatives I’ve seen come out of the corporate drive to leverage AI aren’t actually AI, but just smarter automation tools.

      • discodoubloon
        link
        fedilink
        82 years ago

        It doesn’t “know” anything. It can’t solve that problem. It’s trained on humans so it’s limited to what we have written down.

        I love ChatGPT but if it’s creative it’s because you asked it the right questions and found an oblique answer yourself.

      • @[email protected]
        link
        fedilink
        English
        152 years ago

        if we don’t know, it doesn’t know.

        If we know, but there’s no public text about it, it doesn’t know either.

        it is trained off of stuff that has already been written, and trained to emulate the statistical properties of those words. It cannot and will not tell us anything new

        • FaceDeer
          link
          fedilink
          152 years ago

          That’s not true. These models aren’t just regurgitating text that they were trained on. They learn the patterns and concepts in that text, and they’re able to use those to infer things that weren’t explicitly present in the training data.

          I read recently about some researchers who were experimenting with ChatGPT’s ability to do basic arithmetic. It’s not great at it, but it’s definitely figured out some techniques that allow it to answer math problems that were not in its training set. It gets them wrong sometimes, but it’s like a human doing math in its head rather than a calculator using rigorous algorithms so that’s to be expected.

          • @[email protected]
            link
            fedilink
            English
            52 years ago

            they learn statistical correlations between words. given the last 5000 (or however large the context is) words, and absolutely no other information besides that, what is the most likely word to appear next? It’s a glorified order 5000 markov chain.

            The reason it can “do” some math is that there are tons of examples in the training set using small numbers usually used as examples. it can do basic arithmetic because it has seen “2+2=4” and other examples with simple numbers like that. The studies used test basic arithmetic. The same things that it had millions of pre-worked examples of. And it still gets those wrong, with astonishing frequency. those studies aren’t talking about asking it “what is the square root of pi” or stuff like that. but stuff such as “is 7 greater than 4?”, “what is 10 + 3?”, “is 97 prime?” stuff it has most definitely seen the answers to. ask it about some large prime, and it’ll nay no, and be probably right, because most numbers are composite

            • FaceDeer
              link
              fedilink
              3
              edit-2
              2 years ago

              those studies aren’t talking about asking it “what is the square root of pi” or stuff like that. but stuff such as “is 7 greater than 4?”, “what is 10 + 3?”, “is 97 prime?” stuff it has most definitely seen the answers to.

              No, they very explicitly checked to see whether the training set contains the literal math problem that they asked it for the answer to. ChatGPT is able to answer math questions that it has never seen before. I believe this is the article (though I had to go searching, it’s been a while).

              When people dismiss LLMs as “just prediction engines” they’re really missing the point. Of course they’re prediction engines, that’s not in dispute. The question is about how they go about making those predictions. When I show you the string “18 + 10 =” you can predict what comes next, yes? Well, how did you predict it? Did you memorize that particular specific string, or have you developed heuristics for how to do simple addition problems when you see them?

              • @[email protected]
                link
                fedilink
                English
                1
                edit-2
                2 years ago

                These things are currently infamously bad at math, though.

                I won’t argue that it’ll never get there. I’m confident it will, - though with a lot more perl hacks than elegant emergence.

                But today, these things have an astonishingly high ‘appearance of intelligence’ to ‘incredible stupidity’ ratio.

                • FaceDeer
                  link
                  fedilink
                  32 years ago

                  Humans are also not particularly well known for their math skills. Ask a random stranger to do simple arithmetic in their head, with only a few seconds to think and no outside help, and I wouldn’t expect particularly reliable results.

            • @[email protected]
              link
              fedilink
              English
              1
              edit-2
              2 years ago

              It gives me the giggles that folks speculating are getting more upvotes than your post that demonstrates actual understanding of the implementation details.

              If I were the type to sell sizzle hype and snake oil, now would be the time to do it. The venture capitalists may have learned their lesson, but the general public haven’t.

    • @[email protected]
      link
      fedilink
      English
      8
      edit-2
      2 years ago

      Deepmind is actually working on an AI that improve performances of low level programs. It started with improving sorting algorithm.

      It’s an RL algorithm.

      Main issue is that everything takes time, and expectations on current AI are artificially inflated.

      It will reach the point most are discussing now, it’ll simply take a bit longer than people expect

      Source: https://www.nature.com/articles/d41586-023-01883-4