• conditional_soup@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      16 days ago

      In the Bible’s book of revelations, John (the author) is witnessing the end of the world and sees four horsemen being unleashed upon the world to spread a curse/trial/whatever wherever they ride. Each horseman brings with them something different- famine, disease, war (or strife), and death. Death is the last, IIRC, and rides upon a pale horse. I think that’s what they’re referencing. This person is saying that openAI is going to die soon.

      • jatone@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        16 days ago

        it won’t. its backed by microsoft. they can literally afford to burn the cash on this while it becomes profitable, and it will AI has so many low hanging fruits to optimize its insane.

          • jatone@lemmy.dbzer0.com
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            16 days ago

            quants are pretty basic. switching from floats to ints (faster instruction sets) are the well known issues. both those are related to information theory, but there are other things I legally can’t mention. shrug. suffice to say the model sizes are going to be decreasing dramatically.

            edit: the first two points require reworking the base infrastructure to support which is why they havent hit widespread adoption. but the research showing that 3 bits is as good as 64 is intuitive once you tie the original inspiration for some of the AI designs. that reduction alone means you can get 21x reduction in model size is pretty solid.

            • self@awful.systems
              link
              fedilink
              English
              arrow-up
              0
              ·
              16 days ago

              both those are related to information theory, but there are other things I legally can’t mention. shrug.

              hahahaha fuck off with this. no, the horseshit you’re fetishizing doesn’t fix LLMs. here’s what quantization gets you:

              • the LLM runs on shittier hardware
              • the LLM works worse too
              • that last one’s kinda bad when the technology already works like shit

              anyway speaking of basic information theory:

              but the research showing that 3 bits is as good as 64 is intuitive once you tie the original inspiration for some of the AI designs.

              lol

              • eestileib@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                0
                ·
                edit-2
                16 days ago

                Honestly, the research showing that a schlong that’s 3mm wide is just as satisfying as one that’s 64 is intuitive once you tie the original inspiration for some of the sex positions.

              • khalid_salad@awful.systems
                link
                fedilink
                English
                arrow-up
                0
                ·
                15 days ago

                It’s actually super easy to increase the accuracy of LLMs.

                import pytorch # or ollama or however you fucking dorks use this nonsense
                from decimal import Decimal
                

                I left out all the other details because it’s pretty intuitive why it works if you understand why floats have precision issues.

              • killingspark@feddit.org
                link
                fedilink
                English
                arrow-up
                0
                ·
                16 days ago

                I have seen these 3 bit ai papers on hacker news a few times. And the takeaway apparently is: the current models are being pretty shitty at what we want them to do, and we can reach a similar (but slightly worse) level of shittyness with 3 bits.

                But that doesn’t say anything about how both technologies could progress in the future. I guess you can compensate for having only three bits to pass between nodes by just having more nodes. But that doesn’t really seem helpful, neither for storage nor compute.

                Anyways yeah it always strikes me as a kind of trend that maybe has an application in a very specific niche but is likely bullshit if applied to the general case

                • V0ldek@awful.systems
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  16 days ago

                  If anything that sounds like an indictment? Like, the current models are so incredibly fucking bad that we could achieve the same with three bits and a ham sandwich

                • BlueMonday1984@awful.systems
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  15 days ago

                  Far as I can tell, the only real benefit here is significant energy savings, which would take LLMs from “useless waste of a shitload of power” to “useless waste of power”.

          • froztbyte@awful.systems
            link
            fedilink
            English
            arrow-up
            0
            ·
            15 days ago

            “look, Mme Karen, this is definitely not a rotten tomato. it can’t be a rotten tomato, we don’t sell rotten tomatoes. you can see here on the menu that we don’t have rotten tomatoes on offer. and see here, on your receipt, where it says quinoa salad? absolutely not rotten tomatoes!” explains the manager fervently, avoiding a tableward glance at the pungent red blob with as much will as they can muster

      • froztbyte@awful.systems
        link
        fedilink
        English
        arrow-up
        0
        ·
        16 days ago

        this is correct as to the background of the term itself, the reason ed uses it here is because it is the term that he selected some months ago when he listed “some likely pale horses that signal the bubble popping”

    • ZILtoid1991@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      15 days ago

      Likely they’ll try to sell it to governments, and with Elon Musk proposing goVeRNmeNt eFfIciEnCy, at least xAI can become somewhat profitable.

  • rumba@lemmy.zip
    link
    fedilink
    English
    arrow-up
    0
    ·
    15 days ago

    $200 a month for a user is losing money? There’s no way he’s just including model queries. An entire a6000 server is around $800 / month and you can fit a hell of lot more than 4 peoples worth of queries. He has to include training and or R&D.

    • db0@lemmy.dbzer0.comOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      15 days ago

      I’m honestly fairly surprised as well, but at the same time, they’re not serving a model that can run on an A6000, and the people paying for unlimited, would probably be the ones who setup bots and apps doing thousands of requests per hour.

      • LiveLM@lemmy.zip
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        15 days ago

        And honestly? Those people are 100% right.
        If they can’t deliver true “unlimited” for 200 bucks a month, they shouldn’t market it as such.

        grumble grumble unlimited mobile data grumble grumble

        • db0@lemmy.dbzer0.comOP
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          15 days ago

          To be fair, unlimited is supposed to mean unlimited for a reasonable person. Like someone going to an “all you can eat buffet”. However those purchasing these would immediately set up proxy accounts and use them to serve all their communities, so that one unlimited account, becomes 100 or a 1000 actual users. So like someone going to an “all you can eat” and then sneaking in 5 other people under their trenchcoat.

          If they actually do block this sort of account sharing, and it’s costing them money on just prolific single users, then I don’t know, their scaling is just shite. Like “unlimited” can’t ever be truly unlimited, as there should be a rate limit to prevent these sort of shenanigans. But if the account can’t make money with a reasonable rate limit (like 17280/day which would translate to 1 request per 5 sec) they are fuuuuuucked.

          • LiveLM@lemmy.zip
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            15 days ago

            Yeah, poor wording on my part, proxy accounts being banned is totally fair, but a user using various apps and bots is the type of ‘Power User’ scenario I’d expect a unlimited plan to cover.

            • db0@lemmy.dbzer0.comOP
              link
              fedilink
              English
              arrow-up
              0
              ·
              edit-2
              15 days ago

              Agreed. Like how fucking difficult is it to see “It costs us X per query, what Y rate limit do we need to put on this account so that it doesn’t exceed 200$ per month?”. I bet the answer to is hilariously low rate limit that nobody would buy, so they decided to value below cost and pray people won’t actually use all those queries. Welp. And if they didn’t even put a rate limit, also lol. lmao.

    • Jimmycakes@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      15 days ago

      It includes anything that will keep them from having to pay investors back. Classic tech start up bullshit.

      Silicon valley brain rot formula:

      Losing money, get billions every month

      Making money pay billions back

      Which one do you think they pick

  • DavidGarcia@feddit.nl
    link
    fedilink
    English
    arrow-up
    0
    ·
    16 days ago

    really looking forward to how these multi-billion dollar AI datacenter investments will work out for big tech companies

    that said I’m pretty sure most of that capacity is reserved for the surveillance state anyway

        • BB84@mander.xyz
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          16 days ago

          LLM inference can be batched, reducing the cost per request. If you have too few customers, you can’t fill the optimal batch size.

          That said, the optimal batch size on today’s hardware is not big (<20). I would be very very surprised if they couldn’t fill it for any few-seconds window.

          • flere-imsaho@awful.systems
            link
            fedilink
            English
            arrow-up
            0
            ·
            15 days ago

            i would swear that in an earlier version of this message the optimal batch size was estimated to be as large as twenty.

          • David Gerard@awful.systemsM
            link
            fedilink
            English
            arrow-up
            0
            ·
            16 days ago

            this sounds like an attempt to demand others disprove the assertion that they’re losing money, in a discussion of an article about Sam saying they’re losing money

            • BB84@mander.xyz
              link
              fedilink
              English
              arrow-up
              0
              ·
              16 days ago

              What? I’m not doubting what he said. Just surprised. Look at this. I really hope Sam IPO his company so I can short it.

                • BB84@mander.xyz
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  edit-2
                  16 days ago

                  Can someone explain why I am being downvoted and attacked in this thread? I swear I am not sealioning. Genuinely confused.

                  @[email protected] asked how request frequency might impact cost per request. Batch inference is a reason (ask anyone in the self-hosted LLM community). I noted that this reason only applies at very small scale, probably much smaller than what OpenAI is operating at.

                  @[email protected] why did you say I am demanding someone disprove the assertion? Are you misunderstanding “I would be very very surprised if they couldn’t fill [the optimal batch size] for any few-seconds window” to mean “I would be very very surprised if they are not profitable”?

                  The tweet I linked shows that good LLMs can be much cheaper. I am saying that OpenAI is very inefficient and thus economically “cooked”, as the post title will have it. How does this make me FYGM? @[email protected]

        • EldritchFeminity@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          0
          ·
          16 days ago

          Yeah, the tweet clearly says that the subscribers they have are using it more than they expected, which is costing them more than $200 per month per subscriber just to run it.

          I could see an argument for an economy of scales kind of situation where adding more users would offset the cost per user, but it seems like here that would just increase their overhead, making the problem worse.

      • V0ldek@awful.systems
        link
        fedilink
        English
        arrow-up
        0
        ·
        edit-2
        16 days ago

        Wait but he controls the price, not the subscriber number?

        Like even if the issue was low subscriber number (which it isn’t since they’re losing money per subscriber, more subscribers just makes you lose money faster), that’s still the same category of mistake? You control the price and supply, not the demand, you can’t set a stupid price that loses you money and then be like “ah, not my fault, demand was too low” like bozo it’s your product and you set the price. That’s econ 101, you can move the price to a place where your business is profitable, and if such a price doesn’t exist then maybe your biz is stupid?

        • froztbyte@awful.systems
          link
          fedilink
          English
          arrow-up
          0
          ·
          16 days ago

          I believe our esteemed poster was referencing the oft-seen cloud dynamic of “making just enough in margin” where you can tolerate a handful of big users because you have enough lower-usage subscribers in aggregate to counter the heavies. which, y’know, still requires the margin to exist in the first place

          alas, hard to have margins in Setting The Money On Fire business models

      • froztbyte@awful.systems
        link
        fedilink
        English
        arrow-up
        0
        ·
        16 days ago

        despite that one episode of Leverage where they did some laundering by way of gym memberships, not every shady bullshit business that burns way more than they make can just swizzle the numbers!

        (also if you spend maybe half a second thinking about it you’d realize that economies of scale only apply when you can actually have economies of scale. which they can’t. which is why they’re constantly setting more money on fire the harder they try to make their bad product seem good)

    • Sergio@slrpnk.net
      link
      fedilink
      English
      arrow-up
      0
      ·
      16 days ago

      They’re still in the first stage of enshittification: gaining market share. In fact, this is probably all just a marketing scheme. “Hi! I’m Crazy Sam Altman and my prices are SO LOW that I’m LOSING MONEY!! Tell your friends and subscribe now!”

      • skittle07crusher@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        ·
        16 days ago

        I’m afraid it might be more like Uber, or Funko, apparently, as I just learned tonight.

        Sustained somehow for decades before finally turning any profit. Pumped full of cash like it’s foie gras by Wall Street. Inorganic as fuck, promoted like hell by Wall Street, VC, and/or private equity.

        Shoved down our throats in the end.

    • where_am_i@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      16 days ago

      well, yes. But also this is an extremely difficult to price product. 200$/m is already insane, but now you’re suggesting they should’ve gone even more aggressive. It could turn out almost nobody would use it. An optimal price here is a tricky guess.

      Although they probably should’ve sold a “limited subscription”. That does give you max break-even amount of queries per month, or 2x of that, but not 100x, or unlimited. Otherwise exactly what happened can happen.

        • Saledovil@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          0
          ·
          15 days ago

          What the LLMs do, at the end of the day, is statistics. If you want a more precise model, you need to make it larger. Basically, exponentially scaling marginal costs meet exponentially decaying marginal utility.

            • self@awful.systems
              link
              fedilink
              English
              arrow-up
              0
              ·
              15 days ago

              guess again

              what the locals are probably taking issue with is:

              If you want a more precise model, you need to make it larger.

              this shit doesn’t get more precise for its advertised purpose when you scale it up. LLMs are garbage technology that plateaued a long time ago and are extremely ill-suited for anything but generating spam; any claims of increased precision (like those that openai makes every time they need more money or attention) are marketing that falls apart the moment you dig deeper — unless you’re the kind of promptfondler who needs LLMs to be good and workable just because it’s technology and because you’re all-in on the grift

              • Saledovil@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                0
                ·
                15 days ago

                Well, then let me clear it up. The statistics becomes more precise. As in, for a given prefix A, and token x, the difference between the calculated probability of x following A (P(x|A)) to the actual probability of P(x|A) becomes smaller. Obviously, if you are dealing with a novel problem, then the LLM can’t produce a meaningful answer. And if you’re working on a halfway ambitious project, then you’re virtually guaranteed to encounter a novel problem.

                • self@awful.systems
                  link
                  fedilink
                  English
                  arrow-up
                  0
                  ·
                  15 days ago

                  Obviously, if you are dealing with a novel problem, then the LLM can’t produce a meaningful answer.

                  it doesn’t produce any meaningful answers for non-novel problems either

      • V0ldek@awful.systems
        link
        fedilink
        English
        arrow-up
        0
        ·
        16 days ago

        “Our product that costs metric kilotons of money to produce but provides little-to-no value is extremely difficult to price” oh no, damn, ye, that’s a tricky one

      • confusedbytheBasics@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        15 days ago

        I signed up for API access. I run all my queries through that. I pay per query. I’ve spent about $8.70 since 2021. This seems like a win-win model. I save hundreds of dollars and they make money on every query I run. I’m confused why there are subscriptions at all.

    • V0ldek@awful.systems
      link
      fedilink
      English
      arrow-up
      0
      ·
      16 days ago

      Welcome to the wonderful XXI century where our innovations in communication technology and financial instruments allow a hyperoptimised economy where two tweets are more than enough to cause billion-dollar shifts on the market. Completely organic and based on solid fundamentals I am assured by the same people that assured me of this in 2000 and 2008.

    • Kitathalla@lemy.lol
      link
      fedilink
      English
      arrow-up
      0
      ·
      16 days ago

      I’ve seen more written on one post. People will eat up ‘news’ if presented in the right way. There is a reason the stupid websites and advertisers use the click-bait titles.

    • lars@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      0
      ·
      16 days ago

      According to the Lemmy comment I read about it that is exactly what people are wondering

  • affiliate@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    16 days ago

    sam altman proving once again that he is not only a tech genius but also a business genius. make sure to let him scan your eyeballs before it’s too late.

  • protist@mander.xyz
    link
    fedilink
    English
    arrow-up
    0
    ·
    16 days ago

    “I personally chose the price”

    Is that how well-run companies operate? The CEO unilaterally decides the price rather than delegating that out to the numbers people they employ?

    • brbposting@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      16 days ago

      I’m guessing that means a team or someone presented their pricing analysis to him, and suggested a price range. And this is his way of taking responsibility for making the final judgment call.

      (He’d get blamed either way, anyways)

      • David Gerard@awful.systemsM
        link
        fedilink
        English
        arrow-up
        0
        ·
        16 days ago

        $20/mo sounds like a reasonable subscription-ish price, so he picked that. That OpenAI loses money on every query, well, let’s build up volume!

      • zbyte64@awful.systems
        link
        fedilink
        English
        arrow-up
        0
        ·
        16 days ago

        While the words themselves near an apology, I didn’t read it as taking responsibility. I read it as:

        Anyone could have made this same mistake. In fact, dumber people than I would surely have done worse.

    • froztbyte@awful.systems
      link
      fedilink
      English
      arrow-up
      0
      ·
      16 days ago

      far, far, far, far, far, far, far fewer business people than you’d expect/guess are data-driven decision makers

      and then there’s the whole bayfucker ceo dynamic which adds a whole bunch of extra dumb shit

      it’d be funnier if it weren’t for the tunguska-like effect it’s having on human society both at present and in the coming decades to follow :|

    • lobut@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      16 days ago

      I think I remember Jeff Bezos in “The Everything Store” book seeing a price they charged for AWS and went even lower for growth. So there could be some rationale for that? However, I think switching AI providers is easier than Cloud Providers? Not sure though.

      I can imagine the highest users of this being scam artists and stuff though.

      I want this AI hype train to die.

    • rook@awful.systems
      link
      fedilink
      English
      arrow-up
      0
      ·
      16 days ago

      A real ceo does everything. Delegation is for losers who can’t cope. Can’t move fast enough and break enough things if you’re constantly waiting for your lackeys to catch up.

      If those numbers people were cleverer than the ceo, they’d be the ones in charge, and they aren’t. Checkmate. Do you even read Ayn Rand, bro?

      • Kitathalla@lemy.lol
        link
        fedilink
        English
        arrow-up
        0
        ·
        16 days ago

        Is that what Ayn Rand is about? All I really remember is that having a name you chose yourself is self-fulfilling.

          • Milk_Sheikh@lemm.ee
            link
            fedilink
            English
            arrow-up
            0
            ·
            16 days ago

            A monologue that last SIXTY PAGES of dry exposition. Barely credible characterization from the protagonist and villains and extremely poor world building.

            Anthem is her better book because it keeps to a simple short story format - but still has a very dull plot that shoehorns ideology throughout. There’s far better philosophical fiction writers out there like Camus, Vonnegut, or Koestler. Skip Rand altogether imo

        • sp3ctr4l@lemmy.zip
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          16 days ago

          Ayn Rand is about spending your whole life moralizing a social philosophy based on the impossibility of altruism, perfect meritocratic achievement perfectly distributing wealth, and hatred of government taxation, regulation, and social welfare programs…

          … and then dying alone, almost totally broke, living off of social security and financial charity from your former secretary.

    • azertyfun@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      15 days ago

      In tech? Kinda yeah. When a subscription is 14.99 $£€/month it’s a clear “we just think it’s what people think is a fair price for SaaS”.

      The trick is that tech usually works on really weird economics where the fixed costs (R&D) are astonishingly high and the marginal costs (servers etc) are virtually nil. That’s how successful tech companies are so profitable, even more than oil companies, because once the R&D is paid off every additional user is free money. And this means that companies don’t have to be profitable any time in particular as long as they promise sufficient projected growth to make up for being a money pit until then. You can get away with anything when your investors believe you’ll eventually have a billion users.

      … Of course that doesn’t work when every customer interaction actually costs a buck or two in GPU compute, but I’m sure after a lot of handwaving they were able to explain to their investors how this is totally fine and totally sustainable and they’ll totally make their money back a thousandfold.

  • Honytawk@lemmy.zip
    link
    fedilink
    English
    arrow-up
    0
    ·
    16 days ago

    If they are losing money on $200/month, that does not necessarily mean they lose money on the $20/month.

    One is unlimited, the other is not. You only have to use the $200 subscription more than 10 times the amount the $20 subscription allows for OpenAI to earn less on that subscription.

  • Mikina@programming.dev
    link
    fedilink
    English
    arrow-up
    0
    ·
    16 days ago

    Hmm, we should get together some funds to buy a single unlimited subscription, and then let it continuously generate as large and complex prompts as the rate limitting allows.

    • apotheotic (she/her)@beehaw.org
      link
      fedilink
      English
      arrow-up
      0
      ·
      16 days ago

      On one hand, heck yes. On the other, part of the reason its so expensive is because of the energy and water usage, so sticking it to the man in this way also is harmful to the environment :(

      • areyouevenreal@lemm.ee
        link
        fedilink
        English
        arrow-up
        0
        ·
        15 days ago

        Normally the people talking about water use have no idea what they are talking about. Normally data center cooling is closed loop, much like a cars cooling system. So they don’t use significant amounts of water at all.

  • Viri4thus@feddit.org
    link
    fedilink
    English
    arrow-up
    0
    ·
    15 days ago

    So people are really believing Altman would publish these damning statements without ulterior motives? Are we seriously this gullible? Holy shit, we reached a critical mass of acephalous humans, no turning back now.

  • edgemaster72@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    16 days ago

    losing money because people are using it more than expected

    “I personally chose the price and thought we would make some money.”

    Big MoviePass energy