• 0 Posts
  • 700 Comments
Joined 2 years ago
cake
Cake day: August 24th, 2023

help-circle

  • Australia isn’t the greatest spot to run a data centre in general in terms of heat, but I do understand the need for sovereign data centres, so this obviously can’t work everywhere.

    What makes you think $3.5 million can’t be profitable? A mid sized hospitals heating bill can get into the many hundreds of thousands or into the millions even. Especially if it’s in a colder environment. A 5-6 year payback on that wouldn’t be terrible and would be worth an upfront investment. Even a 10 year payback isn’t terrible.

    These colder locations are the ideal locations for the data centres in the first place because they generally want a cooler climate to begin with, so they will gravitate to them when possible.

    Edit: And if you build a data centre with this ability to recoup heat, you could start building further commercial things in the area and keep the heat redistribution very close. You don’t need to travel very long distances. You do need to put some thought into where they go through and whats around or will be built around.






  • I just wanted to add one other thing on the hardware side.

    These H200’s are power hogs, no doubt about it. But the next generation H300 or whatever it is, will be more efficient as the node process (or whatever its called) gets smaller and the hardware is optimized and can run things faster. I could still see NVIDIA coming out and charging more $/flop or whatever the comparison would be though even if it is more efficient power wise.

    But that could mean that the electricity costs to run these models starts to drop if they truly are plateaued. We might not be following moores law on this anymore (I don’t actually know), but were not completely stagnant either.

    So IF we are plateaued on this one aspect, then costs should start coming down in future years.

    Edit: but they are locking in a lot of overhead costs at today’s prices which could ruin them.





  • Why on earth do you think things can’t be optimized on the LLM level?

    There are constant improvements being made there, they are not in any way shape or form fully optimized yet. Go follow the /r/LocalLlama sub for example and there’s constant breakthroughs happening, and then a few months later you see a LLM utilizing them come out, and they’re suddenly smaller, or you can run a larger model on smaller memory footprint, or you can get a larger context on the same hardware etc.

    This is all so fucking early, to be so naive or ignorant to think that they’re as optimized as they can get is hilarious.



  • These companies have BILLIONS in revenue and millions of customers, and you’re saying very few want to pay…

    The money is there, they just need to optimize the LLMs to run more efficiently (this is continually progressing), and the hardware side work on reducing hardware costs as well (including electricity usage / heat generation). If OpenAI can build a datacenter that re-uses all it’s heat for example to heat a hospital nearby, that’s another step towards reaching profitability.

    I’m not saying this is an easy problem to solve, but you’re making it sound no one wants it and they can never do it.


  • rely on input data, which is running out.

    Thats part of the equation, but there is still a lot of work that can be done to optimize the usage of the llms themselves, and the more optimized and refined they are, the cheaper it becomes to run, and you can also use even bigger datasets that weren’t feasible before.

    I think there’s also a lot of room to still optimize the data in the data set. Ingesting the entire worlds information doesn’t lead to the best output, especially if you’re going into something more factual vs creative like a LLM trained to assist with programming in a specific language.

    And people ARE paying for it today, OpenAI has billions in revenue, the problem is the hardware is so expensive, the data centeres needed to run it are also expensive. They need to continue optimizing things to narrow that gap. Open AI charges $20 USD/month for their base paid plan. They have millions of paying customers, but millions isn’t enough to offset their costs.

    So they can

    1. reduce costs so millions is enough
    2. make it more useful so they can gain more users.

    This is so early that they have room to both improve 1 and 2.

    But like I said, they (and others like them) need to figure that out before they run out of money and everything falls apart and needs to be built back up in a more sustainable way.

    We won’t know if they can or can’t until they do it, or it pops.


  • But the profit absolutely can materialize because it is useful.

    Right now the problem is hardware / data center costs, but those can come down at a per user level.

    They just need to make it useful enough within those cost constants which is 100% without a doubt possible, it’s just a matter of can they do it before they run out of money.

    Edit: for example, nvidia giving OpenAI hardware for ownership helps bring down their costs, which gives them a longer runway to find that sweet spot.