• 0 Posts
  • 12 Comments
Joined 13 days ago
cake
Cake day: February 18th, 2025

help-circle
  • Interesting article.

    Data center buildouts take about three to six years to complete, and the largest hyperscale facilities can easily cost several billion dollars, meaning that these moves are extremely forward-looking. You don’t build a data center for the demand you have now, but for the demand you expect further down the line. This suggests that Microsoft believes its current infrastructure (and its likely scaled-back plans for expansion) will be sufficient for a movement that CEO Satya Nadella called a “golden age for systems” less than a year ago.

    To explain here, TD Cowen is effectively saying that Microsoft is responding to a “major demand signal” and said “major demand signal” is saying “you do not need more data centers.” Said demand signal that Microsoft was responding to, in TD Cowen’s words, is its “appetite for capacity” to provide servers to OpenAI, and it seems that said appetite is waning, and Microsoft no longer wants to build out data centers for OpenAI.

    I believe the reason Microsoft is cutting back is that it does not have the appetite to provide further data center expansion for OpenAI, and it’s having doubts about the future of generative AI as a whole. If Microsoft believed there was a massive opportunity in supporting OpenAI’s further growth, or that it had “massive demand” for generative AI services, there would be no reason to cancel capacity, let alone cancel such a significant amount.






  • Google says that SafetyCore “provides on-device infrastructure for securely and privately performing classification to help users detect unwanted content. Users control SafetyCore, and SafetyCore only classifies specific content when an app requests it through an optionally enabled feature.”

    GrapheneOS — an Android security developer — provides some comfort, that SafetyCore “doesn’t provide client-side scanning used to report things to Google or anyone else. It provides on-device machine learning models usable by applications to classify content as being spam, scams, malware, etc. This allows apps to check content locally without sharing it with a service and mark it with warnings for users.”

    But GrapheneOS also points out that “it’s unfortunate that it’s not open source and released as part of the Android Open Source Project and the models also aren’t open let alone open source… We’d have no problem with having local neural network features for users, but they’d have to be open source.” Which gets to transparency again.