• 0 Posts
  • 27 Comments
Joined 1 year ago
cake
Cake day: August 4th, 2024

help-circle



  • Unlike the dotcom bubble, Another big aspect of it is the unit cost to run the models.

    Traditional web applications scale really well. The incremental cost of adding a new user to your app is basically nothing. Fractions of a cent. With LLMs, scaling is linear. Each machine can only handle a few hundred users and they’re expensive to run:

    Big beefy GPUs are required for inference as well as training and they require a large amount of VRAM. Your typical home gaming GPU might have 16gb vram, 32 if you go high end and spend $2500 on it (just the GPU, not the whole pc). Frontier models need like 128gb VRAM to run and GPUs manufactured for data centre use cost a lot more. A state of the art Nvidia h200 costs $32k. The servers that can host one of these big frontier models cost, at best, $20 an hour to run and can only handle a handful of user requests so you need to scale linearly as your subscriber count increases. If you’re charging $20 a month for access to your model, you are burning a user’s monthly subscription every hour for each of these monster servers you have turned on. That’s generous and assumes they’re not paying the “on-demand” price of $60/hr.

    Sam Altman famously said OpenAI are losing money on their $200/mo subscriptions.

    If/when there is a market correction, a huge factor of the amount of continued interest (like with the internet after dotcom) is whether the quality of output from these models reflects the true, unsubsidized price of running them. I do think local models powered by things like llamacpp and ollama and which can run on high end gaming rigs and macbooks might be a possible direction for these models. Currently though you can’t get the same quality as state-of-the-art models from these small, local LLMs.






  • ErmahgherdDavid@lemmy.dbzer0.comtomemes@lemmy.worldWriting
    link
    fedilink
    English
    arrow-up
    4
    ·
    2 months ago

    I remember a friend telling me their English teacher took them on a field trip to see the Poet Simon Armitage and asked him what the poem “the hitcher” represents and he just said “it’s just about some bloke” - or something to that effect. Anyway it made me chuckle to imagine the dismay of the English teachers in the room


  • If we get a breakthrough moment with quantum, the machines will not be evenly distributed to start with. They will be too expensive to build, power and cool unless you’re a fortune 500 exactly like LLMs right now (aside from small models like llama that can run on consumer hardware). At the moment quantum computers rely on superconductors that have to be cooled near absolute zero which is… somewhat expensive to achieve.

    Unlike LLMs (oh no I can’t talk to waifu without cell coverage waah) Not being able to run quantum algorithms on your phone in this scenario would be bad. It either means your personal comms are, for all intents and purposes decryptable by those who control the quantum machines or that you’ll have to pay rent to the people who control quantum machines to have them encrypt and decrypt stuff for you. Of course you’ll have to trust them too. Also, given governments thirst for spying on our encrypted comms, it’s possible that quantum machines are heavily regulated allowing “the good guys” a back door into our chats without giving “the baddies” a way to encrypt their comms