575 points by albelfio 1 day ago | 334 comments | View on ycombinator
bastawhiz 1 day ago |
ivraatiems 1 day ago |
Still, this is a great idea, and one I hope takes off. I think there's a good argument that the future of AI is in locally-trained models for everyone, rather than relying on a big company's own model.
One thought: The ability to conveniently get this onto a 240v circuit would be nice. Having to find two different 120v circuits to plug this into will be a pain for many folks.
vessenes 1 day ago |
I have no idea who would buy this. Maybe if you think Vera Rubin is three years out? But NV ships, man, they are shipping.
paxys 1 day ago |
alexfromapex 1 day ago |
roarcher 1 day ago |
Has this guy never worked on a B2B product before? Nobody is going to order a $10 million piece of infrastructure through your website's order form. And they are definitely going to want to negotiate something, even if it's just a warranty. And you'll do it because they're waving a $10 million check in your face.
The tone of this website is arrogant to the point of being almost hostile. The guy behind this seems to think that his name carries enough weight to dictate terms like this, among other things like requiring candidates to have already contributed to his product to even be considered for a job. I would be extremely surprised if anyone except him thinks he's that important.
siliconc0w 1 day ago |
mellosouls 1 day ago |
Edit: found a third party referencing the claim but it doesn't belong in the title here I think:
Meet the World’s Smallest ‘Supercomputer’ from Tiiny AI; A Machine Bold Enough to Run 120B AI Models Right in the Palm of Your Hand
https://wccftech.com/meet-the-worlds-smallest-supercomputer-...
hmokiguess 1 day ago |
adrianwaj 1 day ago |
I think Tinygrad should think about recycling. Are they planning ahead in this regard? Is anyone? My thought is if there was a central database of who own what and where, at least when the recycling tech become available, people will know where to source their specific trash (and even pay for it.) Having a database like that in the first place could even fuel the industry.
ekropotin 1 day ago |
I almost sure it’s possible to custom build a machine as powerful as their red v2 within 9k budget. And have a lot of fun along the way.
operatingthetan 1 day ago |
$12,000, $65,000, $10,000,000.
adi_kurian 1 day ago |
mmoustafa 1 day ago |
I'm currently shopping for offline hardware and it is very hard to estimate the performance I will get before dropping $12K, and would love to have a baseline that I can at least always get e.g. 40 tok/s running GPT-OSS-120B using Ollama on Ubuntu out of the box.
mciancia 1 day ago |
With 6 GPUs you have to deal with risers, pcie retimers, dual PSUs and custom case for so value proposition there was much better IMO
the_arun about 12 hours ago |
wongarsu 1 day ago |
Not revolutionary in any way, but nice. Unless I'm missing something here?
comrade1234 1 day ago |
It's funny though... we're using deepseek now for features in our service and based on our customer-type we thought that they would be completely against sending their data to a third-party. We thought we'd have to do everything locally. But they seem ok with deepseek which is practically free. And the few customers that still worry about privacy may not justify such a high price point.
ks2048 1 day ago |
"likely" doesn't inspire much confidence. Surely, they have those numbers, and if it was, they'd publicize the comparisons.
SmartestUnknown 1 day ago |
Can they/someone else give more details as to what workloads pytorch is more than 2x slower than the hardware provides? Most of the papers use standard components and I assume pytorch is already pretty performant at implementing them at 50+% of extractable performance from typical GPUs.
If they mean more esoteric stuff that requires writing custom kernels to get good performance out of the chips, then that's a different issue.
algolint about 17 hours ago |
mayukh 1 day ago |
undefined about 24 hours ago |
triwats about 13 hours ago |
How do you test/generate these numbers?
ilaksh 1 day ago |
I could swear I filed a GitHub issue asking about the plans for that but I don't see it. Anyway I think he mentioned it when explaining tinygrad at one point and I have wondered why that hasn't got more attention.
As far as boxes, I wish that there were more MI355X available for normal hourly rental. Or any.
alasdair_ 1 day ago |
Obviously any Turing machine can run any size of model, so the “120B” claim doesn’t mean much - what actually matters is speed and I just don’t believe this can be speedy enough on models that my $5000 5090-based pc is too slow for and lacks enough vram for.
jmspring 1 day ago |
A single box with those specs without having to build/configure (the red and green) - I could see being useful if you had $ and not time to build/configure/etc yourself.
jeremie_strand 1 day ago |
himata4113 1 day ago |
saidnooneever about 19 hours ago |
the boxes look cool but how good are they really? the cheapest box seems pricey at 12 for a what is essentially a few gaming gpus. i dont see why you couldnt make that like half the price. u could do a PC/server build thats much much faster for way less. size doesnt matter if its more than twice the price i think...
the more expensive box has atleast real processing gpus but afaik also not very popular ones, this one seems maybe more fair priced (there seems a big difference in bang for buck between these???).
the third one suggested looks like a joke.
dont get me wrong, this seems like a really cool idea. But i dont see it taking off as the prices are corporate but the product seems more home use.
maybe in time they will find a better balance, i do respect the fact that the component market now is sour as hell and making good products with stable prices is pretty much i possible.
id love one of these machines someday, maybe when i am less poor, or when they are xD.
(love the styling of everything, this is the most critical i could be from a dumb consumer perspective, which i totally am btw.)
zahirbmirza 1 day ago |
Buttons840 1 day ago |
He's an interesting guy. Seems to be one who does things the way he thinks is right, regardless of corporate profits.
p0w3n3d 1 day ago |
heinternets 1 day ago |
720x RDNA5 AT0 XL 25,920 GB VRAM 23,040 GB System RAM
~ $10 Million
Who is the target market here?
andai 1 day ago |
sudo_cowsay 1 day ago |
operatingthetan 1 day ago |
agnishom about 23 hours ago |
orliesaurus 1 day ago |
jgarzik 1 day ago |
droidjj 1 day ago |
qubex about 20 hours ago |
mememememememo 1 day ago |
ppap3 1 day ago |
raincole 1 day ago |
vlovich123 1 day ago |
DeathArrow about 14 hours ago |
DeathArrow about 15 hours ago |
throwatdem12311 1 day ago |
But let’s be real, 12k is kinda pushing it - what kind of people are gonna spend $65k or even $10M (lmao WTAF) on a boutique thing like this. I dont think these kinds of things go in datacenters (happy to be corrected) and they are way too expensive (and probably way too HOT) to just go in a home or even an office “closet”.
rpastuszak 1 day ago |
gymbeaux about 21 hours ago |
I'm not sure what tinygrad is but I assume the markup is because the customer is making a conscious choice to support the tinygrad project. But what's unusual is there is apparently no reason whatsoever to buy this hardware, even if you plan on using tinygrad exclusively for your project. At least with System76 hardware I get (in theory) first class support for Pop!_OS.
kylehotchkiss 1 day ago |
arunakt 1 day ago |
renewiltord 1 day ago |
aabaker99 1 day ago |
Sorry, what? Is this just a scam?
jauntywundrkind 1 day ago |
Theres a lot there that makes sense & I think needs to be considered. But a lot just seems to be out of the blue, included without connection, in my view. Feels like maybe are in-grouo messages, that I don't understand. How this is headered as against democracy is unclear to me, and revolting. I both think we must grapple with the world as it is, and this post is in that area, strongly, but to let fear be the dominant ruling emotion is one of the main definitions of conservativism, and it's use here to scare us sounds bad.
insane_dreamer about 23 hours ago |
flykespice 1 day ago |
Since when did our perception of tiny blow out of size in tech? Is it the influence of "hello world" eletron apps consuming 100mb of mem while idle setting the new standard? Anyway being an AI bro seems like an expensive hobby...
pugchat about 2 hours ago |
Yanko_11 about 18 hours ago |
WWilliam 1 day ago |
jee599 about 22 hours ago |
jee599 about 24 hours ago |
caijia 1 day ago |
chloecv about 19 hours ago |
aplomb1026 1 day ago |
undefined 1 day ago |
baibai008989 1 day ago |
undefined 1 day ago |
EruditeCoder108 about 15 hours ago |
Heer_J 1 day ago |
pink_eye 1 day ago |
fhn 1 day ago |
I'm running a 70b model now that's okay, but it's still fairly tight. And I've got 16gb more vram then the red v2.
I'm also confused why this is 12U. My whole rig is 4u.
The green v2 has better GPUs. But for $65k, I'd expect a much better CPU and 256gb of RAM. It's not like a threadripper 7000 is going to break the bank.
I'm glad this exists but it's... honestly pretty perplexing