Maybe off topic, but I’m not a fan of watching Yas Marina, or driving it in sims.
Its basically all hairpins with no flow, and the 06/07 hairpins are particularly awkward… I think 02-04 may be the only turns I like.
Maybe off topic, but I’m not a fan of watching Yas Marina, or driving it in sims.
Its basically all hairpins with no flow, and the 06/07 hairpins are particularly awkward… I think 02-04 may be the only turns I like.
The American way is to bail them out financially, then re-elect them.
And a small cluster like Alibaba used to train Qwen 2.5 is basically a drop in the bucket.
The hoard of GPUs Meta, Microsoft/OpenAI, and especially X have are apparently being used extremely inefficiently, or perhaps mostly not used to train AI at all, but do regular ad/engagement optimization stuff.
Ah yeah those were the good old days when vendors were free to do that, before AMD/Nvidia restricted them. It wasn’t even that long ago, I remember some AMD 7970s being double VRAM.
And, again, I’d like to point out how insane this restriction is for AMD given their market struggles…
What could they do, enlist North Korea?
Thats an Onion headline, surely…
Not even that, they really just need to pay more taxes to help Ukraine.
But nope.
And Russia wouldn’t dare bomb NATO unless they want to get totally screwed.
I suppose there hasn’t been much urgency and thr goal of many members is ‘stability,’ but they are now facing a new reality where, according to historical precedent, there will be much less stability after Jan 20th, and there is new urgency to bend rules before then.
Strix Halo is your unicorn, idle power should be very low (assuming AMD VCE is OK over quicksync)
Just that bursts of inference for a small model on a phone or even a desktop is less power hungry than a huge model on A100s/H100s servers. The hardware is already spun up anyway, and (even with the efficiency advantage of batching) Nvidia runs their cloud GPUs in crazy inefficient voltages/power bands just to get more raw performance per chip and squeak out more interactive gains, while phones and such run at extremely efficient voltages.
There are also lots of tricks that can help “local” models like speculative decoding or (theoretically) bitnet models that aren’t great for cloud usage.
Also… GPT-4 is very inefficient. Open 32B models are almost matching it at a fraction of the power usage and cost, even in servers. OpenAI kind of sucks now, but the larger public hasn’t caught on yet.
Oh also you might look at Strix Halo from AMD in 2025?
Its IGP is beefy enough for LLMs, and it will be WAY lower power than any dGPU setup, with enough vram to be “sloppy” and run stuff in parallel with a good LLM.
You could get that with 2x B580s in a single server I guess, though yoi could have already done that with the A770s.
The entire point of NATO is to defend against Russian aggression, so waiving parts of rules seems like fair game if addressing that goal so directly.
Escalate how, exactly? Is Russia going to be at more war with Ukraine?
And WTF do they think is gonna happen after Jan 20th?
Also “ridiculously” is relative lol.
The Llm/workstation crowd would buy a 48GB 4060 without even blinking, if that were possible. These workloads are basically completely vram constrained.
Like the 3060? And 4060 TI?
Its ostensibly because they’re “too powerful” for their vram to be cut in half (so 6GB on the 3060 and 8GB on the 4060 TI), but yes, more generally speaking these are sweetspot for vram heavy workstation/compute workloads. Local LLMs are just the most recent one.
Nvidia cuts vram at the high end to protect their server/workstation cards, AMD does it… Just because?
deleted by creator
I would also ask yourself if this is beneficial at all.
Look at their historical prices. Investors have already bought into this.
There’s no way. I think he’s legit too old, and he only needs 1 term to pardon himself and everyone he wants.
Are the double capacity GDDR6X ICs being sold yet? I thought they weren’t and “double” cards like the 4060 TI were using 2 chips per channel.
Their reputation and past reporting is supposed to back up things they state as facts (like assuming that reviews they cite are real) for practicality and brevity. Imagine having to document every bit of background research in a presentable way.
They could have included screenshots though.
And the skepticism is healthy. I do personally ‘trust’ Axios (which I read almost daily but regularly double check).