Intel Announces Arc Pro B70 and Arc Pro B65 GPUs (techpowerup.com)

by throwaway270925 111 comments 158 points
Read article View on HN

111 comments

[−] genpfault 51d ago
600 GB/s of memory bandwidth isn't anything to sneeze at.

~$1000 for the Pro B70, if Microcenter is to be believed:

https://www.microcenter.com/product/709007/intel-arc-pro-b70...

https://www.microcenter.com/product/708790/asrock-intel-arc-...

[−] hedgehog 50d ago
Recent kernels have SR-IOV support for these chips too. B&H has them listed for $950.

https://www.bhphotovideo.com/c/product/1959142-REG/intel_33p...

When 32GB NVIDIA cards seem to start at around $4000 that's a big enough gap to be motivating for a bunch of applications.

[−] robotnikman 50d ago
I'm probably going to snag one of the Intel cards just for the SR-IOV and use with VM's
[−] scrubs 50d ago
I tried to use SRIOV to virtualize mellanox nics with vlans on redhat Linux. Long story short it did not work. Per Nvidia the os has to also run open switch. This work was on an already complex setup in finance ... so adding open switch was considered too much additionally complexity. This requirement is not something I run across in the docs.

Anybody know better?

[−] hedgehog 48d ago
The situation in networking is a lot different than graphics. I don't know much other than that it depends on what specific protocol, card, firmware, and network topology you're using and there's not really generic advice. If the question is setting up Ethernet switching inside the card so VFs can talk to the network, then I think the Linux switchdev tools can configure that on their own without Open vSwitch but you probably need to find someone who understands your specific type of deployment for better advice.
[−] hedgehog 50d ago
Depending what you're doing AMD's support for VirtIO Native Context might be a useful alternative (I think it gives less isolation which could be good or bad depending on use).
[−] jauntywundrkind 50d ago
I tend to agree that the vram size and bandwidth is the core thing, but this B70 Pro allegedly has 387 int8 tops vs a 5090 having 3400 int8 tops. 600 compares vs 1792GB/s. I'm delighted so see an option with quarter the price! But man, a tenth the performance? https://www.techpowerup.com/347721/sparkle-announces-intel-a... https://www.tomshardware.com/pc-components/gpus/nvidia-annou...
[−] ColonelPhantom 50d ago
838 seems to be the real INT8 TOPS number for the 5090; going from 800 to 3400 takes an x2 speedup for sparsity (so skipping ops) and another x2 speedup for FP4 over INT8.

So it's closer to half the speed than a tenth. Intel also seems to be positioning this card against the RTX PRO 4000 Blackwell, not the 5090, and that one gets more like 300 INT8 TOPS. It also has less memory but at a slightly higher bandwidth. The 5090 is much faster and IIRC priced similarly to the PRO 4000, but is also decidedly a consumer product which, especially for Nvidia, comes with limitations (e.g. no server-friendly form factor cards available, and there are or used to be driver license restrictions that prevented using a consumer card in a data center setup).

[−] jauntywundrkind 50d ago
Thank you for the correction. That seemed way too lopsided to be believed. This assessment balances the memory to tops ratio much much more evenly, which is to be expected! I was low key hoping someone would help me make sense of how wildly disparate figures were, but I wasn't seeing.

AMD R9700 is 378/766 tops int8 dense/sparse. 644GB/s of 32GB memory. ~$1400. To throw one more card into the mix. Intel undercutting that nicely here.

You're right that for companies, the pro grade matters. For us mere mortals, much less so. Features like sr-iov however are just fantastic so see! Good job Intel. AMD has been trickling out such capabilities for a decade (cards fused for "MxGPU" capability) & it makes it such an easier buy to just offer it straight up across the models.

[−] adgjlsfhk1 50d ago
especially for exploratory work 1/10th the perf is fine. Intel isn't able to compete head to head with Nvidia (yet), but vram is capability while speed is capacity. There will be plenty of use cases where the value prop here makes sense.
[−] wmf 50d ago
It's more like a 70 class card with extra VRAM.
[−] qingcharles 51d ago
I think the B65 is priced at $650. Both supported by llamacpp I believe. With that power draw you could run two of them.
[−] giancarlostoro 51d ago
Intel GPU prices have stayed fine, but I do wonder if they are viable for Inference if they will wind up like Nvidia GPUs, severely overpriced.
[−] cmovq 50d ago
I mean it kind of is considering that's comparable to a 5070 which has 672 GB/s? Benefit of NVIDIA being the only one using GDDR7 for now I guess.
[−] daemonologist 50d ago
7800 XT has 624 GB/s as well, and can be found for $400 used. 16 GB of course.
[−] BizarroLand 49d ago
I've heard ROCm is still a crapshoot though. Is that true?
[−] daemonologist 49d ago
If you stick with your OS/package manager-distributed version, installation isn't painful anymore (provided that version approximately overlaps with your generation of GPU). It's okay for inference, and okay for training if you don't stray too far beyond plain torch. If you want to run code from a paper or other more esoteric stuff you're still going to have a bad time.

I don't have an Intel dGPU, but I suspect the situation there is even worse. I mean you go to the torch homepage: https://pytorch.org/get-started/locally/ and Intel isn't even mentioned. (It's here though: https://docs.pytorch.org/docs/stable/notes/get_start_xpu.htm...)

[−] varispeed 50d ago
The product would be excellent in 2024, but now it's a landfill filler. You can run some small models at pedestrian speed, novelty wears off and that's it.

Intel is not looking in the future. If they released Arc Pro B70 with 512GB base RAM, now that could be interesting.

32GB? Meh.

[−] throwaway85825 50d ago
It's true that it's severely late and missed it's market window but 512gb just isn't possible.
[−] kadoban 50d ago
The last go around they looked good on paper and then Intel just didn't make any of them to sell.

Announce all you want, if you don't ever ship anything I could buy, who gives a shit.

[−] cmxch 50d ago
The B60 (and the dual edition) were an entire exercise in how NOT to launch a product.

They let people have the B50 but only released the B60 late in the cycle.

[−] kadoban 50d ago
I wasn't even aware they ever _really_ released the B60. When I got bored of paying attention it was ~months after "release" and they just didn't exist to buy. I do technically see them on ebay, so yeah apparently they're out there.
[−] cmxch 49d ago
The B60 was released but strictly on a B2B basis until a couple of months ago. The B60 dual, a much rarer bird, was scalped heavily enough to be unobtainable.
[−] whalesalad 50d ago
Anyone running an ARC card for desktop Linux who can comment on the experience? I've had smooth sailing with AMD GPU's but have never tried Intel.
[−] oakpond 50d ago
Running dual Pro B60 on Debian stable mostly for AI coding.

I was initially confused what packages were needed (backports kernel + ubuntu kobuk team ppa worksforme). After getting that right I'm now running vllm mostly without issues (though I don't run it 24/7).

At first had major issues with model quality but the vllm xpu guys fixed it fast.

Software capability not as good as nvidia yet (i.e. no fp8 kv cache support last I checked) but with this price difference I don't care. I can basically run a small fp8 local model with almost 100k token context and that's what I wanted.

[−] lostmsu 50d ago

> small fp8 local model with almost 100k token context

Would not fit Qwen3.5 27B would it? That's the SOTA

[−] oakpond 50d ago
This is a fp16 model. That's 54G in weights. I can load it only with fp8 quantization enabled (>= 128k context). I run into this error during generation though: https://github.com/vllm-project/vllm/issues/36350. Looks like an issue with the flash attention backend. But yeah, if you are OK with fp8 quantization on this model, it fits. I expect with 64G VRAM it will fit without quantization
[−] wyre 50d ago
There was the video a little while back where LTT built a computer for Linus Torvalds and they put an Intel Arc card inside, so I'd imagine Linux support is at the very least, acceptable.

[1] https://www.youtube.com/watch?v=mfv0V1SxbNA

[−] toofy 50d ago

> they put an Intel Arc card inside

just add a little bit:

linus requested the card be intel as well.

[−] robertVance 50d ago
Ive ran arc on fedora for years and for general desktop use it’s been perfect. For llm’s/coding it’s getting better but it’s rough around the edges. Had a bug where trying to get vram usage through pytorch would crash the system, ect.
[−] Levitating 50d ago
Afaik driver support is very complete on Linux. You often see Arc GPUs used in media transcoding workloads for that reason.
[−] HerbManic 50d ago
We can all agree that Intel absolutely nailed it with the media encoding on these things. A nice to have for many, vital for others.
[−] whalesalad 50d ago
quicksync has been around for ages its surprising to me that other platorms have not adopted this. no reason a modern cpu can't transcode video.
[−] vel0city 50d ago
Quicksync doesn't do its work on the CPU, it does the work on the integrated GPU. Their processors that did not have on-board graphics did not have Quicksync support. See their P series and many of their Xeon parts which do not carry Quicksync support, while the versions with integrated graphics do have it.

AMD chips that have integrated GPUs (their APU series of chips) often do have support for hardware video encoders. Because, once again, its a function of the GPU and not the CPU.

[−] himata4113 50d ago
Linus Torvalds runs ARC :)
[−] bpye 50d ago
My B580 works fine on Linux. Graphics perf is a bit worse than under Windows, but supposedly compute is pretty much the same.
[−] unethical_ban 50d ago
I'm running A-series Arc for media transcoding and it works just fine.
[−] tbyehl 50d ago
Where's the A310 / A40 successor? Gimme some SR-IOV in a slot-powered, single-width, low-profile card.
[−] thefounder 50d ago
Why don’t they make an GPU optimised for inference/batch jobs with 1 TB of ram ? Everyone wants to run the biggest models locally.
[−] jmward01 50d ago
I think this shows a shift in model architecture. MOE and similar need more memory for the compute available than just one big model with a lot of layers and weights. I think this is likely a trend that will accelerate. You build the trade-off in which encourages even more experts which means more of a tradeoff, so more experts.....
[−] pjmlp 50d ago
New cards in 2026, and targeting Vulkan 1.3?!
[−] cmxch 50d ago
Good to see that Intel learned to release product to more than just resellers.

Now can we have a 64gb B70 that’s worldwide available and not marked to unicorns like the Maxsun B60 Dual model has been?

[−] SkyeCA 50d ago
32GB of vram for a decent price? I wonder if these will work well for VR, because vram is my current main issue.
[−] nickthegreek 51d ago
Both have 32gb vram. Could be a pretty compelling choice.
[−] SmellTheGlove 50d ago
Any idea if it'll be possible to mix these with nvidia cards? Adding 32GB to a single 3090 setup would be pretty nice.
[−] lostmsu 50d ago
Nothing like Crossfire/SLI? Not possible to efficiently connect multiple cards for one large model?
[−] mikelitoris 50d ago
Too little too late, classic Intel
[−] DiabloD3 50d ago
Since they fired the entire Arc team and a lot of the senior engineers already updated their Linkedins to reflect their new positions at AMD, Nvidia, and others, as well as laying off most of their Linux driver team (GPU and non-GPU), uh...

WTF?

[−] jeremie_strand 50d ago
[flagged]
[−] WarmWash 51d ago
Wake me when they wake up and release a middling card with 128GB memory.
[−] mmwelt 50d ago
At the end of the 2nd paragraph:

> Intel will provide certified drivers for Windows 11, Windows 10, and Linux.

Windows 11, OK. Linux, OK. But why Windows 10 for a new product?!

[−] vessenes 51d ago
Not sure why you'd want this over an apple setup. M4 max is 545GB/s of memory bandwidth - $2k for an entire Mac Studio with 48GB of RAM vs 32 for the B70.