Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
irthomasthomas
5 months ago
|
parent
|
context
|
favorite
| on:
Ollama Turbo
Oh, I didn't know that. Weird!
reissbaker
5 months ago
[–]
It was natively trained in FP4. Probably both to reduce VRAM usage at inference time (fits on a single H100), and to allow better utilization of B200s (which are especially fast for FP4).
irthomasthomas
5 months ago
|
parent
[–]
Interesting, thanks. I didn't know you could even train at FP4 on H100s
reissbaker
5 months ago
|
root
|
parent
[–]
It's impressive they got it to work — the lowest I'd heard of this far was native FP8 training.
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: