You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
H100/H200 have dedicated fp8 tensor cores with twice as many FLOPs as fp16. This will halve GPU cost. Developers will use whichever library has lower cost, assuming the cost reduction comes without the dev having to put additional effort.
H100/H200 have dedicated fp8 tensor cores with twice as many FLOPs as fp16. This will halve GPU cost. Developers will use whichever library has lower cost, assuming the cost reduction comes without the dev having to put additional effort.
Please consider prioritising fp8 support, your library risks going outdated otherwise.
The text was updated successfully, but these errors were encountered: