Skip to main content

All Questions

1 vote
0 answers
61 views

Does cuBLAS support mixed precision matrix multiplication in the form C[f32] = A[bf16] * B[f32]?

I'm concerning mixed precision in deep learning LLM. The intermediates are mostly F32 and weights could be any other type like BF16, F16, even quantized type Q8_0, Q4_0. it would be much useful if ...
dentry's user avatar
  • 11
2 votes
2 answers
498 views

How do I print the half-precision / bfloat16 values from in a (binary) file?

This is a variant of: How to print float value from binary file in shell? in that question, we wanted to print IEEE 754 single-precision (i.e. 32-bit) floating-point values from a binary file. Now ...
einpoklum's user avatar
  • 133k
1 vote
0 answers
261 views

How to round up or down when converting f32 to bf16 in rust?

I am converting from f32 to bf16 in rust, and want to control the direction of the rounding error. Is there an easy way to do this? Converting using the standard bf16::to_f32 rounds to the nearest ...
Amir's user avatar
  • 898