All Questions
Tagged with half-precision-float bfloat16
3 questions
1
vote
0
answers
61
views
Does cuBLAS support mixed precision matrix multiplication in the form C[f32] = A[bf16] * B[f32]?
I'm concerning mixed precision in deep learning LLM. The intermediates are mostly F32 and weights could be any other type like BF16, F16, even quantized type Q8_0, Q4_0. it would be much useful if ...
2
votes
2
answers
498
views
How do I print the half-precision / bfloat16 values from in a (binary) file?
This is a variant of:
How to print float value from binary file in shell?
in that question, we wanted to print IEEE 754 single-precision (i.e. 32-bit) floating-point values from a binary file.
Now ...
1
vote
0
answers
261
views
How to round up or down when converting f32 to bf16 in rust?
I am converting from f32 to bf16 in rust, and want to control the direction of the rounding error. Is there an easy way to do this?
Converting using the standard bf16::to_f32 rounds to the nearest ...