Earlier today someone posted the following code

float a = 0.12f;
float b = a * 100f;
Console.WriteLine((int) b); // prints 12
Console.WriteLine((int)(a * 100f)); // prints 11 !!!!!!!!


An (extensive) explanation for this strange behaviour can be found at CLR and floating point: Some answers to common questions… A possible way to force the compiler and runtime to get rid of the additional precision would be the following

Console.WriteLine((int)(float)(a * 100f));