Last time in this series I described how a lifted implicit conversion could be "distributed" to both branches of the conditional operator that is the realization of a lifted arithmetic expression, and then optimized further on each side. Of course, the thing being converted can be any lifted expression in order to take advantage of this optimization. This means that the optimization "composes" nicely; the optimization could be repeatedly applied when lifted operations are nested.
This is a bit of a silly illustrative example: suppose you have expressions
y of type
A? with a lifted addition operator that produces an
A?. There's also a lifted conversion from
B?, and similarly from
C? c = (C?)(B?)(x + y);
As we discussed previously in this series, the compiler realizes the lifted addition as a conditional expression. We know that the lifted conversion to
B? can be "distributed" to the consequence and alternative branches of the conditional expression. That then results in a different conditional expression, but one such that the conversion to
C? can be distributed to each branch of that! That is, the compiler could realize the code above as:
C? c; A? temp1 = x; A? temp2 = y; c = temp1.HasValue & temp2.HasValue ? new C?((C)(B)(temp1.GetValueOrDefault() + temp2.GetValueOrDefault()) : new C?();
... by applying the optimization twice, rather than creating a temporary of type
A? for the sum and a temporary of type
B? for the conversion of the sum, each with its own conditional expression. The aim of the optimization is to reduce the number of temporaries and conditional expressions, and thereby make the code smaller and produce fewer basic blocks.
A lifted conversion is rather like a lifted unary operator, and in fact the compiler could do the analogous optimization for the lifted unary
! operators. Continuing our silly example, suppose we have a lifted
~ operator on
A? that produces an
A?. If you said:
C? c = (C?)(B?)~(x + y);
~ operation can also be "distributed" to each branch of the conditional just as the conversions can be. The insight here is the same as before: if the consequence and alternative are both of the same type then
~(condition ? consequence : alternative)
is the same as
condition ? ~consequence : ~alternative
When we furthermore know that the consequence is of the form
new A?(something) then we know that
~consequence is the same as
new A?(~something). When we know that the alternative is of the form
new A?(), then we know that
~new A?() is going to be a no-op, and just produce
new A?() again. So, to make a long story short, the C# compiler can codegen the code above as:
C? c; A? temp1 = x; A? temp2 = y; c = temp1.HasValue & temp2.HasValue ? new C?((C)(B)(~(temp1.GetValueOrDefault() + temp2.GetValueOrDefault())) : new C?();
Again, we save several temporaries and branches by performing this optimization.
Now, I've been saying "the compiler could" a lot because of course a compiler is not required to perform these optimizations, and in fact, the "original recipe" compiler is not very aggressive about performing these optimizations. I examined the original recipe compiler very closely when implementing nullable arithmetic in Roslyn, and discovered that it suffers from a case of "premature optimization".
Next time on FAIC: We'll digress for the next couple of posts. Then I'll pick up this subject again with a discussion of the evils of "premature optimization" of nullable arithmetic, and how I'm using that loaded term in a subtly different way than Knuth did.