This document describes the details of how WebRender renders text, particularly the blending stage of text rendering. We will go into grayscale text blending, subpixel text blending, and "subpixel text with background color" blending.
The description below assumes you're familiar with regular rgba compositing, operator over, and the concept of premultiplied alpha.
We are going to treat the origin of the text mask as a black box. We're also going to assume we can blend text in the device color space and will not go into the gamma correction and linear pre-blending that happens in some of the backends that produce the text masks.
Grayscale text blending is the simplest form of text blending. Our blending function has three inputs:
- The text color, as a premultiplied rgba color.
- The text mask, as a single-channel alpha texture.
- The existing contents of the framebuffer that we're rendering to, the "destination". This is also a premultiplied rgba buffer.
Note: The word "grayscale" here does not mean that we can only draw gray text. It means that the mask only has a single alpha value per pixel, so we can visualize the mask in our minds as a grayscale image.
We want to mask our text color using the single-channel mask, and composite that to the destination. This compositing step uses operator "over", just like regular compositing of rgba images.
I'll be using GLSL syntax to describe the blend equations, but please consider most of the code below pseudocode.
We can express the blending described above as the following blend equation:
vec4 textblend(vec4 text_color, vec4 mask, vec4 dest) {
return over(in(text_color, mask), dest);
}
with over
being the blend function for (premultiplied) operator "over":
vec4 over(vec4 src, vec4 dest) {
return src + (1.0 - src.a) * dest;
}
and in
being the blend function for (premultiplied) operator "in", i.e. the masking operator:
vec4 in(vec4 src, vec4 mask) {
return src * mask.a;
}
So the complete blending function is:
result.r = text_color.r * mask.a + (1.0 - text_color.a * mask.a) * dest.r;
result.g = text_color.g * mask.a + (1.0 - text_color.a * mask.a) * dest.g;
result.b = text_color.b * mask.a + (1.0 - text_color.a * mask.a) * dest.b;
result.a = text_color.a * mask.a + (1.0 - text_color.a * mask.a) * dest.a;
In general, a fragment shader does not have access to the destination.
So the full blend equation needs to be expressed in a way that the shader only computes values that are independent of the destination,
and the parts of the equation that use the destination values need to be applied by the OpenGL blend pipeline itself.
The OpenGL blend pipeline can be tweaked using the functions glBlendEquation
and glBlendFunc
.
In our example, the fragment shader can output just text_color * mask.a
:
oFragColor = text_color * mask.a;
and the OpenGL blend pipeline can be configured like so:
pub fn set_blend_mode_premultiplied_alpha(&self) {
self.gl.blend_func(gl::ONE, gl::ONE_MINUS_SRC_ALPHA);
self.gl.blend_equation(gl::FUNC_ADD);
}
This results in an overall blend equation of
result.r = 1 * oFragColor.r + (1 - oFragColor.a) * dest.r;
^ ^ ^^^^^^^^^^^^^^^^^
| | |
+--gl::ONE | +-- gl::ONE_MINUS_SRC_ALPHA
|
+-- gl::FUNC_ADD
= 1 * (text_color.r * mask.a) + (1 - (text_color.a * mask.a)) * dest.r
= text_color.r * mask.a + (1 - text_color.a * mask.a) * dest.r
which is exactly what we wanted.
There are two minor differences between the shader code above and the actual code in the text run shader in WebRender:
oFragColor = text_color * mask.a; // (shown above)
// vs.
oFragColor = vColor * mask * alpha; // (actual webrender code)
vColor
is set to the text color. The differences are:
- WebRender multiplies with all components of
mask
instead of just withmask.a
. However, our font rasterization code fills the rgb values ofmask
with the value ofmask.a
, so this is completely equivalent. - WebRender applies another alpha to the text. This is coming from the clip. You can think of this alpha to be a pre-adjustment of the text color for that pixel, or as an additional mask that gets applied to the mask.
Now that we have the blend equation for single-channel text blending, we can look at subpixel text blending.
The main difference between subpixel text blending and grayscale text blending is the fact that, for subpixel text, the text mask contains a separate alpha value for each color component.
Regular painting uses four values per pixel: three color values, and one alpha value. The alpha value applies to all components of the pixel equally.
Imagine for a second a world in which you have three alpha values per pixel, one for each color component.
- Old world: Each pixel has four values:
color.r
,color.g
,color.b
, andcolor.a
. - New world: Each pixel has six values:
color.r
,color.a_r
,color.g
,color.a_g
,color.b
, andcolor.a_b
.
In such a world we can define a component-alpha-aware operator "over":
vec6 over_comp(vec6 src, vec6 dest) {
vec6 result;
result.r = src.r + (1.0 - src.a_r) * dest.r;
result.g = src.g + (1.0 - src.a_g) * dest.g;
result.b = src.b + (1.0 - src.a_b) * dest.b;
result.a_r = src.a_r + (1.0 - src.a_r) * dest.a_r;
result.a_g = src.a_g + (1.0 - src.a_g) * dest.a_g;
result.a_b = src.a_b + (1.0 - src.a_b) * dest.a_b;
return result;
}
and a component-alpha-aware operator "in":
vec6 in_comp(vec6 src, vec6 mask) {
vec6 result;
result.r = src.r * mask.a_r;
result.g = src.g * mask.a_g;
result.b = src.b * mask.a_b;
result.a_r = src.a_r * mask.a_r;
result.a_g = src.a_g * mask.a_g;
result.a_b = src.a_b * mask.a_b;
return result;
}
and even a component-alpha-aware version of textblend
:
vec6 textblend_comp(vec6 text_color, vec6 mask, vec6 dest) {
return over_comp(in_comp(text_color, mask), dest);
}
This results in the following set of equations:
result.r = text_color.r * mask.a_r + (1.0 - text_color.a_r * mask.a_r) * dest.r;
result.g = text_color.g * mask.a_g + (1.0 - text_color.a_g * mask.a_g) * dest.g;
result.b = text_color.b * mask.a_b + (1.0 - text_color.a_b * mask.a_b) * dest.b;
result.a_r = text_color.a_r * mask.a_r + (1.0 - text_color.a_r * mask.a_r) * dest.a_r;
result.a_g = text_color.a_g * mask.a_g + (1.0 - text_color.a_g * mask.a_g) * dest.a_g;
result.a_b = text_color.a_b * mask.a_b + (1.0 - text_color.a_b * mask.a_b) * dest.a_b;
If we want to transfer the component alpha blend equation into the real world, we need to make a few small changes:
- Our text color only needs one alpha value.
So we'll replace all instances of
text_color.a_r/g/b
withtext_color.a
. - We're currently not making use of the mask's
r
,g
andb
values, only of thea_r
,a_g
anda_b
values. So in the real world, we can use the rgb channels ofmask
to store those component alphas and replacemask.a_r/g/b
withmask.r/g/b
.
These two changes give us:
result.r = text_color.r * mask.r + (1.0 - text_color.a * mask.r) * dest.r;
result.g = text_color.g * mask.g + (1.0 - text_color.a * mask.g) * dest.g;
result.b = text_color.b * mask.b + (1.0 - text_color.a * mask.b) * dest.b;
result.a_r = text_color.a * mask.r + (1.0 - text_color.a * mask.r) * dest.a_r;
result.a_g = text_color.a * mask.g + (1.0 - text_color.a * mask.g) * dest.a_g;
result.a_b = text_color.a * mask.b + (1.0 - text_color.a * mask.b) * dest.a_b;
There's a third change we need to make:
- We're rendering to a destination surface that only has one alpha channel instead of three.
So
dest.a_r/g/b
andresult.a_r/g/b
will need to becomedest.a
andresult.a
.
This creates a problem: We're currently assigning different values to result.a_r
, result.a_g
and result.a_b
.
Which of them should we use to compute result.a
?
This question does not have an answer. One alpha value per pixel is simply not sufficient to express the same information as three alpha values.
However, see what happens if the destination is already opaque:
We have dest.a_r == 1
, dest.a_g == 1
, and dest.a_b == 1
.
result.a_r = text_color.a * mask.r + (1 - text_color.a * mask.r) * dest.a_r
= text_color.a * mask.r + (1 - text_color.a * mask.r) * 1
= text_color.a * mask.r + 1 - text_color.a * mask.r
= 1
same for result.a_g and result.a_b
In other words, for opaque destinations, it doesn't matter what which channel of the mask we use when computing result.a
, the result will always be completely opaque anyways. In WebRender we just pick mask.g
(or rather,
have font rasterization set mask.a
to the value of mask.g
) because it's as good as any.
The takeaway here is: Subpixel text blending is only supported for opaque destinations. Attempting to render subpixel text into partially transparent destinations will result in bad alpha values. Or rather, it will result in alpha values which are not anticipated by the r, g, and b values in the same pixel, so that subsequent blend operations, which will mix r and a values from the same pixel, will produce incorrect colors.
Here's the final subpixel blend function:
vec4 subpixeltextblend(vec4 text_color, vec4 mask, vec4 dest) {
vec4 result;
result.r = text_color.r * mask.r + (1.0 - text_color.a * mask.r) * dest.r;
result.g = text_color.g * mask.g + (1.0 - text_color.a * mask.g) * dest.g;
result.b = text_color.b * mask.b + (1.0 - text_color.a * mask.b) * dest.b;
result.a = text_color.a * mask.a + (1.0 - text_color.a * mask.a) * dest.a;
return result;
}
or for short:
vec4 subpixeltextblend(vec4 text_color, vec4 mask, vec4 dest) {
return text_color * mask + (1.0 - text_color.a * mask) * dest;
}
To recap, here's what we gained and lost by making the transition from the full-component-alpha world to the regular rgba world: All colors and textures now only need four values to be represented, we still use a component alpha mask, and the results are equivalent to the full-component-alpha result assuming that the destination is opaque. We lost the ability to draw to partially transparent destinations.
We have the complete subpixel blend function. Now we need to cut it into pieces and mix it with the OpenGL blend pipeline in such a way that the fragment shader does not need to know about the destination.
Compare the equation for the red channel and the alpha channel between the two ways of text blending:
single-channel alpha:
result.r = text_color.r * mask.a + (1.0 - text_color.a * mask.a) * dest.r
result.a = text_color.a * mask.a + (1.0 - text_color.a * mask.a) * dest.r
component alpha:
result.r = text_color.r * mask.r + (1.0 - text_color.a * mask.r) * dest.r
result.a = text_color.a * mask.a + (1.0 - text_color.a * mask.a) * dest.r
Notably, in the single-channel alpha case, all three destination color channels are multiplied with the same thing:
(1.0 - text_color.a * mask.a)
. This factor also happens to be "one minus oFragColor.a
".
So we were able to take advantage of OpenGL's ONE_MINUS_SRC_ALPHA
blend func.
In the component alpha case, we're not so lucky: Each destination color channel
is multiplied with a different factor. We can use ONE_MINUS_SRC_COLOR
instead,
and output text_color.a * mask
from our fragment shader.
But then there's still the problem that the first summand of the computation for result.r
uses
text_color.r * mask.r
and the second summand uses text_color.a * mask.r
.
There are multiple ways to deal with this. They are:
- Making use of
glBlendColor
and theGL_CONSTANT_COLOR
blend func. - Using a two-pass method.
- Using "dual source blending".
Let's look at them in order.
In this approach we return text_color.a * mask
from the shader.
Then we set the blend color to text_color / text_color.a
and use GL_CONSTANT_COLOR
as the source blendfunc.
This results in the following blend equation:
result.r = (text_color.r / text_color.a) * oFragColor.r + (1 - oFragColor.r) * dest.r;
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ^ ^^^^^^^^^^^^^^^^^
| | |
+--gl::CONSTANT_COLOR | +-- gl::ONE_MINUS_SRC_COLOR
|
+-- gl::FUNC_ADD
= (text_color.r / text_color.a) * (text_color.a * mask.r) + (1 - (text_color.a * mask.r)) * dest.r
= text_color.r * mask.r + (1 - text_color.a * mask.r) * dest.r
At the very beginning of this document, we defined text_color
as the premultiplied text color.
So instead of actually doing the calculation text_color.r / text_color.a
when specifying the blend color,
we really just want to use the unpremultiplied text color in that place.
That's usually the representation we start with anyway.
The glBlendColor
method has the disadvantage that the text color is part of the OpenGL state.
So if we want to draw text with different colors, we have two use separate batches / draw calls
to draw the differently-colored parts of text.
Alternatively, we can use a two-pass method which avoids the need to use the GL_CONSTANT_COLOR
blend func:
- The first pass outputs
text_color.a * mask
from the fragment shader and usesgl::ZERO, gl::ONE_MINUS_SRC_COLOR
as the glBlendFuncs. This achieves:
oFragColor = text_color.a * mask;
result_after_pass0.r = 0 * oFragColor.r + (1 - oFragColor.r) * dest.r
= (1 - text_color.a * mask.r) * dest.r
result_after_pass0.g = 0 * oFragColor.g + (1 - oFragColor.g) * dest.r
= (1 - text_color.a * mask.r) * dest.r
...
- The second pass outputs
text_color * mask
from the fragment shader and usesgl::ONE, gl::ONE
as the glBlendFuncs. This results in the correct overall blend equation.
oFragColor = text_color * mask;
result_after_pass1.r
= 1 * oFragColor.r + 1 * result_after_pass0.r
= text_color.r * mask.r + result_after_pass0.r
= text_color.r * mask.r + (1 - text_color.a * mask.r) * dest.r
The third approach is similar to the second approach, but makes use of the ARB_blend_func_extended
extension
in order to fold the two passes into one:
Instead of outputting the two different colors in two separate passes, we output them from the same pass,
as two separate fragment shader outputs.
Those outputs can then be treated as two different sources in the blend equation.