DescriptionReal world impact script: scalable manual rendering QA
Layout Tests can tell you whether rendering has changed, but it's often
hard to determine whether a subtle/controversial change is beneficial
without additional context.
This script aims to provide that context. It takes screenshots of 1000s
of sites, both before and after applying the patch being evaluated,
then diffs each pair of screenshots and sorts them by greatest
difference in rendering, in an HTML report. A human reviewer can then
manually skim through this to review the most impacted sites, rather
than having to browse random sites to see what their patch changed.
This is a script I wrote at the start of the year (but hadn't got round
to uploading a patch, as it used to depend on proprietary binaries;
I fixed that dependency by writing a replacement: crrev.com/67973005).
The script isn't yet smart enough to build content_shell itself, so you
run it in several steps:
1. Build content_shell in out/Release, without the controversial patch.
2. Run: real_world_impact.py before [num sites to test (default 1000)]
3. Apply the controversial patch, and rebuild content_shell in out/Release.
4. Run: real_world_impact.py after [num sites to test (default 1000)]
5. Run: real_world_impact.py compare [num sites to test (default 1000)]
In practice I use the following bash function to drive it:
http://pastebin.com/yJ7KAGdX
BUG=135823
Patch Set 1 #
Total comments: 1
Messages
Total messages: 5 (0 generated)
|