OLD | NEW |
1 #!/bin/bash | 1 #!/bin/bash |
2 | 2 |
3 # Self-tests for gm, based on tools/tests/run.sh | 3 # Self-tests for gm, based on tools/tests/run.sh |
4 # | 4 # |
5 # These tests are run by the Skia_PerCommit_House_Keeping bot at every commit, | 5 # These tests are run by the Skia_PerCommit_House_Keeping bot at every commit, |
6 # so make sure that they still pass when you make changes to gm! | 6 # so make sure that they still pass when you make changes to gm! |
7 # | 7 # |
8 # To generate new baselines when gm behavior changes, run gm/tests/rebaseline.sh | 8 # To generate new baselines when gm behavior changes, run gm/tests/rebaseline.sh |
9 # | 9 # |
10 # TODO: because this is written as a shell script (instead of, say, Python) | 10 # TODO: because this is written as a shell script (instead of, say, Python) |
(...skipping 190 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
201 | 201 |
202 # Test what happens if a subset of the renderModes fail (e.g. pipe) | 202 # Test what happens if a subset of the renderModes fail (e.g. pipe) |
203 gm_test "--simulatePipePlaybackFailure --verbose --hierarchy --match selftest1 $
CONFIGS -r $GM_INPUTS/json/identical-pixels.json" "$GM_OUTPUTS/pipe-playback-fai
lure" | 203 gm_test "--simulatePipePlaybackFailure --verbose --hierarchy --match selftest1 $
CONFIGS -r $GM_INPUTS/json/identical-pixels.json" "$GM_OUTPUTS/pipe-playback-fai
lure" |
204 | 204 |
205 # Confirm that IntentionallySkipped tests are recorded as such. | 205 # Confirm that IntentionallySkipped tests are recorded as such. |
206 gm_test "--verbose --hierarchy --match selftest1 selftest2 $CONFIGS" "$GM_OUTPUT
S/intentionally-skipped-tests" | 206 gm_test "--verbose --hierarchy --match selftest1 selftest2 $CONFIGS" "$GM_OUTPUT
S/intentionally-skipped-tests" |
207 | 207 |
208 # Ignore some error types (including ExpectationsMismatch) | 208 # Ignore some error types (including ExpectationsMismatch) |
209 gm_test "--ignoreErrorTypes ExpectationsMismatch NoGpuContext --verbose --hierar
chy --match selftest1 $CONFIGS -r $GM_INPUTS/json/different-pixels.json" "$GM_OU
TPUTS/ignore-expectations-mismatch" | 209 gm_test "--ignoreErrorTypes ExpectationsMismatch NoGpuContext --verbose --hierar
chy --match selftest1 $CONFIGS -r $GM_INPUTS/json/different-pixels.json" "$GM_OU
TPUTS/ignore-expectations-mismatch" |
210 | 210 |
211 # Exercise confirm_no_failures_in_json.py | 211 # Exercise display_json_results.py |
212 PASSING_CASES="compared-against-identical-bytes-json compared-against-identical-
pixels-json" | 212 PASSING_CASES="compared-against-identical-bytes-json compared-against-identical-
pixels-json" |
213 FAILING_CASES="compared-against-different-pixels-json" | 213 FAILING_CASES="compared-against-different-pixels-json" |
214 for CASE in $PASSING_CASES; do | 214 for CASE in $PASSING_CASES; do |
215 assert_passes "python gm/confirm_no_failures_in_json.py $GM_OUTPUTS/$CASE/$OUT
PUT_EXPECTED_SUBDIR/json-summary.txt" | 215 assert_passes "python gm/display_json_results.py $GM_OUTPUTS/$CASE/$OUTPUT_EXP
ECTED_SUBDIR/json-summary.txt" |
216 done | 216 done |
217 for CASE in $FAILING_CASES; do | 217 for CASE in $FAILING_CASES; do |
218 assert_fails "python gm/confirm_no_failures_in_json.py $GM_OUTPUTS/$CASE/$OUTP
UT_EXPECTED_SUBDIR/json-summary.txt" | 218 assert_fails "python gm/display_json_results.py $GM_OUTPUTS/$CASE/$OUTPUT_EXPE
CTED_SUBDIR/json-summary.txt" |
219 done | 219 done |
220 | 220 |
221 if [ $ENCOUNTERED_ANY_ERRORS == 0 ]; then | 221 if [ $ENCOUNTERED_ANY_ERRORS == 0 ]; then |
222 echo "All tests passed." | 222 echo "All tests passed." |
223 exit 0 | 223 exit 0 |
224 else | 224 else |
225 exit 1 | 225 exit 1 |
226 fi | 226 fi |
OLD | NEW |