| Index: tests_lit/llvm2ice_tests/struct-arith.pnacl.ll | 
| diff --git a/tests_lit/llvm2ice_tests/struct-arith.pnacl.ll b/tests_lit/llvm2ice_tests/struct-arith.pnacl.ll | 
| index de68416c61ccec35cf2b6a53edc91b009edae9fd..3e88b0fd0a35d87046012fd882b6b2e5ee3bd56e 100644 | 
| --- a/tests_lit/llvm2ice_tests/struct-arith.pnacl.ll | 
| +++ b/tests_lit/llvm2ice_tests/struct-arith.pnacl.ll | 
| @@ -1,6 +1,8 @@ | 
| ; RUIN: %llvm2ice -verbose inst %s | FileCheck %s | 
| ; RUIN: %llvm2ice --verbose none %s | FileCheck --check-prefix=ERRORS %s | 
| -; RUN: %szdiff --llvm2ice=%llvm2ice %s | FileCheck --check-prefix=DUMP %s | 
| +; RUN: %llvm2iceinsts %s | %szdiff %s | FileCheck --check-prefix=DUMP %s | 
| +; RUN: %llvm2iceinsts --pnacl %s | %szdiff %s \ | 
| +; RUN:                           | FileCheck --check-prefix=DUMP %s | 
|  | 
| ; This file is lowered from C code that does some simple aritmetic with | 
| ; struct members. It's also built with the PNaCl toolchain so this is the | 
| @@ -9,37 +11,37 @@ | 
|  | 
| define internal i32 @compute_important_function(i32 %v1, i32 %v2) { | 
| entry: | 
| -  %v1.asptr = inttoptr i32 %v1 to i32* | 
| -  %_v0 = load i32* %v1.asptr, align 1 | 
| +  %__2 = inttoptr i32 %v1 to i32* | 
| +  %_v0 = load i32* %__2, align 1 | 
|  | 
| ; CHECK:        entry: | 
| -; CHECK-NEXT:       %v1.asptr = i32 %v1 | 
| +; CHECK-NEXT:       %__2 = i32 %v1 | 
| ; CHECK-NEXT:       %_v0 = load i32* {{.*}}, align 1 | 
|  | 
| -  %v2.asptr = inttoptr i32 %v2 to i32* | 
| -  %_v1 = load i32* %v2.asptr, align 1 | 
| +  %__4 = inttoptr i32 %v2 to i32* | 
| +  %_v1 = load i32* %__4, align 1 | 
| %gep = add i32 %v2, 12 | 
| -  %gep.asptr = inttoptr i32 %gep to i32* | 
| -  %_v2 = load i32* %gep.asptr, align 1 | 
| +  %__7 = inttoptr i32 %gep to i32* | 
| +  %_v2 = load i32* %__7, align 1 | 
| %mul = mul i32 %_v2, %_v1 | 
| %gep6 = add i32 %v1, 4 | 
| -  %gep6.asptr = inttoptr i32 %gep6 to i32* | 
| -  %_v3 = load i32* %gep6.asptr, align 1 | 
| +  %__11 = inttoptr i32 %gep6 to i32* | 
| +  %_v3 = load i32* %__11, align 1 | 
| %gep8 = add i32 %v2, 8 | 
| -  %gep8.asptr = inttoptr i32 %gep8 to i32* | 
| -  %_v4 = load i32* %gep8.asptr, align 1 | 
| +  %__14 = inttoptr i32 %gep8 to i32* | 
| +  %_v4 = load i32* %__14, align 1 | 
| %gep10 = add i32 %v2, 4 | 
| -  %gep10.asptr = inttoptr i32 %gep10 to i32* | 
| -  %_v5 = load i32* %gep10.asptr, align 1 | 
| +  %__17 = inttoptr i32 %gep10 to i32* | 
| +  %_v5 = load i32* %__17, align 1 | 
| %mul3 = mul i32 %_v5, %_v4 | 
| %gep12 = add i32 %v1, 8 | 
| -  %gep12.asptr = inttoptr i32 %gep12 to i32* | 
| -  %_v6 = load i32* %gep12.asptr, align 1 | 
| +  %__21 = inttoptr i32 %gep12 to i32* | 
| +  %_v6 = load i32* %__21, align 1 | 
| %mul7 = mul i32 %_v6, %_v3 | 
| %mul9 = mul i32 %mul7, %_v6 | 
| %gep14 = add i32 %v1, 12 | 
| -  %gep14.asptr = inttoptr i32 %gep14 to i32* | 
| -  %_v7 = load i32* %gep14.asptr, align 1 | 
| +  %__26 = inttoptr i32 %gep14 to i32* | 
| +  %_v7 = load i32* %__26, align 1 | 
| %mul11 = mul i32 %mul9, %_v7 | 
| %add4.neg = add i32 %mul, %_v0 | 
| %add = sub i32 %add4.neg, %_v3 | 
|  |