| // RUN: %clangxx_tsan %s -o %t |
| // RUN: %run %t 2>&1 | FileCheck %s |
| |
| // bench.h needs pthread barriers which are not available on OS X |
| // UNSUPPORTED: darwin |
| |
| #include "bench.h" |
| #include <memory.h> |
| |
| void thread(int tid) { |
| volatile long x = 0; |
| switch (bench_mode) { |
| case 0: |
| for (int i = 0; i < bench_niter; i++) |
| *(volatile char *)&x = 1; |
| break; |
| case 1: |
| for (int i = 0; i < bench_niter; i++) |
| *(volatile short *)&x = 1; |
| break; |
| case 2: |
| for (int i = 0; i < bench_niter; i++) |
| *(volatile int *)&x = 1; |
| break; |
| case 3: |
| for (int i = 0; i < bench_niter; i++) |
| *(volatile long *)&x = 1; |
| break; |
| case 4: |
| for (int i = 0; i < bench_niter; i++) |
| *(volatile char *)&x; |
| break; |
| case 5: |
| for (int i = 0; i < bench_niter; i++) |
| *(volatile short *)&x; |
| break; |
| case 6: |
| for (int i = 0; i < bench_niter; i++) |
| *(volatile int *)&x; |
| break; |
| case 7: |
| for (int i = 0; i < bench_niter; i++) |
| *(volatile long *)&x; |
| case 8: |
| for (int i = 0; i < bench_niter / 10; i++) { |
| ((volatile long *)&x)[0]; |
| ((volatile int *)&x)[0]; |
| ((volatile short *)&x)[2]; |
| ((volatile char *)&x)[6]; |
| ((volatile char *)&x)[7]; |
| ((volatile long *)&x)[0] = 1; |
| ((volatile int *)&x)[0] = 1; |
| ((volatile short *)&x)[2] = 1; |
| ((volatile char *)&x)[6] = 1; |
| ((volatile char *)&x)[7] = 1; |
| } |
| break; |
| case 9: { |
| volatile long size = sizeof(x); |
| for (int i = 0; i < bench_niter; i++) |
| memset((void *)&x, i, size); |
| break; |
| } |
| case 10: { |
| volatile long data[2] = {}; |
| volatile long size = sizeof(data) - 2; |
| for (int i = 0; i < bench_niter; i++) |
| memset(((char *)data) + 1, i, size); |
| break; |
| } |
| case 11: { |
| volatile long data[2] = {}; |
| for (int i = 0; i < bench_niter / 8 / 3; i++) { |
| for (int off = 0; off < 8; off++) { |
| __sanitizer_unaligned_store16(((char *)data) + off, i); |
| __sanitizer_unaligned_store32(((char *)data) + off, i); |
| __sanitizer_unaligned_store64(((char *)data) + off, i); |
| } |
| } |
| break; |
| } |
| #if TSAN_VECTORIZE |
| case 12: { |
| // The compiler wants to optimize all this away. |
| // Use volatile to prevent optimization, but then use kBlock |
| // to avoid the additional non-vector load in the inner loop. |
| // Also use only even indexes to prevent compiler from |
| // inserting memset. |
| const int kBlock = 128; |
| __m128i data[kBlock * 2]; |
| __m128i *volatile vptr = data; |
| for (int i = 0; i < bench_niter / kBlock; i++) { |
| __m128i *ptr = vptr; |
| for (int j = 0; j < kBlock; j++) |
| _mm_store_si128(&ptr[j * 2], _mm_setzero_si128()); |
| } |
| break; |
| } |
| #endif |
| } |
| } |
| |
| void bench() { |
| start_thread_group(bench_nthread, thread); |
| } |
| |
| // CHECK: DONE |