llvm/compiler-rt/test/tsan/bench_memory_access.cpp

// RUN: %clangxx_tsan %s -o %t
// RUN: %run %t 2>&1 | FileCheck %s

// bench.h needs pthread barriers which are not available on OS X
// UNSUPPORTED: darwin

#include "bench.h"
#include <memory.h>

void thread(int tid) {
  volatile long x = 0;
  switch (bench_mode) {
  case 0:
    for (int i = 0; i < bench_niter; i++)
      *(volatile char *)&x = 1;
    break;
  case 1:
    for (int i = 0; i < bench_niter; i++)
      *(volatile short *)&x = 1;
    break;
  case 2:
    for (int i = 0; i < bench_niter; i++)
      *(volatile int *)&x = 1;
    break;
  case 3:
    for (int i = 0; i < bench_niter; i++)
      *(volatile long *)&x = 1;
    break;
  case 4:
    for (int i = 0; i < bench_niter; i++)
      *(volatile char *)&x;
    break;
  case 5:
    for (int i = 0; i < bench_niter; i++)
      *(volatile short *)&x;
    break;
  case 6:
    for (int i = 0; i < bench_niter; i++)
      *(volatile int *)&x;
    break;
  case 7:
    for (int i = 0; i < bench_niter; i++)
      *(volatile long *)&x;
  case 8:
    for (int i = 0; i < bench_niter / 10; i++) {
      ((volatile long *)&x)[0];
      ((volatile int *)&x)[0];
      ((volatile short *)&x)[2];
      ((volatile char *)&x)[6];
      ((volatile char *)&x)[7];
      ((volatile long *)&x)[0] = 1;
      ((volatile int *)&x)[0] = 1;
      ((volatile short *)&x)[2] = 1;
      ((volatile char *)&x)[6] = 1;
      ((volatile char *)&x)[7] = 1;
    }
    break;
  case 9: {
    volatile long size = sizeof(x);
    for (int i = 0; i < bench_niter; i++)
      memset((void *)&x, i, size);
    break;
  }
  case 10: {
    volatile long data[2] = {};
    volatile long size = sizeof(data) - 2;
    for (int i = 0; i < bench_niter; i++)
      memset(((char *)data) + 1, i, size);
    break;
  }
  case 11: {
    volatile long data[2] = {};
    for (int i = 0; i < bench_niter / 8 / 3; i++) {
      for (int off = 0; off < 8; off++) {
        __sanitizer_unaligned_store16(((char *)data) + off, i);
        __sanitizer_unaligned_store32(((char *)data) + off, i);
        __sanitizer_unaligned_store64(((char *)data) + off, i);
      }
    }
    break;
  }
#if TSAN_VECTORIZE
  case 12: {
    // The compiler wants to optimize all this away.
    // Use volatile to prevent optimization, but then use kBlock
    // to avoid the additional non-vector load in the inner loop.
    // Also use only even indexes to prevent compiler from
    // inserting memset.
    const int kBlock = 128;
    __m128i data[kBlock * 2];
    __m128i *volatile vptr = data;
    for (int i = 0; i < bench_niter / kBlock; i++) {
      __m128i *ptr = vptr;
      for (int j = 0; j < kBlock; j++)
        _mm_store_si128(&ptr[j * 2], _mm_setzero_si128());
    }
    break;
  }
#endif
  }
}

void bench() {
  start_thread_group(bench_nthread, thread);
}

// CHECK: DONE