diff options
author | Chandler Carruth <chandlerc@gmail.com> | 2012-12-06 11:14:44 +0000 |
---|---|---|
committer | Chandler Carruth <chandlerc@gmail.com> | 2012-12-06 11:14:44 +0000 |
commit | 72d2dab6058467036df73a5f668036a519043e5b (patch) | |
tree | a0544e418f5ba7594ea2dcff9a6209eddcc7887f /test/CodeGenCXX/bitfield.cpp | |
parent | 3b5943f535eca0e0b91f4bcda9d09c9c275cf855 (diff) |
Rework the bitfield access IR generation to address PR13619 and
generally support the C++11 memory model requirements for bitfield
accesses by relying more heavily on LLVM's memory model.
The primary change this introduces is to move from a manually aligned
and strided access pattern across the bits of the bitfield to a much
simpler lump access of all bits in the bitfield followed by math to
extract the bits relevant for the particular field.
This simplifies the code significantly, but relies on LLVM to
intelligently lowering these integers.
I have tested LLVM's lowering both synthetically and in benchmarks. The
lowering appears to be functional, and there are no really significant
performance regressions. Different code patterns accessing bitfields
will vary in how this impacts them. The only real regressions I'm seeing
are a few patterns where the LLVM code generation for loads that feed
directly into a mask operation don't take advantage of the x86 ability
to do a smaller load and a cheap zero-extension. This doesn't regress
any benchmark in the nightly test suite on my box past the noise
threshold, but my box is quite noisy. I'll be watching the LNT numbers,
and will look into further improvements to the LLVM lowering as needed.
git-svn-id: https://llvm.org/svn/llvm-project/cfe/trunk@169489 91177308-0d34-0410-b5e6-96231b3b80d8
Diffstat (limited to 'test/CodeGenCXX/bitfield.cpp')
-rw-r--r-- | test/CodeGenCXX/bitfield.cpp | 195 |
1 files changed, 195 insertions, 0 deletions
diff --git a/test/CodeGenCXX/bitfield.cpp b/test/CodeGenCXX/bitfield.cpp new file mode 100644 index 0000000000..a737304c98 --- /dev/null +++ b/test/CodeGenCXX/bitfield.cpp @@ -0,0 +1,195 @@ +// RUN: %clang_cc1 -triple x86_64-unknown-unknown -verify -emit-llvm -o - %s | FileCheck %s +// +// Tests for bitfield access patterns in C++ with special attention to +// conformance to C++11 memory model requirements. + +namespace N1 { + // Ensure that neither loads nor stores to bitfields are not widened into + // other memory locations. (PR13691) + // + // NOTE: We could potentially widen loads based on their alignment if we are + // comfortable requiring that subsequent memory locations within the + // alignment-widened load are not volatile. + struct S { + char a; + unsigned b : 1; + char c; + }; + unsigned read(S* s) { + // CHECK: define i32 @_ZN2N14read + // CHECK: %[[ptr:.*]] = getelementptr inbounds %{{.*}}* %{{.*}}, i32 0, i32 1 + // CHECK: %[[val:.*]] = load i8* %[[ptr]] + // CHECK: %[[and:.*]] = and i8 %[[val]], 1 + // CHECK: %[[ext:.*]] = zext i8 %[[and]] to i32 + // CHECK: ret i32 %[[ext]] + return s->b; + } + void write(S* s, unsigned x) { + // CHECK: define void @_ZN2N15write + // CHECK: %[[ptr:.*]] = getelementptr inbounds %{{.*}}* %{{.*}}, i32 0, i32 1 + // CHECK: %[[x_trunc:.*]] = trunc i32 %{{.*}} to i8 + // CHECK: %[[old:.*]] = load i8* %[[ptr]] + // CHECK: %[[x_and:.*]] = and i8 %[[x_trunc]], 1 + // CHECK: %[[old_and:.*]] = and i8 %[[old]], -2 + // CHECK: %[[new:.*]] = or i8 %[[old_and]], %[[x_and]] + // CHECK: store i8 %[[new]], i8* %[[ptr]] + s->b = x; + } +} + +namespace N2 { + // Do widen loads and stores to bitfields when those bitfields have padding + // within the struct following them. + struct S { + unsigned b : 24; + void *p; + }; + unsigned read(S* s) { + // CHECK: define i32 @_ZN2N24read + // CHECK: %[[ptr:.*]] = bitcast %{{.*}}* %{{.*}} to i32* + // CHECK: %[[val:.*]] = load i32* %[[ptr]] + // CHECK: %[[and:.*]] = and i32 %[[val]], 16777215 + // CHECK: ret i32 %[[and]] + return s->b; + } + void write(S* s, unsigned x) { + // CHECK: define void @_ZN2N25write + // CHECK: %[[ptr:.*]] = bitcast %{{.*}}* %{{.*}} to i32* + // CHECK: %[[old:.*]] = load i32* %[[ptr]] + // CHECK: %[[x_and:.*]] = and i32 %{{.*}}, 16777215 + // CHECK: %[[old_and:.*]] = and i32 %[[old]], -16777216 + // CHECK: %[[new:.*]] = or i32 %[[old_and]], %[[x_and]] + // CHECK: store i32 %[[new]], i32* %[[ptr]] + s->b = x; + } +} + +namespace N3 { + // Do widen loads and stores to bitfields through the trailing padding at the + // end of a struct. + struct S { + unsigned b : 24; + }; + unsigned read(S* s) { + // CHECK: define i32 @_ZN2N34read + // CHECK: %[[ptr:.*]] = bitcast %{{.*}}* %{{.*}} to i32* + // CHECK: %[[val:.*]] = load i32* %[[ptr]] + // CHECK: %[[and:.*]] = and i32 %[[val]], 16777215 + // CHECK: ret i32 %[[and]] + return s->b; + } + void write(S* s, unsigned x) { + // CHECK: define void @_ZN2N35write + // CHECK: %[[ptr:.*]] = bitcast %{{.*}}* %{{.*}} to i32* + // CHECK: %[[old:.*]] = load i32* %[[ptr]] + // CHECK: %[[x_and:.*]] = and i32 %{{.*}}, 16777215 + // CHECK: %[[old_and:.*]] = and i32 %[[old]], -16777216 + // CHECK: %[[new:.*]] = or i32 %[[old_and]], %[[x_and]] + // CHECK: store i32 %[[new]], i32* %[[ptr]] + s->b = x; + } +} + +namespace N4 { + // Do NOT widen loads and stores to bitfields into padding at the end of + // a class which might end up with members inside of it when inside a derived + // class. + struct Base { + virtual ~Base() {} + + unsigned b : 24; + }; + // Imagine some other translation unit introduces: +#if 0 + struct Derived : public Base { + char c; + }; +#endif + unsigned read(Base* s) { + // FIXME: We should widen this load as long as the function isn't being + // instrumented by thread-sanitizer. + // + // CHECK: define i32 @_ZN2N44read + // CHECK: %[[ptr:.*]] = bitcast {{.*}}* %{{.*}} to i24* + // CHECK: %[[val:.*]] = load i24* %[[ptr]] + // CHECK: %[[ext:.*]] = zext i24 %[[val]] to i32 + // CHECK: ret i32 %[[ext]] + return s->b; + } + void write(Base* s, unsigned x) { + // CHECK: define void @_ZN2N45write + // CHECK: %[[ptr:.*]] = bitcast {{.*}}* %{{.*}} to i24* + // CHECK: %[[new:.*]] = trunc i32 %{{.*}} to i24 + // CHECK: store i24 %[[new]], i24* %[[ptr]] + s->b = x; + } +} + +namespace N5 { + // Widen through padding at the end of a struct even if that struct + // participates in a union with another struct which has a separate field in + // that location. The reasoning is that if the operation is storing to that + // member of the union, it must be the active member, and thus we can write + // through the padding. If it is a load, it might be a load of a common + // prefix through a non-active member, but in such a case the extra bits + // loaded are masked off anyways. + union U { + struct X { unsigned b : 24; char c; } x; + struct Y { unsigned b : 24; } y; + }; + unsigned read(U* u) { + // CHECK: define i32 @_ZN2N54read + // CHECK: %[[ptr:.*]] = bitcast %{{.*}}* %{{.*}} to i32* + // CHECK: %[[val:.*]] = load i32* %[[ptr]] + // CHECK: %[[and:.*]] = and i32 %[[val]], 16777215 + // CHECK: ret i32 %[[and]] + return u->y.b; + } + void write(U* u, unsigned x) { + // CHECK: define void @_ZN2N55write + // CHECK: %[[ptr:.*]] = bitcast %{{.*}}* %{{.*}} to i32* + // CHECK: %[[old:.*]] = load i32* %[[ptr]] + // CHECK: %[[x_and:.*]] = and i32 %{{.*}}, 16777215 + // CHECK: %[[old_and:.*]] = and i32 %[[old]], -16777216 + // CHECK: %[[new:.*]] = or i32 %[[old_and]], %[[x_and]] + // CHECK: store i32 %[[new]], i32* %[[ptr]] + u->y.b = x; + } +} + +namespace N6 { + // Zero-length bitfields partition the memory locations of bitfields for the + // purposes of the memory model. That means stores must not span zero-length + // bitfields and loads may only span them when we are not instrumenting with + // thread sanitizer. + // FIXME: We currently don't widen loads even without thread sanitizer, even + // though we could. + struct S { + unsigned b1 : 24; + unsigned char : 0; + unsigned char b2 : 8; + }; + unsigned read(S* s) { + // CHECK: define i32 @_ZN2N64read + // CHECK: %[[ptr1:.*]] = bitcast {{.*}}* %{{.*}} to i24* + // CHECK: %[[val1:.*]] = load i24* %[[ptr1]] + // CHECK: %[[ext1:.*]] = zext i24 %[[val1]] to i32 + // CHECK: %[[ptr2:.*]] = getelementptr inbounds {{.*}}* %{{.*}}, i32 0, i32 1 + // CHECK: %[[val2:.*]] = load i8* %[[ptr2]] + // CHECK: %[[ext2:.*]] = zext i8 %[[val2]] to i32 + // CHECK: %[[add:.*]] = add nsw i32 %[[ext1]], %[[ext2]] + // CHECK: ret i32 %[[add]] + return s->b1 + s->b2; + } + void write(S* s, unsigned x) { + // CHECK: define void @_ZN2N65write + // CHECK: %[[ptr1:.*]] = bitcast {{.*}}* %{{.*}} to i24* + // CHECK: %[[new1:.*]] = trunc i32 %{{.*}} to i24 + // CHECK: store i24 %[[new1]], i24* %[[ptr1]] + // CHECK: %[[new2:.*]] = trunc i32 %{{.*}} to i8 + // CHECK: %[[ptr2:.*]] = getelementptr inbounds {{.*}}* %{{.*}}, i32 0, i32 1 + // CHECK: store i8 %[[new2]], i8* %[[ptr2]] + s->b1 = x; + s->b2 = x; + } +} |