aboutsummaryrefslogtreecommitdiff
path: root/test/CodeGenCXX/bitfield.cpp
blob: a737304c98d4b340b7775953e3e70c990e4a0b90 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
// RUN: %clang_cc1 -triple x86_64-unknown-unknown -verify -emit-llvm -o - %s | FileCheck %s
//
// Tests for bitfield access patterns in C++ with special attention to
// conformance to C++11 memory model requirements.

namespace N1 {
  // Ensure that neither loads nor stores to bitfields are not widened into
  // other memory locations. (PR13691)
  //
  // NOTE: We could potentially widen loads based on their alignment if we are
  // comfortable requiring that subsequent memory locations within the
  // alignment-widened load are not volatile.
  struct S {
    char a;
    unsigned b : 1;
    char c;
  };
  unsigned read(S* s) {
    // CHECK: define i32 @_ZN2N14read
    // CHECK:   %[[ptr:.*]] = getelementptr inbounds %{{.*}}* %{{.*}}, i32 0, i32 1
    // CHECK:   %[[val:.*]] = load i8* %[[ptr]]
    // CHECK:   %[[and:.*]] = and i8 %[[val]], 1
    // CHECK:   %[[ext:.*]] = zext i8 %[[and]] to i32
    // CHECK:                 ret i32 %[[ext]]
    return s->b;
  }
  void write(S* s, unsigned x) {
    // CHECK: define void @_ZN2N15write
    // CHECK:   %[[ptr:.*]]     = getelementptr inbounds %{{.*}}* %{{.*}}, i32 0, i32 1
    // CHECK:   %[[x_trunc:.*]] = trunc i32 %{{.*}} to i8
    // CHECK:   %[[old:.*]]     = load i8* %[[ptr]]
    // CHECK:   %[[x_and:.*]]   = and i8 %[[x_trunc]], 1
    // CHECK:   %[[old_and:.*]] = and i8 %[[old]], -2
    // CHECK:   %[[new:.*]]     = or i8 %[[old_and]], %[[x_and]]
    // CHECK:                     store i8 %[[new]], i8* %[[ptr]]
    s->b = x;
  }
}

namespace N2 {
  // Do widen loads and stores to bitfields when those bitfields have padding
  // within the struct following them.
  struct S {
    unsigned b : 24;
    void *p;
  };
  unsigned read(S* s) {
    // CHECK: define i32 @_ZN2N24read
    // CHECK:   %[[ptr:.*]] = bitcast %{{.*}}* %{{.*}} to i32*
    // CHECK:   %[[val:.*]] = load i32* %[[ptr]]
    // CHECK:   %[[and:.*]] = and i32 %[[val]], 16777215
    // CHECK:                 ret i32 %[[and]]
    return s->b;
  }
  void write(S* s, unsigned x) {
    // CHECK: define void @_ZN2N25write
    // CHECK:   %[[ptr:.*]]     = bitcast %{{.*}}* %{{.*}} to i32*
    // CHECK:   %[[old:.*]]     = load i32* %[[ptr]]
    // CHECK:   %[[x_and:.*]]   = and i32 %{{.*}}, 16777215
    // CHECK:   %[[old_and:.*]] = and i32 %[[old]], -16777216
    // CHECK:   %[[new:.*]]     = or i32 %[[old_and]], %[[x_and]]
    // CHECK:                     store i32 %[[new]], i32* %[[ptr]]
    s->b = x;
  }
}

namespace N3 {
  // Do widen loads and stores to bitfields through the trailing padding at the
  // end of a struct.
  struct S {
    unsigned b : 24;
  };
  unsigned read(S* s) {
    // CHECK: define i32 @_ZN2N34read
    // CHECK:   %[[ptr:.*]] = bitcast %{{.*}}* %{{.*}} to i32*
    // CHECK:   %[[val:.*]] = load i32* %[[ptr]]
    // CHECK:   %[[and:.*]] = and i32 %[[val]], 16777215
    // CHECK:                 ret i32 %[[and]]
    return s->b;
  }
  void write(S* s, unsigned x) {
    // CHECK: define void @_ZN2N35write
    // CHECK:   %[[ptr:.*]]     = bitcast %{{.*}}* %{{.*}} to i32*
    // CHECK:   %[[old:.*]]     = load i32* %[[ptr]]
    // CHECK:   %[[x_and:.*]]   = and i32 %{{.*}}, 16777215
    // CHECK:   %[[old_and:.*]] = and i32 %[[old]], -16777216
    // CHECK:   %[[new:.*]]     = or i32 %[[old_and]], %[[x_and]]
    // CHECK:                     store i32 %[[new]], i32* %[[ptr]]
    s->b = x;
  }
}

namespace N4 {
  // Do NOT widen loads and stores to bitfields into padding at the end of
  // a class which might end up with members inside of it when inside a derived
  // class.
  struct Base {
    virtual ~Base() {}

    unsigned b : 24;
  };
  // Imagine some other translation unit introduces:
#if 0
  struct Derived : public Base {
    char c;
  };
#endif
  unsigned read(Base* s) {
    // FIXME: We should widen this load as long as the function isn't being
    // instrumented by thread-sanitizer.
    //
    // CHECK: define i32 @_ZN2N44read
    // CHECK:   %[[ptr:.*]] = bitcast {{.*}}* %{{.*}} to i24*
    // CHECK:   %[[val:.*]] = load i24* %[[ptr]]
    // CHECK:   %[[ext:.*]] = zext i24 %[[val]] to i32
    // CHECK:                 ret i32 %[[ext]]
    return s->b;
  }
  void write(Base* s, unsigned x) {
    // CHECK: define void @_ZN2N45write
    // CHECK:   %[[ptr:.*]] = bitcast {{.*}}* %{{.*}} to i24*
    // CHECK:   %[[new:.*]] = trunc i32 %{{.*}} to i24
    // CHECK:                 store i24 %[[new]], i24* %[[ptr]]
    s->b = x;
  }
}

namespace N5 {
  // Widen through padding at the end of a struct even if that struct
  // participates in a union with another struct which has a separate field in
  // that location. The reasoning is that if the operation is storing to that
  // member of the union, it must be the active member, and thus we can write
  // through the padding. If it is a load, it might be a load of a common
  // prefix through a non-active member, but in such a case the extra bits
  // loaded are masked off anyways.
  union U {
    struct X { unsigned b : 24; char c; } x;
    struct Y { unsigned b : 24; } y;
  };
  unsigned read(U* u) {
    // CHECK: define i32 @_ZN2N54read
    // CHECK:   %[[ptr:.*]] = bitcast %{{.*}}* %{{.*}} to i32*
    // CHECK:   %[[val:.*]] = load i32* %[[ptr]]
    // CHECK:   %[[and:.*]] = and i32 %[[val]], 16777215
    // CHECK:                 ret i32 %[[and]]
    return u->y.b;
  }
  void write(U* u, unsigned x) {
    // CHECK: define void @_ZN2N55write
    // CHECK:   %[[ptr:.*]]     = bitcast %{{.*}}* %{{.*}} to i32*
    // CHECK:   %[[old:.*]]     = load i32* %[[ptr]]
    // CHECK:   %[[x_and:.*]]   = and i32 %{{.*}}, 16777215
    // CHECK:   %[[old_and:.*]] = and i32 %[[old]], -16777216
    // CHECK:   %[[new:.*]]     = or i32 %[[old_and]], %[[x_and]]
    // CHECK:                     store i32 %[[new]], i32* %[[ptr]]
    u->y.b = x;
  }
}

namespace N6 {
  // Zero-length bitfields partition the memory locations of bitfields for the
  // purposes of the memory model. That means stores must not span zero-length
  // bitfields and loads may only span them when we are not instrumenting with
  // thread sanitizer.
  // FIXME: We currently don't widen loads even without thread sanitizer, even
  // though we could.
  struct S {
    unsigned b1 : 24;
    unsigned char : 0;
    unsigned char b2 : 8;
  };
  unsigned read(S* s) {
    // CHECK: define i32 @_ZN2N64read
    // CHECK:   %[[ptr1:.*]] = bitcast {{.*}}* %{{.*}} to i24*
    // CHECK:   %[[val1:.*]] = load i24* %[[ptr1]]
    // CHECK:   %[[ext1:.*]] = zext i24 %[[val1]] to i32
    // CHECK:   %[[ptr2:.*]] = getelementptr inbounds {{.*}}* %{{.*}}, i32 0, i32 1
    // CHECK:   %[[val2:.*]] = load i8* %[[ptr2]]
    // CHECK:   %[[ext2:.*]] = zext i8 %[[val2]] to i32
    // CHECK:   %[[add:.*]]  = add nsw i32 %[[ext1]], %[[ext2]]
    // CHECK:                  ret i32 %[[add]]
    return s->b1 + s->b2;
  }
  void write(S* s, unsigned x) {
    // CHECK: define void @_ZN2N65write
    // CHECK:   %[[ptr1:.*]] = bitcast {{.*}}* %{{.*}} to i24*
    // CHECK:   %[[new1:.*]] = trunc i32 %{{.*}} to i24
    // CHECK:                  store i24 %[[new1]], i24* %[[ptr1]]
    // CHECK:   %[[new2:.*]] = trunc i32 %{{.*}} to i8
    // CHECK:   %[[ptr2:.*]] = getelementptr inbounds {{.*}}* %{{.*}}, i32 0, i32 1
    // CHECK:                  store i8 %[[new2]], i8* %[[ptr2]]
    s->b1 = x;
    s->b2 = x;
  }
}