2 * Copyright (C) 2011 The Android Open Source Project
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
8 * http://www.apache.org/licenses/LICENSE-2.0
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
17 #ifndef ART_RUNTIME_MIRROR_OBJECT_READBARRIER_INL_H_
18 #define ART_RUNTIME_MIRROR_OBJECT_READBARRIER_INL_H_
23 #include "lock_word-inl.h"
24 #include "object_reference-inl.h"
25 #include "read_barrier.h"
31 template<VerifyObjectFlags kVerifyFlags>
32 inline LockWord Object::GetLockWord(bool as_volatile) {
34 return LockWord(GetField32Volatile<kVerifyFlags>(OFFSET_OF_OBJECT_MEMBER(Object, monitor_)));
36 return LockWord(GetField32<kVerifyFlags>(OFFSET_OF_OBJECT_MEMBER(Object, monitor_)));
39 template<bool kTransactionActive, bool kCheckTransaction, VerifyObjectFlags kVerifyFlags>
40 inline bool Object::CasFieldWeakRelaxed32(MemberOffset field_offset,
41 int32_t old_value, int32_t new_value) {
42 if (kCheckTransaction) {
43 DCHECK_EQ(kTransactionActive, Runtime::Current()->IsActiveTransaction());
45 if (kTransactionActive) {
46 Runtime::Current()->RecordWriteField32(this, field_offset, old_value, true);
48 if (kVerifyFlags & kVerifyThis) {
51 uint8_t* raw_addr = reinterpret_cast<uint8_t*>(this) + field_offset.Int32Value();
52 AtomicInteger* atomic_addr = reinterpret_cast<AtomicInteger*>(raw_addr);
54 return atomic_addr->CompareExchangeWeakRelaxed(old_value, new_value);
57 inline bool Object::CasLockWordWeakRelaxed(LockWord old_val, LockWord new_val) {
58 // Force use of non-transactional mode and do not check.
59 return CasFieldWeakRelaxed32<false, false>(
60 OFFSET_OF_OBJECT_MEMBER(Object, monitor_), old_val.GetValue(), new_val.GetValue());
63 inline bool Object::CasLockWordWeakRelease(LockWord old_val, LockWord new_val) {
64 // Force use of non-transactional mode and do not check.
65 return CasFieldWeakRelease32<false, false>(
66 OFFSET_OF_OBJECT_MEMBER(Object, monitor_), old_val.GetValue(), new_val.GetValue());
69 inline uint32_t Object::GetReadBarrierState(uintptr_t* fake_address_dependency) {
70 if (!kUseBakerReadBarrier) {
71 LOG(FATAL) << "Unreachable";
75 uintptr_t obj = reinterpret_cast<uintptr_t>(this);
77 DCHECK_EQ(OFFSETOF_MEMBER(Object, monitor_), 4U);
78 // Use inline assembly to prevent the compiler from optimizing away the false dependency.
80 "ldr %[result], [%[obj], #4]\n\t"
81 // This instruction is enough to "fool the compiler and the CPU" by having `fad` always be
82 // null, without them being able to assume that fact.
83 "eor %[fad], %[result], %[result]\n\t"
84 : [result] "+r" (result), [fad] "=r" (*fake_address_dependency)
86 DCHECK_EQ(*fake_address_dependency, 0U);
87 LockWord lw(static_cast<uint32_t>(result));
88 uint32_t rb_state = lw.ReadBarrierState();
90 #elif defined(__aarch64__)
91 uintptr_t obj = reinterpret_cast<uintptr_t>(this);
93 DCHECK_EQ(OFFSETOF_MEMBER(Object, monitor_), 4U);
94 // Use inline assembly to prevent the compiler from optimizing away the false dependency.
96 "ldr %w[result], [%[obj], #4]\n\t"
97 // This instruction is enough to "fool the compiler and the CPU" by having `fad` always be
98 // null, without them being able to assume that fact.
99 "eor %[fad], %[result], %[result]\n\t"
100 : [result] "+r" (result), [fad] "=r" (*fake_address_dependency)
102 DCHECK_EQ(*fake_address_dependency, 0U);
103 LockWord lw(static_cast<uint32_t>(result));
104 uint32_t rb_state = lw.ReadBarrierState();
106 #elif defined(__i386__) || defined(__x86_64__)
107 LockWord lw = GetLockWord(false);
108 // i386/x86_64 don't need fake address dependency. Use a compiler fence to avoid compiler
110 *fake_address_dependency = 0;
111 std::atomic_signal_fence(std::memory_order_acquire);
112 uint32_t rb_state = lw.ReadBarrierState();
115 // MIPS32/MIPS64: use a memory barrier to prevent load-load reordering.
116 LockWord lw = GetLockWord(false);
117 *fake_address_dependency = 0;
118 std::atomic_thread_fence(std::memory_order_acquire);
119 uint32_t rb_state = lw.ReadBarrierState();
124 inline uint32_t Object::GetReadBarrierState() {
125 if (!kUseBakerReadBarrier) {
126 LOG(FATAL) << "Unreachable";
129 DCHECK(kUseBakerReadBarrier);
130 LockWord lw(GetField<uint32_t, /*kIsVolatile*/false>(OFFSET_OF_OBJECT_MEMBER(Object, monitor_)));
131 uint32_t rb_state = lw.ReadBarrierState();
132 DCHECK(ReadBarrier::IsValidReadBarrierState(rb_state)) << rb_state;
136 inline uint32_t Object::GetReadBarrierStateAcquire() {
137 if (!kUseBakerReadBarrier) {
138 LOG(FATAL) << "Unreachable";
141 LockWord lw(GetFieldAcquire<uint32_t>(OFFSET_OF_OBJECT_MEMBER(Object, monitor_)));
142 uint32_t rb_state = lw.ReadBarrierState();
143 DCHECK(ReadBarrier::IsValidReadBarrierState(rb_state)) << rb_state;
147 template<bool kCasRelease>
148 inline bool Object::AtomicSetReadBarrierState(uint32_t expected_rb_state, uint32_t rb_state) {
149 if (!kUseBakerReadBarrier) {
150 LOG(FATAL) << "Unreachable";
153 DCHECK(ReadBarrier::IsValidReadBarrierState(expected_rb_state)) << expected_rb_state;
154 DCHECK(ReadBarrier::IsValidReadBarrierState(rb_state)) << rb_state;
155 LockWord expected_lw;
158 LockWord lw = GetLockWord(false);
159 if (UNLIKELY(lw.ReadBarrierState() != expected_rb_state)) {
164 expected_lw.SetReadBarrierState(expected_rb_state);
166 new_lw.SetReadBarrierState(rb_state);
167 // ConcurrentCopying::ProcessMarkStackRef uses this with kCasRelease == true.
168 // If kCasRelease == true, use a CAS release so that when GC updates all the fields of
169 // an object and then changes the object from gray to black, the field updates (stores) will be
170 // visible (won't be reordered after this CAS.)
171 } while (!(kCasRelease ?
172 CasLockWordWeakRelease(expected_lw, new_lw) :
173 CasLockWordWeakRelaxed(expected_lw, new_lw)));
177 inline bool Object::AtomicSetMarkBit(uint32_t expected_mark_bit, uint32_t mark_bit) {
178 LockWord expected_lw;
181 LockWord lw = GetLockWord(false);
182 if (UNLIKELY(lw.MarkBitState() != expected_mark_bit)) {
188 new_lw.SetMarkBitState(mark_bit);
189 // Since this is only set from the mutator, we can use the non release Cas.
190 } while (!CasLockWordWeakRelaxed(expected_lw, new_lw));
194 template<bool kTransactionActive, bool kCheckTransaction, VerifyObjectFlags kVerifyFlags>
195 inline bool Object::CasFieldStrongRelaxedObjectWithoutWriteBarrier(
196 MemberOffset field_offset,
197 ObjPtr<Object> old_value,
198 ObjPtr<Object> new_value) {
199 if (kCheckTransaction) {
200 DCHECK_EQ(kTransactionActive, Runtime::Current()->IsActiveTransaction());
202 if (kVerifyFlags & kVerifyThis) {
205 if (kVerifyFlags & kVerifyWrites) {
206 VerifyObject(new_value);
208 if (kVerifyFlags & kVerifyReads) {
209 VerifyObject(old_value);
211 if (kTransactionActive) {
212 Runtime::Current()->RecordWriteFieldReference(this, field_offset, old_value, true);
214 HeapReference<Object> old_ref(HeapReference<Object>::FromObjPtr(old_value));
215 HeapReference<Object> new_ref(HeapReference<Object>::FromObjPtr(new_value));
216 uint8_t* raw_addr = reinterpret_cast<uint8_t*>(this) + field_offset.Int32Value();
217 Atomic<uint32_t>* atomic_addr = reinterpret_cast<Atomic<uint32_t>*>(raw_addr);
219 bool success = atomic_addr->CompareExchangeStrongRelaxed(old_ref.reference_,
224 } // namespace mirror
227 #endif // ART_RUNTIME_MIRROR_OBJECT_READBARRIER_INL_H_