| OLD | NEW |
| 1 /* | 1 /* |
| 2 * Copyright 2012 Google Inc. | 2 * Copyright 2012 Google Inc. |
| 3 * | 3 * |
| 4 * Use of this source code is governed by a BSD-style license that can be | 4 * Use of this source code is governed by a BSD-style license that can be |
| 5 * found in the LICENSE file. | 5 * found in the LICENSE file. |
| 6 */ | 6 */ |
| 7 | 7 |
| 8 #ifndef SkWeakRefCnt_DEFINED | 8 #ifndef SkWeakRefCnt_DEFINED |
| 9 #define SkWeakRefCnt_DEFINED | 9 #define SkWeakRefCnt_DEFINED |
| 10 | 10 |
| 11 #include "SkRefCnt.h" | 11 #include "SkRefCnt.h" |
| 12 #include "../private/SkAtomics.h" | 12 #include <atomic> |
| 13 | 13 |
| 14 /** \class SkWeakRefCnt | 14 /** \class SkWeakRefCnt |
| 15 | 15 |
| 16 SkWeakRefCnt is the base class for objects that may be shared by multiple | 16 SkWeakRefCnt is the base class for objects that may be shared by multiple |
| 17 objects. When an existing strong owner wants to share a reference, it calls | 17 objects. When an existing strong owner wants to share a reference, it calls |
| 18 ref(). When a strong owner wants to release its reference, it calls | 18 ref(). When a strong owner wants to release its reference, it calls |
| 19 unref(). When the shared object's strong reference count goes to zero as | 19 unref(). When the shared object's strong reference count goes to zero as |
| 20 the result of an unref() call, its (virtual) weak_dispose method is called. | 20 the result of an unref() call, its (virtual) weak_dispose method is called. |
| 21 It is an error for the destructor to be called explicitly (or via the | 21 It is an error for the destructor to be called explicitly (or via the |
| 22 object going out of scope on the stack or calling delete) if | 22 object going out of scope on the stack or calling delete) if |
| (...skipping 32 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 55 The strong references collectively hold one weak reference. When the | 55 The strong references collectively hold one weak reference. When the |
| 56 strong reference count goes to zero, the collectively held weak | 56 strong reference count goes to zero, the collectively held weak |
| 57 reference is released. | 57 reference is released. |
| 58 */ | 58 */ |
| 59 SkWeakRefCnt() : SkRefCnt(), fWeakCnt(1) {} | 59 SkWeakRefCnt() : SkRefCnt(), fWeakCnt(1) {} |
| 60 | 60 |
| 61 /** Destruct, asserting that the weak reference count is 1. | 61 /** Destruct, asserting that the weak reference count is 1. |
| 62 */ | 62 */ |
| 63 virtual ~SkWeakRefCnt() { | 63 virtual ~SkWeakRefCnt() { |
| 64 #ifdef SK_DEBUG | 64 #ifdef SK_DEBUG |
| 65 SkASSERT(fWeakCnt == 1); | 65 SkASSERT(getWeakCnt() == 1); |
| 66 fWeakCnt = 0; | 66 fWeakCnt.store(0, std::memory_order_relaxed); |
| 67 #endif | 67 #endif |
| 68 } | 68 } |
| 69 | 69 |
| 70 /** Return the weak reference count. | 70 #ifdef SK_DEBUG |
| 71 */ | 71 /** Return the weak reference count. */ |
| 72 int32_t getWeakCnt() const { return fWeakCnt; } | 72 int32_t getWeakCnt() const { |
| 73 return fWeakCnt.load(std::memory_order_relaxed); |
| 74 } |
| 73 | 75 |
| 74 #ifdef SK_DEBUG | |
| 75 void validate() const { | 76 void validate() const { |
| 76 this->INHERITED::validate(); | 77 this->INHERITED::validate(); |
| 77 SkASSERT(fWeakCnt > 0); | 78 SkASSERT(getWeakCnt() > 0); |
| 78 } | 79 } |
| 79 #endif | 80 #endif |
| 80 | 81 |
| 82 private: |
| 83 /** If fRefCnt is 0, returns 0. |
| 84 * Otherwise increments fRefCnt, acquires, and returns the old value. |
| 85 */ |
| 86 int32_t atomic_conditional_acquire_strong_ref() const { |
| 87 int32_t prev = fRefCnt.load(std::memory_order_relaxed); |
| 88 do { |
| 89 if (0 == prev) { |
| 90 break; |
| 91 } |
| 92 } while(!fRefCnt.compare_exchange_weak(prev, prev+1, std::memory_order_a
cquire, |
| 93 std::memory_order_r
elaxed)); |
| 94 return prev; |
| 95 } |
| 96 |
| 97 public: |
| 81 /** Creates a strong reference from a weak reference, if possible. The | 98 /** Creates a strong reference from a weak reference, if possible. The |
| 82 caller must already be an owner. If try_ref() returns true the owner | 99 caller must already be an owner. If try_ref() returns true the owner |
| 83 is in posession of an additional strong reference. Both the original | 100 is in posession of an additional strong reference. Both the original |
| 84 reference and new reference must be properly unreferenced. If try_ref() | 101 reference and new reference must be properly unreferenced. If try_ref() |
| 85 returns false, no strong reference could be created and the owner's | 102 returns false, no strong reference could be created and the owner's |
| 86 reference is in the same state as before the call. | 103 reference is in the same state as before the call. |
| 87 */ | 104 */ |
| 88 bool SK_WARN_UNUSED_RESULT try_ref() const { | 105 bool SK_WARN_UNUSED_RESULT try_ref() const { |
| 89 if (sk_atomic_conditional_inc(&fRefCnt) != 0) { | 106 if (atomic_conditional_acquire_strong_ref() != 0) { |
| 90 // Acquire barrier (L/SL), if not provided above. | 107 // Acquire barrier (L/SL), if not provided above. |
| 91 // Prevents subsequent code from happening before the increment. | 108 // Prevents subsequent code from happening before the increment. |
| 92 sk_membar_acquire__after_atomic_conditional_inc(); | |
| 93 return true; | 109 return true; |
| 94 } | 110 } |
| 95 return false; | 111 return false; |
| 96 } | 112 } |
| 97 | 113 |
| 98 /** Increment the weak reference count. Must be balanced by a call to | 114 /** Increment the weak reference count. Must be balanced by a call to |
| 99 weak_unref(). | 115 weak_unref(). |
| 100 */ | 116 */ |
| 101 void weak_ref() const { | 117 void weak_ref() const { |
| 102 SkASSERT(fRefCnt > 0); | 118 SkASSERT(getRefCnt() > 0); |
| 103 SkASSERT(fWeakCnt > 0); | 119 SkASSERT(getWeakCnt() > 0); |
| 104 sk_atomic_inc(&fWeakCnt); // No barrier required. | 120 // No barrier required. |
| 121 (void)fWeakCnt.fetch_add(+1, std::memory_order_relaxed); |
| 105 } | 122 } |
| 106 | 123 |
| 107 /** Decrement the weak reference count. If the weak reference count is 1 | 124 /** Decrement the weak reference count. If the weak reference count is 1 |
| 108 before the decrement, then call delete on the object. Note that if this | 125 before the decrement, then call delete on the object. Note that if this |
| 109 is the case, then the object needs to have been allocated via new, and | 126 is the case, then the object needs to have been allocated via new, and |
| 110 not on the stack. | 127 not on the stack. |
| 111 */ | 128 */ |
| 112 void weak_unref() const { | 129 void weak_unref() const { |
| 113 SkASSERT(fWeakCnt > 0); | 130 SkASSERT(getWeakCnt() > 0); |
| 114 // Release barrier (SL/S), if not provided below. | 131 // A release here acts in place of all releases we "should" have been do
ing in ref(). |
| 115 if (sk_atomic_dec(&fWeakCnt) == 1) { | 132 if (1 == fWeakCnt.fetch_add(-1, std::memory_order_acq_rel)) { |
| 116 // Acquire barrier (L/SL), if not provided above. | 133 // Like try_ref(), the acquire is only needed on success, to make su
re |
| 117 // Prevents code in destructor from happening before the decrement. | 134 // code in internal_dispose() doesn't happen before the decrement. |
| 118 sk_membar_acquire__after_atomic_dec(); | |
| 119 #ifdef SK_DEBUG | 135 #ifdef SK_DEBUG |
| 120 // so our destructor won't complain | 136 // so our destructor won't complain |
| 121 fWeakCnt = 1; | 137 fWeakCnt.store(1, std::memory_order_relaxed); |
| 122 #endif | 138 #endif |
| 123 this->INHERITED::internal_dispose(); | 139 this->INHERITED::internal_dispose(); |
| 124 } | 140 } |
| 125 } | 141 } |
| 126 | 142 |
| 127 /** Returns true if there are no strong references to the object. When this | 143 /** Returns true if there are no strong references to the object. When this |
| 128 is the case all future calls to try_ref() will return false. | 144 is the case all future calls to try_ref() will return false. |
| 129 */ | 145 */ |
| 130 bool weak_expired() const { | 146 bool weak_expired() const { |
| 131 return fRefCnt == 0; | 147 return fRefCnt.load(std::memory_order_relaxed) == 0; |
| 132 } | 148 } |
| 133 | 149 |
| 134 protected: | 150 protected: |
| 135 /** Called when the strong reference count goes to zero. This allows the | 151 /** Called when the strong reference count goes to zero. This allows the |
| 136 object to free any resources it may be holding. Weak references may | 152 object to free any resources it may be holding. Weak references may |
| 137 still exist and their level of allowed access to the object is defined | 153 still exist and their level of allowed access to the object is defined |
| 138 by the object's class. | 154 by the object's class. |
| 139 */ | 155 */ |
| 140 virtual void weak_dispose() const { | 156 virtual void weak_dispose() const { |
| 141 } | 157 } |
| 142 | 158 |
| 143 private: | 159 private: |
| 144 /** Called when the strong reference count goes to zero. Calls weak_dispose | 160 /** Called when the strong reference count goes to zero. Calls weak_dispose |
| 145 on the object and releases the implicit weak reference held | 161 on the object and releases the implicit weak reference held |
| 146 collectively by the strong references. | 162 collectively by the strong references. |
| 147 */ | 163 */ |
| 148 void internal_dispose() const override { | 164 void internal_dispose() const override { |
| 149 weak_dispose(); | 165 weak_dispose(); |
| 150 weak_unref(); | 166 weak_unref(); |
| 151 } | 167 } |
| 152 | 168 |
| 153 /* Invariant: fWeakCnt = #weak + (fRefCnt > 0 ? 1 : 0) */ | 169 /* Invariant: fWeakCnt = #weak + (fRefCnt > 0 ? 1 : 0) */ |
| 154 mutable int32_t fWeakCnt; | 170 mutable std::atomic<int32_t> fWeakCnt; |
| 155 | 171 |
| 156 typedef SkRefCnt INHERITED; | 172 typedef SkRefCnt INHERITED; |
| 157 }; | 173 }; |
| 158 | 174 |
| 159 #endif | 175 #endif |
| OLD | NEW |