OLD | NEW |
1 // Copyright 2010 the V8 project authors. All rights reserved. | 1 // Copyright 2010 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 12 matching lines...) Expand all Loading... |
23 // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY | 23 // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
24 // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | 24 // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
25 // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | 25 // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
26 // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | 26 // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
27 | 27 |
28 #ifndef V8_UNBOUND_QUEUE_INL_H_ | 28 #ifndef V8_UNBOUND_QUEUE_INL_H_ |
29 #define V8_UNBOUND_QUEUE_INL_H_ | 29 #define V8_UNBOUND_QUEUE_INL_H_ |
30 | 30 |
31 #include "unbound-queue.h" | 31 #include "unbound-queue.h" |
32 | 32 |
| 33 #include "atomicops.h" |
| 34 |
33 namespace v8 { | 35 namespace v8 { |
34 namespace internal { | 36 namespace internal { |
35 | 37 |
36 template<typename Record> | 38 template<typename Record> |
37 struct UnboundQueue<Record>::Node: public Malloced { | 39 struct UnboundQueue<Record>::Node: public Malloced { |
38 explicit Node(const Record& value) | 40 explicit Node(const Record& value) |
39 : value(value), next(NULL) { | 41 : value(value), next(NULL) { |
40 } | 42 } |
41 | 43 |
42 Record value; | 44 Record value; |
(...skipping 20 matching lines...) Expand all Loading... |
63 first_ = tmp->next; | 65 first_ = tmp->next; |
64 delete tmp; | 66 delete tmp; |
65 } | 67 } |
66 | 68 |
67 | 69 |
68 template<typename Record> | 70 template<typename Record> |
69 void UnboundQueue<Record>::Dequeue(Record* rec) { | 71 void UnboundQueue<Record>::Dequeue(Record* rec) { |
70 ASSERT(divider_ != last_); | 72 ASSERT(divider_ != last_); |
71 Node* next = reinterpret_cast<Node*>(divider_)->next; | 73 Node* next = reinterpret_cast<Node*>(divider_)->next; |
72 *rec = next->value; | 74 *rec = next->value; |
73 OS::ReleaseStore(÷r_, reinterpret_cast<AtomicWord>(next)); | 75 Release_Store(÷r_, reinterpret_cast<AtomicWord>(next)); |
74 } | 76 } |
75 | 77 |
76 | 78 |
77 template<typename Record> | 79 template<typename Record> |
78 void UnboundQueue<Record>::Enqueue(const Record& rec) { | 80 void UnboundQueue<Record>::Enqueue(const Record& rec) { |
79 Node*& next = reinterpret_cast<Node*>(last_)->next; | 81 Node*& next = reinterpret_cast<Node*>(last_)->next; |
80 next = new Node(rec); | 82 next = new Node(rec); |
81 OS::ReleaseStore(&last_, reinterpret_cast<AtomicWord>(next)); | 83 Release_Store(&last_, reinterpret_cast<AtomicWord>(next)); |
82 while (first_ != reinterpret_cast<Node*>(divider_)) DeleteFirst(); | 84 while (first_ != reinterpret_cast<Node*>(divider_)) DeleteFirst(); |
83 } | 85 } |
84 | 86 |
85 | 87 |
86 template<typename Record> | 88 template<typename Record> |
87 Record* UnboundQueue<Record>::Peek() { | 89 Record* UnboundQueue<Record>::Peek() { |
88 ASSERT(divider_ != last_); | 90 ASSERT(divider_ != last_); |
89 Node* next = reinterpret_cast<Node*>(divider_)->next; | 91 Node* next = reinterpret_cast<Node*>(divider_)->next; |
90 return &next->value; | 92 return &next->value; |
91 } | 93 } |
92 | 94 |
93 } } // namespace v8::internal | 95 } } // namespace v8::internal |
94 | 96 |
95 #endif // V8_UNBOUND_QUEUE_INL_H_ | 97 #endif // V8_UNBOUND_QUEUE_INL_H_ |
OLD | NEW |