97
98 inline bool PSParallelCompact::mark_obj(oop obj) {
99 const int obj_size = obj->size();
100 if (mark_bitmap()->mark_obj(obj, obj_size)) {
101 _summary_data.add_obj(obj, obj_size);
102 return true;
103 } else {
104 return false;
105 }
106 }
107
108 template <class T>
109 inline void PSParallelCompact::adjust_pointer(T* p, ParCompactionManager* cm) {
110 T heap_oop = RawAccess<>::oop_load(p);
111 if (!CompressedOops::is_null(heap_oop)) {
112 oop obj = CompressedOops::decode_not_null(heap_oop);
113 assert(ParallelScavengeHeap::heap()->is_in(obj), "should be in heap");
114
115 oop new_obj = (oop)summary_data().calc_new_pointer(obj, cm);
116 assert(new_obj != NULL, // is forwarding ptr?
117 "should be forwarded");
118 // Just always do the update unconditionally?
119 if (new_obj != NULL) {
120 assert(ParallelScavengeHeap::heap()->is_in_reserved(new_obj),
121 "should be in object space");
122 RawAccess<IS_NOT_NULL>::oop_store(p, new_obj);
123 }
124 }
125 }
126
127 class PCAdjustPointerClosure: public BasicOopIterateClosure {
128 public:
129 PCAdjustPointerClosure(ParCompactionManager* cm) {
130 assert(cm != NULL, "associate ParCompactionManage should not be NULL");
131 _cm = cm;
132 }
133 template <typename T> void do_oop_nv(T* p) { PSParallelCompact::adjust_pointer(p, _cm); }
134 virtual void do_oop(oop* p) { do_oop_nv(p); }
135 virtual void do_oop(narrowOop* p) { do_oop_nv(p); }
136
137 // This closure provides its own oop verification code.
138 debug_only(virtual bool should_verify_oops() { return false; })
139 virtual ReferenceIterationMode reference_iteration_mode() { return DO_FIELDS; }
140 private:
141 ParCompactionManager* _cm;
|
97
98 inline bool PSParallelCompact::mark_obj(oop obj) {
99 const int obj_size = obj->size();
100 if (mark_bitmap()->mark_obj(obj, obj_size)) {
101 _summary_data.add_obj(obj, obj_size);
102 return true;
103 } else {
104 return false;
105 }
106 }
107
108 template <class T>
109 inline void PSParallelCompact::adjust_pointer(T* p, ParCompactionManager* cm) {
110 T heap_oop = RawAccess<>::oop_load(p);
111 if (!CompressedOops::is_null(heap_oop)) {
112 oop obj = CompressedOops::decode_not_null(heap_oop);
113 assert(ParallelScavengeHeap::heap()->is_in(obj), "should be in heap");
114
115 oop new_obj = (oop)summary_data().calc_new_pointer(obj, cm);
116 assert(new_obj != NULL, // is forwarding ptr?
117 "should be forwarded");
118 // Just always do the update unconditionally?
119 if (new_obj != NULL) {
120 assert(ParallelScavengeHeap::heap()->is_in_reserved(new_obj),
121 "should be in object space");
122 RawAccess<IS_NOT_NULL>::oop_store(p, new_obj);
123 }
124 }
125 }
126
127 class PCAdjustPointerClosure: public BasicOopIterateClosure {
128 public:
129 PCAdjustPointerClosure(ParCompactionManager* cm) {
130 assert(cm != NULL, "associate ParCompactionManage should not be NULL");
131 _cm = cm;
132 }
133 template <typename T> void do_oop_nv(T* p) { PSParallelCompact::adjust_pointer(p, _cm); }
134 virtual void do_oop(oop* p) { do_oop_nv(p); }
135 virtual void do_oop(narrowOop* p) { do_oop_nv(p); }
136
137 // This closure provides its own oop verification code.
138 debug_only(virtual bool should_verify_oops() { return false; })
139 virtual ReferenceIterationMode reference_iteration_mode() { return DO_FIELDS; }
140 private:
141 ParCompactionManager* _cm;
|