/home/uke/oil/mycpp/mark_sweep_heap.h
Line | Count | Source (jump to first uncovered line) |
1 | | #ifndef MARKSWEEP_HEAP_H |
2 | | #define MARKSWEEP_HEAP_H |
3 | | |
4 | | #include <stdlib.h> |
5 | | |
6 | | #include <vector> |
7 | | |
8 | | #include "mycpp/common.h" |
9 | | #include "mycpp/gc_obj.h" |
10 | | |
11 | | #if GC_ALWAYS |
12 | | #define VALIDATE_ROOTS 1 |
13 | | #else |
14 | | #define VALIDATE_ROOTS 0 // flip this manually to diagnose bugs |
15 | | #endif |
16 | | |
17 | | #if VALIDATE_ROOTS |
18 | | static void ValidateRoot(const RawObject* obj) { |
19 | | if (obj == nullptr) { |
20 | | return; |
21 | | } |
22 | | |
23 | | // Assuming 64-bit == 8 byte alignment |
24 | | if (reinterpret_cast<uintptr_t>(obj) & 0x3) { |
25 | | log("Misaligned object %p", obj); |
26 | | FAIL(kShouldNotGetHere); |
27 | | return; |
28 | | } |
29 | | |
30 | | ObjHeader* header = ObjHeader::FromObject(obj); |
31 | | // log("obj %p header %p", obj, header); |
32 | | |
33 | | if (reinterpret_cast<uintptr_t>(header) & 0x3) { |
34 | | log("Misaligned header %p", header); |
35 | | FAIL(kShouldNotGetHere); |
36 | | return; |
37 | | } |
38 | | |
39 | | switch (header->heap_tag) { |
40 | | case HeapTag::Global: |
41 | | case HeapTag::Opaque: |
42 | | case HeapTag::Scanned: |
43 | | case HeapTag::FixedSize: |
44 | | break; |
45 | | |
46 | | default: |
47 | | log("root %p heap %d type %d mask %d len %d", obj, header->heap_tag, |
48 | | header->type_tag, header->u_mask_npointers); |
49 | | FAIL(kShouldNotGetHere); |
50 | | break; |
51 | | } |
52 | | } |
53 | | #endif |
54 | | |
55 | | class MarkSet { |
56 | | public: |
57 | 96 | MarkSet() : bits_() { |
58 | 96 | } |
59 | | |
60 | | // ReInit() must be called at the start of MarkObjects(). Allocate() should |
61 | | // keep track of the maximum object ID. |
62 | 123 | void ReInit(int max_obj_id) { |
63 | | // https://stackoverflow.com/questions/8848575/fastest-way-to-reset-every-value-of-stdvectorint-to-0 |
64 | 123 | std::fill(bits_.begin(), bits_.end(), 0); |
65 | 123 | int max_byte_index = (max_obj_id >> 3) + 1; // round up |
66 | | // log("ReInit max_byte_index %d", max_byte_index); |
67 | 123 | bits_.resize(max_byte_index); |
68 | 123 | } |
69 | | |
70 | | // Called by MarkObjects() |
71 | 1.61k | void Mark(int obj_id) { |
72 | 1.61k | DCHECK(obj_id >= 0); |
73 | | // log("obj id %d", obj_id); |
74 | 1.61k | DCHECK(!IsMarked(obj_id)); |
75 | 0 | int byte_index = obj_id >> 3; // 8 bits per byte |
76 | 1.61k | int bit_index = obj_id & 0b111; |
77 | | // log("byte_index %d %d", byte_index, bit_index); |
78 | 1.61k | bits_[byte_index] |= (1 << bit_index); |
79 | 1.61k | } |
80 | | |
81 | | // Called by Sweep() |
82 | 47.6k | bool IsMarked(int obj_id) { |
83 | 47.6k | DCHECK(obj_id >= 0); |
84 | 0 | int byte_index = obj_id >> 3; |
85 | 47.6k | int bit_index = obj_id & 0b111; |
86 | 47.6k | return bits_[byte_index] & (1 << bit_index); |
87 | 47.6k | } |
88 | | |
89 | 0 | void Debug() { |
90 | 0 | int n = bits_.size(); |
91 | 0 | dprintf(2, "[ "); |
92 | 0 | for (int i = 0; i < n; ++i) { |
93 | 0 | dprintf(2, "%02x ", bits_[i]); |
94 | 0 | } |
95 | 0 | dprintf(2, "] (%d bytes) \n", n); |
96 | 0 | dprintf(2, "[ "); |
97 | 0 | int num_bits = 0; |
98 | 0 | for (int i = 0; i < n; ++i) { |
99 | 0 | for (int j = 0; j < 8; ++j) { |
100 | 0 | int bit = (bits_[i] & (1 << j)) != 0; |
101 | 0 | dprintf(2, "%d", bit); |
102 | 0 | num_bits += bit; |
103 | 0 | } |
104 | 0 | } |
105 | 0 | dprintf(2, " ] (%d bits set)\n", num_bits); |
106 | 0 | } |
107 | | |
108 | | std::vector<uint8_t> bits_; // bit vector indexed by obj_id |
109 | | }; |
110 | | |
111 | | // A simple Pool allocator for allocating small objects. It maintains an ever |
112 | | // growing number of Blocks each consisting of a number of fixed size Cells. |
113 | | // Memory is handed out one Cell at a time. |
114 | | // Note: within the context of the Pool allocator we refer to object IDs as cell |
115 | | // IDs because in addition to identifying an object they're also used to index |
116 | | // into the Cell storage. |
117 | | template <int CellsPerBlock, size_t CellSize> |
118 | | class Pool { |
119 | | public: |
120 | | static constexpr size_t kMaxObjSize = CellSize; |
121 | | static constexpr int kBlockSize = CellSize * CellsPerBlock; |
122 | | |
123 | 64 | Pool() = default; _ZN4PoolILi682ELm24EEC2Ev Line | Count | Source | 123 | 32 | Pool() = default; |
_ZN4PoolILi341ELm48EEC2Ev Line | Count | Source | 123 | 32 | Pool() = default; |
|
124 | | |
125 | 10.0k | void* Allocate(int* obj_id) { |
126 | 10.0k | num_allocated_++; |
127 | | |
128 | 10.0k | if (!free_list_) { |
129 | | // Allocate a new Block and add every new Cell to the free list. |
130 | 57 | Block* block = static_cast<Block*>(malloc(sizeof(Block))); |
131 | 57 | blocks_.push_back(block); |
132 | 57 | bytes_allocated_ += kBlockSize; |
133 | 57 | num_free_ += CellsPerBlock; |
134 | | |
135 | | // The starting cell_id for Cells in this block. |
136 | 57 | int cell_id = (blocks_.size() - 1) * CellsPerBlock; |
137 | 28.6k | for (Cell& cell : block->cells) { |
138 | 28.6k | FreeCell* free_cell = reinterpret_cast<FreeCell*>(cell); |
139 | 28.6k | free_cell->id = cell_id++; |
140 | 28.6k | free_cell->next = free_list_; |
141 | 28.6k | free_list_ = free_cell; |
142 | 28.6k | } |
143 | 57 | } |
144 | | |
145 | 10.0k | FreeCell* cell = free_list_; |
146 | 10.0k | free_list_ = free_list_->next; |
147 | 10.0k | num_free_--; |
148 | 10.0k | *obj_id = cell->id; |
149 | 10.0k | return cell; |
150 | 10.0k | } _ZN4PoolILi682ELm24EE8AllocateEPi Line | Count | Source | 125 | 6.94k | void* Allocate(int* obj_id) { | 126 | 6.94k | num_allocated_++; | 127 | | | 128 | 6.94k | if (!free_list_) { | 129 | | // Allocate a new Block and add every new Cell to the free list. | 130 | 27 | Block* block = static_cast<Block*>(malloc(sizeof(Block))); | 131 | 27 | blocks_.push_back(block); | 132 | 27 | bytes_allocated_ += kBlockSize; | 133 | 27 | num_free_ += CellsPerBlock; | 134 | | | 135 | | // The starting cell_id for Cells in this block. | 136 | 27 | int cell_id = (blocks_.size() - 1) * CellsPerBlock; | 137 | 18.4k | for (Cell& cell : block->cells) { | 138 | 18.4k | FreeCell* free_cell = reinterpret_cast<FreeCell*>(cell); | 139 | 18.4k | free_cell->id = cell_id++; | 140 | 18.4k | free_cell->next = free_list_; | 141 | 18.4k | free_list_ = free_cell; | 142 | 18.4k | } | 143 | 27 | } | 144 | | | 145 | 6.94k | FreeCell* cell = free_list_; | 146 | 6.94k | free_list_ = free_list_->next; | 147 | 6.94k | num_free_--; | 148 | 6.94k | *obj_id = cell->id; | 149 | 6.94k | return cell; | 150 | 6.94k | } |
_ZN4PoolILi341ELm48EE8AllocateEPi Line | Count | Source | 125 | 3.15k | void* Allocate(int* obj_id) { | 126 | 3.15k | num_allocated_++; | 127 | | | 128 | 3.15k | if (!free_list_) { | 129 | | // Allocate a new Block and add every new Cell to the free list. | 130 | 30 | Block* block = static_cast<Block*>(malloc(sizeof(Block))); | 131 | 30 | blocks_.push_back(block); | 132 | 30 | bytes_allocated_ += kBlockSize; | 133 | 30 | num_free_ += CellsPerBlock; | 134 | | | 135 | | // The starting cell_id for Cells in this block. | 136 | 30 | int cell_id = (blocks_.size() - 1) * CellsPerBlock; | 137 | 10.2k | for (Cell& cell : block->cells) { | 138 | 10.2k | FreeCell* free_cell = reinterpret_cast<FreeCell*>(cell); | 139 | 10.2k | free_cell->id = cell_id++; | 140 | 10.2k | free_cell->next = free_list_; | 141 | 10.2k | free_list_ = free_cell; | 142 | 10.2k | } | 143 | 30 | } | 144 | | | 145 | 3.15k | FreeCell* cell = free_list_; | 146 | 3.15k | free_list_ = free_list_->next; | 147 | 3.15k | num_free_--; | 148 | 3.15k | *obj_id = cell->id; | 149 | 3.15k | return cell; | 150 | 3.15k | } |
|
151 | | |
152 | 82 | void PrepareForGc() { |
153 | 82 | DCHECK(!gc_underway_); |
154 | 0 | gc_underway_ = true; |
155 | 82 | mark_set_.ReInit(blocks_.size() * CellsPerBlock); |
156 | 82 | } _ZN4PoolILi682ELm24EE12PrepareForGcEv Line | Count | Source | 152 | 41 | void PrepareForGc() { | 153 | 41 | DCHECK(!gc_underway_); | 154 | 0 | gc_underway_ = true; | 155 | 41 | mark_set_.ReInit(blocks_.size() * CellsPerBlock); | 156 | 41 | } |
_ZN4PoolILi341ELm48EE12PrepareForGcEv Line | Count | Source | 152 | 41 | void PrepareForGc() { | 153 | 41 | DCHECK(!gc_underway_); | 154 | 0 | gc_underway_ = true; | 155 | 41 | mark_set_.ReInit(blocks_.size() * CellsPerBlock); | 156 | 41 | } |
|
157 | | |
158 | 1.77k | bool IsMarked(int cell_id) { |
159 | 1.77k | DCHECK(gc_underway_); |
160 | 0 | return mark_set_.IsMarked(cell_id); |
161 | 1.77k | } _ZN4PoolILi682ELm24EE8IsMarkedEi Line | Count | Source | 158 | 1.50k | bool IsMarked(int cell_id) { | 159 | 1.50k | DCHECK(gc_underway_); | 160 | 0 | return mark_set_.IsMarked(cell_id); | 161 | 1.50k | } |
_ZN4PoolILi341ELm48EE8IsMarkedEi Line | Count | Source | 158 | 266 | bool IsMarked(int cell_id) { | 159 | 266 | DCHECK(gc_underway_); | 160 | 0 | return mark_set_.IsMarked(cell_id); | 161 | 266 | } |
|
162 | | |
163 | 1.59k | void Mark(int cell_id) { |
164 | 1.59k | DCHECK(gc_underway_); |
165 | 0 | mark_set_.Mark(cell_id); |
166 | 1.59k | } _ZN4PoolILi682ELm24EE4MarkEi Line | Count | Source | 163 | 1.38k | void Mark(int cell_id) { | 164 | 1.38k | DCHECK(gc_underway_); | 165 | 0 | mark_set_.Mark(cell_id); | 166 | 1.38k | } |
_ZN4PoolILi341ELm48EE4MarkEi Line | Count | Source | 163 | 210 | void Mark(int cell_id) { | 164 | 210 | DCHECK(gc_underway_); | 165 | 0 | mark_set_.Mark(cell_id); | 166 | 210 | } |
|
167 | | |
168 | 82 | void Sweep() { |
169 | 82 | DCHECK(gc_underway_); |
170 | | // Iterate over every Cell linking the free ones into a new free list. |
171 | 0 | num_free_ = 0; |
172 | 82 | free_list_ = nullptr; |
173 | 82 | int cell_id = 0; |
174 | 85 | for (Block* block : blocks_) { |
175 | 43.9k | for (Cell& cell : block->cells) { |
176 | 43.9k | if (!mark_set_.IsMarked(cell_id)) { |
177 | 42.3k | num_free_++; |
178 | 42.3k | FreeCell* free_cell = reinterpret_cast<FreeCell*>(cell); |
179 | 42.3k | free_cell->id = cell_id; |
180 | 42.3k | free_cell->next = free_list_; |
181 | 42.3k | free_list_ = free_cell; |
182 | 42.3k | } |
183 | 43.9k | cell_id++; |
184 | 43.9k | } |
185 | 85 | } |
186 | 82 | gc_underway_ = false; |
187 | 82 | } _ZN4PoolILi682ELm24EE5SweepEv Line | Count | Source | 168 | 41 | void Sweep() { | 169 | 41 | DCHECK(gc_underway_); | 170 | | // Iterate over every Cell linking the free ones into a new free list. | 171 | 0 | num_free_ = 0; | 172 | 41 | free_list_ = nullptr; | 173 | 41 | int cell_id = 0; | 174 | 44 | for (Block* block : blocks_) { | 175 | 30.0k | for (Cell& cell : block->cells) { | 176 | 30.0k | if (!mark_set_.IsMarked(cell_id)) { | 177 | 28.6k | num_free_++; | 178 | 28.6k | FreeCell* free_cell = reinterpret_cast<FreeCell*>(cell); | 179 | 28.6k | free_cell->id = cell_id; | 180 | 28.6k | free_cell->next = free_list_; | 181 | 28.6k | free_list_ = free_cell; | 182 | 28.6k | } | 183 | 30.0k | cell_id++; | 184 | 30.0k | } | 185 | 44 | } | 186 | 41 | gc_underway_ = false; | 187 | 41 | } |
_ZN4PoolILi341ELm48EE5SweepEv Line | Count | Source | 168 | 41 | void Sweep() { | 169 | 41 | DCHECK(gc_underway_); | 170 | | // Iterate over every Cell linking the free ones into a new free list. | 171 | 0 | num_free_ = 0; | 172 | 41 | free_list_ = nullptr; | 173 | 41 | int cell_id = 0; | 174 | 41 | for (Block* block : blocks_) { | 175 | 13.9k | for (Cell& cell : block->cells) { | 176 | 13.9k | if (!mark_set_.IsMarked(cell_id)) { | 177 | 13.7k | num_free_++; | 178 | 13.7k | FreeCell* free_cell = reinterpret_cast<FreeCell*>(cell); | 179 | 13.7k | free_cell->id = cell_id; | 180 | 13.7k | free_cell->next = free_list_; | 181 | 13.7k | free_list_ = free_cell; | 182 | 13.7k | } | 183 | 13.9k | cell_id++; | 184 | 13.9k | } | 185 | 41 | } | 186 | 41 | gc_underway_ = false; | 187 | 41 | } |
|
188 | | |
189 | 64 | void Free() { |
190 | 64 | for (Block* block : blocks_) { |
191 | 57 | free(block); |
192 | 57 | } |
193 | 64 | blocks_.clear(); |
194 | 64 | num_free_ = 0; |
195 | 64 | } _ZN4PoolILi682ELm24EE4FreeEv Line | Count | Source | 189 | 32 | void Free() { | 190 | 32 | for (Block* block : blocks_) { | 191 | 27 | free(block); | 192 | 27 | } | 193 | 32 | blocks_.clear(); | 194 | 32 | num_free_ = 0; | 195 | 32 | } |
_ZN4PoolILi341ELm48EE4FreeEv Line | Count | Source | 189 | 32 | void Free() { | 190 | 32 | for (Block* block : blocks_) { | 191 | 30 | free(block); | 192 | 30 | } | 193 | 32 | blocks_.clear(); | 194 | 32 | num_free_ = 0; | 195 | 32 | } |
|
196 | | |
197 | 0 | int num_allocated() { |
198 | 0 | return num_allocated_; |
199 | 0 | } Unexecuted instantiation: _ZN4PoolILi682ELm24EE13num_allocatedEv Unexecuted instantiation: _ZN4PoolILi341ELm48EE13num_allocatedEv |
200 | | |
201 | 0 | int64_t bytes_allocated() { |
202 | 0 | return bytes_allocated_; |
203 | 0 | } Unexecuted instantiation: _ZN4PoolILi682ELm24EE15bytes_allocatedEv Unexecuted instantiation: _ZN4PoolILi341ELm48EE15bytes_allocatedEv |
204 | | |
205 | 4.39k | int num_live() { |
206 | 4.39k | #ifndef OPTIMIZED |
207 | 4.39k | int capacity = blocks_.size() * CellsPerBlock; |
208 | | // log("Pool capacity = %d", capacity); |
209 | | // log("Pool num_free_ = %d", num_free_); |
210 | 4.39k | DCHECK(num_free_ <= capacity); |
211 | 0 | #endif |
212 | 0 | return blocks_.size() * CellsPerBlock - num_free_; |
213 | 4.39k | } _ZN4PoolILi682ELm24EE8num_liveEv Line | Count | Source | 205 | 2.19k | int num_live() { | 206 | 2.19k | #ifndef OPTIMIZED | 207 | 2.19k | int capacity = blocks_.size() * CellsPerBlock; | 208 | | // log("Pool capacity = %d", capacity); | 209 | | // log("Pool num_free_ = %d", num_free_); | 210 | 2.19k | DCHECK(num_free_ <= capacity); | 211 | 0 | #endif | 212 | 0 | return blocks_.size() * CellsPerBlock - num_free_; | 213 | 2.19k | } |
_ZN4PoolILi341ELm48EE8num_liveEv Line | Count | Source | 205 | 2.19k | int num_live() { | 206 | 2.19k | #ifndef OPTIMIZED | 207 | 2.19k | int capacity = blocks_.size() * CellsPerBlock; | 208 | | // log("Pool capacity = %d", capacity); | 209 | | // log("Pool num_free_ = %d", num_free_); | 210 | 2.19k | DCHECK(num_free_ <= capacity); | 211 | 0 | #endif | 212 | 0 | return blocks_.size() * CellsPerBlock - num_free_; | 213 | 2.19k | } |
|
214 | | |
215 | | private: |
216 | | using Cell = uint8_t[CellSize]; |
217 | | |
218 | | struct Block { |
219 | | Cell cells[CellsPerBlock]; |
220 | | }; |
221 | | |
222 | | // Unused/free cells are tracked via a linked list of FreeCells. The FreeCells |
223 | | // are stored in the unused Cells, so it takes no extra memory to track them. |
224 | | struct FreeCell { |
225 | | int id; |
226 | | FreeCell* next; |
227 | | }; |
228 | | static_assert(CellSize >= sizeof(FreeCell), "CellSize is too small"); |
229 | | |
230 | | // Whether a GC is underway, for asserting that calls are in order. |
231 | | bool gc_underway_ = false; |
232 | | |
233 | | FreeCell* free_list_ = nullptr; |
234 | | int num_free_ = 0; |
235 | | int num_allocated_ = 0; |
236 | | int64_t bytes_allocated_ = 0; |
237 | | std::vector<Block*> blocks_; |
238 | | MarkSet mark_set_; |
239 | | |
240 | | DISALLOW_COPY_AND_ASSIGN(Pool); |
241 | | }; |
242 | | |
243 | | class MarkSweepHeap { |
244 | | public: |
245 | | // reserve 32 frames to start |
246 | 32 | MarkSweepHeap() { |
247 | 32 | } |
248 | | |
249 | | void Init(); // use default threshold |
250 | | void Init(int gc_threshold); |
251 | | |
252 | 15.9k | void PushRoot(RawObject** p) { |
253 | | #if VALIDATE_ROOTS |
254 | | ValidateRoot(*p); |
255 | | #endif |
256 | 15.9k | roots_.push_back(p); |
257 | 15.9k | } |
258 | | |
259 | 15.9k | void PopRoot() { |
260 | 15.9k | roots_.pop_back(); |
261 | 15.9k | } |
262 | | |
263 | 3 | void RootGlobalVar(void* root) { |
264 | 3 | global_roots_.push_back(reinterpret_cast<RawObject*>(root)); |
265 | 3 | } |
266 | | |
267 | | void* Allocate(size_t num_bytes, int* obj_id, int* pool_id); |
268 | | |
269 | | #if 0 |
270 | | void* Reallocate(void* p, size_t num_bytes); |
271 | | #endif |
272 | | int MaybeCollect(); |
273 | | int Collect(); |
274 | | |
275 | | void MaybeMarkAndPush(RawObject* obj); |
276 | | void TraceChildren(); |
277 | | |
278 | | void Sweep(); |
279 | | |
280 | | void PrintStats(int fd); // public for testing |
281 | | void PrintShortStats(); |
282 | | |
283 | | void CleanProcessExit(); // do one last GC, used in unit tests |
284 | | void ProcessExit(); // main() lets OS clean up, except ASAN variant |
285 | | |
286 | 2.19k | int num_live() { |
287 | 2.19k | return num_live_ |
288 | 2.19k | #ifndef NO_POOL_ALLOC |
289 | 2.19k | + pool1_.num_live() + pool2_.num_live() |
290 | 2.19k | #endif |
291 | 2.19k | ; |
292 | 2.19k | } |
293 | | |
294 | | bool is_initialized_ = true; // mark/sweep doesn't need to be initialized |
295 | | |
296 | | // Runtime params |
297 | | |
298 | | // Threshold is a number of live objects, since we aren't keeping track of |
299 | | // total bytes |
300 | | int gc_threshold_; |
301 | | |
302 | | // Show debug logging |
303 | | bool gc_verbose_ = false; |
304 | | |
305 | | // Current stats |
306 | | int num_live_ = 0; |
307 | | // Should we keep track of sizes? |
308 | | // int64_t bytes_live_ = 0; |
309 | | |
310 | | // Cumulative stats |
311 | | int max_survived_ = 0; // max # live after a collection |
312 | | int num_allocated_ = 0; |
313 | | int64_t bytes_allocated_ = 0; // avoid overflow |
314 | | int num_gc_points_ = 0; // manual collection points |
315 | | int num_collections_ = 0; |
316 | | int num_growths_; |
317 | | double max_gc_millis_ = 0.0; |
318 | | double total_gc_millis_ = 0.0; |
319 | | |
320 | | #ifndef NO_POOL_ALLOC |
321 | | // 16,384 / 24 bytes = 682 cells (rounded), 16,368 bytes |
322 | | // 16,384 / 48 bytes = 341 cells (rounded), 16,368 bytes |
323 | | // Conveniently, the glibc malloc header is 16 bytes, giving exactly 16 Ki |
324 | | // differences |
325 | | Pool<682, 24> pool1_; |
326 | | Pool<341, 48> pool2_; |
327 | | #endif |
328 | | |
329 | | std::vector<RawObject**> roots_; |
330 | | std::vector<RawObject*> global_roots_; |
331 | | |
332 | | // Allocate() appends live objects, and Sweep() compacts it |
333 | | std::vector<ObjHeader*> live_objs_; |
334 | | // Allocate lazily frees these, and Sweep() replenishes it |
335 | | std::vector<ObjHeader*> to_free_; |
336 | | |
337 | | std::vector<ObjHeader*> gray_stack_; |
338 | | MarkSet mark_set_; |
339 | | |
340 | | int greatest_obj_id_ = 0; |
341 | | |
342 | | private: |
343 | | void FreeEverything(); |
344 | | void MaybePrintStats(); |
345 | | |
346 | | DISALLOW_COPY_AND_ASSIGN(MarkSweepHeap); |
347 | | }; |
348 | | |
349 | | #endif // MARKSWEEP_HEAP_H |