|
19 | 19 | import bisect |
20 | 20 | import collections |
21 | 21 | import random |
22 | | -import sys |
23 | 22 |
|
24 | 23 | # |
25 | 24 | # each line of index events source is this syntax: |
@@ -94,10 +93,7 @@ def __repr__(self): |
94 | 93 | return f"<segment {self.name} in_ram={self.in_ram} size={mult * self.size_in_bytes / 1024 / 1024:.1f} MB>" |
95 | 94 |
|
96 | 95 | def add_document(self, docid, size_bytes): |
97 | | - """ |
98 | | - Only used for in-memory segments. |
99 | | - """ |
100 | | - |
| 96 | + """Only used for in-memory segments.""" |
101 | 97 | assert docid not in self.docs |
102 | 98 | assert docid not in self.deletes |
103 | 99 |
|
@@ -163,10 +159,7 @@ def get_doc_counts(self): |
163 | 159 | return max_doc, del_count |
164 | 160 |
|
165 | 161 | def refresh(self): |
166 | | - """ |
167 | | - Returns frozen segments for searching. |
168 | | - """ |
169 | | - |
| 162 | + """Returns frozen segments for searching.""" |
170 | 163 | for index_thread, seg in list(self.index_thread_to_segment.items()): |
171 | 164 | self.flush(index_thread, seg, "refresh") |
172 | 165 | print(f"after refresh {self.ram_bytes_used} {len(self.index_thread_to_segment)}") |
@@ -254,14 +247,12 @@ def finish_merge(self, merged_seg, to_merge_segments, del_docs_to_reclaim, final |
254 | 247 | self.maybe_merge("finish-merge") |
255 | 248 |
|
256 | 249 | def launch_merge(self, to_merge_segments, reason): |
257 | | - """ |
258 | | - Simulates a merge running, scheduling the end of the merge to commit / reclaim deletes. |
| 250 | + """Simulates a merge running, scheduling the end of the merge to commit / reclaim deletes. |
259 | 251 | The merge runs in the background ... once the clock advances to the merge finish time, |
260 | 252 | we commit the merge. |
261 | 253 |
|
262 | 254 | We model merge run-time as simple linear multiplier on size of merged segment. |
263 | 255 | """ |
264 | | - |
265 | 256 | for seg in to_merge_segments: |
266 | 257 | assert seg not in self.merging_segments |
267 | 258 | self.merging_segments.add(seg) |
@@ -412,7 +403,7 @@ def main(): |
412 | 403 | search_net_deletes = 0 |
413 | 404 | net_replicate_bytes = 0 |
414 | 405 |
|
415 | | - with open(index_events_source, "r") as f: |
| 406 | + with open(index_events_source) as f: |
416 | 407 | while True: |
417 | 408 | line = f.readline() |
418 | 409 | if line == "": |
@@ -468,7 +459,7 @@ def main(): |
468 | 459 | next_print_sec += print_every_sec |
469 | 460 |
|
470 | 461 | # summary stats |
471 | | - print(f"\nDONE!") |
| 462 | + print("\nDONE!") |
472 | 463 | print(f" {search_net_docs=:,} {search_net_deletes=:,} ({100.0 * search_net_deletes / search_net_docs:.1f} %)") |
473 | 464 | print(f" replicated {net_replicate_bytes / 1024 / 1024 / 1024.0:,.1f} GB") |
474 | 465 |
|
|
0 commit comments