|
| 1 | +import typing as t |
| 2 | +from graph_net.torch.rp_expr.rp_expr_parser import RpExprParser |
| 3 | +from graph_net.torch.rp_expr.rp_expr import PrimitiveId, LetsListTokenRpExpr |
| 4 | +import numpy as np |
| 5 | +import sys |
| 6 | + |
| 7 | + |
| 8 | +class LongestRpExprParser: |
| 9 | + def __init__(self, max_window_size=1024, min_window_size=4): |
| 10 | + self.max_window_size = max_window_size |
| 11 | + self.min_window_size = min_window_size |
| 12 | + |
| 13 | + def __call__(self, primitive_id_lists: t.List[t.List[PrimitiveId]]): |
| 14 | + fold_policy = "default" |
| 15 | + rp_expr_parser = RpExprParser( |
| 16 | + self.max_window_size, |
| 17 | + fold_policy=fold_policy, |
| 18 | + fold_times=1, |
| 19 | + ) |
| 20 | + lets_list_rp_expr, token_id2primitive_id = rp_expr_parser(primitive_id_lists) |
| 21 | + for window_size in self._get_sub_window_sizes(): |
| 22 | + rp_expr_parser = RpExprParser( |
| 23 | + window_size, |
| 24 | + fold_policy=fold_policy, |
| 25 | + fold_times=1, |
| 26 | + ) |
| 27 | + cur_primitive_id_lists = [ |
| 28 | + [token_id2primitive_id[token_id] for token_id in tensor.tolist()] |
| 29 | + for tensor in lets_list_rp_expr.get_pure_primitive_binding_tensors( |
| 30 | + token_id2primitive_id |
| 31 | + ) |
| 32 | + ] |
| 33 | + cur_lets_list_rp_expr, cur_token_id2primitive_id = rp_expr_parser( |
| 34 | + cur_primitive_id_lists |
| 35 | + ) |
| 36 | + # cur_lets_list_rp_expr.try_unwrap_body_of_sole_symbol_token() |
| 37 | + lets_list_rp_expr = self._merge_lets_list_rp_expr( |
| 38 | + inner=cur_lets_list_rp_expr, |
| 39 | + outer=lets_list_rp_expr, |
| 40 | + inner_token_id2primitive_id=cur_token_id2primitive_id, |
| 41 | + outer_token_id2primitive_id=token_id2primitive_id, |
| 42 | + ) |
| 43 | + lets_list_rp_expr.try_recursive_inline_symbol_sole_used( |
| 44 | + token_id2primitive_id=token_id2primitive_id |
| 45 | + ) |
| 46 | + # lets_list_rp_expr.try_unwrap_body_of_sole_symbol_token() |
| 47 | + return lets_list_rp_expr, token_id2primitive_id |
| 48 | + |
| 49 | + def _merge_lets_list_rp_expr( |
| 50 | + self, |
| 51 | + inner, |
| 52 | + outer, |
| 53 | + inner_token_id2primitive_id, |
| 54 | + outer_token_id2primitive_id, |
| 55 | + ): |
| 56 | + def get_inner_token_id2outer_token_id(): |
| 57 | + primitive_id2outer_token_id = {} |
| 58 | + for token_id, primitive_id in enumerate(outer_token_id2primitive_id): |
| 59 | + assert primitive_id not in primitive_id2outer_token_id |
| 60 | + primitive_id2outer_token_id[primitive_id] = token_id |
| 61 | + return [ |
| 62 | + primitive_id2outer_token_id[primitive_id] |
| 63 | + for primitive_id in inner_token_id2primitive_id |
| 64 | + ] |
| 65 | + |
| 66 | + kInner = "inner" |
| 67 | + kOuter = "outer" |
| 68 | + uid2new_symbol_token = self._make_uid2new_symbol_token_id( |
| 69 | + inner=inner, |
| 70 | + outer=outer, |
| 71 | + inner_uid_prefix=kInner, |
| 72 | + outer_uid_prefix=kOuter, |
| 73 | + outer_primitive_table_size=len(outer_token_id2primitive_id), |
| 74 | + ) |
| 75 | + inner_symbol_token_ids = self._convert_symbol_token_ids( |
| 76 | + symbol_token_ids=inner.symbol_token_ids, |
| 77 | + new_token4old_token=( |
| 78 | + lambda old_token: uid2new_symbol_token[f"{kInner}{old_token}"] |
| 79 | + ), |
| 80 | + ) |
| 81 | + inner_token_id2outer_token_id = get_inner_token_id2outer_token_id() |
| 82 | + inner_symbol_token_tensors = self._convert_token_tensors( |
| 83 | + inner.symbol_token_tensors, |
| 84 | + new_token4old_primitive_token=( |
| 85 | + lambda old_token: inner_token_id2outer_token_id[old_token] |
| 86 | + ), |
| 87 | + new_token4old_symbol_token=( |
| 88 | + lambda old_token: uid2new_symbol_token[f"{kInner}{old_token}"] |
| 89 | + ), |
| 90 | + primitive_ids_table_size=len(inner_token_id2primitive_id), |
| 91 | + ) |
| 92 | + |
| 93 | + inner_body_rp_expr = self._convert_token_tensors( |
| 94 | + inner.body_rp_expr, |
| 95 | + new_token4old_primitive_token=( |
| 96 | + lambda old_token: inner_token_id2outer_token_id[old_token] |
| 97 | + ), |
| 98 | + new_token4old_symbol_token=( |
| 99 | + lambda old_token: uid2new_symbol_token[f"{kInner}{old_token}"] |
| 100 | + ), |
| 101 | + primitive_ids_table_size=len(inner_token_id2primitive_id), |
| 102 | + ) |
| 103 | + |
| 104 | + inner_symbol_token2token_tensor = { |
| 105 | + symbol_token: token_tensor |
| 106 | + for symbol_token, token_tensor in zip( |
| 107 | + inner_symbol_token_ids, inner_symbol_token_tensors |
| 108 | + ) |
| 109 | + } |
| 110 | + |
| 111 | + outer_symbol_token_tensors = self._convert_outer_symbol_binding_token_tensors( |
| 112 | + inner_body_rp_expr=inner_body_rp_expr, |
| 113 | + inner_symbol_token2token_tensor=inner_symbol_token2token_tensor, |
| 114 | + outer_lets_list_rp_expr=outer, |
| 115 | + new_token4old_primitive_token=lambda x: x, |
| 116 | + new_token4old_symbol_token=( |
| 117 | + lambda old_token: uid2new_symbol_token[f"{kOuter}{old_token}"] |
| 118 | + ), |
| 119 | + outer_token_id2primitive_id=outer_token_id2primitive_id, |
| 120 | + ) |
| 121 | + |
| 122 | + symbol_token_ids = inner_symbol_token_ids + self._convert_symbol_token_ids( |
| 123 | + symbol_token_ids=outer.symbol_token_ids, |
| 124 | + new_token4old_token=( |
| 125 | + lambda old_token: uid2new_symbol_token[f"{kOuter}{old_token}"] |
| 126 | + ), |
| 127 | + ) |
| 128 | + |
| 129 | + symbol_token_tensors = inner_symbol_token_tensors + outer_symbol_token_tensors |
| 130 | + |
| 131 | + body_rp_expr = self._convert_token_tensors( |
| 132 | + outer.body_rp_expr, |
| 133 | + new_token4old_primitive_token=lambda x: x, |
| 134 | + new_token4old_symbol_token=( |
| 135 | + lambda old_token: uid2new_symbol_token[f"{kOuter}{old_token}"] |
| 136 | + ), |
| 137 | + primitive_ids_table_size=len(outer_token_id2primitive_id), |
| 138 | + ) |
| 139 | + ret_lets_list_token_rp_expr = LetsListTokenRpExpr( |
| 140 | + symbol_token_ids=symbol_token_ids, |
| 141 | + symbol_token_tensors=symbol_token_tensors, |
| 142 | + body_rp_expr=body_rp_expr, |
| 143 | + ) |
| 144 | + ret_lets_list_token_rp_expr.move_pure_primitive_bindings_front( |
| 145 | + outer_token_id2primitive_id |
| 146 | + ) |
| 147 | + return ret_lets_list_token_rp_expr |
| 148 | + |
| 149 | + def _convert_outer_symbol_binding_token_tensors( |
| 150 | + self, |
| 151 | + inner_body_rp_expr, |
| 152 | + inner_symbol_token2token_tensor, |
| 153 | + outer_lets_list_rp_expr, |
| 154 | + new_token4old_primitive_token, |
| 155 | + new_token4old_symbol_token, |
| 156 | + outer_token_id2primitive_id, |
| 157 | + ): |
| 158 | + indexes = outer_lets_list_rp_expr.get_pure_primitive_binding_indexes( |
| 159 | + outer_token_id2primitive_id |
| 160 | + ) |
| 161 | + assert len(inner_body_rp_expr) == len(indexes) |
| 162 | + index2inner_body_rp_expr_idx = { |
| 163 | + index: inner_body_rp_expr_idx |
| 164 | + for inner_body_rp_expr_idx, index in enumerate(indexes) |
| 165 | + } |
| 166 | + old_tensors = outer_lets_list_rp_expr.symbol_token_tensors |
| 167 | + return [ |
| 168 | + ( |
| 169 | + inner_body_rp_expr[index2inner_body_rp_expr_idx[index]] |
| 170 | + if index in index2inner_body_rp_expr_idx |
| 171 | + else self._convert_token_tensor( |
| 172 | + tensor=old_tensors[index], |
| 173 | + new_token4old_primitive_token=new_token4old_primitive_token, |
| 174 | + new_token4old_symbol_token=new_token4old_symbol_token, |
| 175 | + primitive_ids_table_size=len(outer_token_id2primitive_id), |
| 176 | + ) |
| 177 | + ) |
| 178 | + for index in range(len(old_tensors)) |
| 179 | + ] |
| 180 | + |
| 181 | + def _convert_token_tensors( |
| 182 | + self, |
| 183 | + tensors, |
| 184 | + new_token4old_primitive_token, |
| 185 | + new_token4old_symbol_token, |
| 186 | + primitive_ids_table_size, |
| 187 | + ): |
| 188 | + return [ |
| 189 | + self._convert_token_tensor( |
| 190 | + tensor, |
| 191 | + new_token4old_primitive_token, |
| 192 | + new_token4old_symbol_token, |
| 193 | + primitive_ids_table_size, |
| 194 | + ) |
| 195 | + for tensor in tensors |
| 196 | + ] |
| 197 | + |
| 198 | + def _convert_token_tensor( |
| 199 | + self, |
| 200 | + tensor, |
| 201 | + new_token4old_primitive_token, |
| 202 | + new_token4old_symbol_token, |
| 203 | + primitive_ids_table_size, |
| 204 | + ): |
| 205 | + return np.array( |
| 206 | + [ |
| 207 | + ( |
| 208 | + new_token4old_primitive_token(token_id) |
| 209 | + if token_id < primitive_ids_table_size |
| 210 | + else new_token4old_symbol_token(token_id) |
| 211 | + ) |
| 212 | + for token_id in tensor.tolist() |
| 213 | + ], |
| 214 | + dtype=np.int64, |
| 215 | + ) |
| 216 | + |
| 217 | + def _make_uid2new_symbol_token_id( |
| 218 | + self, |
| 219 | + inner, |
| 220 | + outer, |
| 221 | + inner_uid_prefix, |
| 222 | + outer_uid_prefix, |
| 223 | + outer_primitive_table_size, |
| 224 | + ): |
| 225 | + new_symbol_token_id = outer_primitive_table_size |
| 226 | + |
| 227 | + def get_new_symbol_token_id(): |
| 228 | + nonlocal new_symbol_token_id |
| 229 | + ret = new_symbol_token_id |
| 230 | + new_symbol_token_id += 1 |
| 231 | + return ret |
| 232 | + |
| 233 | + uid2new_symbol_token_id = {} |
| 234 | + for inner_symbol_token_id in inner.symbol_token_ids: |
| 235 | + uid = f"{inner_uid_prefix}{inner_symbol_token_id}" |
| 236 | + uid2new_symbol_token_id[uid] = get_new_symbol_token_id() |
| 237 | + for outer_symbol_token_id in outer.symbol_token_ids: |
| 238 | + uid = f"{outer_uid_prefix}{outer_symbol_token_id}" |
| 239 | + uid2new_symbol_token_id[uid] = get_new_symbol_token_id() |
| 240 | + return uid2new_symbol_token_id |
| 241 | + |
| 242 | + def _convert_symbol_token_ids(self, symbol_token_ids, new_token4old_token): |
| 243 | + return [ |
| 244 | + new_token4old_token(symbol_token_id) for symbol_token_id in symbol_token_ids |
| 245 | + ] |
| 246 | + |
| 247 | + def _get_sub_window_sizes(self): |
| 248 | + min_window_size = max(1, self.min_window_size) |
| 249 | + window_size = self.max_window_size // 2 |
| 250 | + while window_size > min_window_size: |
| 251 | + yield window_size |
| 252 | + window_size = window_size // 2 |
0 commit comments