From 1335b259cff697d3882da4a1b9c75cf5e682f4ae Mon Sep 17 00:00:00 2001 From: pompolic <pompolic@special-circumstanc.es> Date: Thu, 20 Oct 2022 14:36:52 +0200 Subject: [PATCH] More cleanup --- gdb-port/top-level-parse.py | 36 ------------------------------------ 1 file changed, 36 deletions(-) diff --git a/gdb-port/top-level-parse.py b/gdb-port/top-level-parse.py index 111458e..c0aa1ae 100644 --- a/gdb-port/top-level-parse.py +++ b/gdb-port/top-level-parse.py @@ -401,47 +401,13 @@ class TopLevelParse: # TODO: remove first printing loop def print_input_chunk(self, token, rows, start, end): #print("print_input_chunk(): start: ", start, "end: ", end) - #h = min(rows, len(token.children)) - tokenmap_values = list(self.input_token_map.values()) - #tokenmap_val_list = [token for token in list(self.input_token_map.values()) if (token['start'] <= start and token['start'] <= end)] tokenmap_val_list = [dict(token, addr=hex(key)) for key, token in self.input_token_map.items() if (token['start'] >= start and token['start'] <= end)] numrows = min(rows, len(tokenmap_val_list)) - #row_lengths_for_columns = [len([token for token in list(tokenmap_values) if (token['start'] <= x and x >= token['end'])]) for x in range(start,end+1)] - #numrows = min(rows, max(row_lengths_for_columns)) - #print("numrows: ", numrows) h = numrows w = end-start if w == 0: print("all tokens passed to display are zero-width") return - charbuf = [[' '] * w for i in range(0,h)] - #print("h=len(charbuf):", len(charbuf), "w :", w) - tokenmap_values_sorted = sorted(tokenmap_val_list, key=lambda token: (token['start'], token['end'])) - #print("len(tokenmap_values_sorted):", len(tokenmap_values_sorted), "rows:", numrows) - hpr_strings = [] - for i in range(0, numrows): - try: # try-block for debugging - token_length = tokenmap_values_sorted[i]['end'] - tokenmap_values_sorted[i]['start'] # TODO: save tokenmap_values_sorted[i] - start_col = max(tokenmap_values_sorted[i]['start'] - start, 0) - end_col = min(tokenmap_values_sorted[i]['end'] - start, w-1) - charbuf[i][start_col:end_col] = ['X'] * min(token_length, w) #TODO: this appends to the row in question -- charbuf[0] is input[start] - hparseresult_addr = tokenmap_values_sorted[i]['hparseresult'] - except IndexError as ie: #DEBUG - print(hpr_strings) - print(i) - print(ie) - try: - hpr = HParseResult(int(hparseresult_addr, 16)) - hpr_strings.append(hpr.str_no_deref()) - except: - hpr_strings.append("invalid") - charbufrows_token_debug = [ "".join(row) + "\n" + " ".join([hpr_strings[index], str(tokenmap_values_sorted[index])]) for index, row in enumerate(charbuf)] - charbufrows = ["".join(row) for row in charbuf] - #charbuf_final = "\n".join(charbufrows) - charbuf_final = "\n".join(charbufrows_token_debug) - #print(charbuf_final) - #print(hpr_strings) - #print(len(hpr_strings), len(charbuf), len(tokenmap_values_sorted)) # Print by getting a token, and recursively walking its children if applicable charbuf = [[' '] * w for i in range(0,h)] @@ -451,12 +417,10 @@ class TopLevelParse: if not self.printing_tokens: # This ignores the passed-in token parameter if there are still tokens left to print self.printing_tokens = [token] while current_row < h and self.printing_tokens: - #print(self.printing_tokens) token = self.printing_tokens[-1] #token = self.printing_tokens.pop() tokenmap_entry = self.input_token_map[int(token.address)] if tokenmap_entry['start'] > end: # Next token won't fit on the screen - #print("Token start past chunk end, break, tokenmap_entry['start']:", tokenmap_entry['start'], "end:", end) break self.printing_tokens.pop() token_length = tokenmap_entry['end'] - tokenmap_entry['start'] # TODO: save tokenmap_values_sorted[i] -- GitLab