From a5e65b2ecb46718dc4c3a12e8fc219cececfc285 Mon Sep 17 00:00:00 2001 From: pompolic <pompolic@special-circumstanc.es> Date: Tue, 11 Oct 2022 18:18:52 +0200 Subject: [PATCH] Fix condition when filtering token map --- gdb-port/top-level-parse.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/gdb-port/top-level-parse.py b/gdb-port/top-level-parse.py index 962d35b..1c7e9ad 100644 --- a/gdb-port/top-level-parse.py +++ b/gdb-port/top-level-parse.py @@ -394,7 +394,7 @@ class TopLevelParse: #h = min(rows, len(token.children)) tokenmap_values = list(self.input_token_map.values()) #tokenmap_val_list = [token for token in list(self.input_token_map.values()) if (token['start'] <= start and token['start'] <= end)] - tokenmap_val_list = [dict(token, addr=hex(key)) for key, token in self.input_token_map.items() if (token['start'] <= start and token['start'] <= end)] + tokenmap_val_list = [dict(token, addr=hex(key)) for key, token in self.input_token_map.items() if (token['start'] >= start and token['start'] <= end)] numrows = min(rows, len(tokenmap_val_list)) #row_lengths_for_columns = [len([token for token in list(tokenmap_values) if (token['start'] <= x and x >= token['end'])]) for x in range(start,end+1)] #numrows = min(rows, max(row_lengths_for_columns)) -- GitLab