From a5e65b2ecb46718dc4c3a12e8fc219cececfc285 Mon Sep 17 00:00:00 2001
From: pompolic <pompolic@special-circumstanc.es>
Date: Tue, 11 Oct 2022 18:18:52 +0200
Subject: [PATCH] Fix condition when filtering token map

---
 gdb-port/top-level-parse.py | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/gdb-port/top-level-parse.py b/gdb-port/top-level-parse.py
index 962d35b..1c7e9ad 100644
--- a/gdb-port/top-level-parse.py
+++ b/gdb-port/top-level-parse.py
@@ -394,7 +394,7 @@ class TopLevelParse:
 		#h = min(rows, len(token.children))
 		tokenmap_values = list(self.input_token_map.values())
 		#tokenmap_val_list = [token for token in list(self.input_token_map.values()) if (token['start'] <= start and token['start'] <= end)]
-		tokenmap_val_list = [dict(token, addr=hex(key)) for key, token in self.input_token_map.items() if (token['start'] <= start and token['start'] <= end)]
+		tokenmap_val_list = [dict(token, addr=hex(key)) for key, token in self.input_token_map.items() if (token['start'] >= start and token['start'] <= end)]
 		numrows = min(rows, len(tokenmap_val_list))
 		#row_lengths_for_columns = [len([token for token in list(tokenmap_values) if (token['start'] <= x and x >= token['end'])]) for x in range(start,end+1)]
 		#numrows = min(rows, max(row_lengths_for_columns))
-- 
GitLab