32
32
import ctypes
33
33
import functools
34
34
from colorama import Fore , Back , Style
35
- from prettytable import PrettyTable
35
+ from prettytable import PrettyTable , HEADER
36
36
from copy import copy , deepcopy
37
37
38
38
from time import sleep , time
@@ -766,7 +766,7 @@ def generate_test_summary_by_target(self, test_summary, shuffle_seed=None):
766
766
result_dict [test [TEST_INDEX ]][test [TOOLCHAIN_INDEX ]] = test [RESULT_INDEX ]
767
767
768
768
pt_cols = ["Target" , "Test ID" , "Test Description" ] + unique_target_toolchains
769
- pt = PrettyTable (pt_cols )
769
+ pt = PrettyTable (pt_cols , junction_char = "|" , hrules = HEADER )
770
770
for col in pt_cols :
771
771
pt .align [col ] = "l"
772
772
pt .padding_width = 1 # One space between column edges and contents (default)
@@ -794,7 +794,7 @@ def generate_test_summary(self, test_summary, shuffle_seed=None):
794
794
result = "Test summary:\n "
795
795
# Pretty table package is used to print results
796
796
pt = PrettyTable (["Result" , "Target" , "Toolchain" , "Test ID" , "Test Description" ,
797
- "Elapsed Time (sec)" , "Timeout (sec)" , "Loops" ])
797
+ "Elapsed Time (sec)" , "Timeout (sec)" , "Loops" ], junction_char = "|" , hrules = HEADER )
798
798
pt .align ["Result" ] = "l" # Left align
799
799
pt .align ["Target" ] = "l" # Left align
800
800
pt .align ["Toolchain" ] = "l" # Left align
@@ -1328,7 +1328,7 @@ def print_muts_configuration_from_json(json_data, join_delim=", ", platform_filt
1328
1328
1329
1329
# Prepare pretty table object to display all MUTs
1330
1330
pt_cols = ["index" ] + muts_info_cols
1331
- pt = PrettyTable (pt_cols )
1331
+ pt = PrettyTable (pt_cols , junction_char = "|" , hrules = HEADER )
1332
1332
for col in pt_cols :
1333
1333
pt .align [col ] = "l"
1334
1334
@@ -1366,7 +1366,7 @@ def print_test_configuration_from_json(json_data, join_delim=", "):
1366
1366
1367
1367
# Prepare pretty table object to display test specification
1368
1368
pt_cols = ["mcu" ] + sorted (toolchains_info_cols )
1369
- pt = PrettyTable (pt_cols )
1369
+ pt = PrettyTable (pt_cols , junction_char = "|" , hrules = HEADER )
1370
1370
for col in pt_cols :
1371
1371
pt .align [col ] = "l"
1372
1372
@@ -1455,7 +1455,7 @@ def get_avail_tests_summary_table(cols=None, result_summary=True, join_delim=','
1455
1455
'duration' ] if cols is None else cols
1456
1456
1457
1457
# All tests status table print
1458
- pt = PrettyTable (test_properties )
1458
+ pt = PrettyTable (test_properties , junction_char = "|" , hrules = HEADER )
1459
1459
for col in test_properties :
1460
1460
pt .align [col ] = "l"
1461
1461
pt .align ['duration' ] = "r"
@@ -1495,7 +1495,7 @@ def get_avail_tests_summary_table(cols=None, result_summary=True, join_delim=','
1495
1495
if result_summary and not platform_filter :
1496
1496
# Automation result summary
1497
1497
test_id_cols = ['automated' , 'all' , 'percent [%]' , 'progress' ]
1498
- pt = PrettyTable (test_id_cols )
1498
+ pt = PrettyTable (test_id_cols , junction_char = "|" , hrules = HEADER )
1499
1499
pt .align ['automated' ] = "r"
1500
1500
pt .align ['all' ] = "r"
1501
1501
pt .align ['percent [%]' ] = "r"
@@ -1509,7 +1509,7 @@ def get_avail_tests_summary_table(cols=None, result_summary=True, join_delim=','
1509
1509
1510
1510
# Test automation coverage table print
1511
1511
test_id_cols = ['id' , 'automated' , 'all' , 'percent [%]' , 'progress' ]
1512
- pt = PrettyTable (test_id_cols )
1512
+ pt = PrettyTable (test_id_cols , junction_char = "|" , hrules = HEADER )
1513
1513
pt .align ['id' ] = "l"
1514
1514
pt .align ['automated' ] = "r"
1515
1515
pt .align ['all' ] = "r"
0 commit comments