2018-06-21 11:32:29 +00:00
|
|
|
#!/usr/bin/env python
|
|
|
|
#
|
|
|
|
# Copyright 2018 Espressif Systems (Shanghai) PTE LTD
|
|
|
|
#
|
|
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
# you may not use this file except in compliance with the License.
|
|
|
|
# You may obtain a copy of the License at
|
|
|
|
#
|
|
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
#
|
|
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
# See the License for the specific language governing permissions and
|
|
|
|
# limitations under the License.
|
|
|
|
|
2018-01-31 10:59:10 +00:00
|
|
|
"""
|
|
|
|
Test script for unit test case.
|
|
|
|
"""
|
|
|
|
|
|
|
|
import re
|
|
|
|
import os
|
|
|
|
import sys
|
|
|
|
import time
|
2018-06-21 11:32:29 +00:00
|
|
|
import argparse
|
2018-01-31 10:59:10 +00:00
|
|
|
|
|
|
|
import threading
|
|
|
|
|
|
|
|
# if we want to run test case outside `tiny-test-fw` folder,
|
|
|
|
# we need to insert tiny-test-fw path into sys path
|
|
|
|
test_fw_path = os.getenv("TEST_FW_PATH")
|
|
|
|
if test_fw_path and test_fw_path not in sys.path:
|
|
|
|
sys.path.insert(0, test_fw_path)
|
|
|
|
|
|
|
|
import TinyFW
|
|
|
|
import IDF
|
|
|
|
import Utility
|
2018-06-21 11:32:29 +00:00
|
|
|
import Env
|
2018-01-31 10:59:10 +00:00
|
|
|
from DUT import ExpectTimeout
|
|
|
|
from IDF.IDFApp import UT
|
|
|
|
|
|
|
|
|
|
|
|
UT_APP_BOOT_UP_DONE = "Press ENTER to see the list of tests."
|
2018-03-09 06:50:34 +00:00
|
|
|
RESET_PATTERN = re.compile(r"(ets [\w]{3}\s+[\d]{1,2} [\d]{4} [\d]{2}:[\d]{2}:[\d]{2}[^()]*\([\w].*?\))")
|
|
|
|
EXCEPTION_PATTERN = re.compile(r"(Guru Meditation Error: Core\s+\d panic'ed \([\w].*?\))")
|
2018-08-07 05:37:43 +00:00
|
|
|
ABORT_PATTERN = re.compile(r"(abort\(\) was called at PC 0x[a-fA-F\d]{8} on core \d)")
|
2018-03-09 06:50:34 +00:00
|
|
|
FINISH_PATTERN = re.compile(r"1 Tests (\d) Failures (\d) Ignored")
|
2018-06-21 11:32:29 +00:00
|
|
|
END_LIST_STR = r'\r?\nEnter test for running'
|
|
|
|
TEST_PATTERN = re.compile(r'\((\d+)\)\s+"([^"]+)" ([^\r]+)\r?\n(' + END_LIST_STR + r')?')
|
|
|
|
TEST_SUBMENU_PATTERN = re.compile(r'\s+\((\d+)\)\s+"[^"]+"\r?\n(?=(?=\()|(' + END_LIST_STR + r'))')
|
2018-05-21 04:50:27 +00:00
|
|
|
|
2018-06-21 11:32:29 +00:00
|
|
|
SIMPLE_TEST_ID = 0
|
|
|
|
MULTI_STAGE_ID = 1
|
|
|
|
MULTI_DEVICE_ID = 2
|
2018-01-31 10:59:10 +00:00
|
|
|
|
2018-06-21 11:32:29 +00:00
|
|
|
STARTUP_TIMEOUT=10
|
|
|
|
DEFAULT_TIMEOUT=20
|
2018-03-09 06:50:34 +00:00
|
|
|
|
2018-01-31 10:59:10 +00:00
|
|
|
def format_test_case_config(test_case_data):
|
|
|
|
"""
|
|
|
|
convert the test case data to unified format.
|
|
|
|
We need to following info to run unit test cases:
|
|
|
|
|
|
|
|
1. unit test app config
|
|
|
|
2. test case name
|
|
|
|
3. test case reset info
|
|
|
|
|
|
|
|
the formatted case config is a dict, with ut app config as keys. The value is a list of test cases.
|
|
|
|
Each test case is a dict with "name" and "reset" as keys. For example::
|
|
|
|
|
|
|
|
case_config = {
|
|
|
|
"default": [{"name": "restart from PRO CPU", "reset": "SW_CPU_RESET"}, {...}],
|
|
|
|
"psram": [{"name": "restart from PRO CPU", "reset": "SW_CPU_RESET"}],
|
|
|
|
}
|
|
|
|
|
|
|
|
If config is not specified for test case, then
|
|
|
|
|
|
|
|
:param test_case_data: string, list, or a dictionary list
|
|
|
|
:return: formatted data
|
|
|
|
"""
|
|
|
|
|
|
|
|
case_config = dict()
|
|
|
|
|
|
|
|
def parse_case(one_case_data):
|
|
|
|
""" parse and format one case """
|
|
|
|
|
|
|
|
def process_reset_list(reset_list):
|
|
|
|
# strip space and remove white space only items
|
|
|
|
_output = list()
|
|
|
|
for _r in reset_list:
|
|
|
|
_data = _r.strip(" ")
|
|
|
|
if _data:
|
|
|
|
_output.append(_data)
|
|
|
|
return _output
|
|
|
|
|
|
|
|
_case = dict()
|
|
|
|
if isinstance(one_case_data, str):
|
|
|
|
_temp = one_case_data.split(" [reset=")
|
|
|
|
_case["name"] = _temp[0]
|
|
|
|
try:
|
|
|
|
_case["reset"] = process_reset_list(_temp[1][0:-1].split(","))
|
|
|
|
except IndexError:
|
|
|
|
_case["reset"] = list()
|
|
|
|
elif isinstance(one_case_data, dict):
|
|
|
|
_case = one_case_data.copy()
|
|
|
|
assert "name" in _case
|
|
|
|
if "reset" not in _case:
|
|
|
|
_case["reset"] = list()
|
|
|
|
else:
|
|
|
|
if isinstance(_case["reset"], str):
|
|
|
|
_case["reset"] = process_reset_list(_case["reset"].split(","))
|
|
|
|
else:
|
|
|
|
raise TypeError("Not supported type during parsing unit test case")
|
|
|
|
|
|
|
|
if "config" not in _case:
|
|
|
|
_case["config"] = "default"
|
|
|
|
|
|
|
|
return _case
|
|
|
|
|
|
|
|
if not isinstance(test_case_data, list):
|
|
|
|
test_case_data = [test_case_data]
|
|
|
|
|
|
|
|
for case_data in test_case_data:
|
|
|
|
parsed_case = parse_case(case_data)
|
|
|
|
try:
|
|
|
|
case_config[parsed_case["config"]].append(parsed_case)
|
|
|
|
except KeyError:
|
|
|
|
case_config[parsed_case["config"]] = [parsed_case]
|
|
|
|
|
|
|
|
return case_config
|
|
|
|
|
2018-06-21 11:32:29 +00:00
|
|
|
def replace_app_bin(dut, name, new_app_bin):
|
|
|
|
if new_app_bin is None:
|
|
|
|
return
|
|
|
|
search_pattern = '/{}.bin'.format(name)
|
|
|
|
for i, config in enumerate(dut.download_config):
|
|
|
|
if config.endswith(search_pattern):
|
|
|
|
dut.download_config[i] = new_app_bin
|
|
|
|
Utility.console_log("The replaced application binary is {}".format(new_app_bin), "O")
|
|
|
|
break
|
2018-01-31 10:59:10 +00:00
|
|
|
|
2018-09-03 10:33:05 +00:00
|
|
|
|
|
|
|
def reset_dut(dut):
|
|
|
|
dut.reset()
|
|
|
|
# esptool ``run`` cmd takes quite long time.
|
|
|
|
# before reset finish, serial port is closed. therefore DUT could already bootup before serial port opened.
|
|
|
|
# this could cause checking bootup print failed.
|
|
|
|
# now we input cmd `-`, and check either bootup print or test history,
|
|
|
|
# to determine if DUT is ready to test.
|
|
|
|
dut.write("-", flush=False)
|
|
|
|
dut.expect_any(UT_APP_BOOT_UP_DONE,
|
|
|
|
"0 Tests 0 Failures 0 Ignored", timeout=STARTUP_TIMEOUT)
|
|
|
|
|
|
|
|
|
2018-07-26 07:07:32 +00:00
|
|
|
@IDF.idf_unit_test(env_tag="UT_T1_1")
|
2018-03-09 06:50:34 +00:00
|
|
|
def run_unit_test_cases(env, extra_data):
|
2018-01-31 10:59:10 +00:00
|
|
|
"""
|
|
|
|
extra_data can be three types of value
|
|
|
|
1. as string:
|
|
|
|
1. "case_name"
|
|
|
|
2. "case_name [reset=RESET_REASON]"
|
|
|
|
2. as dict:
|
|
|
|
1. with key like {"name": "Intr_alloc test, shared ints"}
|
|
|
|
2. with key like {"name": "restart from PRO CPU", "reset": "SW_CPU_RESET", "config": "psram"}
|
|
|
|
3. as list of string or dict:
|
|
|
|
[case1, case2, case3, {"name": "restart from PRO CPU", "reset": "SW_CPU_RESET"}, ...]
|
|
|
|
|
|
|
|
:param extra_data: the case name or case list or case dictionary
|
|
|
|
:return: None
|
|
|
|
"""
|
|
|
|
|
|
|
|
case_config = format_test_case_config(extra_data)
|
|
|
|
|
|
|
|
# we don't want stop on failed case (unless some special scenarios we can't handle)
|
|
|
|
# this flag is used to log if any of the case failed during executing
|
|
|
|
# Before exit test function this flag is used to log if the case fails
|
|
|
|
failed_cases = []
|
|
|
|
|
|
|
|
for ut_config in case_config:
|
2018-05-15 03:51:56 +00:00
|
|
|
Utility.console_log("Running unit test for config: " + ut_config, "O")
|
2018-01-31 10:59:10 +00:00
|
|
|
dut = env.get_dut("unit-test-app", app_path=ut_config)
|
2018-06-21 11:32:29 +00:00
|
|
|
if len(case_config[ut_config]) > 0:
|
|
|
|
replace_app_bin(dut, "unit-test-app", case_config[ut_config][0].get('app_bin'))
|
2018-01-31 10:59:10 +00:00
|
|
|
dut.start_app()
|
|
|
|
|
|
|
|
for one_case in case_config[ut_config]:
|
2018-09-03 10:33:05 +00:00
|
|
|
reset_dut(dut)
|
2018-01-31 10:59:10 +00:00
|
|
|
|
|
|
|
# run test case
|
|
|
|
dut.write("\"{}\"".format(one_case["name"]))
|
|
|
|
dut.expect("Running " + one_case["name"] + "...")
|
|
|
|
|
|
|
|
exception_reset_list = []
|
|
|
|
|
|
|
|
# we want to set this flag in callbacks (inner functions)
|
|
|
|
# use list here so we can use append to set this flag
|
|
|
|
test_finish = list()
|
|
|
|
|
|
|
|
# expect callbacks
|
|
|
|
def one_case_finish(result):
|
|
|
|
""" one test finished, let expect loop break and log result """
|
|
|
|
test_finish.append(True)
|
|
|
|
if result:
|
|
|
|
Utility.console_log("Success: " + one_case["name"], color="green")
|
|
|
|
else:
|
|
|
|
failed_cases.append(one_case["name"])
|
|
|
|
Utility.console_log("Failed: " + one_case["name"], color="red")
|
|
|
|
|
|
|
|
def handle_exception_reset(data):
|
|
|
|
"""
|
|
|
|
just append data to exception list.
|
|
|
|
exception list will be checked in ``handle_reset_finish``, once reset finished.
|
|
|
|
"""
|
|
|
|
exception_reset_list.append(data[0])
|
|
|
|
|
|
|
|
def handle_test_finish(data):
|
|
|
|
""" test finished without reset """
|
|
|
|
# in this scenario reset should not happen
|
|
|
|
assert not exception_reset_list
|
|
|
|
if int(data[1]):
|
|
|
|
# case ignored
|
|
|
|
Utility.console_log("Ignored: " + one_case["name"], color="orange")
|
|
|
|
one_case_finish(not int(data[0]))
|
|
|
|
|
|
|
|
def handle_reset_finish(data):
|
|
|
|
""" reset happened and reboot finished """
|
|
|
|
assert exception_reset_list # reboot but no exception/reset logged. should never happen
|
|
|
|
result = False
|
|
|
|
if len(one_case["reset"]) == len(exception_reset_list):
|
|
|
|
for i, exception in enumerate(exception_reset_list):
|
|
|
|
if one_case["reset"][i] not in exception:
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
result = True
|
|
|
|
if not result:
|
|
|
|
Utility.console_log("""Reset Check Failed: \r\n\tExpected: {}\r\n\tGet: {}"""
|
|
|
|
.format(one_case["reset"], exception_reset_list),
|
|
|
|
color="orange")
|
|
|
|
one_case_finish(result)
|
|
|
|
|
|
|
|
while not test_finish:
|
|
|
|
try:
|
2018-03-09 06:50:34 +00:00
|
|
|
dut.expect_any((RESET_PATTERN, handle_exception_reset),
|
|
|
|
(EXCEPTION_PATTERN, handle_exception_reset),
|
|
|
|
(ABORT_PATTERN, handle_exception_reset),
|
|
|
|
(FINISH_PATTERN, handle_test_finish),
|
|
|
|
(UT_APP_BOOT_UP_DONE, handle_reset_finish),
|
2018-05-21 04:50:27 +00:00
|
|
|
timeout=one_case["timeout"])
|
2018-01-31 10:59:10 +00:00
|
|
|
except ExpectTimeout:
|
|
|
|
Utility.console_log("Timeout in expect", color="orange")
|
|
|
|
one_case_finish(False)
|
|
|
|
break
|
|
|
|
|
|
|
|
# raise exception if any case fails
|
|
|
|
if failed_cases:
|
|
|
|
Utility.console_log("Failed Cases:", color="red")
|
|
|
|
for _case_name in failed_cases:
|
|
|
|
Utility.console_log("\t" + _case_name, color="red")
|
|
|
|
raise AssertionError("Unit Test Failed")
|
|
|
|
|
|
|
|
|
2018-08-20 11:21:10 +00:00
|
|
|
|
2018-01-31 10:59:10 +00:00
|
|
|
class Handler(threading.Thread):
|
|
|
|
|
|
|
|
WAIT_SIGNAL_PATTERN = re.compile(r'Waiting for signal: \[(.+)\]!')
|
|
|
|
SEND_SIGNAL_PATTERN = re.compile(r'Send signal: \[(.+)\]!')
|
|
|
|
FINISH_PATTERN = re.compile(r"1 Tests (\d) Failures (\d) Ignored")
|
|
|
|
|
2018-05-21 04:50:27 +00:00
|
|
|
def __init__(self, dut, sent_signal_list, lock, parent_case_name, child_case_index, timeout):
|
2018-01-31 10:59:10 +00:00
|
|
|
self.dut = dut
|
|
|
|
self.sent_signal_list = sent_signal_list
|
|
|
|
self.lock = lock
|
|
|
|
self.parent_case_name = parent_case_name
|
|
|
|
self.child_case_name = ""
|
|
|
|
self.child_case_index = child_case_index + 1
|
|
|
|
self.finish = False
|
|
|
|
self.result = False
|
|
|
|
self.fail_name = None
|
|
|
|
self.timeout = timeout
|
2018-08-20 11:21:10 +00:00
|
|
|
self.force_stop = threading.Event() # it show the running status
|
2018-09-03 10:33:05 +00:00
|
|
|
|
|
|
|
reset_dut(self.dut) # reset the board to make it start from begining
|
|
|
|
|
2018-01-31 10:59:10 +00:00
|
|
|
threading.Thread.__init__(self, name="{} Handler".format(dut))
|
|
|
|
|
|
|
|
def run(self):
|
|
|
|
def get_child_case_name(data):
|
|
|
|
self.child_case_name = data[0]
|
|
|
|
time.sleep(1)
|
|
|
|
self.dut.write(str(self.child_case_index))
|
|
|
|
|
|
|
|
def one_device_case_finish(result):
|
|
|
|
""" one test finished, let expect loop break and log result """
|
|
|
|
self.finish = True
|
|
|
|
self.result = result
|
|
|
|
if not result:
|
|
|
|
self.fail_name = self.child_case_name
|
|
|
|
|
|
|
|
def device_wait_action(data):
|
|
|
|
start_time = time.time()
|
|
|
|
expected_signal = data[0]
|
2018-08-20 11:21:10 +00:00
|
|
|
while 1:
|
2018-01-31 10:59:10 +00:00
|
|
|
if time.time() > start_time + self.timeout:
|
|
|
|
Utility.console_log("Timeout in device for function: %s"%self.child_case_name, color="orange")
|
|
|
|
break
|
|
|
|
with self.lock:
|
|
|
|
if expected_signal in self.sent_signal_list:
|
|
|
|
self.dut.write(" ")
|
|
|
|
self.sent_signal_list.remove(expected_signal)
|
|
|
|
break
|
|
|
|
time.sleep(0.01)
|
|
|
|
|
|
|
|
def device_send_action(data):
|
|
|
|
with self.lock:
|
|
|
|
self.sent_signal_list.append(data[0].encode('utf-8'))
|
|
|
|
|
|
|
|
def handle_device_test_finish(data):
|
|
|
|
""" test finished without reset """
|
|
|
|
# in this scenario reset should not happen
|
|
|
|
if int(data[1]):
|
|
|
|
# case ignored
|
|
|
|
Utility.console_log("Ignored: " + self.child_case_name, color="orange")
|
|
|
|
one_device_case_finish(not int(data[0]))
|
|
|
|
|
2018-09-03 10:33:05 +00:00
|
|
|
|
2018-06-26 01:45:09 +00:00
|
|
|
try:
|
|
|
|
time.sleep(1)
|
|
|
|
self.dut.write("\"{}\"".format(self.parent_case_name))
|
|
|
|
self.dut.expect("Running " + self.parent_case_name + "...")
|
|
|
|
except ExpectTimeout:
|
|
|
|
Utility.console_log("No case detected!", color="orange")
|
2018-08-20 11:21:10 +00:00
|
|
|
while not self.finish and not self.force_stop.isSet():
|
2018-01-31 10:59:10 +00:00
|
|
|
try:
|
|
|
|
self.dut.expect_any((re.compile('\(' + str(self.child_case_index) + '\)\s"(\w+)"'), get_child_case_name),
|
|
|
|
(self.WAIT_SIGNAL_PATTERN, device_wait_action), # wait signal pattern
|
|
|
|
(self.SEND_SIGNAL_PATTERN, device_send_action), # send signal pattern
|
|
|
|
(self.FINISH_PATTERN, handle_device_test_finish), # test finish pattern
|
2018-05-21 04:50:27 +00:00
|
|
|
timeout=self.timeout)
|
2018-01-31 10:59:10 +00:00
|
|
|
except ExpectTimeout:
|
|
|
|
Utility.console_log("Timeout in expect", color="orange")
|
|
|
|
one_device_case_finish(False)
|
|
|
|
break
|
|
|
|
|
2018-08-20 11:21:10 +00:00
|
|
|
def stop(self):
|
|
|
|
self.force_stop.set()
|
|
|
|
|
2018-01-31 10:59:10 +00:00
|
|
|
|
|
|
|
def get_case_info(one_case):
|
|
|
|
parent_case = one_case["name"]
|
|
|
|
child_case_num = one_case["child case num"]
|
|
|
|
return parent_case, child_case_num
|
|
|
|
|
|
|
|
|
2018-06-21 11:32:29 +00:00
|
|
|
def get_dut(duts, env, name, ut_config, app_bin=None):
|
2018-01-31 10:59:10 +00:00
|
|
|
if name in duts:
|
|
|
|
dut = duts[name]
|
|
|
|
else:
|
|
|
|
dut = env.get_dut(name, app_path=ut_config)
|
|
|
|
duts[name] = dut
|
2018-06-21 11:32:29 +00:00
|
|
|
replace_app_bin(dut, "unit-test-app", app_bin)
|
2018-09-03 10:33:05 +00:00
|
|
|
dut.start_app() # download bin to board
|
2018-01-31 10:59:10 +00:00
|
|
|
return dut
|
|
|
|
|
|
|
|
|
2018-06-21 11:32:29 +00:00
|
|
|
def case_run(duts, ut_config, env, one_case, failed_cases, app_bin):
|
2018-01-31 10:59:10 +00:00
|
|
|
lock = threading.RLock()
|
|
|
|
threads = []
|
|
|
|
send_signal_list = []
|
|
|
|
result = True
|
|
|
|
parent_case, case_num = get_case_info(one_case)
|
2018-06-26 01:45:09 +00:00
|
|
|
|
2018-01-31 10:59:10 +00:00
|
|
|
for i in range(case_num):
|
2018-06-21 11:32:29 +00:00
|
|
|
dut = get_dut(duts, env, "dut%d" % i, ut_config, app_bin)
|
2018-01-31 10:59:10 +00:00
|
|
|
threads.append(Handler(dut, send_signal_list, lock,
|
2018-05-21 04:50:27 +00:00
|
|
|
parent_case, i, one_case["timeout"]))
|
2018-01-31 10:59:10 +00:00
|
|
|
for thread in threads:
|
|
|
|
thread.setDaemon(True)
|
|
|
|
thread.start()
|
|
|
|
for thread in threads:
|
|
|
|
thread.join()
|
|
|
|
result = result and thread.result
|
|
|
|
if not thread.result:
|
2018-08-20 11:21:10 +00:00
|
|
|
[thd.stop() for thd in threads]
|
2018-06-26 01:45:09 +00:00
|
|
|
|
2018-01-31 10:59:10 +00:00
|
|
|
if result:
|
|
|
|
Utility.console_log("Success: " + one_case["name"], color="green")
|
|
|
|
else:
|
|
|
|
failed_cases.append(one_case["name"])
|
|
|
|
Utility.console_log("Failed: " + one_case["name"], color="red")
|
|
|
|
|
|
|
|
|
2018-07-26 07:07:32 +00:00
|
|
|
@IDF.idf_unit_test(env_tag="UT_T2_1")
|
2018-03-09 06:50:34 +00:00
|
|
|
def run_multiple_devices_cases(env, extra_data):
|
2018-01-31 10:59:10 +00:00
|
|
|
"""
|
|
|
|
extra_data can be two types of value
|
|
|
|
1. as dict:
|
|
|
|
e.g.
|
|
|
|
{"name": "gpio master/slave test example",
|
|
|
|
"child case num": 2,
|
|
|
|
"config": "release",
|
|
|
|
"env_tag": "UT_T2_1"}
|
|
|
|
2. as list dict:
|
|
|
|
e.g.
|
|
|
|
[{"name": "gpio master/slave test example1",
|
|
|
|
"child case num": 2,
|
|
|
|
"config": "release",
|
|
|
|
"env_tag": "UT_T2_1"},
|
|
|
|
{"name": "gpio master/slave test example2",
|
|
|
|
"child case num": 2,
|
|
|
|
"config": "release",
|
|
|
|
"env_tag": "UT_T2_1"}]
|
|
|
|
|
|
|
|
"""
|
|
|
|
failed_cases = []
|
|
|
|
case_config = format_test_case_config(extra_data)
|
|
|
|
DUTS = {}
|
|
|
|
for ut_config in case_config:
|
2018-05-15 03:51:56 +00:00
|
|
|
Utility.console_log("Running unit test for config: " + ut_config, "O")
|
2018-01-31 10:59:10 +00:00
|
|
|
for one_case in case_config[ut_config]:
|
2018-06-21 11:32:29 +00:00
|
|
|
case_run(DUTS, ut_config, env, one_case, failed_cases, one_case.get('app_bin'))
|
2018-01-31 10:59:10 +00:00
|
|
|
|
|
|
|
if failed_cases:
|
|
|
|
Utility.console_log("Failed Cases:", color="red")
|
|
|
|
for _case_name in failed_cases:
|
|
|
|
Utility.console_log("\t" + _case_name, color="red")
|
|
|
|
raise AssertionError("Unit Test Failed")
|
|
|
|
|
|
|
|
|
2018-07-26 07:07:32 +00:00
|
|
|
@IDF.idf_unit_test(env_tag="UT_T1_1")
|
2018-03-09 06:50:34 +00:00
|
|
|
def run_multiple_stage_cases(env, extra_data):
|
|
|
|
"""
|
|
|
|
extra_data can be 2 types of value
|
|
|
|
1. as dict: Mandantory keys: "name" and "child case num", optional keys: "reset" and others
|
|
|
|
3. as list of string or dict:
|
|
|
|
[case1, case2, case3, {"name": "restart from PRO CPU", "child case num": 2}, ...]
|
|
|
|
|
|
|
|
:param extra_data: the case name or case list or case dictionary
|
|
|
|
:return: None
|
|
|
|
"""
|
|
|
|
|
|
|
|
case_config = format_test_case_config(extra_data)
|
|
|
|
|
|
|
|
# we don't want stop on failed case (unless some special scenarios we can't handle)
|
|
|
|
# this flag is used to log if any of the case failed during executing
|
|
|
|
# Before exit test function this flag is used to log if the case fails
|
|
|
|
failed_cases = []
|
|
|
|
|
|
|
|
for ut_config in case_config:
|
2018-05-15 03:51:56 +00:00
|
|
|
Utility.console_log("Running unit test for config: " + ut_config, "O")
|
2018-03-09 06:50:34 +00:00
|
|
|
dut = env.get_dut("unit-test-app", app_path=ut_config)
|
2018-06-21 11:32:29 +00:00
|
|
|
if len(case_config[ut_config]) > 0:
|
|
|
|
replace_app_bin(dut, "unit-test-app", case_config[ut_config][0].get('app_bin'))
|
2018-03-09 06:50:34 +00:00
|
|
|
dut.start_app()
|
|
|
|
|
|
|
|
for one_case in case_config[ut_config]:
|
2018-09-03 10:33:05 +00:00
|
|
|
reset_dut(dut)
|
2018-03-09 06:50:34 +00:00
|
|
|
exception_reset_list = []
|
2018-01-31 10:59:10 +00:00
|
|
|
|
2018-03-09 06:50:34 +00:00
|
|
|
for test_stage in range(one_case["child case num"]):
|
|
|
|
# select multi stage test case name
|
|
|
|
dut.write("\"{}\"".format(one_case["name"]))
|
|
|
|
dut.expect("Running " + one_case["name"] + "...")
|
|
|
|
# select test function for current stage
|
|
|
|
dut.write(str(test_stage + 1))
|
|
|
|
|
|
|
|
# we want to set this flag in callbacks (inner functions)
|
|
|
|
# use list here so we can use append to set this flag
|
|
|
|
stage_finish = list()
|
|
|
|
|
|
|
|
def last_stage():
|
|
|
|
return test_stage == one_case["child case num"] - 1
|
|
|
|
|
|
|
|
def check_reset():
|
|
|
|
if one_case["reset"]:
|
|
|
|
assert exception_reset_list # reboot but no exception/reset logged. should never happen
|
|
|
|
result = False
|
|
|
|
if len(one_case["reset"]) == len(exception_reset_list):
|
|
|
|
for i, exception in enumerate(exception_reset_list):
|
|
|
|
if one_case["reset"][i] not in exception:
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
result = True
|
|
|
|
if not result:
|
|
|
|
Utility.console_log("""Reset Check Failed: \r\n\tExpected: {}\r\n\tGet: {}"""
|
|
|
|
.format(one_case["reset"], exception_reset_list),
|
|
|
|
color="orange")
|
|
|
|
else:
|
|
|
|
# we allow omit reset in multi stage cases
|
|
|
|
result = True
|
|
|
|
return result
|
|
|
|
|
|
|
|
# expect callbacks
|
|
|
|
def one_case_finish(result):
|
|
|
|
""" one test finished, let expect loop break and log result """
|
|
|
|
# handle test finish
|
|
|
|
result = result and check_reset()
|
|
|
|
if result:
|
|
|
|
Utility.console_log("Success: " + one_case["name"], color="green")
|
|
|
|
else:
|
|
|
|
failed_cases.append(one_case["name"])
|
|
|
|
Utility.console_log("Failed: " + one_case["name"], color="red")
|
|
|
|
stage_finish.append("break")
|
|
|
|
|
|
|
|
def handle_exception_reset(data):
|
|
|
|
"""
|
|
|
|
just append data to exception list.
|
|
|
|
exception list will be checked in ``handle_reset_finish``, once reset finished.
|
|
|
|
"""
|
|
|
|
exception_reset_list.append(data[0])
|
|
|
|
|
|
|
|
def handle_test_finish(data):
|
|
|
|
""" test finished without reset """
|
|
|
|
# in this scenario reset should not happen
|
|
|
|
if int(data[1]):
|
|
|
|
# case ignored
|
|
|
|
Utility.console_log("Ignored: " + one_case["name"], color="orange")
|
|
|
|
# only passed in last stage will be regarded as real pass
|
|
|
|
if last_stage():
|
|
|
|
one_case_finish(not int(data[0]))
|
|
|
|
else:
|
|
|
|
Utility.console_log("test finished before enter last stage", color="orange")
|
|
|
|
one_case_finish(False)
|
|
|
|
|
|
|
|
def handle_next_stage(data):
|
|
|
|
""" reboot finished. we goto next stage """
|
|
|
|
if last_stage():
|
|
|
|
# already last stage, should never goto next stage
|
|
|
|
Utility.console_log("didn't finish at last stage", color="orange")
|
|
|
|
one_case_finish(False)
|
|
|
|
else:
|
|
|
|
stage_finish.append("continue")
|
|
|
|
|
|
|
|
while not stage_finish:
|
|
|
|
try:
|
|
|
|
dut.expect_any((RESET_PATTERN, handle_exception_reset),
|
|
|
|
(EXCEPTION_PATTERN, handle_exception_reset),
|
|
|
|
(ABORT_PATTERN, handle_exception_reset),
|
|
|
|
(FINISH_PATTERN, handle_test_finish),
|
|
|
|
(UT_APP_BOOT_UP_DONE, handle_next_stage),
|
2018-05-21 04:50:27 +00:00
|
|
|
timeout=one_case["timeout"])
|
2018-03-09 06:50:34 +00:00
|
|
|
except ExpectTimeout:
|
|
|
|
Utility.console_log("Timeout in expect", color="orange")
|
|
|
|
one_case_finish(False)
|
|
|
|
break
|
|
|
|
if stage_finish[0] == "break":
|
|
|
|
# test breaks on current stage
|
|
|
|
break
|
|
|
|
|
|
|
|
# raise exception if any case fails
|
|
|
|
if failed_cases:
|
|
|
|
Utility.console_log("Failed Cases:", color="red")
|
|
|
|
for _case_name in failed_cases:
|
|
|
|
Utility.console_log("\t" + _case_name, color="red")
|
|
|
|
raise AssertionError("Unit Test Failed")
|
|
|
|
|
2018-06-21 11:32:29 +00:00
|
|
|
def detect_update_unit_test_info(env, extra_data, app_bin):
|
|
|
|
|
|
|
|
case_config = format_test_case_config(extra_data)
|
|
|
|
|
|
|
|
for ut_config in case_config:
|
|
|
|
dut = env.get_dut("unit-test-app", app_path=ut_config)
|
|
|
|
replace_app_bin(dut, "unit-test-app", app_bin)
|
|
|
|
dut.start_app()
|
|
|
|
|
2018-09-03 10:33:05 +00:00
|
|
|
reset_dut(dut)
|
2018-06-21 11:32:29 +00:00
|
|
|
|
|
|
|
# get the list of test cases
|
|
|
|
dut.write("")
|
|
|
|
dut.expect("Here's the test menu, pick your combo:", timeout=DEFAULT_TIMEOUT)
|
|
|
|
|
|
|
|
def find_update_dic(name, t, timeout, child_case_num=None):
|
|
|
|
for dic in extra_data:
|
|
|
|
if dic['name'] == name:
|
|
|
|
dic['type'] = t
|
|
|
|
if 'timeout' not in dic:
|
|
|
|
dic['timeout'] = timeout
|
|
|
|
if child_case_num:
|
|
|
|
dic['child case num'] = child_case_num
|
|
|
|
|
|
|
|
try:
|
|
|
|
while True:
|
|
|
|
data = dut.expect(TEST_PATTERN, timeout=DEFAULT_TIMEOUT)
|
|
|
|
test_case_name = data[1]
|
|
|
|
m = re.search(r'\[timeout=(\d+)\]', data[2])
|
|
|
|
if m:
|
|
|
|
timeout = int(m.group(1))
|
|
|
|
else:
|
|
|
|
timeout = 30
|
|
|
|
m = re.search(r'\[multi_stage\]', data[2])
|
|
|
|
if m:
|
|
|
|
test_case_type = MULTI_STAGE_ID
|
|
|
|
else:
|
|
|
|
m = re.search(r'\[multi_device\]', data[2])
|
|
|
|
if m:
|
|
|
|
test_case_type = MULTI_DEVICE_ID
|
|
|
|
else:
|
|
|
|
test_case_type = SIMPLE_TEST_ID
|
|
|
|
find_update_dic(test_case_name, test_case_type, timeout)
|
|
|
|
if data[3] and re.search(END_LIST_STR, data[3]):
|
|
|
|
break
|
|
|
|
continue
|
|
|
|
# find the last submenu item
|
|
|
|
data = dut.expect(TEST_SUBMENU_PATTERN, timeout=DEFAULT_TIMEOUT)
|
|
|
|
find_update_dic(test_case_name, test_case_type, timeout, child_case_num=int(data[0]))
|
|
|
|
if data[1] and re.search(END_LIST_STR, data[1]):
|
|
|
|
break
|
|
|
|
# check if the unit test case names are correct, i.e. they could be found in the device
|
|
|
|
for dic in extra_data:
|
|
|
|
if 'type' not in dic:
|
|
|
|
raise ValueError("Unit test \"{}\" doesn't exist in the flashed device!".format(dic.get('name')))
|
|
|
|
except ExpectTimeout:
|
|
|
|
Utility.console_log("Timeout during getting the test list", color="red")
|
|
|
|
finally:
|
|
|
|
dut.close()
|
|
|
|
|
|
|
|
# These options are the same for all configs, therefore there is no need to continue
|
|
|
|
break
|
2018-03-09 06:50:34 +00:00
|
|
|
|
|
|
|
if __name__ == '__main__':
|
2018-06-21 11:32:29 +00:00
|
|
|
parser = argparse.ArgumentParser()
|
|
|
|
parser.add_argument(
|
|
|
|
'--repeat', '-r',
|
|
|
|
help='Number of repetitions for the test(s). Default is 1.',
|
|
|
|
type=int,
|
|
|
|
default=1
|
|
|
|
)
|
|
|
|
parser.add_argument("--env_config_file", "-e",
|
|
|
|
help="test env config file",
|
|
|
|
default=None
|
|
|
|
)
|
|
|
|
parser.add_argument("--app_bin", "-b",
|
|
|
|
help="application binary file for flashing the chip",
|
|
|
|
default=None
|
|
|
|
)
|
|
|
|
parser.add_argument(
|
|
|
|
'test',
|
|
|
|
help='Comma separated list of <option>:<argument> where option can be "name" (default), "child case num", \
|
|
|
|
"config", "timeout".',
|
|
|
|
nargs='+'
|
|
|
|
)
|
|
|
|
args = parser.parse_args()
|
|
|
|
list_of_dicts = []
|
|
|
|
for test in args.test:
|
|
|
|
test_args = test.split(r',')
|
|
|
|
test_dict = dict()
|
|
|
|
for test_item in test_args:
|
|
|
|
if len(test_item) == 0:
|
|
|
|
continue
|
|
|
|
pair = test_item.split(r':')
|
|
|
|
if len(pair) == 1 or pair[0] is 'name':
|
|
|
|
test_dict['name'] = pair[0]
|
|
|
|
elif len(pair) == 2:
|
|
|
|
if pair[0] == 'timeout' or pair[0] == 'child case num':
|
|
|
|
test_dict[pair[0]] = int(pair[1])
|
|
|
|
else:
|
|
|
|
test_dict[pair[0]] = pair[1]
|
|
|
|
else:
|
|
|
|
raise ValueError('Error in argument item {} of {}'.format(test_item, test))
|
|
|
|
test_dict['app_bin'] = args.app_bin
|
|
|
|
list_of_dicts.append(test_dict)
|
|
|
|
|
|
|
|
TinyFW.set_default_config(env_config_file=args.env_config_file)
|
|
|
|
|
|
|
|
env_config = TinyFW.get_default_config()
|
|
|
|
env_config['app'] = UT
|
|
|
|
env_config['dut'] = IDF.IDFDUT
|
|
|
|
env_config['test_suite_name'] = 'unit_test_parsing'
|
|
|
|
env = Env.Env(**env_config)
|
|
|
|
detect_update_unit_test_info(env, extra_data=list_of_dicts, app_bin=args.app_bin)
|
|
|
|
|
|
|
|
for i in range(1, args.repeat+1):
|
|
|
|
if args.repeat > 1:
|
|
|
|
Utility.console_log("Repetition {}".format(i), color="green")
|
|
|
|
for dic in list_of_dicts:
|
|
|
|
t = dic.get('type', SIMPLE_TEST_ID)
|
|
|
|
if t == SIMPLE_TEST_ID:
|
|
|
|
run_unit_test_cases(extra_data=dic)
|
|
|
|
elif t == MULTI_STAGE_ID:
|
|
|
|
run_multiple_stage_cases(extra_data=dic)
|
|
|
|
elif t == MULTI_DEVICE_ID:
|
|
|
|
run_multiple_devices_cases(extra_data=dic)
|
|
|
|
else:
|
|
|
|
raise ValueError('Unknown type {} of {}'.format(t, dic.get('name')))
|