1
0
mirror of https://github.com/trezor/trezor-firmware.git synced 2024-11-18 13:38:12 +00:00
trezor-firmware/tests/ui_tests/__init__.py

226 lines
7.0 KiB
Python
Raw Normal View History

2019-12-30 11:20:51 +00:00
import hashlib
import json
import os
2019-12-30 11:20:51 +00:00
import re
import shutil
from contextlib import contextmanager
from pathlib import Path
from typing import Dict, Generator, Optional, Set
2019-12-30 11:20:51 +00:00
import pytest
from _pytest.outcomes import Failed
from PIL import Image
2019-12-30 11:20:51 +00:00
from trezorlib.debuglink import TrezorClientDebugLink as Client
from .reporting import testreport
2019-12-30 11:20:51 +00:00
UI_TESTS_DIR = Path(__file__).resolve().parent
SCREENS_DIR = UI_TESTS_DIR / "screens"
HASH_FILE = UI_TESTS_DIR / "fixtures.json"
SUGGESTION_FILE = UI_TESTS_DIR / "fixtures.suggestion.json"
FILE_HASHES: Dict[str, str] = {}
ACTUAL_HASHES: Dict[str, str] = {}
PROCESSED: Set[str] = set()
FAILED_TESTS: Set[str] = set()
2022-02-08 15:37:34 +00:00
# T1/TT, to be set in screen_recording(), as we do not know it beforehand
# TODO: it is not the cleanest, we could create a class out of this file
MODEL = ""
2019-12-30 11:20:51 +00:00
def get_test_name(node_id: str) -> str:
2019-12-30 11:20:51 +00:00
# Test item name is usually function name, but when parametrization is used,
# parameters are also part of the name. Some functions have very long parameter
# names (tx hashes etc) that run out of maximum allowable filename length, so
# we limit the name to first 100 chars. This is not a problem with txhashes.
2020-01-09 11:29:45 +00:00
new_name = node_id.replace("tests/device_tests/", "")
# remove ::TestClass:: if present because it is usually the same as the test file name
new_name = re.sub(r"::.*?::", "-", new_name)
new_name = new_name.replace("/", "-") # in case there is "/"
if len(new_name) <= 100:
return new_name
return new_name[:91] + "-" + hashlib.sha256(new_name.encode()).hexdigest()[:8]
2019-12-30 11:20:51 +00:00
def _process_recorded(screen_path: Path, test_name: str) -> None:
# calculate hash
actual_hash = _hash_files(screen_path)
FILE_HASHES[test_name] = actual_hash
ACTUAL_HASHES[test_name] = actual_hash
2019-12-30 11:20:51 +00:00
_rename_records(screen_path)
testreport.recorded(screen_path, test_name, actual_hash)
2019-12-30 11:20:51 +00:00
def _rename_records(screen_path: Path) -> None:
2019-12-30 11:20:51 +00:00
# rename screenshots
for index, record in enumerate(sorted(screen_path.iterdir())):
2020-01-06 14:44:30 +00:00
record.replace(screen_path / f"{index:08}.png")
2019-12-30 11:20:51 +00:00
def _hash_files(path: Path) -> str:
2020-01-06 14:44:30 +00:00
files = path.iterdir()
2019-12-30 11:20:51 +00:00
hasher = hashlib.sha256()
for file in sorted(files):
hasher.update(_get_bytes_from_png(str(file)))
2019-12-30 11:20:51 +00:00
return hasher.digest().hex()
def _get_bytes_from_png(png_file: str) -> bytes:
"""Decode a PNG file into bytes representing all the pixels.
Is necessary because Linux and Mac are using different PNG encoding libraries,
and we need the file hashes to be the same on both platforms.
"""
return Image.open(png_file).tobytes()
def _process_tested(fixture_test_path: Path, test_name: str) -> None:
2019-12-30 11:20:51 +00:00
actual_path = fixture_test_path / "actual"
2020-01-06 14:44:30 +00:00
actual_hash = _hash_files(actual_path)
ACTUAL_HASHES[test_name] = actual_hash
2019-12-30 11:20:51 +00:00
2020-01-06 14:44:30 +00:00
_rename_records(actual_path)
2019-12-30 11:20:51 +00:00
expected_hash = FILE_HASHES.get(test_name)
if expected_hash is None:
pytest.fail(f"Hash of {test_name} not found in fixtures.json")
2019-12-30 11:20:51 +00:00
if actual_hash != expected_hash:
assert expected_hash is not None
file_path = testreport.failed(
fixture_test_path, test_name, actual_hash, expected_hash
)
2019-12-30 11:20:51 +00:00
pytest.fail(
f"Hash of {test_name} differs.\n"
f"Expected: {expected_hash}\n"
f"Actual: {actual_hash}\n"
f"Diff file: {file_path}"
2019-12-30 11:20:51 +00:00
)
else:
testreport.passed(fixture_test_path, test_name, actual_hash)
2019-12-30 11:20:51 +00:00
def get_last_call_test_result(request: pytest.FixtureRequest) -> Optional[bool]:
# if test did not finish, e.g. interrupted by Ctrl+C, the pytest_runtest_makereport
# did not create the attribute we need
if not hasattr(request.node, "rep_call"):
return None
return request.node.rep_call.passed
2019-12-30 11:20:51 +00:00
@contextmanager
def screen_recording(
client: Client, request: pytest.FixtureRequest
) -> Generator[None, None, None]:
2020-01-07 09:13:08 +00:00
test_ui = request.config.getoption("ui")
2020-01-09 11:29:45 +00:00
test_name = get_test_name(request.node.nodeid)
2022-02-08 15:37:34 +00:00
# Differentiating test names between T1 and TT
# Making the model global for other functions
global MODEL
MODEL = f"T{client.features.model}"
if os.getenv("UI2") == "1":
MODEL += "ui2"
2022-02-08 15:37:34 +00:00
test_name = f"{MODEL}_{test_name}"
screens_test_path = SCREENS_DIR / test_name
2019-12-30 11:20:51 +00:00
2020-01-07 09:13:08 +00:00
if test_ui == "record":
screen_path = screens_test_path / "recorded"
2019-12-30 11:20:51 +00:00
else:
screen_path = screens_test_path / "actual"
2019-12-30 11:20:51 +00:00
if not screens_test_path.exists():
screens_test_path.mkdir()
# remove previous files
shutil.rmtree(screen_path, ignore_errors=True)
screen_path.mkdir()
2019-12-30 11:20:51 +00:00
try:
client.debug.start_recording(str(screen_path))
yield
finally:
# Wait for response to Initialize, which gives the emulator time to catch up
# and redraw the homescreen. Otherwise there's a race condition between that
# and stopping recording.
client.init_device()
client.debug.stop_recording()
if test_ui:
PROCESSED.add(test_name)
if get_last_call_test_result(request) is False:
FAILED_TESTS.add(test_name)
if test_ui == "record":
_process_recorded(screen_path, test_name)
else:
_process_tested(screens_test_path, test_name)
def list_missing() -> Set[str]:
2022-02-08 15:37:34 +00:00
# Only listing the ones for the current model
relevant_cases = {
case for case in FILE_HASHES.keys() if case.startswith(f"{MODEL}_")
}
return relevant_cases - PROCESSED
def read_fixtures() -> None:
if not HASH_FILE.exists():
raise ValueError("File fixtures.json not found.")
global FILE_HASHES
FILE_HASHES = json.loads(HASH_FILE.read_text())
def write_fixtures(remove_missing: bool) -> None:
HASH_FILE.write_text(_get_fixtures_content(FILE_HASHES, remove_missing))
def write_fixtures_suggestion(
remove_missing: bool, only_passed_tests: bool = False
) -> None:
SUGGESTION_FILE.write_text(
_get_fixtures_content(ACTUAL_HASHES, remove_missing, only_passed_tests)
)
def _get_fixtures_content(
fixtures: Dict[str, str], remove_missing: bool, only_passed_tests: bool = False
) -> str:
if remove_missing:
2022-02-08 15:37:34 +00:00
# Not removing the ones for different model
nonrelevant_cases = {
f: h for f, h in FILE_HASHES.items() if not f.startswith(f"{MODEL}_")
}
filtered_processed_tests = PROCESSED
if only_passed_tests:
filtered_processed_tests = PROCESSED - FAILED_TESTS
processed_fixtures = {i: fixtures[i] for i in filtered_processed_tests}
2022-02-08 15:37:34 +00:00
fixtures = {**nonrelevant_cases, **processed_fixtures}
else:
fixtures = fixtures
return json.dumps(fixtures, indent="", sort_keys=True) + "\n"
def main() -> None:
read_fixtures()
for record in SCREENS_DIR.iterdir():
if not (record / "actual").exists():
continue
try:
_process_tested(record, record.name)
print("PASSED:", record.name)
except Failed:
print("FAILED:", record.name)
testreport.index()