mirror of
https://github.com/trezor/trezor-firmware.git
synced 2024-11-22 23:48:12 +00:00
271 lines
9.2 KiB
Python
271 lines
9.2 KiB
Python
# This file is part of the Trezor project.
|
|
#
|
|
# Copyright (C) 2012-2019 SatoshiLabs and contributors
|
|
#
|
|
# This library is free software: you can redistribute it and/or modify
|
|
# it under the terms of the GNU Lesser General Public License version 3
|
|
# as published by the Free Software Foundation.
|
|
#
|
|
# This library is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU Lesser General Public License for more details.
|
|
#
|
|
# You should have received a copy of the License along with this library.
|
|
# If not, see <https://www.gnu.org/licenses/lgpl-3.0.html>.
|
|
|
|
import os
|
|
|
|
import pytest
|
|
|
|
from trezorlib import debuglink, log
|
|
from trezorlib.debuglink import TrezorClientDebugLink
|
|
from trezorlib.device import wipe as wipe_device
|
|
from trezorlib.transport import enumerate_devices, get_transport
|
|
|
|
from . import ui_tests
|
|
from .device_handler import BackgroundDeviceHandler
|
|
from .ui_tests.reporting import testreport
|
|
|
|
|
|
def get_device():
|
|
path = os.environ.get("TREZOR_PATH")
|
|
interact = int(os.environ.get("INTERACT", 0))
|
|
if path:
|
|
try:
|
|
transport = get_transport(path)
|
|
return TrezorClientDebugLink(transport, auto_interact=not interact)
|
|
except Exception as e:
|
|
raise RuntimeError("Failed to open debuglink for {}".format(path)) from e
|
|
|
|
else:
|
|
devices = enumerate_devices()
|
|
for device in devices:
|
|
try:
|
|
return TrezorClientDebugLink(device, auto_interact=not interact)
|
|
except Exception:
|
|
pass
|
|
else:
|
|
raise RuntimeError("No debuggable device found")
|
|
|
|
|
|
@pytest.fixture(scope="function")
|
|
def client(request):
|
|
"""Client fixture.
|
|
|
|
Every test function that requires a client instance will get it from here.
|
|
If we can't connect to a debuggable device, the test will fail.
|
|
If 'skip_t2' is used and TT is connected, the test is skipped. Vice versa with T1
|
|
and 'skip_t1'.
|
|
|
|
The client instance is wiped and preconfigured with "all all all..." mnemonic, no
|
|
password and no pin. It is possible to customize this with the `setup_client`
|
|
marker.
|
|
|
|
To specify a custom mnemonic and/or custom pin and/or enable passphrase:
|
|
|
|
@pytest.mark.setup_client(mnemonic=MY_MNEMONIC, pin="9999", passphrase=True)
|
|
|
|
To receive a client instance that was not initialized:
|
|
|
|
@pytest.mark.setup_client(uninitialized=True)
|
|
"""
|
|
try:
|
|
client = get_device()
|
|
except RuntimeError:
|
|
request.session.shouldstop = "No debuggable Trezor is available"
|
|
pytest.fail("No debuggable Trezor is available")
|
|
|
|
if request.node.get_closest_marker("skip_t2") and client.features.model == "T":
|
|
pytest.skip("Test excluded on Trezor T")
|
|
if request.node.get_closest_marker("skip_t1") and client.features.model == "1":
|
|
pytest.skip("Test excluded on Trezor 1")
|
|
|
|
sd_marker = request.node.get_closest_marker("sd_card")
|
|
if sd_marker and not client.features.sd_card_present:
|
|
raise RuntimeError(
|
|
"This test requires SD card.\n"
|
|
"To skip all such tests, run:\n"
|
|
" pytest -m 'not sd_card' <test path>"
|
|
)
|
|
|
|
test_ui = request.config.getoption("ui")
|
|
run_ui_tests = not request.node.get_closest_marker("skip_ui") and test_ui
|
|
|
|
client.open()
|
|
if run_ui_tests:
|
|
# we need to reseed before the wipe
|
|
client.debug.reseed(0)
|
|
|
|
if sd_marker:
|
|
should_format = sd_marker.kwargs.get("formatted", True)
|
|
client.debug.erase_sd_card(format=should_format)
|
|
|
|
wipe_device(client)
|
|
|
|
setup_params = dict(
|
|
uninitialized=False,
|
|
mnemonic=" ".join(["all"] * 12),
|
|
pin=None,
|
|
passphrase=False,
|
|
needs_backup=False,
|
|
no_backup=False,
|
|
)
|
|
|
|
marker = request.node.get_closest_marker("setup_client")
|
|
if marker:
|
|
setup_params.update(marker.kwargs)
|
|
|
|
if not setup_params["uninitialized"]:
|
|
debuglink.load_device(
|
|
client,
|
|
mnemonic=setup_params["mnemonic"],
|
|
pin=setup_params["pin"],
|
|
passphrase_protection=setup_params["passphrase"],
|
|
label="test",
|
|
language="en-US",
|
|
needs_backup=setup_params["needs_backup"],
|
|
no_backup=setup_params["no_backup"],
|
|
)
|
|
|
|
if setup_params["pin"]:
|
|
# ClearSession locks the device. We only do that if the PIN is set.
|
|
client.clear_session()
|
|
|
|
if run_ui_tests:
|
|
with ui_tests.screen_recording(client, request):
|
|
yield client
|
|
else:
|
|
yield client
|
|
|
|
client.close()
|
|
|
|
|
|
def pytest_sessionstart(session):
|
|
ui_tests.read_fixtures()
|
|
if session.config.getoption("ui") == "test":
|
|
testreport.clear_dir()
|
|
|
|
|
|
def _should_write_ui_report(exitstatus):
|
|
# generate UI report and check missing only if pytest is exitting cleanly
|
|
# I.e., the test suite passed or failed (as opposed to ctrl+c break, internal error,
|
|
# etc.)
|
|
return exitstatus in (pytest.ExitCode.OK, pytest.ExitCode.TESTS_FAILED)
|
|
|
|
|
|
def pytest_sessionfinish(session, exitstatus):
|
|
if not _should_write_ui_report(exitstatus):
|
|
return
|
|
|
|
if session.config.getoption("ui") == "test":
|
|
if session.config.getoption("ui_check_missing") and ui_tests.list_missing():
|
|
session.exitstatus = pytest.ExitCode.TESTS_FAILED
|
|
testreport.index()
|
|
if session.config.getoption("ui") == "record":
|
|
ui_tests.write_fixtures(session.config.getoption("ui_check_missing"))
|
|
|
|
|
|
def pytest_terminal_summary(terminalreporter, exitstatus, config):
|
|
println = terminalreporter.write_line
|
|
println("")
|
|
|
|
ui_option = config.getoption("ui")
|
|
missing_tests = ui_tests.list_missing()
|
|
if ui_option and _should_write_ui_report(exitstatus) and missing_tests:
|
|
println(f"{len(missing_tests)} expected UI tests did not run.")
|
|
if config.getoption("ui_check_missing"):
|
|
println("List of missing tests follows:")
|
|
for test in missing_tests:
|
|
println("\t" + test)
|
|
|
|
if ui_option == "test":
|
|
println("UI test failed.")
|
|
elif ui_option == "record":
|
|
println("Removing missing tests from record.")
|
|
println("")
|
|
|
|
if _should_write_ui_report(exitstatus):
|
|
println(f"UI tests summary: {testreport.REPORTS_PATH / 'index.html'}")
|
|
|
|
|
|
def pytest_addoption(parser):
|
|
parser.addoption(
|
|
"--ui",
|
|
action="store",
|
|
choices=["test", "record"],
|
|
help="Enable UI intergration tests: 'record' or 'test'",
|
|
)
|
|
parser.addoption(
|
|
"--ui-check-missing",
|
|
action="store_true",
|
|
default=False,
|
|
help="Check UI fixtures are containing the appropriate test cases (fails on `test`,"
|
|
"deletes old ones on `record`).",
|
|
)
|
|
|
|
|
|
def pytest_configure(config):
|
|
"""Called at testsuite setup time.
|
|
|
|
Registers known markers, enables verbose output if requested.
|
|
"""
|
|
# register known markers
|
|
config.addinivalue_line("markers", "skip_t1: skip the test on Trezor One")
|
|
config.addinivalue_line("markers", "skip_t2: skip the test on Trezor T")
|
|
config.addinivalue_line(
|
|
"markers",
|
|
'setup_client(mnemonic="all all all...", pin=None, passphrase=False, uninitialized=False): configure the client instance',
|
|
)
|
|
config.addinivalue_line(
|
|
"markers", "skip_ui: skip UI integration checks for this test"
|
|
)
|
|
with open(os.path.join(os.path.dirname(__file__), "REGISTERED_MARKERS")) as f:
|
|
for line in f:
|
|
config.addinivalue_line("markers", line.strip())
|
|
|
|
# enable debug
|
|
if config.getoption("verbose"):
|
|
log.enable_debug_output()
|
|
|
|
|
|
def pytest_runtest_setup(item):
|
|
"""Called for each test item (class, individual tests).
|
|
|
|
Ensures that altcoin tests are skipped, and that no test is skipped on
|
|
both T1 and TT.
|
|
"""
|
|
if item.get_closest_marker("skip_t1") and item.get_closest_marker("skip_t2"):
|
|
raise RuntimeError("Don't skip tests for both trezors!")
|
|
|
|
skip_altcoins = int(os.environ.get("TREZOR_PYTEST_SKIP_ALTCOINS", 0))
|
|
if item.get_closest_marker("altcoin") and skip_altcoins:
|
|
pytest.skip("Skipping altcoin test")
|
|
|
|
|
|
@pytest.hookimpl(tryfirst=True, hookwrapper=True)
|
|
def pytest_runtest_makereport(item, call):
|
|
# Make test results available in fixtures.
|
|
# See https://docs.pytest.org/en/latest/example/simple.html#making-test-result-information-available-in-fixtures
|
|
# The device_handler fixture uses this as 'request.node.rep_call.passed' attribute,
|
|
# in order to raise error only if the test passed.
|
|
outcome = yield
|
|
rep = outcome.get_result()
|
|
setattr(item, f"rep_{rep.when}", rep)
|
|
|
|
|
|
@pytest.fixture
|
|
def device_handler(client, request):
|
|
device_handler = BackgroundDeviceHandler(client)
|
|
yield device_handler
|
|
|
|
# if test did not finish, e.g. interrupted by Ctrl+C, the pytest_runtest_makereport
|
|
# did not create the attribute we need
|
|
if not hasattr(request.node, "rep_call"):
|
|
return
|
|
|
|
# if test finished, make sure all background tasks are done
|
|
finalized_ok = device_handler.check_finalize()
|
|
if request.node.rep_call.passed and not finalized_ok:
|
|
raise RuntimeError("Test did not check result of background task")
|