Skip to content

conftest.py is updated [refactoring] #226

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
146 changes: 98 additions & 48 deletions tests/conftest.py
Original file line number Diff line number Diff line change
@@ -1,17 +1,18 @@
# /////////////////////////////////////////////////////////////////////////////
# PyTest Configuration

import _pytest.outcomes

import pluggy
import pytest
import _pytest
import os
import logging
import pathlib
import math
import datetime

import _pytest.outcomes
import _pytest.unittest
import _pytest.logging

# /////////////////////////////////////////////////////////////////////////////

C_ROOT_DIR__RELATIVE = ".."
Expand Down Expand Up @@ -91,10 +92,6 @@ def GetCurrentTestWorkerSignature() -> str:
return __class__.sm_CurrentTestWorkerSignature


# /////////////////////////////////////////////////////////////////////////////# /////////////////////////////////////////////////////////////////////////////
# Fixtures


# /////////////////////////////////////////////////////////////////////////////
# TEST_PROCESS_STATS

Expand All @@ -109,10 +106,12 @@ class TEST_PROCESS_STATS:
cSkippedTests: int = 0
cNotXFailedTests: int = 0
cUnexpectedTests: int = 0
cAchtungTests: int = 0

FailedTests = list[str]()
XFailedTests = list[str]()
NotXFailedTests = list[str]()
AchtungTests = list[str]()

# --------------------------------------------------------------------
def incrementTotalTestCount() -> None:
Expand Down Expand Up @@ -163,6 +162,14 @@ def incrementNotXFailedTests(testID: str) -> None:
def incrementUnexpectedTests() -> None:
__class__.cUnexpectedTests += 1

# --------------------------------------------------------------------
def incrementAchtungTestCount(testID: str) -> None:
assert type(testID) == str # noqa: E721
assert type(__class__.AchtungTests) == list # noqa: E721

__class__.AchtungTests.append(testID) # raise?
__class__.cAchtungTests += 1


# /////////////////////////////////////////////////////////////////////////////

Expand Down Expand Up @@ -198,10 +205,13 @@ def helper__makereport__setup(
assert item is not None
assert call is not None
assert outcome is not None
assert type(item) == pytest.Function # noqa: E721
# it may be pytest.Function or _pytest.unittest.TestCaseFunction
assert isinstance(item, pytest.Function)
assert type(call) == pytest.CallInfo # noqa: E721
assert type(outcome) == pluggy.Result # noqa: E721

C_LINE1 = "******************************************************"

# logging.info("pytest_runtest_makereport - setup")

TEST_PROCESS_STATS.incrementTotalTestCount()
Expand All @@ -214,26 +224,42 @@ def helper__makereport__setup(
TEST_PROCESS_STATS.incrementNotExecutedTestCount()
return

assert rep.outcome == "passed"

testNumber = TEST_PROCESS_STATS.incrementExecutedTestCount()

testID = ""

if item.cls is not None:
testID = item.cls.__module__ + "." + item.cls.__name__ + "::"

testID = testID + item.name

if testNumber > 1:
logging.info("")
if rep.outcome == "passed":
testNumber = TEST_PROCESS_STATS.incrementExecutedTestCount()

logging.info(C_LINE1)
logging.info("* START TEST {0}".format(testID))
logging.info("*")
logging.info("* Path : {0}".format(item.path))
logging.info("* Number: {0}".format(testNumber))
logging.info("*")
return

assert rep.outcome != "passed"

TEST_PROCESS_STATS.incrementAchtungTestCount(testID)

logging.info("******************************************************")
logging.info("* START TEST {0}".format(testID))
logging.info(C_LINE1)
logging.info("* ACHTUNG TEST {0}".format(testID))
logging.info("*")
logging.info("* Path : {0}".format(item.path))
logging.info("* Number: {0}".format(testNumber))
logging.info("* Outcome is [{0}]".format(rep.outcome))

if rep.outcome == "failed":
assert call.excinfo is not None
assert call.excinfo.value is not None
logging.info("*")
logging.error(call.excinfo.value)

logging.info("*")
return


# ------------------------------------------------------------------------
Expand All @@ -243,12 +269,11 @@ def helper__makereport__call(
assert item is not None
assert call is not None
assert outcome is not None
assert type(item) == pytest.Function # noqa: E721
# it may be pytest.Function or _pytest.unittest.TestCaseFunction
assert isinstance(item, pytest.Function)
assert type(call) == pytest.CallInfo # noqa: E721
assert type(outcome) == pluggy.Result # noqa: E721

# logging.info("pytest_runtest_makereport - call")

rep = outcome.get_result()
assert rep is not None
assert type(rep) == pytest.TestReport # noqa: E721
Expand Down Expand Up @@ -341,7 +366,8 @@ def helper__makereport__call(
else:
TEST_PROCESS_STATS.incrementUnexpectedTests()
exitStatus = "UNEXPECTED [{0}]".format(rep.outcome)
assert False
# [2025-03-28] It may create a useless problem in new environment.
# assert False

# --------
logging.info("*")
Expand All @@ -350,6 +376,7 @@ def helper__makereport__call(
logging.info("* EXIT STATUS : {0}".format(exitStatus))
logging.info("*")
logging.info("* STOP TEST {0}".format(testID))
logging.info("*")


# /////////////////////////////////////////////////////////////////////////////
Expand All @@ -359,17 +386,14 @@ def helper__makereport__call(
def pytest_runtest_makereport(item: pytest.Function, call: pytest.CallInfo):
assert item is not None
assert call is not None
assert type(item) == pytest.Function # noqa: E721
# it may be pytest.Function or _pytest.unittest.TestCaseFunction
assert isinstance(item, pytest.Function)
assert type(call) == pytest.CallInfo # noqa: E721

# logging.info("[pytest_runtest_makereport][#001][{0}][{1}]".format(item.name, call.when))

outcome: pluggy.Result = yield
assert outcome is not None
assert type(outcome) == pluggy.Result # noqa: E721

# logging.info("[pytest_runtest_makereport][#002][{0}][{1}]".format(item.name, call.when))

rep: pytest.TestReport = outcome.get_result()
assert rep is not None
assert type(rep) == pytest.TestReport # noqa: E721
Expand Down Expand Up @@ -440,41 +464,61 @@ def run_after_tests(request: pytest.FixtureRequest):

yield

logging.info("--------------------------- [FAILED TESTS]")
logging.info("")

assert len(TEST_PROCESS_STATS.FailedTests) == TEST_PROCESS_STATS.cFailedTests

if len(TEST_PROCESS_STATS.FailedTests) > 0:
helper__print_test_list(TEST_PROCESS_STATS.FailedTests)
logging.info("")
C_LINE1 = "---------------------------"

logging.info("--------------------------- [XFAILED TESTS]")
logging.info("")
def LOCAL__print_line1_with_header(header: str):
assert type(C_LINE1) == str # noqa: E721
assert type(header) == str # noqa: E721
assert header != ""
logging.info(C_LINE1 + " [" + header + "]")

assert len(TEST_PROCESS_STATS.XFailedTests) == TEST_PROCESS_STATS.cXFailedTests
def LOCAL__print_test_list(header: str, test_count: int, test_list: list[str]):
assert type(header) == str # noqa: E721
assert type(test_count) == int # noqa: E721
assert type(test_list) == list # noqa: E721
assert header != ""
assert test_count >= 0
assert len(test_list) == test_count

if len(TEST_PROCESS_STATS.XFailedTests) > 0:
helper__print_test_list(TEST_PROCESS_STATS.XFailedTests)
LOCAL__print_line1_with_header(header)
logging.info("")
if len(test_list) > 0:
helper__print_test_list(test_list)
logging.info("")

# fmt: off
LOCAL__print_test_list(
"ACHTUNG TESTS",
TEST_PROCESS_STATS.cAchtungTests,
TEST_PROCESS_STATS.AchtungTests,
)

logging.info("--------------------------- [NOT XFAILED TESTS]")
logging.info("")
LOCAL__print_test_list(
"FAILED TESTS",
TEST_PROCESS_STATS.cFailedTests,
TEST_PROCESS_STATS.FailedTests
)

assert (
len(TEST_PROCESS_STATS.NotXFailedTests) == TEST_PROCESS_STATS.cNotXFailedTests
LOCAL__print_test_list(
"XFAILED TESTS",
TEST_PROCESS_STATS.cXFailedTests,
TEST_PROCESS_STATS.XFailedTests,
)

if len(TEST_PROCESS_STATS.NotXFailedTests) > 0:
helper__print_test_list(TEST_PROCESS_STATS.NotXFailedTests)
logging.info("")
LOCAL__print_test_list(
"NOT XFAILED TESTS",
TEST_PROCESS_STATS.cNotXFailedTests,
TEST_PROCESS_STATS.NotXFailedTests,
)
# fmt: on

logging.info("--------------------------- [SUMMARY STATISTICS]")
LOCAL__print_line1_with_header("SUMMARY STATISTICS")
logging.info("")
logging.info("[TESTS]")
logging.info(" TOTAL : {0}".format(TEST_PROCESS_STATS.cTotalTests))
logging.info(" EXECUTED : {0}".format(TEST_PROCESS_STATS.cExecutedTests))
logging.info(" NOT EXECUTED: {0}".format(TEST_PROCESS_STATS.cNotExecutedTests))
logging.info(" ACHTUNG : {0}".format(TEST_PROCESS_STATS.cAchtungTests))
logging.info("")
logging.info(" PASSED : {0}".format(TEST_PROCESS_STATS.cPassedTests))
logging.info(" FAILED : {0}".format(TEST_PROCESS_STATS.cFailedTests))
Expand Down Expand Up @@ -503,7 +547,13 @@ def pytest_configure(config: pytest.Config) -> None:

log_path.mkdir(exist_ok=True)

logging_plugin = config.pluginmanager.get_plugin("logging-plugin")
logging_plugin: _pytest.logging.LoggingPlugin = config.pluginmanager.get_plugin(
"logging-plugin"
)

assert logging_plugin is not None
assert isinstance(logging_plugin, _pytest.logging.LoggingPlugin)

logging_plugin.set_log_path(str(log_path / log_name))


Expand Down