forked from Giskard-AI/giskard
-
Notifications
You must be signed in to change notification settings - Fork 0
/
conftest.py
110 lines (91 loc) · 3.63 KB
/
conftest.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
from typing import List
import gc
import json
import os
import subprocess
import sys
import time
from pathlib import Path
from tempfile import NamedTemporaryFile
import psutil
import pytest
from _pytest.config.argparsing import Parser
from _pytest.python import Function
from _pytest.reports import TestReport
@pytest.hookimpl(hookwrapper=True)
def pytest_fixture_setup(fixturedef, request):
start = time.time()
process = psutil.Process(os.getpid())
base_memory_usage = process.memory_info().rss
yield
memory_usage = process.memory_info().rss
fixture_memory_usage = memory_usage - base_memory_usage # in bytes
fixture_memory_usage = fixture_memory_usage / (1024 * 1024) # in mo
end = time.time()
file = Path("memory_fixtures.csv")
write_header = False
if not file.exists():
write_header = True
with file.open("a", encoding="utf-8") as writer:
if write_header:
writer.write("fixture_name,execution_time,memory_usage(mo)\n")
# Add overall test results
writer.write(f"{request.fixturename},{end - start:.3f},{fixture_memory_usage:.3f}\n")
@pytest.hookimpl(hookwrapper=True)
def pytest_runtest_protocol(item: Function, nextitem: Function):
process = psutil.Process(os.getpid())
file = Path("memory_tests.csv")
activate_gc = False
if activate_gc:
gc.collect()
gc.collect()
base_memory_usage = process.memory_info().rss
yield
if activate_gc:
gc.collect()
gc.collect()
memory_usage = process.memory_info().rss
test_memory_usage = memory_usage - base_memory_usage # in bytes
test_memory_usage = test_memory_usage / (1024 * 1024) # in mo
full_memory_usage = memory_usage / (1024 * 1024)
write_header = False
if not file.exists():
write_header = True
with file.open("a", encoding="utf-8") as writer:
if write_header:
writer.write("test_name,memory_usage(mo),total_usage(mo)\n")
# Add overall test results
writer.write(f"{item.nodeid},{test_memory_usage:.3f},{full_memory_usage:.3f}\n")
# we know this bit is bad, but we cant help it with the current pytest setup
def pytest_addoption(parser: Parser):
parser.addoption("--use-subprocess", action="store_true", default=False, help="Whether to use subprocess")
def separate_process(item: Function) -> List[TestReport]:
with NamedTemporaryFile(delete=False) as fp:
subprocess.run(
shell=True,
check=False,
stdout=sys.stdout,
stderr=sys.stderr,
args=f"{sys.executable} -m pytest {item.nodeid} -vvv --tb=long --report-log={fp.name} --no-header --no-summary",
cwd=Path(__file__).parent,
)
reports = []
try:
for line in Path(fp.name).read_text().splitlines():
report_dict = json.loads(line)
if report_dict["$report_type"] == "TestReport":
reports.append(TestReport._from_json(report_dict))
return reports
finally:
# Force deletion of the temp file
Path(fp.name).unlink(missing_ok=True)
# https://docs.pytest.org/en/7.1.x/reference/reference.html#pytest.hookspec.pytest_runtest_protocol
@pytest.hookimpl(tryfirst=True)
def pytest_runtest_call(item: Function):
mark = item.get_closest_marker("skip")
skip = mark is not None
mark = item.get_closest_marker("skipif")
skip |= mark is not None and ((len(mark.args) == 1 and mark.args[0]) or mark.kwargs.get("condition", False))
if not skip and item.get_closest_marker("memory_expensive") and item.config.getoption("--use-subprocess"):
reports = separate_process(item)
return reports