982 lines
		
	
	
		
			29 KiB
		
	
	
	
		
			Python
		
	
	
	
			
		
		
	
	
			982 lines
		
	
	
		
			29 KiB
		
	
	
	
		
			Python
		
	
	
	
| # -*- coding: utf-8 -*-
 | |
| from __future__ import absolute_import
 | |
| from __future__ import division
 | |
| from __future__ import print_function
 | |
| 
 | |
| import inspect
 | |
| import os
 | |
| import sys
 | |
| import types
 | |
| 
 | |
| import py
 | |
| 
 | |
| import _pytest._code
 | |
| import pytest
 | |
| from _pytest import main
 | |
| from _pytest import outcomes
 | |
| from _pytest import reports
 | |
| from _pytest import runner
 | |
| 
 | |
| 
 | |
| class TestSetupState(object):
 | |
|     def test_setup(self, testdir):
 | |
|         ss = runner.SetupState()
 | |
|         item = testdir.getitem("def test_func(): pass")
 | |
|         values = [1]
 | |
|         ss.prepare(item)
 | |
|         ss.addfinalizer(values.pop, colitem=item)
 | |
|         assert values
 | |
|         ss._pop_and_teardown()
 | |
|         assert not values
 | |
| 
 | |
|     def test_teardown_exact_stack_empty(self, testdir):
 | |
|         item = testdir.getitem("def test_func(): pass")
 | |
|         ss = runner.SetupState()
 | |
|         ss.teardown_exact(item, None)
 | |
|         ss.teardown_exact(item, None)
 | |
|         ss.teardown_exact(item, None)
 | |
| 
 | |
|     def test_setup_fails_and_failure_is_cached(self, testdir):
 | |
|         item = testdir.getitem(
 | |
|             """
 | |
|             def setup_module(mod):
 | |
|                 raise ValueError(42)
 | |
|             def test_func(): pass
 | |
|         """
 | |
|         )
 | |
|         ss = runner.SetupState()
 | |
|         pytest.raises(ValueError, lambda: ss.prepare(item))
 | |
|         pytest.raises(ValueError, lambda: ss.prepare(item))
 | |
| 
 | |
|     def test_teardown_multiple_one_fails(self, testdir):
 | |
|         r = []
 | |
| 
 | |
|         def fin1():
 | |
|             r.append("fin1")
 | |
| 
 | |
|         def fin2():
 | |
|             raise Exception("oops")
 | |
| 
 | |
|         def fin3():
 | |
|             r.append("fin3")
 | |
| 
 | |
|         item = testdir.getitem("def test_func(): pass")
 | |
|         ss = runner.SetupState()
 | |
|         ss.addfinalizer(fin1, item)
 | |
|         ss.addfinalizer(fin2, item)
 | |
|         ss.addfinalizer(fin3, item)
 | |
|         with pytest.raises(Exception) as err:
 | |
|             ss._callfinalizers(item)
 | |
|         assert err.value.args == ("oops",)
 | |
|         assert r == ["fin3", "fin1"]
 | |
| 
 | |
|     def test_teardown_multiple_fail(self, testdir):
 | |
|         # Ensure the first exception is the one which is re-raised.
 | |
|         # Ideally both would be reported however.
 | |
|         def fin1():
 | |
|             raise Exception("oops1")
 | |
| 
 | |
|         def fin2():
 | |
|             raise Exception("oops2")
 | |
| 
 | |
|         item = testdir.getitem("def test_func(): pass")
 | |
|         ss = runner.SetupState()
 | |
|         ss.addfinalizer(fin1, item)
 | |
|         ss.addfinalizer(fin2, item)
 | |
|         with pytest.raises(Exception) as err:
 | |
|             ss._callfinalizers(item)
 | |
|         assert err.value.args == ("oops2",)
 | |
| 
 | |
|     def test_teardown_multiple_scopes_one_fails(self, testdir):
 | |
|         module_teardown = []
 | |
| 
 | |
|         def fin_func():
 | |
|             raise Exception("oops1")
 | |
| 
 | |
|         def fin_module():
 | |
|             module_teardown.append("fin_module")
 | |
| 
 | |
|         item = testdir.getitem("def test_func(): pass")
 | |
|         ss = runner.SetupState()
 | |
|         ss.addfinalizer(fin_module, item.listchain()[-2])
 | |
|         ss.addfinalizer(fin_func, item)
 | |
|         ss.prepare(item)
 | |
|         with pytest.raises(Exception, match="oops1"):
 | |
|             ss.teardown_exact(item, None)
 | |
|         assert module_teardown
 | |
| 
 | |
| 
 | |
| class BaseFunctionalTests(object):
 | |
|     def test_passfunction(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             def test_func():
 | |
|                 pass
 | |
|         """
 | |
|         )
 | |
|         rep = reports[1]
 | |
|         assert rep.passed
 | |
|         assert not rep.failed
 | |
|         assert rep.outcome == "passed"
 | |
|         assert not rep.longrepr
 | |
| 
 | |
|     def test_failfunction(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             def test_func():
 | |
|                 assert 0
 | |
|         """
 | |
|         )
 | |
|         rep = reports[1]
 | |
|         assert not rep.passed
 | |
|         assert not rep.skipped
 | |
|         assert rep.failed
 | |
|         assert rep.when == "call"
 | |
|         assert rep.outcome == "failed"
 | |
|         # assert isinstance(rep.longrepr, ReprExceptionInfo)
 | |
| 
 | |
|     def test_skipfunction(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             import pytest
 | |
|             def test_func():
 | |
|                 pytest.skip("hello")
 | |
|         """
 | |
|         )
 | |
|         rep = reports[1]
 | |
|         assert not rep.failed
 | |
|         assert not rep.passed
 | |
|         assert rep.skipped
 | |
|         assert rep.outcome == "skipped"
 | |
|         # assert rep.skipped.when == "call"
 | |
|         # assert rep.skipped.when == "call"
 | |
|         # assert rep.skipped == "%sreason == "hello"
 | |
|         # assert rep.skipped.location.lineno == 3
 | |
|         # assert rep.skipped.location.path
 | |
|         # assert not rep.skipped.failurerepr
 | |
| 
 | |
|     def test_skip_in_setup_function(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             import pytest
 | |
|             def setup_function(func):
 | |
|                 pytest.skip("hello")
 | |
|             def test_func():
 | |
|                 pass
 | |
|         """
 | |
|         )
 | |
|         print(reports)
 | |
|         rep = reports[0]
 | |
|         assert not rep.failed
 | |
|         assert not rep.passed
 | |
|         assert rep.skipped
 | |
|         # assert rep.skipped.reason == "hello"
 | |
|         # assert rep.skipped.location.lineno == 3
 | |
|         # assert rep.skipped.location.lineno == 3
 | |
|         assert len(reports) == 2
 | |
|         assert reports[1].passed  # teardown
 | |
| 
 | |
|     def test_failure_in_setup_function(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             import pytest
 | |
|             def setup_function(func):
 | |
|                 raise ValueError(42)
 | |
|             def test_func():
 | |
|                 pass
 | |
|         """
 | |
|         )
 | |
|         rep = reports[0]
 | |
|         assert not rep.skipped
 | |
|         assert not rep.passed
 | |
|         assert rep.failed
 | |
|         assert rep.when == "setup"
 | |
|         assert len(reports) == 2
 | |
| 
 | |
|     def test_failure_in_teardown_function(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             import pytest
 | |
|             def teardown_function(func):
 | |
|                 raise ValueError(42)
 | |
|             def test_func():
 | |
|                 pass
 | |
|         """
 | |
|         )
 | |
|         print(reports)
 | |
|         assert len(reports) == 3
 | |
|         rep = reports[2]
 | |
|         assert not rep.skipped
 | |
|         assert not rep.passed
 | |
|         assert rep.failed
 | |
|         assert rep.when == "teardown"
 | |
|         # assert rep.longrepr.reprcrash.lineno == 3
 | |
|         # assert rep.longrepr.reprtraceback.reprentries
 | |
| 
 | |
|     def test_custom_failure_repr(self, testdir):
 | |
|         testdir.makepyfile(
 | |
|             conftest="""
 | |
|             import pytest
 | |
|             class Function(pytest.Function):
 | |
|                 def repr_failure(self, excinfo):
 | |
|                     return "hello"
 | |
|         """
 | |
|         )
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             import pytest
 | |
|             def test_func():
 | |
|                 assert 0
 | |
|         """
 | |
|         )
 | |
|         rep = reports[1]
 | |
|         assert not rep.skipped
 | |
|         assert not rep.passed
 | |
|         assert rep.failed
 | |
|         # assert rep.outcome.when == "call"
 | |
|         # assert rep.failed.where.lineno == 3
 | |
|         # assert rep.failed.where.path.basename == "test_func.py"
 | |
|         # assert rep.failed.failurerepr == "hello"
 | |
| 
 | |
|     def test_teardown_final_returncode(self, testdir):
 | |
|         rec = testdir.inline_runsource(
 | |
|             """
 | |
|             def test_func():
 | |
|                 pass
 | |
|             def teardown_function(func):
 | |
|                 raise ValueError(42)
 | |
|         """
 | |
|         )
 | |
|         assert rec.ret == 1
 | |
| 
 | |
|     def test_logstart_logfinish_hooks(self, testdir):
 | |
|         rec = testdir.inline_runsource(
 | |
|             """
 | |
|             import pytest
 | |
|             def test_func():
 | |
|                 pass
 | |
|         """
 | |
|         )
 | |
|         reps = rec.getcalls("pytest_runtest_logstart pytest_runtest_logfinish")
 | |
|         assert [x._name for x in reps] == [
 | |
|             "pytest_runtest_logstart",
 | |
|             "pytest_runtest_logfinish",
 | |
|         ]
 | |
|         for rep in reps:
 | |
|             assert rep.nodeid == "test_logstart_logfinish_hooks.py::test_func"
 | |
|             assert rep.location == ("test_logstart_logfinish_hooks.py", 1, "test_func")
 | |
| 
 | |
|     def test_exact_teardown_issue90(self, testdir):
 | |
|         rec = testdir.inline_runsource(
 | |
|             """
 | |
|             import pytest
 | |
| 
 | |
|             class TestClass(object):
 | |
|                 def test_method(self):
 | |
|                     pass
 | |
|                 def teardown_class(cls):
 | |
|                     raise Exception()
 | |
| 
 | |
|             def test_func():
 | |
|                 import sys
 | |
|                 # on python2 exc_info is keept till a function exits
 | |
|                 # so we would end up calling test functions while
 | |
|                 # sys.exc_info would return the indexerror
 | |
|                 # from guessing the lastitem
 | |
|                 excinfo = sys.exc_info()
 | |
|                 import traceback
 | |
|                 assert excinfo[0] is None, \
 | |
|                        traceback.format_exception(*excinfo)
 | |
|             def teardown_function(func):
 | |
|                 raise ValueError(42)
 | |
|         """
 | |
|         )
 | |
|         reps = rec.getreports("pytest_runtest_logreport")
 | |
|         print(reps)
 | |
|         for i in range(2):
 | |
|             assert reps[i].nodeid.endswith("test_method")
 | |
|             assert reps[i].passed
 | |
|         assert reps[2].when == "teardown"
 | |
|         assert reps[2].failed
 | |
|         assert len(reps) == 6
 | |
|         for i in range(3, 5):
 | |
|             assert reps[i].nodeid.endswith("test_func")
 | |
|             assert reps[i].passed
 | |
|         assert reps[5].when == "teardown"
 | |
|         assert reps[5].nodeid.endswith("test_func")
 | |
|         assert reps[5].failed
 | |
| 
 | |
|     def test_exact_teardown_issue1206(self, testdir):
 | |
|         """issue shadowing error with wrong number of arguments on teardown_method."""
 | |
|         rec = testdir.inline_runsource(
 | |
|             """
 | |
|             import pytest
 | |
| 
 | |
|             class TestClass(object):
 | |
|                 def teardown_method(self, x, y, z):
 | |
|                     pass
 | |
| 
 | |
|                 def test_method(self):
 | |
|                     assert True
 | |
|         """
 | |
|         )
 | |
|         reps = rec.getreports("pytest_runtest_logreport")
 | |
|         print(reps)
 | |
|         assert len(reps) == 3
 | |
|         #
 | |
|         assert reps[0].nodeid.endswith("test_method")
 | |
|         assert reps[0].passed
 | |
|         assert reps[0].when == "setup"
 | |
|         #
 | |
|         assert reps[1].nodeid.endswith("test_method")
 | |
|         assert reps[1].passed
 | |
|         assert reps[1].when == "call"
 | |
|         #
 | |
|         assert reps[2].nodeid.endswith("test_method")
 | |
|         assert reps[2].failed
 | |
|         assert reps[2].when == "teardown"
 | |
|         assert reps[2].longrepr.reprcrash.message in (
 | |
|             # python3 error
 | |
|             "TypeError: teardown_method() missing 2 required positional arguments: 'y' and 'z'",
 | |
|             # python2 error
 | |
|             "TypeError: teardown_method() takes exactly 4 arguments (2 given)",
 | |
|         )
 | |
| 
 | |
|     def test_failure_in_setup_function_ignores_custom_repr(self, testdir):
 | |
|         testdir.makepyfile(
 | |
|             conftest="""
 | |
|             import pytest
 | |
|             class Function(pytest.Function):
 | |
|                 def repr_failure(self, excinfo):
 | |
|                     assert 0
 | |
|         """
 | |
|         )
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             def setup_function(func):
 | |
|                 raise ValueError(42)
 | |
|             def test_func():
 | |
|                 pass
 | |
|         """
 | |
|         )
 | |
|         assert len(reports) == 2
 | |
|         rep = reports[0]
 | |
|         print(rep)
 | |
|         assert not rep.skipped
 | |
|         assert not rep.passed
 | |
|         assert rep.failed
 | |
|         # assert rep.outcome.when == "setup"
 | |
|         # assert rep.outcome.where.lineno == 3
 | |
|         # assert rep.outcome.where.path.basename == "test_func.py"
 | |
|         # assert instanace(rep.failed.failurerepr, PythonFailureRepr)
 | |
| 
 | |
|     def test_systemexit_does_not_bail_out(self, testdir):
 | |
|         try:
 | |
|             reports = testdir.runitem(
 | |
|                 """
 | |
|                 def test_func():
 | |
|                     raise SystemExit(42)
 | |
|             """
 | |
|             )
 | |
|         except SystemExit:
 | |
|             pytest.fail("runner did not catch SystemExit")
 | |
|         rep = reports[1]
 | |
|         assert rep.failed
 | |
|         assert rep.when == "call"
 | |
| 
 | |
|     def test_exit_propagates(self, testdir):
 | |
|         try:
 | |
|             testdir.runitem(
 | |
|                 """
 | |
|                 import pytest
 | |
|                 def test_func():
 | |
|                     raise pytest.exit.Exception()
 | |
|             """
 | |
|             )
 | |
|         except pytest.exit.Exception:
 | |
|             pass
 | |
|         else:
 | |
|             pytest.fail("did not raise")
 | |
| 
 | |
| 
 | |
| class TestExecutionNonForked(BaseFunctionalTests):
 | |
|     def getrunner(self):
 | |
|         def f(item):
 | |
|             return runner.runtestprotocol(item, log=False)
 | |
| 
 | |
|         return f
 | |
| 
 | |
|     def test_keyboardinterrupt_propagates(self, testdir):
 | |
|         try:
 | |
|             testdir.runitem(
 | |
|                 """
 | |
|                 def test_func():
 | |
|                     raise KeyboardInterrupt("fake")
 | |
|             """
 | |
|             )
 | |
|         except KeyboardInterrupt:
 | |
|             pass
 | |
|         else:
 | |
|             pytest.fail("did not raise")
 | |
| 
 | |
| 
 | |
| class TestExecutionForked(BaseFunctionalTests):
 | |
|     pytestmark = pytest.mark.skipif("not hasattr(os, 'fork')")
 | |
| 
 | |
|     def getrunner(self):
 | |
|         # XXX re-arrange this test to live in pytest-xdist
 | |
|         boxed = pytest.importorskip("xdist.boxed")
 | |
|         return boxed.forked_run_report
 | |
| 
 | |
|     def test_suicide(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             def test_func():
 | |
|                 import os
 | |
|                 os.kill(os.getpid(), 15)
 | |
|         """
 | |
|         )
 | |
|         rep = reports[0]
 | |
|         assert rep.failed
 | |
|         assert rep.when == "???"
 | |
| 
 | |
| 
 | |
| class TestSessionReports(object):
 | |
|     def test_collect_result(self, testdir):
 | |
|         col = testdir.getmodulecol(
 | |
|             """
 | |
|             def test_func1():
 | |
|                 pass
 | |
|             class TestClass(object):
 | |
|                 pass
 | |
|         """
 | |
|         )
 | |
|         rep = runner.collect_one_node(col)
 | |
|         assert not rep.failed
 | |
|         assert not rep.skipped
 | |
|         assert rep.passed
 | |
|         locinfo = rep.location
 | |
|         assert locinfo[0] == col.fspath.basename
 | |
|         assert not locinfo[1]
 | |
|         assert locinfo[2] == col.fspath.basename
 | |
|         res = rep.result
 | |
|         assert len(res) == 2
 | |
|         assert res[0].name == "test_func1"
 | |
|         assert res[1].name == "TestClass"
 | |
| 
 | |
| 
 | |
| reporttypes = [reports.BaseReport, reports.TestReport, reports.CollectReport]
 | |
| 
 | |
| 
 | |
| @pytest.mark.parametrize(
 | |
|     "reporttype", reporttypes, ids=[x.__name__ for x in reporttypes]
 | |
| )
 | |
| def test_report_extra_parameters(reporttype):
 | |
|     if hasattr(inspect, "signature"):
 | |
|         args = list(inspect.signature(reporttype.__init__).parameters.keys())[1:]
 | |
|     else:
 | |
|         args = inspect.getargspec(reporttype.__init__)[0][1:]
 | |
|     basekw = dict.fromkeys(args, [])
 | |
|     report = reporttype(newthing=1, **basekw)
 | |
|     assert report.newthing == 1
 | |
| 
 | |
| 
 | |
| def test_callinfo():
 | |
|     ci = runner.CallInfo.from_call(lambda: 0, "123")
 | |
|     assert ci.when == "123"
 | |
|     assert ci.result == 0
 | |
|     assert "result" in repr(ci)
 | |
|     assert repr(ci) == "<CallInfo when='123' result: 0>"
 | |
| 
 | |
|     ci = runner.CallInfo.from_call(lambda: 0 / 0, "123")
 | |
|     assert ci.when == "123"
 | |
|     assert not hasattr(ci, "result")
 | |
|     assert repr(ci) == "<CallInfo when='123' exception: division by zero>"
 | |
|     assert ci.excinfo
 | |
|     assert "exc" in repr(ci)
 | |
| 
 | |
| 
 | |
| # design question: do we want general hooks in python files?
 | |
| # then something like the following functional tests makes sense
 | |
| 
 | |
| 
 | |
| @pytest.mark.xfail
 | |
| def test_runtest_in_module_ordering(testdir):
 | |
|     p1 = testdir.makepyfile(
 | |
|         """
 | |
|         import pytest
 | |
|         def pytest_runtest_setup(item): # runs after class-level!
 | |
|             item.function.mylist.append("module")
 | |
|         class TestClass(object):
 | |
|             def pytest_runtest_setup(self, item):
 | |
|                 assert not hasattr(item.function, 'mylist')
 | |
|                 item.function.mylist = ['class']
 | |
|             @pytest.fixture
 | |
|             def mylist(self, request):
 | |
|                 return request.function.mylist
 | |
|             def pytest_runtest_call(self, item, __multicall__):
 | |
|                 try:
 | |
|                     __multicall__.execute()
 | |
|                 except ValueError:
 | |
|                     pass
 | |
|             def test_hello1(self, mylist):
 | |
|                 assert mylist == ['class', 'module'], mylist
 | |
|                 raise ValueError()
 | |
|             def test_hello2(self, mylist):
 | |
|                 assert mylist == ['class', 'module'], mylist
 | |
|         def pytest_runtest_teardown(item):
 | |
|             del item.function.mylist
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest(p1)
 | |
|     result.stdout.fnmatch_lines(["*2 passed*"])
 | |
| 
 | |
| 
 | |
| def test_outcomeexception_exceptionattributes():
 | |
|     outcome = outcomes.OutcomeException("test")
 | |
|     assert outcome.args[0] == outcome.msg
 | |
| 
 | |
| 
 | |
| def test_outcomeexception_passes_except_Exception():
 | |
|     with pytest.raises(outcomes.OutcomeException):
 | |
|         try:
 | |
|             raise outcomes.OutcomeException("test")
 | |
|         except Exception:
 | |
|             pass
 | |
| 
 | |
| 
 | |
| def test_pytest_exit():
 | |
|     with pytest.raises(pytest.exit.Exception) as excinfo:
 | |
|         pytest.exit("hello")
 | |
|     assert excinfo.errisinstance(pytest.exit.Exception)
 | |
| 
 | |
| 
 | |
| def test_pytest_fail():
 | |
|     with pytest.raises(pytest.fail.Exception) as excinfo:
 | |
|         pytest.fail("hello")
 | |
|     s = excinfo.exconly(tryshort=True)
 | |
|     assert s.startswith("Failed")
 | |
| 
 | |
| 
 | |
| def test_pytest_exit_msg(testdir):
 | |
|     testdir.makeconftest(
 | |
|         """
 | |
|     import pytest
 | |
| 
 | |
|     def pytest_configure(config):
 | |
|         pytest.exit('oh noes')
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest()
 | |
|     result.stderr.fnmatch_lines(["Exit: oh noes"])
 | |
| 
 | |
| 
 | |
| def test_pytest_exit_returncode(testdir):
 | |
|     testdir.makepyfile(
 | |
|         """
 | |
|         import pytest
 | |
|         def test_foo():
 | |
|             pytest.exit("some exit msg", 99)
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest()
 | |
|     assert result.ret == 99
 | |
| 
 | |
| 
 | |
| def test_pytest_fail_notrace_runtest(testdir):
 | |
|     """Test pytest.fail(..., pytrace=False) does not show tracebacks during test run."""
 | |
|     testdir.makepyfile(
 | |
|         """
 | |
|         import pytest
 | |
|         def test_hello():
 | |
|             pytest.fail("hello", pytrace=False)
 | |
|         def teardown_function(function):
 | |
|             pytest.fail("world", pytrace=False)
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest()
 | |
|     result.stdout.fnmatch_lines(["world", "hello"])
 | |
|     assert "def teardown_function" not in result.stdout.str()
 | |
| 
 | |
| 
 | |
| def test_pytest_fail_notrace_collection(testdir):
 | |
|     """Test pytest.fail(..., pytrace=False) does not show tracebacks during collection."""
 | |
|     testdir.makepyfile(
 | |
|         """
 | |
|         import pytest
 | |
|         def some_internal_function():
 | |
|             pytest.fail("hello", pytrace=False)
 | |
|         some_internal_function()
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest()
 | |
|     result.stdout.fnmatch_lines(["hello"])
 | |
|     assert "def some_internal_function()" not in result.stdout.str()
 | |
| 
 | |
| 
 | |
| @pytest.mark.parametrize("str_prefix", ["u", ""])
 | |
| def test_pytest_fail_notrace_non_ascii(testdir, str_prefix):
 | |
|     """Fix pytest.fail with pytrace=False with non-ascii characters (#1178).
 | |
| 
 | |
|     This tests with native and unicode strings containing non-ascii chars.
 | |
|     """
 | |
|     testdir.makepyfile(
 | |
|         u"""
 | |
|         # coding: utf-8
 | |
|         import pytest
 | |
| 
 | |
|         def test_hello():
 | |
|             pytest.fail(%s'oh oh: ☺', pytrace=False)
 | |
|     """
 | |
|         % str_prefix
 | |
|     )
 | |
|     result = testdir.runpytest()
 | |
|     if sys.version_info[0] >= 3:
 | |
|         result.stdout.fnmatch_lines(["*test_hello*", "oh oh: ☺"])
 | |
|     else:
 | |
|         result.stdout.fnmatch_lines(["*test_hello*", "oh oh: *"])
 | |
|     assert "def test_hello" not in result.stdout.str()
 | |
| 
 | |
| 
 | |
| def test_pytest_no_tests_collected_exit_status(testdir):
 | |
|     result = testdir.runpytest()
 | |
|     result.stdout.fnmatch_lines(["*collected 0 items*"])
 | |
|     assert result.ret == main.EXIT_NOTESTSCOLLECTED
 | |
| 
 | |
|     testdir.makepyfile(
 | |
|         test_foo="""
 | |
|         def test_foo():
 | |
|             assert 1
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest()
 | |
|     result.stdout.fnmatch_lines(["*collected 1 item*"])
 | |
|     result.stdout.fnmatch_lines(["*1 passed*"])
 | |
|     assert result.ret == main.EXIT_OK
 | |
| 
 | |
|     result = testdir.runpytest("-k nonmatch")
 | |
|     result.stdout.fnmatch_lines(["*collected 1 item*"])
 | |
|     result.stdout.fnmatch_lines(["*1 deselected*"])
 | |
|     assert result.ret == main.EXIT_NOTESTSCOLLECTED
 | |
| 
 | |
| 
 | |
| def test_exception_printing_skip():
 | |
|     try:
 | |
|         pytest.skip("hello")
 | |
|     except pytest.skip.Exception:
 | |
|         excinfo = _pytest._code.ExceptionInfo.from_current()
 | |
|         s = excinfo.exconly(tryshort=True)
 | |
|         assert s.startswith("Skipped")
 | |
| 
 | |
| 
 | |
| def test_importorskip(monkeypatch):
 | |
|     importorskip = pytest.importorskip
 | |
| 
 | |
|     def f():
 | |
|         importorskip("asdlkj")
 | |
| 
 | |
|     try:
 | |
|         sysmod = importorskip("sys")
 | |
|         assert sysmod is sys
 | |
|         # path = pytest.importorskip("os.path")
 | |
|         # assert path == os.path
 | |
|         excinfo = pytest.raises(pytest.skip.Exception, f)
 | |
|         path = py.path.local(excinfo.getrepr().reprcrash.path)
 | |
|         # check that importorskip reports the actual call
 | |
|         # in this test the test_runner.py file
 | |
|         assert path.purebasename == "test_runner"
 | |
|         pytest.raises(SyntaxError, pytest.importorskip, "x y z")
 | |
|         pytest.raises(SyntaxError, pytest.importorskip, "x=y")
 | |
|         mod = types.ModuleType("hello123")
 | |
|         mod.__version__ = "1.3"
 | |
|         monkeypatch.setitem(sys.modules, "hello123", mod)
 | |
|         with pytest.raises(pytest.skip.Exception):
 | |
|             pytest.importorskip("hello123", minversion="1.3.1")
 | |
|         mod2 = pytest.importorskip("hello123", minversion="1.3")
 | |
|         assert mod2 == mod
 | |
|     except pytest.skip.Exception:
 | |
|         print(_pytest._code.ExceptionInfo.from_current())
 | |
|         pytest.fail("spurious skip")
 | |
| 
 | |
| 
 | |
| def test_importorskip_imports_last_module_part():
 | |
|     ospath = pytest.importorskip("os.path")
 | |
|     assert os.path == ospath
 | |
| 
 | |
| 
 | |
| def test_importorskip_dev_module(monkeypatch):
 | |
|     try:
 | |
|         mod = types.ModuleType("mockmodule")
 | |
|         mod.__version__ = "0.13.0.dev-43290"
 | |
|         monkeypatch.setitem(sys.modules, "mockmodule", mod)
 | |
|         mod2 = pytest.importorskip("mockmodule", minversion="0.12.0")
 | |
|         assert mod2 == mod
 | |
|         with pytest.raises(pytest.skip.Exception):
 | |
|             pytest.importorskip("mockmodule1", minversion="0.14.0")
 | |
|     except pytest.skip.Exception:
 | |
|         print(_pytest._code.ExceptionInfo.from_current())
 | |
|         pytest.fail("spurious skip")
 | |
| 
 | |
| 
 | |
| def test_importorskip_module_level(testdir):
 | |
|     """importorskip must be able to skip entire modules when used at module level"""
 | |
|     testdir.makepyfile(
 | |
|         """
 | |
|         import pytest
 | |
|         foobarbaz = pytest.importorskip("foobarbaz")
 | |
| 
 | |
|         def test_foo():
 | |
|             pass
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest()
 | |
|     result.stdout.fnmatch_lines(["*collected 0 items / 1 skipped*"])
 | |
| 
 | |
| 
 | |
| def test_importorskip_custom_reason(testdir):
 | |
|     """make sure custom reasons are used"""
 | |
|     testdir.makepyfile(
 | |
|         """
 | |
|         import pytest
 | |
|         foobarbaz = pytest.importorskip("foobarbaz2", reason="just because")
 | |
| 
 | |
|         def test_foo():
 | |
|             pass
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest("-ra")
 | |
|     result.stdout.fnmatch_lines(["*just because*"])
 | |
|     result.stdout.fnmatch_lines(["*collected 0 items / 1 skipped*"])
 | |
| 
 | |
| 
 | |
| def test_pytest_cmdline_main(testdir):
 | |
|     p = testdir.makepyfile(
 | |
|         """
 | |
|         import pytest
 | |
|         def test_hello():
 | |
|             assert 1
 | |
|         if __name__ == '__main__':
 | |
|            pytest.cmdline.main([__file__])
 | |
|     """
 | |
|     )
 | |
|     import subprocess
 | |
| 
 | |
|     popen = subprocess.Popen([sys.executable, str(p)], stdout=subprocess.PIPE)
 | |
|     popen.communicate()
 | |
|     ret = popen.wait()
 | |
|     assert ret == 0
 | |
| 
 | |
| 
 | |
| def test_unicode_in_longrepr(testdir):
 | |
|     testdir.makeconftest(
 | |
|         """
 | |
|         # -*- coding: utf-8 -*-
 | |
|         import pytest
 | |
|         @pytest.hookimpl(hookwrapper=True)
 | |
|         def pytest_runtest_makereport():
 | |
|             outcome = yield
 | |
|             rep = outcome.get_result()
 | |
|             if rep.when == "call":
 | |
|                 rep.longrepr = u'ä'
 | |
|     """
 | |
|     )
 | |
|     testdir.makepyfile(
 | |
|         """
 | |
|         def test_out():
 | |
|             assert 0
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest()
 | |
|     assert result.ret == 1
 | |
|     assert "UnicodeEncodeError" not in result.stderr.str()
 | |
| 
 | |
| 
 | |
| def test_failure_in_setup(testdir):
 | |
|     testdir.makepyfile(
 | |
|         """
 | |
|         def setup_module():
 | |
|             0/0
 | |
|         def test_func():
 | |
|             pass
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest("--tb=line")
 | |
|     assert "def setup_module" not in result.stdout.str()
 | |
| 
 | |
| 
 | |
| def test_makereport_getsource(testdir):
 | |
|     testdir.makepyfile(
 | |
|         """
 | |
|         def test_foo():
 | |
|             if False: pass
 | |
|             else: assert False
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest()
 | |
|     assert "INTERNALERROR" not in result.stdout.str()
 | |
|     result.stdout.fnmatch_lines(["*else: assert False*"])
 | |
| 
 | |
| 
 | |
| def test_makereport_getsource_dynamic_code(testdir, monkeypatch):
 | |
|     """Test that exception in dynamically generated code doesn't break getting the source line."""
 | |
|     import inspect
 | |
| 
 | |
|     original_findsource = inspect.findsource
 | |
| 
 | |
|     def findsource(obj, *args, **kwargs):
 | |
|         # Can be triggered by dynamically created functions
 | |
|         if obj.__name__ == "foo":
 | |
|             raise IndexError()
 | |
|         return original_findsource(obj, *args, **kwargs)
 | |
| 
 | |
|     monkeypatch.setattr(inspect, "findsource", findsource)
 | |
| 
 | |
|     testdir.makepyfile(
 | |
|         """
 | |
|         import pytest
 | |
| 
 | |
|         @pytest.fixture
 | |
|         def foo(missing):
 | |
|             pass
 | |
| 
 | |
|         def test_fix(foo):
 | |
|             assert False
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest("-vv")
 | |
|     assert "INTERNALERROR" not in result.stdout.str()
 | |
|     result.stdout.fnmatch_lines(["*test_fix*", "*fixture*'missing'*not found*"])
 | |
| 
 | |
| 
 | |
| def test_store_except_info_on_error():
 | |
|     """ Test that upon test failure, the exception info is stored on
 | |
|     sys.last_traceback and friends.
 | |
|     """
 | |
|     # Simulate item that might raise a specific exception, depending on `raise_error` class var
 | |
|     class ItemMightRaise(object):
 | |
|         nodeid = "item_that_raises"
 | |
|         raise_error = True
 | |
| 
 | |
|         def runtest(self):
 | |
|             if self.raise_error:
 | |
|                 raise IndexError("TEST")
 | |
| 
 | |
|     try:
 | |
|         runner.pytest_runtest_call(ItemMightRaise())
 | |
|     except IndexError:
 | |
|         pass
 | |
|     # Check that exception info is stored on sys
 | |
|     assert sys.last_type is IndexError
 | |
|     assert sys.last_value.args[0] == "TEST"
 | |
|     assert sys.last_traceback
 | |
| 
 | |
|     # The next run should clear the exception info stored by the previous run
 | |
|     ItemMightRaise.raise_error = False
 | |
|     runner.pytest_runtest_call(ItemMightRaise())
 | |
|     assert sys.last_type is None
 | |
|     assert sys.last_value is None
 | |
|     assert sys.last_traceback is None
 | |
| 
 | |
| 
 | |
| def test_current_test_env_var(testdir, monkeypatch):
 | |
|     pytest_current_test_vars = []
 | |
|     monkeypatch.setattr(
 | |
|         sys, "pytest_current_test_vars", pytest_current_test_vars, raising=False
 | |
|     )
 | |
|     testdir.makepyfile(
 | |
|         """
 | |
|         import pytest
 | |
|         import sys
 | |
|         import os
 | |
| 
 | |
|         @pytest.fixture
 | |
|         def fix():
 | |
|             sys.pytest_current_test_vars.append(('setup', os.environ['PYTEST_CURRENT_TEST']))
 | |
|             yield
 | |
|             sys.pytest_current_test_vars.append(('teardown', os.environ['PYTEST_CURRENT_TEST']))
 | |
| 
 | |
|         def test(fix):
 | |
|             sys.pytest_current_test_vars.append(('call', os.environ['PYTEST_CURRENT_TEST']))
 | |
|     """
 | |
|     )
 | |
|     result = testdir.runpytest_inprocess()
 | |
|     assert result.ret == 0
 | |
|     test_id = "test_current_test_env_var.py::test"
 | |
|     assert pytest_current_test_vars == [
 | |
|         ("setup", test_id + " (setup)"),
 | |
|         ("call", test_id + " (call)"),
 | |
|         ("teardown", test_id + " (teardown)"),
 | |
|     ]
 | |
|     assert "PYTEST_CURRENT_TEST" not in os.environ
 | |
| 
 | |
| 
 | |
| class TestReportContents(object):
 | |
|     """
 | |
|     Test user-level API of ``TestReport`` objects.
 | |
|     """
 | |
| 
 | |
|     def getrunner(self):
 | |
|         return lambda item: runner.runtestprotocol(item, log=False)
 | |
| 
 | |
|     def test_longreprtext_pass(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             def test_func():
 | |
|                 pass
 | |
|         """
 | |
|         )
 | |
|         rep = reports[1]
 | |
|         assert rep.longreprtext == ""
 | |
| 
 | |
|     def test_longreprtext_failure(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             def test_func():
 | |
|                 x = 1
 | |
|                 assert x == 4
 | |
|         """
 | |
|         )
 | |
|         rep = reports[1]
 | |
|         assert "assert 1 == 4" in rep.longreprtext
 | |
| 
 | |
|     def test_captured_text(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             import pytest
 | |
|             import sys
 | |
| 
 | |
|             @pytest.fixture
 | |
|             def fix():
 | |
|                 sys.stdout.write('setup: stdout\\n')
 | |
|                 sys.stderr.write('setup: stderr\\n')
 | |
|                 yield
 | |
|                 sys.stdout.write('teardown: stdout\\n')
 | |
|                 sys.stderr.write('teardown: stderr\\n')
 | |
|                 assert 0
 | |
| 
 | |
|             def test_func(fix):
 | |
|                 sys.stdout.write('call: stdout\\n')
 | |
|                 sys.stderr.write('call: stderr\\n')
 | |
|                 assert 0
 | |
|         """
 | |
|         )
 | |
|         setup, call, teardown = reports
 | |
|         assert setup.capstdout == "setup: stdout\n"
 | |
|         assert call.capstdout == "setup: stdout\ncall: stdout\n"
 | |
|         assert teardown.capstdout == "setup: stdout\ncall: stdout\nteardown: stdout\n"
 | |
| 
 | |
|         assert setup.capstderr == "setup: stderr\n"
 | |
|         assert call.capstderr == "setup: stderr\ncall: stderr\n"
 | |
|         assert teardown.capstderr == "setup: stderr\ncall: stderr\nteardown: stderr\n"
 | |
| 
 | |
|     def test_no_captured_text(self, testdir):
 | |
|         reports = testdir.runitem(
 | |
|             """
 | |
|             def test_func():
 | |
|                 pass
 | |
|         """
 | |
|         )
 | |
|         rep = reports[1]
 | |
|         assert rep.capstdout == ""
 | |
|         assert rep.capstderr == ""
 |