254 lines
8.5 KiB
Python
254 lines
8.5 KiB
Python
import py
|
|
|
|
EXPECTTIMEOUT=10.0
|
|
|
|
class TestGeneralUsage:
|
|
def test_config_error(self, testdir):
|
|
testdir.makeconftest("""
|
|
def pytest_configure(config):
|
|
raise config.Error("hello")
|
|
""")
|
|
result = testdir.runpytest(testdir.tmpdir)
|
|
assert result.ret != 0
|
|
assert result.stderr.fnmatch_lines([
|
|
'*ERROR: hello'
|
|
])
|
|
|
|
def test_config_preparse_plugin_option(self, testdir):
|
|
testdir.makepyfile(pytest_xyz="""
|
|
def pytest_addoption(parser):
|
|
parser.addoption("--xyz", dest="xyz", action="store")
|
|
""")
|
|
testdir.makepyfile(test_one="""
|
|
import py
|
|
def test_option():
|
|
assert py.test.config.option.xyz == "123"
|
|
""")
|
|
result = testdir.runpytest("-p", "xyz", "--xyz=123")
|
|
assert result.ret == 0
|
|
assert result.stdout.fnmatch_lines([
|
|
'*1 passed*',
|
|
])
|
|
|
|
def test_basetemp(self, testdir):
|
|
mytemp = testdir.tmpdir.mkdir("mytemp")
|
|
p = testdir.makepyfile("""
|
|
import py
|
|
def test_1():
|
|
py.test.ensuretemp('xyz')
|
|
""")
|
|
result = testdir.runpytest(p, '--basetemp=%s' %mytemp)
|
|
assert result.ret == 0
|
|
assert mytemp.join('xyz').check(dir=1)
|
|
|
|
def test_assertion_magic(self, testdir):
|
|
p = testdir.makepyfile("""
|
|
def test_this():
|
|
x = 0
|
|
assert x
|
|
""")
|
|
result = testdir.runpytest(p)
|
|
extra = result.stdout.fnmatch_lines([
|
|
"> assert x",
|
|
"E assert 0",
|
|
])
|
|
assert result.ret == 1
|
|
|
|
def test_nested_import_error(self, testdir):
|
|
p = testdir.makepyfile("""
|
|
import import_fails
|
|
def test_this():
|
|
assert import_fails.a == 1
|
|
""")
|
|
testdir.makepyfile(import_fails="import does_not_work")
|
|
result = testdir.runpytest(p)
|
|
extra = result.stdout.fnmatch_lines([
|
|
"> import import_fails",
|
|
"E ImportError: No module named does_not_work",
|
|
])
|
|
assert result.ret == 1
|
|
|
|
class TestDistribution:
|
|
def test_dist_conftest_options(self, testdir):
|
|
p1 = testdir.tmpdir.ensure("dir", 'p1.py')
|
|
p1.dirpath("__init__.py").write("")
|
|
p1.dirpath("conftest.py").write(py.code.Source("""
|
|
print "importing conftest", __file__
|
|
import py
|
|
Option = py.test.config.Option
|
|
option = py.test.config.addoptions("someopt",
|
|
Option('--someopt', action="store_true", dest="someopt", default=False))
|
|
dist_rsync_roots = ['../dir']
|
|
print "added options", option
|
|
print "config file seen from conftest", py.test.config
|
|
"""))
|
|
p1.write(py.code.Source("""
|
|
import py, conftest
|
|
def test_1():
|
|
print "config from test_1", py.test.config
|
|
print "conftest from test_1", conftest.__file__
|
|
print "test_1: py.test.config.option.someopt", py.test.config.option.someopt
|
|
print "test_1: conftest", conftest
|
|
print "test_1: conftest.option.someopt", conftest.option.someopt
|
|
assert conftest.option.someopt
|
|
"""))
|
|
result = testdir.runpytest('-d', '--tx=popen', p1, '--someopt')
|
|
assert result.ret == 0
|
|
extra = result.stdout.fnmatch_lines([
|
|
"*1 passed*",
|
|
])
|
|
|
|
def test_manytests_to_one_popen(self, testdir):
|
|
p1 = testdir.makepyfile("""
|
|
import py
|
|
def test_fail0():
|
|
assert 0
|
|
def test_fail1():
|
|
raise ValueError()
|
|
def test_ok():
|
|
pass
|
|
def test_skip():
|
|
py.test.skip("hello")
|
|
""",
|
|
)
|
|
result = testdir.runpytest(p1, '-d', '--tx=popen', '--tx=popen')
|
|
result.stdout.fnmatch_lines([
|
|
"*1*popen*Python*",
|
|
"*2*popen*Python*",
|
|
"*2 failed, 1 passed, 1 skipped*",
|
|
])
|
|
assert result.ret == 1
|
|
|
|
def test_dist_conftest_specified(self, testdir):
|
|
p1 = testdir.makepyfile("""
|
|
import py
|
|
def test_fail0():
|
|
assert 0
|
|
def test_fail1():
|
|
raise ValueError()
|
|
def test_ok():
|
|
pass
|
|
def test_skip():
|
|
py.test.skip("hello")
|
|
""",
|
|
)
|
|
testdir.makeconftest("""
|
|
pytest_option_tx = 'popen popen popen'.split()
|
|
""")
|
|
result = testdir.runpytest(p1, '-d')
|
|
result.stdout.fnmatch_lines([
|
|
"*1*popen*Python*",
|
|
"*2*popen*Python*",
|
|
"*3*popen*Python*",
|
|
"*2 failed, 1 passed, 1 skipped*",
|
|
])
|
|
assert result.ret == 1
|
|
|
|
def test_dist_tests_with_crash(self, testdir):
|
|
if not hasattr(py.std.os, 'kill'):
|
|
py.test.skip("no os.kill")
|
|
|
|
p1 = testdir.makepyfile("""
|
|
import py
|
|
def test_fail0():
|
|
assert 0
|
|
def test_fail1():
|
|
raise ValueError()
|
|
def test_ok():
|
|
pass
|
|
def test_skip():
|
|
py.test.skip("hello")
|
|
def test_crash():
|
|
import time
|
|
import os
|
|
time.sleep(0.5)
|
|
os.kill(os.getpid(), 15)
|
|
"""
|
|
)
|
|
result = testdir.runpytest(p1, '-d', '--tx=3*popen')
|
|
result.stdout.fnmatch_lines([
|
|
"*popen*Python*",
|
|
"*popen*Python*",
|
|
"*popen*Python*",
|
|
"*node down*",
|
|
"*3 failed, 1 passed, 1 skipped*"
|
|
])
|
|
assert result.ret == 1
|
|
|
|
def test_distribution_rsyncdirs_example(self, testdir):
|
|
source = testdir.mkdir("source")
|
|
dest = testdir.mkdir("dest")
|
|
subdir = source.mkdir("example_pkg")
|
|
subdir.ensure("__init__.py")
|
|
p = subdir.join("test_one.py")
|
|
p.write("def test_5(): assert not __file__.startswith(%r)" % str(p))
|
|
result = testdir.runpytest("-d", "--rsyncdir=%(subdir)s" % locals(),
|
|
"--tx=popen//chdir=%(dest)s" % locals(), p)
|
|
assert result.ret == 0
|
|
result.stdout.fnmatch_lines([
|
|
"*1* *popen*platform*",
|
|
#"RSyncStart: [G1]",
|
|
#"RSyncFinished: [G1]",
|
|
"*1 passed*"
|
|
])
|
|
assert dest.join(subdir.basename).check(dir=1)
|
|
|
|
def test_dist_each(self, testdir):
|
|
interpreters = []
|
|
for name in ("python2.4", "python2.5"):
|
|
interp = py.path.local.sysfind(name)
|
|
if interp is None:
|
|
py.test.skip("%s not found" % name)
|
|
interpreters.append(interp)
|
|
|
|
testdir.makepyfile(__init__="", test_one="""
|
|
import sys
|
|
def test_hello():
|
|
print "%s...%s" % sys.version_info[:2]
|
|
assert 0
|
|
""")
|
|
args = ["--dist=each"]
|
|
args += ["--tx", "popen//python=%s" % interpreters[0]]
|
|
args += ["--tx", "popen//python=%s" % interpreters[1]]
|
|
result = testdir.runpytest(*args)
|
|
result.stdout.fnmatch_lines(["2...4"])
|
|
result.stdout.fnmatch_lines(["2...5"])
|
|
|
|
|
|
class TestInteractive:
|
|
def test_simple_looponfail_interaction(self, testdir):
|
|
p1 = testdir.makepyfile("""
|
|
def test_1():
|
|
assert 1 == 0
|
|
""")
|
|
p1.setmtime(p1.mtime() - 50.0)
|
|
child = testdir.spawn_pytest("--looponfail %s" % p1)
|
|
child.expect("assert 1 == 0")
|
|
child.expect("test_simple_looponfail_interaction.py:")
|
|
child.expect("1 failed")
|
|
child.expect("waiting for changes")
|
|
p1.write(py.code.Source("""
|
|
def test_1():
|
|
assert 1 == 1
|
|
"""))
|
|
child.expect("MODIFIED.*test_simple_looponfail_interaction.py", timeout=4.0)
|
|
child.expect("1 passed", timeout=5.0)
|
|
child.kill(15)
|
|
|
|
class TestKeyboardInterrupt:
|
|
def test_raised_in_testfunction(self, testdir):
|
|
p1 = testdir.makepyfile("""
|
|
import py
|
|
def test_fail():
|
|
raise ValueError()
|
|
def test_inter():
|
|
raise KeyboardInterrupt()
|
|
""")
|
|
result = testdir.runpytest(p1)
|
|
result.stdout.fnmatch_lines([
|
|
#"*test_inter() INTERRUPTED",
|
|
"*KEYBOARD INTERRUPT*",
|
|
"*1 failed*",
|
|
])
|
|
|