summaryrefslogtreecommitdiff
path: root/parts/django/tests/runtests.py
diff options
context:
space:
mode:
Diffstat (limited to 'parts/django/tests/runtests.py')
-rwxr-xr-xparts/django/tests/runtests.py336
1 files changed, 0 insertions, 336 deletions
diff --git a/parts/django/tests/runtests.py b/parts/django/tests/runtests.py
deleted file mode 100755
index 3dde214..0000000
--- a/parts/django/tests/runtests.py
+++ /dev/null
@@ -1,336 +0,0 @@
-#!/usr/bin/env python
-
-import os, subprocess, sys, traceback
-import unittest
-
-import django.contrib as contrib
-
-CONTRIB_DIR_NAME = 'django.contrib'
-MODEL_TESTS_DIR_NAME = 'modeltests'
-REGRESSION_TESTS_DIR_NAME = 'regressiontests'
-
-TEST_TEMPLATE_DIR = 'templates'
-
-CONTRIB_DIR = os.path.dirname(contrib.__file__)
-MODEL_TEST_DIR = os.path.join(os.path.dirname(__file__), MODEL_TESTS_DIR_NAME)
-REGRESSION_TEST_DIR = os.path.join(os.path.dirname(__file__), REGRESSION_TESTS_DIR_NAME)
-
-REGRESSION_SUBDIRS_TO_SKIP = ['locale']
-
-ALWAYS_INSTALLED_APPS = [
- 'django.contrib.contenttypes',
- 'django.contrib.auth',
- 'django.contrib.sites',
- 'django.contrib.flatpages',
- 'django.contrib.redirects',
- 'django.contrib.sessions',
- 'django.contrib.messages',
- 'django.contrib.comments',
- 'django.contrib.admin',
- 'django.contrib.admindocs',
-]
-
-def geodjango(settings):
- # All databases must have spatial backends to run GeoDjango tests.
- spatial_dbs = [name for name, db_dict in settings.DATABASES.items()
- if db_dict['ENGINE'].startswith('django.contrib.gis')]
- return len(spatial_dbs) == len(settings.DATABASES)
-
-def get_test_models():
- models = []
- for loc, dirpath in (MODEL_TESTS_DIR_NAME, MODEL_TEST_DIR), (REGRESSION_TESTS_DIR_NAME, REGRESSION_TEST_DIR), (CONTRIB_DIR_NAME, CONTRIB_DIR):
- for f in os.listdir(dirpath):
- if f.startswith('__init__') or f.startswith('.') or \
- f.startswith('sql') or f.startswith('invalid') or \
- os.path.basename(f) in REGRESSION_SUBDIRS_TO_SKIP:
- continue
- models.append((loc, f))
- return models
-
-def get_invalid_models():
- models = []
- for loc, dirpath in (MODEL_TESTS_DIR_NAME, MODEL_TEST_DIR), (REGRESSION_TESTS_DIR_NAME, REGRESSION_TEST_DIR), (CONTRIB_DIR_NAME, CONTRIB_DIR):
- for f in os.listdir(dirpath):
- if f.startswith('__init__') or f.startswith('.') or f.startswith('sql'):
- continue
- if f.startswith('invalid'):
- models.append((loc, f))
- return models
-
-class InvalidModelTestCase(unittest.TestCase):
- def __init__(self, model_label):
- unittest.TestCase.__init__(self)
- self.model_label = model_label
-
- def runTest(self):
- from django.core.management.validation import get_validation_errors
- from django.db.models.loading import load_app
- from cStringIO import StringIO
-
- try:
- module = load_app(self.model_label)
- except Exception, e:
- self.fail('Unable to load invalid model module')
-
- # Make sure sys.stdout is not a tty so that we get errors without
- # coloring attached (makes matching the results easier). We restore
- # sys.stderr afterwards.
- orig_stdout = sys.stdout
- s = StringIO()
- sys.stdout = s
- count = get_validation_errors(s, module)
- sys.stdout = orig_stdout
- s.seek(0)
- error_log = s.read()
- actual = error_log.split('\n')
- expected = module.model_errors.split('\n')
-
- unexpected = [err for err in actual if err not in expected]
- missing = [err for err in expected if err not in actual]
-
- self.assert_(not unexpected, "Unexpected Errors: " + '\n'.join(unexpected))
- self.assert_(not missing, "Missing Errors: " + '\n'.join(missing))
-
-def setup(verbosity, test_labels):
- from django.conf import settings
- state = {
- 'INSTALLED_APPS': settings.INSTALLED_APPS,
- 'ROOT_URLCONF': getattr(settings, "ROOT_URLCONF", ""),
- 'TEMPLATE_DIRS': settings.TEMPLATE_DIRS,
- 'USE_I18N': settings.USE_I18N,
- 'LOGIN_URL': settings.LOGIN_URL,
- 'LANGUAGE_CODE': settings.LANGUAGE_CODE,
- 'MIDDLEWARE_CLASSES': settings.MIDDLEWARE_CLASSES,
- }
-
- # Redirect some settings for the duration of these tests.
- settings.INSTALLED_APPS = ALWAYS_INSTALLED_APPS
- settings.ROOT_URLCONF = 'urls'
- settings.TEMPLATE_DIRS = (os.path.join(os.path.dirname(__file__), TEST_TEMPLATE_DIR),)
- settings.USE_I18N = True
- settings.LANGUAGE_CODE = 'en'
- settings.LOGIN_URL = '/accounts/login/'
- settings.MIDDLEWARE_CLASSES = (
- 'django.contrib.sessions.middleware.SessionMiddleware',
- 'django.contrib.auth.middleware.AuthenticationMiddleware',
- 'django.contrib.messages.middleware.MessageMiddleware',
- 'django.middleware.common.CommonMiddleware',
- )
- settings.SITE_ID = 1
- # For testing comment-utils, we require the MANAGERS attribute
- # to be set, so that a test email is sent out which we catch
- # in our tests.
- settings.MANAGERS = ("admin@djangoproject.com",)
-
- # Load all the ALWAYS_INSTALLED_APPS.
- # (This import statement is intentionally delayed until after we
- # access settings because of the USE_I18N dependency.)
- from django.db.models.loading import get_apps, load_app
- get_apps()
-
- # Load all the test model apps.
- test_labels_set = set([label.split('.')[0] for label in test_labels])
- test_models = get_test_models()
-
- # If GeoDjango, then we'll want to add in the test applications
- # that are a part of its test suite.
- if geodjango(settings):
- from django.contrib.gis.tests import geo_apps
- test_models.extend(geo_apps(runtests=True))
-
- for model_dir, model_name in test_models:
- model_label = '.'.join([model_dir, model_name])
- # if the model was named on the command line, or
- # no models were named (i.e., run all), import
- # this model and add it to the list to test.
- if not test_labels or model_name in test_labels_set:
- if verbosity >= 1:
- print "Importing model %s" % model_name
- mod = load_app(model_label)
- if mod:
- if model_label not in settings.INSTALLED_APPS:
- settings.INSTALLED_APPS.append(model_label)
-
- return state
-
-def teardown(state):
- from django.conf import settings
- # Restore the old settings.
- for key, value in state.items():
- setattr(settings, key, value)
-
-def django_tests(verbosity, interactive, failfast, test_labels):
- from django.conf import settings
- state = setup(verbosity, test_labels)
-
- # Add tests for invalid models.
- extra_tests = []
- for model_dir, model_name in get_invalid_models():
- model_label = '.'.join([model_dir, model_name])
- if not test_labels or model_name in test_labels:
- extra_tests.append(InvalidModelTestCase(model_label))
- try:
- # Invalid models are not working apps, so we cannot pass them into
- # the test runner with the other test_labels
- test_labels.remove(model_name)
- except ValueError:
- pass
-
- # If GeoDjango is used, add it's tests that aren't a part of
- # an application (e.g., GEOS, GDAL, Distance objects).
- if geodjango(settings):
- from django.contrib.gis.tests import geodjango_suite
- extra_tests.append(geodjango_suite(apps=False))
-
- # Run the test suite, including the extra validation tests.
- from django.test.utils import get_runner
- if not hasattr(settings, 'TEST_RUNNER'):
- settings.TEST_RUNNER = 'django.test.simple.DjangoTestSuiteRunner'
- TestRunner = get_runner(settings)
-
- if hasattr(TestRunner, 'func_name'):
- # Pre 1.2 test runners were just functions,
- # and did not support the 'failfast' option.
- import warnings
- warnings.warn(
- 'Function-based test runners are deprecated. Test runners should be classes with a run_tests() method.',
- PendingDeprecationWarning
- )
- failures = TestRunner(test_labels, verbosity=verbosity, interactive=interactive,
- extra_tests=extra_tests)
- else:
- test_runner = TestRunner(verbosity=verbosity, interactive=interactive, failfast=failfast)
- failures = test_runner.run_tests(test_labels, extra_tests=extra_tests)
-
- teardown(state)
- return failures
-
-
-def bisect_tests(bisection_label, options, test_labels):
- state = setup(int(options.verbosity), test_labels)
-
- if not test_labels:
- # Get the full list of test labels to use for bisection
- from django.db.models.loading import get_apps
- test_labels = [app.__name__.split('.')[-2] for app in get_apps()]
-
- print '***** Bisecting test suite:',' '.join(test_labels)
-
- # Make sure the bisection point isn't in the test list
- # Also remove tests that need to be run in specific combinations
- for label in [bisection_label, 'model_inheritance_same_model_name']:
- try:
- test_labels.remove(label)
- except ValueError:
- pass
-
- subprocess_args = ['python','runtests.py', '--settings=%s' % options.settings]
- if options.failfast:
- subprocess_args.append('--failfast')
- if options.verbosity:
- subprocess_args.append('--verbosity=%s' % options.verbosity)
- if not options.interactive:
- subprocess_args.append('--noinput')
-
- iteration = 1
- while len(test_labels) > 1:
- midpoint = len(test_labels)/2
- test_labels_a = test_labels[:midpoint] + [bisection_label]
- test_labels_b = test_labels[midpoint:] + [bisection_label]
- print '***** Pass %da: Running the first half of the test suite' % iteration
- print '***** Test labels:',' '.join(test_labels_a)
- failures_a = subprocess.call(subprocess_args + test_labels_a)
-
- print '***** Pass %db: Running the second half of the test suite' % iteration
- print '***** Test labels:',' '.join(test_labels_b)
- print
- failures_b = subprocess.call(subprocess_args + test_labels_b)
-
- if failures_a and not failures_b:
- print "***** Problem found in first half. Bisecting again..."
- iteration = iteration + 1
- test_labels = test_labels_a[:-1]
- elif failures_b and not failures_a:
- print "***** Problem found in second half. Bisecting again..."
- iteration = iteration + 1
- test_labels = test_labels_b[:-1]
- elif failures_a and failures_b:
- print "***** Multiple sources of failure found"
- break
- else:
- print "***** No source of failure found... try pair execution (--pair)"
- break
-
- if len(test_labels) == 1:
- print "***** Source of error:",test_labels[0]
- teardown(state)
-
-def paired_tests(paired_test, options, test_labels):
- state = setup(int(options.verbosity), test_labels)
-
- if not test_labels:
- print ""
- # Get the full list of test labels to use for bisection
- from django.db.models.loading import get_apps
- test_labels = [app.__name__.split('.')[-2] for app in get_apps()]
-
- print '***** Trying paired execution'
-
- # Make sure the bisection point isn't in the test list
- # Also remove tests that need to be run in specific combinations
- for label in [paired_test, 'model_inheritance_same_model_name']:
- try:
- test_labels.remove(label)
- except ValueError:
- pass
-
- subprocess_args = ['python','runtests.py', '--settings=%s' % options.settings]
- if options.failfast:
- subprocess_args.append('--failfast')
- if options.verbosity:
- subprocess_args.append('--verbosity=%s' % options.verbosity)
- if not options.interactive:
- subprocess_args.append('--noinput')
-
- for i, label in enumerate(test_labels):
- print '***** %d of %d: Check test pairing with %s' % (i+1, len(test_labels), label)
- failures = subprocess.call(subprocess_args + [label, paired_test])
- if failures:
- print '***** Found problem pair with',label
- return
-
- print '***** No problem pair found'
- teardown(state)
-
-if __name__ == "__main__":
- from optparse import OptionParser
- usage = "%prog [options] [model model model ...]"
- parser = OptionParser(usage=usage)
- parser.add_option('-v','--verbosity', action='store', dest='verbosity', default='0',
- type='choice', choices=['0', '1', '2'],
- help='Verbosity level; 0=minimal output, 1=normal output, 2=all output')
- parser.add_option('--noinput', action='store_false', dest='interactive', default=True,
- help='Tells Django to NOT prompt the user for input of any kind.')
- parser.add_option('--failfast', action='store_true', dest='failfast', default=False,
- help='Tells Django to stop running the test suite after first failed test.')
- parser.add_option('--settings',
- help='Python path to settings module, e.g. "myproject.settings". If this isn\'t provided, the DJANGO_SETTINGS_MODULE environment variable will be used.')
- parser.add_option('--bisect', action='store', dest='bisect', default=None,
- help="Bisect the test suite to discover a test that causes a test failure when combined with the named test.")
- parser.add_option('--pair', action='store', dest='pair', default=None,
- help="Run the test suite in pairs with the named test to find problem pairs.")
- options, args = parser.parse_args()
- if options.settings:
- os.environ['DJANGO_SETTINGS_MODULE'] = options.settings
- elif "DJANGO_SETTINGS_MODULE" not in os.environ:
- parser.error("DJANGO_SETTINGS_MODULE is not set in the environment. "
- "Set it or use --settings.")
-
- if options.bisect:
- bisect_tests(options.bisect, options, args)
- elif options.pair:
- paired_tests(options.pair, options, args)
- else:
- failures = django_tests(int(options.verbosity), options.interactive, options.failfast, args)
- if failures:
- sys.exit(bool(failures))