| # TestRun.py |
| # |
| # Copyright (C) 2011 Carlos Garcia Campos <carlosgc@gnome.org> |
| # |
| # This program is free software; you can redistribute it and/or modify |
| # it under the terms of the GNU General Public License as published by |
| # the Free Software Foundation; either version 2 of the License, or |
| # (at your option) any later version. |
| # |
| # This program is distributed in the hope that it will be useful, |
| # but WITHOUT ANY WARRANTY; without even the implied warranty of |
| # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| # GNU General Public License for more details. |
| # |
| # You should have received a copy of the GNU General Public License |
| # along with this program; if not, write to the Free Software |
| # Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA |
| |
| from backends import get_backend, get_all_backends |
| from Config import Config |
| from Utils import get_document_paths_from_dir, get_skipped_tests |
| import sys |
| import os |
| import errno |
| |
| class TestRun: |
| |
| def __init__(self, docsdir, refsdir, outdir): |
| self._docsdir = docsdir |
| self._refsdir = refsdir |
| self._outdir = outdir |
| self._skip = get_skipped_tests(docsdir) |
| self.config = Config() |
| |
| # Results |
| self._n_tests = 0 |
| self._n_passed = 0 |
| self._failed = [] |
| self._crashed = [] |
| self._failed_status_error = [] |
| self._stderr = [] |
| self._skipped = [] |
| |
| try: |
| os.makedirs(self._outdir); |
| except OSError as e: |
| if e.errno != errno.EEXIST: |
| raise |
| except: |
| raise |
| |
| def test(self, refs_path, doc_path, test_path, backend, n_doc, total_docs): |
| # First check whether there are test results for the backend |
| ref_has_md5 = backend.has_md5(refs_path) |
| ref_is_crashed = backend.is_crashed(refs_path) |
| ref_is_failed = backend.is_failed(refs_path) |
| if not ref_has_md5 and not ref_is_crashed and not ref_is_failed: |
| self._skipped.append("%s (%s)" % (doc_path, backend.get_name())) |
| print("Reference files not found, skipping '%s' for %s backend" % (doc_path, backend.get_name())) |
| return |
| |
| self._n_tests += 1 |
| sys.stdout.write("Testing '%s' using %s backend (%d/%d): " % (doc_path, backend.get_name(), n_doc, total_docs)) |
| sys.stdout.flush() |
| test_has_md5 = backend.create_refs(doc_path, test_path) |
| |
| if backend.has_stderr(test_path): |
| self._stderr.append("%s (%s)" % (doc_path, backend.get_name())) |
| |
| if ref_has_md5 and test_has_md5: |
| if backend.compare_checksums(refs_path, test_path, not self.config.keep_results, self.config.create_diffs, self.config.update_refs): |
| # FIXME: remove dir if it's empty? |
| print("PASS") |
| self._n_passed += 1 |
| else: |
| print("FAIL") |
| self._failed.append("%s (%s)" % (doc_path, backend.get_name())) |
| return |
| elif test_has_md5: |
| if ref_is_crashed: |
| print("DOES NOT CRASH") |
| elif ref_is_failed: |
| print("DOES NOT FAIL") |
| |
| return |
| |
| test_is_crashed = backend.is_crashed(test_path) |
| if ref_is_crashed and test_is_crashed: |
| print("PASS (Expected crash)") |
| self._n_passed += 1 |
| return |
| |
| test_is_failed = backend.is_failed(test_path) |
| if ref_is_failed and test_is_failed: |
| # FIXME: compare status errors |
| print("PASS (Expected fail with status error %d)" % (test_is_failed)) |
| self._n_passed += 1 |
| return |
| |
| if test_is_crashed: |
| print("CRASH") |
| self._crashed.append("%s (%s)" % (doc_path, backend.get_name())) |
| return |
| |
| if test_is_failed: |
| print("FAIL (status error %d)" % (test_is_failed)) |
| self._failed_status_error("%s (%s)" % (doc_path, backend.get_name())) |
| return |
| |
| def run_test(self, filename, n_doc = 1, total_docs = 1): |
| if filename in self._skip: |
| doc_path = os.path.join(self._docsdir, filename) |
| self._skipped.append("%s" % (doc_path)) |
| print("Skipping test '%s' (%d/%d)" % (doc_path, n_doc, total_docs)) |
| return |
| |
| out_path = os.path.join(self._outdir, filename) |
| try: |
| os.makedirs(out_path) |
| except OSError as e: |
| if e.errno != errno.EEXIST: |
| raise |
| except: |
| raise |
| doc_path = os.path.join(self._docsdir, filename) |
| refs_path = os.path.join(self._refsdir, filename) |
| |
| if not os.path.isdir(refs_path): |
| self._skipped.append("%s" % (doc_path)) |
| print("Reference dir not found for %s, skipping (%d/%d)" % (doc_path, n_doc, total_docs)) |
| return |
| |
| if self.config.backends: |
| backends = [get_backend(name) for name in self.config.backends] |
| else: |
| backends = get_all_backends() |
| |
| for backend in backends: |
| self.test(refs_path, doc_path, out_path, backend, n_doc, total_docs) |
| |
| def run_tests(self): |
| docs, total_docs = get_document_paths_from_dir(self._docsdir) |
| n_doc = 0 |
| for doc in docs: |
| n_doc += 1 |
| self.run_test(doc, n_doc, total_docs) |
| |
| def summary(self): |
| if not self._n_tests: |
| print("No tests run") |
| return |
| |
| print("Total %d tests" % (self._n_tests)) |
| print("%d tests passed (%.2f%%)" % (self._n_passed, (self._n_passed * 100.) / self._n_tests)) |
| def report_tests(test_list, test_type): |
| n_tests = len(test_list) |
| if not n_tests: |
| return |
| print("%d tests %s (%.2f%%): %s" % (n_tests, test_type, (n_tests * 100.) / self._n_tests, ", ".join(test_list))) |
| report_tests(self._failed, "failed") |
| report_tests(self._crashed, "crashed") |
| report_tests(self._failed_status_error, "failed to run") |
| report_tests(self._stderr, "have stderr output") |
| report_tests(self._skipped, "skipped") |
| |
| |