| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122 |
- import compileall
- import contextlib
- import filecmp
- import importlib.util
- import io
- import os
- import pathlib
- import py_compile
- import shutil
- import struct
- import sys
- import tempfile
- import test.test_importlib.util
- import time
- import unittest
- from unittest import mock, skipUnless
- try:
- # compileall relies on ProcessPoolExecutor if ProcessPoolExecutor exists
- # and it can function.
- from concurrent.futures import ProcessPoolExecutor
- from concurrent.futures.process import _check_system_limits
- _check_system_limits()
- _have_multiprocessing = True
- except (NotImplementedError, ModuleNotFoundError):
- _have_multiprocessing = False
- from test import support
- from test.support import os_helper
- from test.support import script_helper
- from test.test_py_compile import without_source_date_epoch
- from test.test_py_compile import SourceDateEpochTestMeta
- def get_pyc(script, opt):
- if not opt:
- # Replace None and 0 with ''
- opt = ''
- return importlib.util.cache_from_source(script, optimization=opt)
- def get_pycs(script):
- return [get_pyc(script, opt) for opt in (0, 1, 2)]
- def is_hardlink(filename1, filename2):
- """Returns True if two files have the same inode (hardlink)"""
- inode1 = os.stat(filename1).st_ino
- inode2 = os.stat(filename2).st_ino
- return inode1 == inode2
- class CompileallTestsBase:
- def setUp(self):
- self.directory = tempfile.mkdtemp()
- self.source_path = os.path.join(self.directory, '_test.py')
- self.bc_path = importlib.util.cache_from_source(self.source_path)
- with open(self.source_path, 'w', encoding="utf-8") as file:
- file.write('x = 123\n')
- self.source_path2 = os.path.join(self.directory, '_test2.py')
- self.bc_path2 = importlib.util.cache_from_source(self.source_path2)
- shutil.copyfile(self.source_path, self.source_path2)
- self.subdirectory = os.path.join(self.directory, '_subdir')
- os.mkdir(self.subdirectory)
- self.source_path3 = os.path.join(self.subdirectory, '_test3.py')
- shutil.copyfile(self.source_path, self.source_path3)
- def tearDown(self):
- shutil.rmtree(self.directory)
- def add_bad_source_file(self):
- self.bad_source_path = os.path.join(self.directory, '_test_bad.py')
- with open(self.bad_source_path, 'w', encoding="utf-8") as file:
- file.write('x (\n')
- def timestamp_metadata(self):
- with open(self.bc_path, 'rb') as file:
- data = file.read(12)
- mtime = int(os.stat(self.source_path).st_mtime)
- compare = struct.pack('<4sLL', importlib.util.MAGIC_NUMBER, 0,
- mtime & 0xFFFF_FFFF)
- return data, compare
- def test_year_2038_mtime_compilation(self):
- # Test to make sure we can handle mtimes larger than what a 32-bit
- # signed number can hold as part of bpo-34990
- try:
- os.utime(self.source_path, (2**32 - 1, 2**32 - 1))
- except (OverflowError, OSError):
- self.skipTest("filesystem doesn't support timestamps near 2**32")
- with contextlib.redirect_stdout(io.StringIO()):
- self.assertTrue(compileall.compile_file(self.source_path))
- def test_larger_than_32_bit_times(self):
- # This is similar to the test above but we skip it if the OS doesn't
- # support modification times larger than 32-bits.
- try:
- os.utime(self.source_path, (2**35, 2**35))
- except (OverflowError, OSError):
- self.skipTest("filesystem doesn't support large timestamps")
- with contextlib.redirect_stdout(io.StringIO()):
- self.assertTrue(compileall.compile_file(self.source_path))
- def recreation_check(self, metadata):
- """Check that compileall recreates bytecode when the new metadata is
- used."""
- if os.environ.get('SOURCE_DATE_EPOCH'):
- raise unittest.SkipTest('SOURCE_DATE_EPOCH is set')
- py_compile.compile(self.source_path)
- self.assertEqual(*self.timestamp_metadata())
- with open(self.bc_path, 'rb') as file:
- bc = file.read()[len(metadata):]
- with open(self.bc_path, 'wb') as file:
- file.write(metadata)
- file.write(bc)
- self.assertNotEqual(*self.timestamp_metadata())
- compileall.compile_dir(self.directory, force=False, quiet=True)
- self.assertTrue(*self.timestamp_metadata())
- def test_mtime(self):
- # Test a change in mtime leads to a new .pyc.
- self.recreation_check(struct.pack('<4sLL', importlib.util.MAGIC_NUMBER,
- 0, 1))
- def test_magic_number(self):
- # Test a change in mtime leads to a new .pyc.
- self.recreation_check(b'\0\0\0\0')
- def test_compile_files(self):
- # Test compiling a single file, and complete directory
- for fn in (self.bc_path, self.bc_path2):
- try:
- os.unlink(fn)
- except:
- pass
- self.assertTrue(compileall.compile_file(self.source_path,
- force=False, quiet=True))
- self.assertTrue(os.path.isfile(self.bc_path) and
- not os.path.isfile(self.bc_path2))
- os.unlink(self.bc_path)
- self.assertTrue(compileall.compile_dir(self.directory, force=False,
- quiet=True))
- self.assertTrue(os.path.isfile(self.bc_path) and
- os.path.isfile(self.bc_path2))
- os.unlink(self.bc_path)
- os.unlink(self.bc_path2)
- # Test against bad files
- self.add_bad_source_file()
- self.assertFalse(compileall.compile_file(self.bad_source_path,
- force=False, quiet=2))
- self.assertFalse(compileall.compile_dir(self.directory,
- force=False, quiet=2))
- def test_compile_file_pathlike(self):
- self.assertFalse(os.path.isfile(self.bc_path))
- # we should also test the output
- with support.captured_stdout() as stdout:
- self.assertTrue(compileall.compile_file(pathlib.Path(self.source_path)))
- self.assertRegex(stdout.getvalue(), r'Compiling ([^WindowsPath|PosixPath].*)')
- self.assertTrue(os.path.isfile(self.bc_path))
- def test_compile_file_pathlike_ddir(self):
- self.assertFalse(os.path.isfile(self.bc_path))
- self.assertTrue(compileall.compile_file(pathlib.Path(self.source_path),
- ddir=pathlib.Path('ddir_path'),
- quiet=2))
- self.assertTrue(os.path.isfile(self.bc_path))
- def test_compile_file_pathlike_stripdir(self):
- self.assertFalse(os.path.isfile(self.bc_path))
- self.assertTrue(compileall.compile_file(pathlib.Path(self.source_path),
- stripdir=pathlib.Path('stripdir_path'),
- quiet=2))
- self.assertTrue(os.path.isfile(self.bc_path))
- def test_compile_file_pathlike_prependdir(self):
- self.assertFalse(os.path.isfile(self.bc_path))
- self.assertTrue(compileall.compile_file(pathlib.Path(self.source_path),
- prependdir=pathlib.Path('prependdir_path'),
- quiet=2))
- self.assertTrue(os.path.isfile(self.bc_path))
- def test_compile_path(self):
- with test.test_importlib.util.import_state(path=[self.directory]):
- self.assertTrue(compileall.compile_path(quiet=2))
- with test.test_importlib.util.import_state(path=[self.directory]):
- self.add_bad_source_file()
- self.assertFalse(compileall.compile_path(skip_curdir=False,
- force=True, quiet=2))
- def test_no_pycache_in_non_package(self):
- # Bug 8563 reported that __pycache__ directories got created by
- # compile_file() for non-.py files.
- data_dir = os.path.join(self.directory, 'data')
- data_file = os.path.join(data_dir, 'file')
- os.mkdir(data_dir)
- # touch data/file
- with open(data_file, 'wb'):
- pass
- compileall.compile_file(data_file)
- self.assertFalse(os.path.exists(os.path.join(data_dir, '__pycache__')))
- def test_compile_file_encoding_fallback(self):
- # Bug 44666 reported that compile_file failed when sys.stdout.encoding is None
- self.add_bad_source_file()
- with contextlib.redirect_stdout(io.StringIO()):
- self.assertFalse(compileall.compile_file(self.bad_source_path))
- def test_optimize(self):
- # make sure compiling with different optimization settings than the
- # interpreter's creates the correct file names
- optimize, opt = (1, 1) if __debug__ else (0, '')
- compileall.compile_dir(self.directory, quiet=True, optimize=optimize)
- cached = importlib.util.cache_from_source(self.source_path,
- optimization=opt)
- self.assertTrue(os.path.isfile(cached))
- cached2 = importlib.util.cache_from_source(self.source_path2,
- optimization=opt)
- self.assertTrue(os.path.isfile(cached2))
- cached3 = importlib.util.cache_from_source(self.source_path3,
- optimization=opt)
- self.assertTrue(os.path.isfile(cached3))
- def test_compile_dir_pathlike(self):
- self.assertFalse(os.path.isfile(self.bc_path))
- with support.captured_stdout() as stdout:
- compileall.compile_dir(pathlib.Path(self.directory))
- line = stdout.getvalue().splitlines()[0]
- self.assertRegex(line, r'Listing ([^WindowsPath|PosixPath].*)')
- self.assertTrue(os.path.isfile(self.bc_path))
- def test_compile_dir_pathlike_stripdir(self):
- self.assertFalse(os.path.isfile(self.bc_path))
- self.assertTrue(compileall.compile_dir(pathlib.Path(self.directory),
- stripdir=pathlib.Path('stripdir_path'),
- quiet=2))
- self.assertTrue(os.path.isfile(self.bc_path))
- def test_compile_dir_pathlike_prependdir(self):
- self.assertFalse(os.path.isfile(self.bc_path))
- self.assertTrue(compileall.compile_dir(pathlib.Path(self.directory),
- prependdir=pathlib.Path('prependdir_path'),
- quiet=2))
- self.assertTrue(os.path.isfile(self.bc_path))
- @skipUnless(_have_multiprocessing, "requires multiprocessing")
- @mock.patch('concurrent.futures.ProcessPoolExecutor')
- def test_compile_pool_called(self, pool_mock):
- compileall.compile_dir(self.directory, quiet=True, workers=5)
- self.assertTrue(pool_mock.called)
- def test_compile_workers_non_positive(self):
- with self.assertRaisesRegex(ValueError,
- "workers must be greater or equal to 0"):
- compileall.compile_dir(self.directory, workers=-1)
- @skipUnless(_have_multiprocessing, "requires multiprocessing")
- @mock.patch('concurrent.futures.ProcessPoolExecutor')
- def test_compile_workers_cpu_count(self, pool_mock):
- compileall.compile_dir(self.directory, quiet=True, workers=0)
- self.assertEqual(pool_mock.call_args[1]['max_workers'], None)
- @skipUnless(_have_multiprocessing, "requires multiprocessing")
- @mock.patch('concurrent.futures.ProcessPoolExecutor')
- @mock.patch('compileall.compile_file')
- def test_compile_one_worker(self, compile_file_mock, pool_mock):
- compileall.compile_dir(self.directory, quiet=True)
- self.assertFalse(pool_mock.called)
- self.assertTrue(compile_file_mock.called)
- @skipUnless(_have_multiprocessing, "requires multiprocessing")
- @mock.patch('concurrent.futures.ProcessPoolExecutor', new=None)
- @mock.patch('compileall.compile_file')
- def test_compile_missing_multiprocessing(self, compile_file_mock):
- compileall.compile_dir(self.directory, quiet=True, workers=5)
- self.assertTrue(compile_file_mock.called)
- def test_compile_dir_maxlevels(self):
- # Test the actual impact of maxlevels parameter
- depth = 3
- path = self.directory
- for i in range(1, depth + 1):
- path = os.path.join(path, f"dir_{i}")
- source = os.path.join(path, 'script.py')
- os.mkdir(path)
- shutil.copyfile(self.source_path, source)
- pyc_filename = importlib.util.cache_from_source(source)
- compileall.compile_dir(self.directory, quiet=True, maxlevels=depth - 1)
- self.assertFalse(os.path.isfile(pyc_filename))
- compileall.compile_dir(self.directory, quiet=True, maxlevels=depth)
- self.assertTrue(os.path.isfile(pyc_filename))
- def _test_ddir_only(self, *, ddir, parallel=True):
- """Recursive compile_dir ddir must contain package paths; bpo39769."""
- fullpath = ["test", "foo"]
- path = self.directory
- mods = []
- for subdir in fullpath:
- path = os.path.join(path, subdir)
- os.mkdir(path)
- script_helper.make_script(path, "__init__", "")
- mods.append(script_helper.make_script(path, "mod",
- "def fn(): 1/0\nfn()\n"))
- compileall.compile_dir(
- self.directory, quiet=True, ddir=ddir,
- workers=2 if parallel else 1)
- self.assertTrue(mods)
- for mod in mods:
- self.assertTrue(mod.startswith(self.directory), mod)
- modcode = importlib.util.cache_from_source(mod)
- modpath = mod[len(self.directory+os.sep):]
- _, _, err = script_helper.assert_python_failure(modcode)
- expected_in = os.path.join(ddir, modpath)
- mod_code_obj = test.test_importlib.util.get_code_from_pyc(modcode)
- self.assertEqual(mod_code_obj.co_filename, expected_in)
- self.assertIn(f'"{expected_in}"', os.fsdecode(err))
- def test_ddir_only_one_worker(self):
- """Recursive compile_dir ddir= contains package paths; bpo39769."""
- return self._test_ddir_only(ddir="<a prefix>", parallel=False)
- @skipUnless(_have_multiprocessing, "requires multiprocessing")
- def test_ddir_multiple_workers(self):
- """Recursive compile_dir ddir= contains package paths; bpo39769."""
- return self._test_ddir_only(ddir="<a prefix>", parallel=True)
- def test_ddir_empty_only_one_worker(self):
- """Recursive compile_dir ddir='' contains package paths; bpo39769."""
- return self._test_ddir_only(ddir="", parallel=False)
- @skipUnless(_have_multiprocessing, "requires multiprocessing")
- def test_ddir_empty_multiple_workers(self):
- """Recursive compile_dir ddir='' contains package paths; bpo39769."""
- return self._test_ddir_only(ddir="", parallel=True)
- def test_strip_only(self):
- fullpath = ["test", "build", "real", "path"]
- path = os.path.join(self.directory, *fullpath)
- os.makedirs(path)
- script = script_helper.make_script(path, "test", "1 / 0")
- bc = importlib.util.cache_from_source(script)
- stripdir = os.path.join(self.directory, *fullpath[:2])
- compileall.compile_dir(path, quiet=True, stripdir=stripdir)
- rc, out, err = script_helper.assert_python_failure(bc)
- expected_in = os.path.join(*fullpath[2:])
- self.assertIn(
- expected_in,
- str(err, encoding=sys.getdefaultencoding())
- )
- self.assertNotIn(
- stripdir,
- str(err, encoding=sys.getdefaultencoding())
- )
- def test_prepend_only(self):
- fullpath = ["test", "build", "real", "path"]
- path = os.path.join(self.directory, *fullpath)
- os.makedirs(path)
- script = script_helper.make_script(path, "test", "1 / 0")
- bc = importlib.util.cache_from_source(script)
- prependdir = "/foo"
- compileall.compile_dir(path, quiet=True, prependdir=prependdir)
- rc, out, err = script_helper.assert_python_failure(bc)
- expected_in = os.path.join(prependdir, self.directory, *fullpath)
- self.assertIn(
- expected_in,
- str(err, encoding=sys.getdefaultencoding())
- )
- def test_strip_and_prepend(self):
- fullpath = ["test", "build", "real", "path"]
- path = os.path.join(self.directory, *fullpath)
- os.makedirs(path)
- script = script_helper.make_script(path, "test", "1 / 0")
- bc = importlib.util.cache_from_source(script)
- stripdir = os.path.join(self.directory, *fullpath[:2])
- prependdir = "/foo"
- compileall.compile_dir(path, quiet=True,
- stripdir=stripdir, prependdir=prependdir)
- rc, out, err = script_helper.assert_python_failure(bc)
- expected_in = os.path.join(prependdir, *fullpath[2:])
- self.assertIn(
- expected_in,
- str(err, encoding=sys.getdefaultencoding())
- )
- self.assertNotIn(
- stripdir,
- str(err, encoding=sys.getdefaultencoding())
- )
- def test_strip_prepend_and_ddir(self):
- fullpath = ["test", "build", "real", "path", "ddir"]
- path = os.path.join(self.directory, *fullpath)
- os.makedirs(path)
- script_helper.make_script(path, "test", "1 / 0")
- with self.assertRaises(ValueError):
- compileall.compile_dir(path, quiet=True, ddir="/bar",
- stripdir="/foo", prependdir="/bar")
- def test_multiple_optimization_levels(self):
- script = script_helper.make_script(self.directory,
- "test_optimization",
- "a = 0")
- bc = []
- for opt_level in "", 1, 2, 3:
- bc.append(importlib.util.cache_from_source(script,
- optimization=opt_level))
- test_combinations = [[0, 1], [1, 2], [0, 2], [0, 1, 2]]
- for opt_combination in test_combinations:
- compileall.compile_file(script, quiet=True,
- optimize=opt_combination)
- for opt_level in opt_combination:
- self.assertTrue(os.path.isfile(bc[opt_level]))
- try:
- os.unlink(bc[opt_level])
- except Exception:
- pass
- @os_helper.skip_unless_symlink
- def test_ignore_symlink_destination(self):
- # Create folders for allowed files, symlinks and prohibited area
- allowed_path = os.path.join(self.directory, "test", "dir", "allowed")
- symlinks_path = os.path.join(self.directory, "test", "dir", "symlinks")
- prohibited_path = os.path.join(self.directory, "test", "dir", "prohibited")
- os.makedirs(allowed_path)
- os.makedirs(symlinks_path)
- os.makedirs(prohibited_path)
- # Create scripts and symlinks and remember their byte-compiled versions
- allowed_script = script_helper.make_script(allowed_path, "test_allowed", "a = 0")
- prohibited_script = script_helper.make_script(prohibited_path, "test_prohibited", "a = 0")
- allowed_symlink = os.path.join(symlinks_path, "test_allowed.py")
- prohibited_symlink = os.path.join(symlinks_path, "test_prohibited.py")
- os.symlink(allowed_script, allowed_symlink)
- os.symlink(prohibited_script, prohibited_symlink)
- allowed_bc = importlib.util.cache_from_source(allowed_symlink)
- prohibited_bc = importlib.util.cache_from_source(prohibited_symlink)
- compileall.compile_dir(symlinks_path, quiet=True, limit_sl_dest=allowed_path)
- self.assertTrue(os.path.isfile(allowed_bc))
- self.assertFalse(os.path.isfile(prohibited_bc))
- class CompileallTestsWithSourceEpoch(CompileallTestsBase,
- unittest.TestCase,
- metaclass=SourceDateEpochTestMeta,
- source_date_epoch=True):
- pass
- class CompileallTestsWithoutSourceEpoch(CompileallTestsBase,
- unittest.TestCase,
- metaclass=SourceDateEpochTestMeta,
- source_date_epoch=False):
- pass
- # WASI does not have a temp directory and uses cwd instead. The cwd contains
- # non-ASCII chars, so _walk_dir() fails to encode self.directory.
- @unittest.skipIf(support.is_wasi, "tempdir is not encodable on WASI")
- class EncodingTest(unittest.TestCase):
- """Issue 6716: compileall should escape source code when printing errors
- to stdout."""
- def setUp(self):
- self.directory = tempfile.mkdtemp()
- self.source_path = os.path.join(self.directory, '_test.py')
- with open(self.source_path, 'w', encoding='utf-8') as file:
- file.write('# -*- coding: utf-8 -*-\n')
- file.write('print u"\u20ac"\n')
- def tearDown(self):
- shutil.rmtree(self.directory)
- def test_error(self):
- try:
- orig_stdout = sys.stdout
- sys.stdout = io.TextIOWrapper(io.BytesIO(),encoding='ascii')
- compileall.compile_dir(self.directory)
- finally:
- sys.stdout = orig_stdout
- class CommandLineTestsBase:
- """Test compileall's CLI."""
- def setUp(self):
- self.directory = tempfile.mkdtemp()
- self.addCleanup(os_helper.rmtree, self.directory)
- self.pkgdir = os.path.join(self.directory, 'foo')
- os.mkdir(self.pkgdir)
- self.pkgdir_cachedir = os.path.join(self.pkgdir, '__pycache__')
- # Create the __init__.py and a package module.
- self.initfn = script_helper.make_script(self.pkgdir, '__init__', '')
- self.barfn = script_helper.make_script(self.pkgdir, 'bar', '')
- @contextlib.contextmanager
- def temporary_pycache_prefix(self):
- """Adjust and restore sys.pycache_prefix."""
- old_prefix = sys.pycache_prefix
- new_prefix = os.path.join(self.directory, '__testcache__')
- try:
- sys.pycache_prefix = new_prefix
- yield {
- 'PYTHONPATH': self.directory,
- 'PYTHONPYCACHEPREFIX': new_prefix,
- }
- finally:
- sys.pycache_prefix = old_prefix
- def _get_run_args(self, args):
- return [*support.optim_args_from_interpreter_flags(),
- '-S', '-m', 'compileall',
- *args]
- def assertRunOK(self, *args, **env_vars):
- rc, out, err = script_helper.assert_python_ok(
- *self._get_run_args(args), **env_vars,
- PYTHONIOENCODING='utf-8')
- self.assertEqual(b'', err)
- return out
- def assertRunNotOK(self, *args, **env_vars):
- rc, out, err = script_helper.assert_python_failure(
- *self._get_run_args(args), **env_vars,
- PYTHONIOENCODING='utf-8')
- return rc, out, err
- def assertCompiled(self, fn):
- path = importlib.util.cache_from_source(fn)
- self.assertTrue(os.path.exists(path))
- def assertNotCompiled(self, fn):
- path = importlib.util.cache_from_source(fn)
- self.assertFalse(os.path.exists(path))
- def test_no_args_compiles_path(self):
- # Note that -l is implied for the no args case.
- bazfn = script_helper.make_script(self.directory, 'baz', '')
- with self.temporary_pycache_prefix() as env:
- self.assertRunOK(**env)
- self.assertCompiled(bazfn)
- self.assertNotCompiled(self.initfn)
- self.assertNotCompiled(self.barfn)
- @without_source_date_epoch # timestamp invalidation test
- def test_no_args_respects_force_flag(self):
- bazfn = script_helper.make_script(self.directory, 'baz', '')
- with self.temporary_pycache_prefix() as env:
- self.assertRunOK(**env)
- pycpath = importlib.util.cache_from_source(bazfn)
- # Set atime/mtime backward to avoid file timestamp resolution issues
- os.utime(pycpath, (time.time()-60,)*2)
- mtime = os.stat(pycpath).st_mtime
- # Without force, no recompilation
- self.assertRunOK(**env)
- mtime2 = os.stat(pycpath).st_mtime
- self.assertEqual(mtime, mtime2)
- # Now force it.
- self.assertRunOK('-f', **env)
- mtime2 = os.stat(pycpath).st_mtime
- self.assertNotEqual(mtime, mtime2)
- def test_no_args_respects_quiet_flag(self):
- script_helper.make_script(self.directory, 'baz', '')
- with self.temporary_pycache_prefix() as env:
- noisy = self.assertRunOK(**env)
- self.assertIn(b'Listing ', noisy)
- quiet = self.assertRunOK('-q', **env)
- self.assertNotIn(b'Listing ', quiet)
- # Ensure that the default behavior of compileall's CLI is to create
- # PEP 3147/PEP 488 pyc files.
- for name, ext, switch in [
- ('normal', 'pyc', []),
- ('optimize', 'opt-1.pyc', ['-O']),
- ('doubleoptimize', 'opt-2.pyc', ['-OO']),
- ]:
- def f(self, ext=ext, switch=switch):
- script_helper.assert_python_ok(*(switch +
- ['-m', 'compileall', '-q', self.pkgdir]))
- # Verify the __pycache__ directory contents.
- self.assertTrue(os.path.exists(self.pkgdir_cachedir))
- expected = sorted(base.format(sys.implementation.cache_tag, ext)
- for base in ('__init__.{}.{}', 'bar.{}.{}'))
- self.assertEqual(sorted(os.listdir(self.pkgdir_cachedir)), expected)
- # Make sure there are no .pyc files in the source directory.
- self.assertFalse([fn for fn in os.listdir(self.pkgdir)
- if fn.endswith(ext)])
- locals()['test_pep3147_paths_' + name] = f
- def test_legacy_paths(self):
- # Ensure that with the proper switch, compileall leaves legacy
- # pyc files, and no __pycache__ directory.
- self.assertRunOK('-b', '-q', self.pkgdir)
- # Verify the __pycache__ directory contents.
- self.assertFalse(os.path.exists(self.pkgdir_cachedir))
- expected = sorted(['__init__.py', '__init__.pyc', 'bar.py',
- 'bar.pyc'])
- self.assertEqual(sorted(os.listdir(self.pkgdir)), expected)
- def test_multiple_runs(self):
- # Bug 8527 reported that multiple calls produced empty
- # __pycache__/__pycache__ directories.
- self.assertRunOK('-q', self.pkgdir)
- # Verify the __pycache__ directory contents.
- self.assertTrue(os.path.exists(self.pkgdir_cachedir))
- cachecachedir = os.path.join(self.pkgdir_cachedir, '__pycache__')
- self.assertFalse(os.path.exists(cachecachedir))
- # Call compileall again.
- self.assertRunOK('-q', self.pkgdir)
- self.assertTrue(os.path.exists(self.pkgdir_cachedir))
- self.assertFalse(os.path.exists(cachecachedir))
- @without_source_date_epoch # timestamp invalidation test
- def test_force(self):
- self.assertRunOK('-q', self.pkgdir)
- pycpath = importlib.util.cache_from_source(self.barfn)
- # set atime/mtime backward to avoid file timestamp resolution issues
- os.utime(pycpath, (time.time()-60,)*2)
- mtime = os.stat(pycpath).st_mtime
- # without force, no recompilation
- self.assertRunOK('-q', self.pkgdir)
- mtime2 = os.stat(pycpath).st_mtime
- self.assertEqual(mtime, mtime2)
- # now force it.
- self.assertRunOK('-q', '-f', self.pkgdir)
- mtime2 = os.stat(pycpath).st_mtime
- self.assertNotEqual(mtime, mtime2)
- def test_recursion_control(self):
- subpackage = os.path.join(self.pkgdir, 'spam')
- os.mkdir(subpackage)
- subinitfn = script_helper.make_script(subpackage, '__init__', '')
- hamfn = script_helper.make_script(subpackage, 'ham', '')
- self.assertRunOK('-q', '-l', self.pkgdir)
- self.assertNotCompiled(subinitfn)
- self.assertFalse(os.path.exists(os.path.join(subpackage, '__pycache__')))
- self.assertRunOK('-q', self.pkgdir)
- self.assertCompiled(subinitfn)
- self.assertCompiled(hamfn)
- def test_recursion_limit(self):
- subpackage = os.path.join(self.pkgdir, 'spam')
- subpackage2 = os.path.join(subpackage, 'ham')
- subpackage3 = os.path.join(subpackage2, 'eggs')
- for pkg in (subpackage, subpackage2, subpackage3):
- script_helper.make_pkg(pkg)
- subinitfn = os.path.join(subpackage, '__init__.py')
- hamfn = script_helper.make_script(subpackage, 'ham', '')
- spamfn = script_helper.make_script(subpackage2, 'spam', '')
- eggfn = script_helper.make_script(subpackage3, 'egg', '')
- self.assertRunOK('-q', '-r 0', self.pkgdir)
- self.assertNotCompiled(subinitfn)
- self.assertFalse(
- os.path.exists(os.path.join(subpackage, '__pycache__')))
- self.assertRunOK('-q', '-r 1', self.pkgdir)
- self.assertCompiled(subinitfn)
- self.assertCompiled(hamfn)
- self.assertNotCompiled(spamfn)
- self.assertRunOK('-q', '-r 2', self.pkgdir)
- self.assertCompiled(subinitfn)
- self.assertCompiled(hamfn)
- self.assertCompiled(spamfn)
- self.assertNotCompiled(eggfn)
- self.assertRunOK('-q', '-r 5', self.pkgdir)
- self.assertCompiled(subinitfn)
- self.assertCompiled(hamfn)
- self.assertCompiled(spamfn)
- self.assertCompiled(eggfn)
- @os_helper.skip_unless_symlink
- def test_symlink_loop(self):
- # Currently, compileall ignores symlinks to directories.
- # If that limitation is ever lifted, it should protect against
- # recursion in symlink loops.
- pkg = os.path.join(self.pkgdir, 'spam')
- script_helper.make_pkg(pkg)
- os.symlink('.', os.path.join(pkg, 'evil'))
- os.symlink('.', os.path.join(pkg, 'evil2'))
- self.assertRunOK('-q', self.pkgdir)
- self.assertCompiled(os.path.join(
- self.pkgdir, 'spam', 'evil', 'evil2', '__init__.py'
- ))
- def test_quiet(self):
- noisy = self.assertRunOK(self.pkgdir)
- quiet = self.assertRunOK('-q', self.pkgdir)
- self.assertNotEqual(b'', noisy)
- self.assertEqual(b'', quiet)
- def test_silent(self):
- script_helper.make_script(self.pkgdir, 'crunchyfrog', 'bad(syntax')
- _, quiet, _ = self.assertRunNotOK('-q', self.pkgdir)
- _, silent, _ = self.assertRunNotOK('-qq', self.pkgdir)
- self.assertNotEqual(b'', quiet)
- self.assertEqual(b'', silent)
- def test_regexp(self):
- self.assertRunOK('-q', '-x', r'ba[^\\/]*$', self.pkgdir)
- self.assertNotCompiled(self.barfn)
- self.assertCompiled(self.initfn)
- def test_multiple_dirs(self):
- pkgdir2 = os.path.join(self.directory, 'foo2')
- os.mkdir(pkgdir2)
- init2fn = script_helper.make_script(pkgdir2, '__init__', '')
- bar2fn = script_helper.make_script(pkgdir2, 'bar2', '')
- self.assertRunOK('-q', self.pkgdir, pkgdir2)
- self.assertCompiled(self.initfn)
- self.assertCompiled(self.barfn)
- self.assertCompiled(init2fn)
- self.assertCompiled(bar2fn)
- def test_d_compile_error(self):
- script_helper.make_script(self.pkgdir, 'crunchyfrog', 'bad(syntax')
- rc, out, err = self.assertRunNotOK('-q', '-d', 'dinsdale', self.pkgdir)
- self.assertRegex(out, b'File "dinsdale')
- def test_d_runtime_error(self):
- bazfn = script_helper.make_script(self.pkgdir, 'baz', 'raise Exception')
- self.assertRunOK('-q', '-d', 'dinsdale', self.pkgdir)
- fn = script_helper.make_script(self.pkgdir, 'bing', 'import baz')
- pyc = importlib.util.cache_from_source(bazfn)
- os.rename(pyc, os.path.join(self.pkgdir, 'baz.pyc'))
- os.remove(bazfn)
- rc, out, err = script_helper.assert_python_failure(fn, __isolated=False)
- self.assertRegex(err, b'File "dinsdale')
- def test_include_bad_file(self):
- rc, out, err = self.assertRunNotOK(
- '-i', os.path.join(self.directory, 'nosuchfile'), self.pkgdir)
- self.assertRegex(out, b'rror.*nosuchfile')
- self.assertNotRegex(err, b'Traceback')
- self.assertFalse(os.path.exists(importlib.util.cache_from_source(
- self.pkgdir_cachedir)))
- def test_include_file_with_arg(self):
- f1 = script_helper.make_script(self.pkgdir, 'f1', '')
- f2 = script_helper.make_script(self.pkgdir, 'f2', '')
- f3 = script_helper.make_script(self.pkgdir, 'f3', '')
- f4 = script_helper.make_script(self.pkgdir, 'f4', '')
- with open(os.path.join(self.directory, 'l1'), 'w', encoding="utf-8") as l1:
- l1.write(os.path.join(self.pkgdir, 'f1.py')+os.linesep)
- l1.write(os.path.join(self.pkgdir, 'f2.py')+os.linesep)
- self.assertRunOK('-i', os.path.join(self.directory, 'l1'), f4)
- self.assertCompiled(f1)
- self.assertCompiled(f2)
- self.assertNotCompiled(f3)
- self.assertCompiled(f4)
- def test_include_file_no_arg(self):
- f1 = script_helper.make_script(self.pkgdir, 'f1', '')
- f2 = script_helper.make_script(self.pkgdir, 'f2', '')
- f3 = script_helper.make_script(self.pkgdir, 'f3', '')
- f4 = script_helper.make_script(self.pkgdir, 'f4', '')
- with open(os.path.join(self.directory, 'l1'), 'w', encoding="utf-8") as l1:
- l1.write(os.path.join(self.pkgdir, 'f2.py')+os.linesep)
- self.assertRunOK('-i', os.path.join(self.directory, 'l1'))
- self.assertNotCompiled(f1)
- self.assertCompiled(f2)
- self.assertNotCompiled(f3)
- self.assertNotCompiled(f4)
- def test_include_on_stdin(self):
- f1 = script_helper.make_script(self.pkgdir, 'f1', '')
- f2 = script_helper.make_script(self.pkgdir, 'f2', '')
- f3 = script_helper.make_script(self.pkgdir, 'f3', '')
- f4 = script_helper.make_script(self.pkgdir, 'f4', '')
- p = script_helper.spawn_python(*(self._get_run_args(()) + ['-i', '-']))
- p.stdin.write((f3+os.linesep).encode('ascii'))
- script_helper.kill_python(p)
- self.assertNotCompiled(f1)
- self.assertNotCompiled(f2)
- self.assertCompiled(f3)
- self.assertNotCompiled(f4)
- def test_compiles_as_much_as_possible(self):
- bingfn = script_helper.make_script(self.pkgdir, 'bing', 'syntax(error')
- rc, out, err = self.assertRunNotOK('nosuchfile', self.initfn,
- bingfn, self.barfn)
- self.assertRegex(out, b'rror')
- self.assertNotCompiled(bingfn)
- self.assertCompiled(self.initfn)
- self.assertCompiled(self.barfn)
- def test_invalid_arg_produces_message(self):
- out = self.assertRunOK('badfilename')
- self.assertRegex(out, b"Can't list 'badfilename'")
- def test_pyc_invalidation_mode(self):
- script_helper.make_script(self.pkgdir, 'f1', '')
- pyc = importlib.util.cache_from_source(
- os.path.join(self.pkgdir, 'f1.py'))
- self.assertRunOK('--invalidation-mode=checked-hash', self.pkgdir)
- with open(pyc, 'rb') as fp:
- data = fp.read()
- self.assertEqual(int.from_bytes(data[4:8], 'little'), 0b11)
- self.assertRunOK('--invalidation-mode=unchecked-hash', self.pkgdir)
- with open(pyc, 'rb') as fp:
- data = fp.read()
- self.assertEqual(int.from_bytes(data[4:8], 'little'), 0b01)
- @skipUnless(_have_multiprocessing, "requires multiprocessing")
- def test_workers(self):
- bar2fn = script_helper.make_script(self.directory, 'bar2', '')
- files = []
- for suffix in range(5):
- pkgdir = os.path.join(self.directory, 'foo{}'.format(suffix))
- os.mkdir(pkgdir)
- fn = script_helper.make_script(pkgdir, '__init__', '')
- files.append(script_helper.make_script(pkgdir, 'bar2', ''))
- self.assertRunOK(self.directory, '-j', '0')
- self.assertCompiled(bar2fn)
- for file in files:
- self.assertCompiled(file)
- @mock.patch('compileall.compile_dir')
- def test_workers_available_cores(self, compile_dir):
- with mock.patch("sys.argv",
- new=[sys.executable, self.directory, "-j0"]):
- compileall.main()
- self.assertTrue(compile_dir.called)
- self.assertEqual(compile_dir.call_args[-1]['workers'], 0)
- def test_strip_and_prepend(self):
- fullpath = ["test", "build", "real", "path"]
- path = os.path.join(self.directory, *fullpath)
- os.makedirs(path)
- script = script_helper.make_script(path, "test", "1 / 0")
- bc = importlib.util.cache_from_source(script)
- stripdir = os.path.join(self.directory, *fullpath[:2])
- prependdir = "/foo"
- self.assertRunOK("-s", stripdir, "-p", prependdir, path)
- rc, out, err = script_helper.assert_python_failure(bc)
- expected_in = os.path.join(prependdir, *fullpath[2:])
- self.assertIn(
- expected_in,
- str(err, encoding=sys.getdefaultencoding())
- )
- self.assertNotIn(
- stripdir,
- str(err, encoding=sys.getdefaultencoding())
- )
- def test_multiple_optimization_levels(self):
- path = os.path.join(self.directory, "optimizations")
- os.makedirs(path)
- script = script_helper.make_script(path,
- "test_optimization",
- "a = 0")
- bc = []
- for opt_level in "", 1, 2, 3:
- bc.append(importlib.util.cache_from_source(script,
- optimization=opt_level))
- test_combinations = [["0", "1"],
- ["1", "2"],
- ["0", "2"],
- ["0", "1", "2"]]
- for opt_combination in test_combinations:
- self.assertRunOK(path, *("-o" + str(n) for n in opt_combination))
- for opt_level in opt_combination:
- self.assertTrue(os.path.isfile(bc[int(opt_level)]))
- try:
- os.unlink(bc[opt_level])
- except Exception:
- pass
- @os_helper.skip_unless_symlink
- def test_ignore_symlink_destination(self):
- # Create folders for allowed files, symlinks and prohibited area
- allowed_path = os.path.join(self.directory, "test", "dir", "allowed")
- symlinks_path = os.path.join(self.directory, "test", "dir", "symlinks")
- prohibited_path = os.path.join(self.directory, "test", "dir", "prohibited")
- os.makedirs(allowed_path)
- os.makedirs(symlinks_path)
- os.makedirs(prohibited_path)
- # Create scripts and symlinks and remember their byte-compiled versions
- allowed_script = script_helper.make_script(allowed_path, "test_allowed", "a = 0")
- prohibited_script = script_helper.make_script(prohibited_path, "test_prohibited", "a = 0")
- allowed_symlink = os.path.join(symlinks_path, "test_allowed.py")
- prohibited_symlink = os.path.join(symlinks_path, "test_prohibited.py")
- os.symlink(allowed_script, allowed_symlink)
- os.symlink(prohibited_script, prohibited_symlink)
- allowed_bc = importlib.util.cache_from_source(allowed_symlink)
- prohibited_bc = importlib.util.cache_from_source(prohibited_symlink)
- self.assertRunOK(symlinks_path, "-e", allowed_path)
- self.assertTrue(os.path.isfile(allowed_bc))
- self.assertFalse(os.path.isfile(prohibited_bc))
- def test_hardlink_bad_args(self):
- # Bad arguments combination, hardlink deduplication make sense
- # only for more than one optimization level
- self.assertRunNotOK(self.directory, "-o 1", "--hardlink-dupes")
- def test_hardlink(self):
- # 'a = 0' code produces the same bytecode for the 3 optimization
- # levels. All three .pyc files must have the same inode (hardlinks).
- #
- # If deduplication is disabled, all pyc files must have different
- # inodes.
- for dedup in (True, False):
- with tempfile.TemporaryDirectory() as path:
- with self.subTest(dedup=dedup):
- script = script_helper.make_script(path, "script", "a = 0")
- pycs = get_pycs(script)
- args = ["-q", "-o 0", "-o 1", "-o 2"]
- if dedup:
- args.append("--hardlink-dupes")
- self.assertRunOK(path, *args)
- self.assertEqual(is_hardlink(pycs[0], pycs[1]), dedup)
- self.assertEqual(is_hardlink(pycs[1], pycs[2]), dedup)
- self.assertEqual(is_hardlink(pycs[0], pycs[2]), dedup)
- class CommandLineTestsWithSourceEpoch(CommandLineTestsBase,
- unittest.TestCase,
- metaclass=SourceDateEpochTestMeta,
- source_date_epoch=True):
- pass
- class CommandLineTestsNoSourceEpoch(CommandLineTestsBase,
- unittest.TestCase,
- metaclass=SourceDateEpochTestMeta,
- source_date_epoch=False):
- pass
- @unittest.skipUnless(hasattr(os, 'link'), 'requires os.link')
- class HardlinkDedupTestsBase:
- # Test hardlink_dupes parameter of compileall.compile_dir()
- def setUp(self):
- self.path = None
- @contextlib.contextmanager
- def temporary_directory(self):
- with tempfile.TemporaryDirectory() as path:
- self.path = path
- yield path
- self.path = None
- def make_script(self, code, name="script"):
- return script_helper.make_script(self.path, name, code)
- def compile_dir(self, *, dedup=True, optimize=(0, 1, 2), force=False):
- compileall.compile_dir(self.path, quiet=True, optimize=optimize,
- hardlink_dupes=dedup, force=force)
- def test_bad_args(self):
- # Bad arguments combination, hardlink deduplication make sense
- # only for more than one optimization level
- with self.temporary_directory():
- self.make_script("pass")
- with self.assertRaises(ValueError):
- compileall.compile_dir(self.path, quiet=True, optimize=0,
- hardlink_dupes=True)
- with self.assertRaises(ValueError):
- # same optimization level specified twice:
- # compile_dir() removes duplicates
- compileall.compile_dir(self.path, quiet=True, optimize=[0, 0],
- hardlink_dupes=True)
- def create_code(self, docstring=False, assertion=False):
- lines = []
- if docstring:
- lines.append("'module docstring'")
- lines.append('x = 1')
- if assertion:
- lines.append("assert x == 1")
- return '\n'.join(lines)
- def iter_codes(self):
- for docstring in (False, True):
- for assertion in (False, True):
- code = self.create_code(docstring=docstring, assertion=assertion)
- yield (code, docstring, assertion)
- def test_disabled(self):
- # Deduplication disabled, no hardlinks
- for code, docstring, assertion in self.iter_codes():
- with self.subTest(docstring=docstring, assertion=assertion):
- with self.temporary_directory():
- script = self.make_script(code)
- pycs = get_pycs(script)
- self.compile_dir(dedup=False)
- self.assertFalse(is_hardlink(pycs[0], pycs[1]))
- self.assertFalse(is_hardlink(pycs[0], pycs[2]))
- self.assertFalse(is_hardlink(pycs[1], pycs[2]))
- def check_hardlinks(self, script, docstring=False, assertion=False):
- pycs = get_pycs(script)
- self.assertEqual(is_hardlink(pycs[0], pycs[1]),
- not assertion)
- self.assertEqual(is_hardlink(pycs[0], pycs[2]),
- not assertion and not docstring)
- self.assertEqual(is_hardlink(pycs[1], pycs[2]),
- not docstring)
- def test_hardlink(self):
- # Test deduplication on all combinations
- for code, docstring, assertion in self.iter_codes():
- with self.subTest(docstring=docstring, assertion=assertion):
- with self.temporary_directory():
- script = self.make_script(code)
- self.compile_dir()
- self.check_hardlinks(script, docstring, assertion)
- def test_only_two_levels(self):
- # Don't build the 3 optimization levels, but only 2
- for opts in ((0, 1), (1, 2), (0, 2)):
- with self.subTest(opts=opts):
- with self.temporary_directory():
- # code with no dostring and no assertion:
- # same bytecode for all optimization levels
- script = self.make_script(self.create_code())
- self.compile_dir(optimize=opts)
- pyc1 = get_pyc(script, opts[0])
- pyc2 = get_pyc(script, opts[1])
- self.assertTrue(is_hardlink(pyc1, pyc2))
- def test_duplicated_levels(self):
- # compile_dir() must not fail if optimize contains duplicated
- # optimization levels and/or if optimization levels are not sorted.
- with self.temporary_directory():
- # code with no dostring and no assertion:
- # same bytecode for all optimization levels
- script = self.make_script(self.create_code())
- self.compile_dir(optimize=[1, 0, 1, 0])
- pyc1 = get_pyc(script, 0)
- pyc2 = get_pyc(script, 1)
- self.assertTrue(is_hardlink(pyc1, pyc2))
- def test_recompilation(self):
- # Test compile_dir() when pyc files already exists and the script
- # content changed
- with self.temporary_directory():
- script = self.make_script("a = 0")
- self.compile_dir()
- # All three levels have the same inode
- self.check_hardlinks(script)
- pycs = get_pycs(script)
- inode = os.stat(pycs[0]).st_ino
- # Change of the module content
- script = self.make_script("print(0)")
- # Recompilation without -o 1
- self.compile_dir(optimize=[0, 2], force=True)
- # opt-1.pyc should have the same inode as before and others should not
- self.assertEqual(inode, os.stat(pycs[1]).st_ino)
- self.assertTrue(is_hardlink(pycs[0], pycs[2]))
- self.assertNotEqual(inode, os.stat(pycs[2]).st_ino)
- # opt-1.pyc and opt-2.pyc have different content
- self.assertFalse(filecmp.cmp(pycs[1], pycs[2], shallow=True))
- def test_import(self):
- # Test that import updates a single pyc file when pyc files already
- # exists and the script content changed
- with self.temporary_directory():
- script = self.make_script(self.create_code(), name="module")
- self.compile_dir()
- # All three levels have the same inode
- self.check_hardlinks(script)
- pycs = get_pycs(script)
- inode = os.stat(pycs[0]).st_ino
- # Change of the module content
- script = self.make_script("print(0)", name="module")
- # Import the module in Python with -O (optimization level 1)
- script_helper.assert_python_ok(
- "-O", "-c", "import module", __isolated=False, PYTHONPATH=self.path
- )
- # Only opt-1.pyc is changed
- self.assertEqual(inode, os.stat(pycs[0]).st_ino)
- self.assertEqual(inode, os.stat(pycs[2]).st_ino)
- self.assertFalse(is_hardlink(pycs[1], pycs[2]))
- # opt-1.pyc and opt-2.pyc have different content
- self.assertFalse(filecmp.cmp(pycs[1], pycs[2], shallow=True))
- class HardlinkDedupTestsWithSourceEpoch(HardlinkDedupTestsBase,
- unittest.TestCase,
- metaclass=SourceDateEpochTestMeta,
- source_date_epoch=True):
- pass
- class HardlinkDedupTestsNoSourceEpoch(HardlinkDedupTestsBase,
- unittest.TestCase,
- metaclass=SourceDateEpochTestMeta,
- source_date_epoch=False):
- pass
- if __name__ == "__main__":
- unittest.main()
|