mirror of
				https://github.com/python/cpython.git
				synced 2025-11-04 07:31:38 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			1072 lines
		
	
	
	
		
			45 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			1072 lines
		
	
	
	
		
			45 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
import compileall
 | 
						|
import contextlib
 | 
						|
import filecmp
 | 
						|
import importlib.util
 | 
						|
import io
 | 
						|
import itertools
 | 
						|
import os
 | 
						|
import pathlib
 | 
						|
import py_compile
 | 
						|
import shutil
 | 
						|
import struct
 | 
						|
import sys
 | 
						|
import tempfile
 | 
						|
import test.test_importlib.util
 | 
						|
import time
 | 
						|
import unittest
 | 
						|
 | 
						|
from unittest import mock, skipUnless
 | 
						|
from concurrent.futures import ProcessPoolExecutor
 | 
						|
try:
 | 
						|
    # compileall relies on ProcessPoolExecutor if ProcessPoolExecutor exists
 | 
						|
    # and it can function.
 | 
						|
    from concurrent.futures.process import _check_system_limits
 | 
						|
    _check_system_limits()
 | 
						|
    _have_multiprocessing = True
 | 
						|
except NotImplementedError:
 | 
						|
    _have_multiprocessing = False
 | 
						|
 | 
						|
from test import support
 | 
						|
from test.support import os_helper
 | 
						|
from test.support import script_helper
 | 
						|
 | 
						|
from .test_py_compile import without_source_date_epoch
 | 
						|
from .test_py_compile import SourceDateEpochTestMeta
 | 
						|
 | 
						|
 | 
						|
def get_pyc(script, opt):
 | 
						|
    if not opt:
 | 
						|
        # Replace None and 0 with ''
 | 
						|
        opt = ''
 | 
						|
    return importlib.util.cache_from_source(script, optimization=opt)
 | 
						|
 | 
						|
 | 
						|
def get_pycs(script):
 | 
						|
    return [get_pyc(script, opt) for opt in (0, 1, 2)]
 | 
						|
 | 
						|
 | 
						|
def is_hardlink(filename1, filename2):
 | 
						|
    """Returns True if two files have the same inode (hardlink)"""
 | 
						|
    inode1 = os.stat(filename1).st_ino
 | 
						|
    inode2 = os.stat(filename2).st_ino
 | 
						|
    return inode1 == inode2
 | 
						|
 | 
						|
 | 
						|
class CompileallTestsBase:
 | 
						|
 | 
						|
    def setUp(self):
 | 
						|
        self.directory = tempfile.mkdtemp()
 | 
						|
        self.source_path = os.path.join(self.directory, '_test.py')
 | 
						|
        self.bc_path = importlib.util.cache_from_source(self.source_path)
 | 
						|
        with open(self.source_path, 'w', encoding="utf-8") as file:
 | 
						|
            file.write('x = 123\n')
 | 
						|
        self.source_path2 = os.path.join(self.directory, '_test2.py')
 | 
						|
        self.bc_path2 = importlib.util.cache_from_source(self.source_path2)
 | 
						|
        shutil.copyfile(self.source_path, self.source_path2)
 | 
						|
        self.subdirectory = os.path.join(self.directory, '_subdir')
 | 
						|
        os.mkdir(self.subdirectory)
 | 
						|
        self.source_path3 = os.path.join(self.subdirectory, '_test3.py')
 | 
						|
        shutil.copyfile(self.source_path, self.source_path3)
 | 
						|
 | 
						|
    def tearDown(self):
 | 
						|
        shutil.rmtree(self.directory)
 | 
						|
 | 
						|
    def add_bad_source_file(self):
 | 
						|
        self.bad_source_path = os.path.join(self.directory, '_test_bad.py')
 | 
						|
        with open(self.bad_source_path, 'w', encoding="utf-8") as file:
 | 
						|
            file.write('x (\n')
 | 
						|
 | 
						|
    def timestamp_metadata(self):
 | 
						|
        with open(self.bc_path, 'rb') as file:
 | 
						|
            data = file.read(12)
 | 
						|
        mtime = int(os.stat(self.source_path).st_mtime)
 | 
						|
        compare = struct.pack('<4sll', importlib.util.MAGIC_NUMBER, 0, mtime)
 | 
						|
        return data, compare
 | 
						|
 | 
						|
    def recreation_check(self, metadata):
 | 
						|
        """Check that compileall recreates bytecode when the new metadata is
 | 
						|
        used."""
 | 
						|
        if os.environ.get('SOURCE_DATE_EPOCH'):
 | 
						|
            raise unittest.SkipTest('SOURCE_DATE_EPOCH is set')
 | 
						|
        py_compile.compile(self.source_path)
 | 
						|
        self.assertEqual(*self.timestamp_metadata())
 | 
						|
        with open(self.bc_path, 'rb') as file:
 | 
						|
            bc = file.read()[len(metadata):]
 | 
						|
        with open(self.bc_path, 'wb') as file:
 | 
						|
            file.write(metadata)
 | 
						|
            file.write(bc)
 | 
						|
        self.assertNotEqual(*self.timestamp_metadata())
 | 
						|
        compileall.compile_dir(self.directory, force=False, quiet=True)
 | 
						|
        self.assertTrue(*self.timestamp_metadata())
 | 
						|
 | 
						|
    def test_mtime(self):
 | 
						|
        # Test a change in mtime leads to a new .pyc.
 | 
						|
        self.recreation_check(struct.pack('<4sll', importlib.util.MAGIC_NUMBER,
 | 
						|
                                          0, 1))
 | 
						|
 | 
						|
    def test_magic_number(self):
 | 
						|
        # Test a change in mtime leads to a new .pyc.
 | 
						|
        self.recreation_check(b'\0\0\0\0')
 | 
						|
 | 
						|
    def test_compile_files(self):
 | 
						|
        # Test compiling a single file, and complete directory
 | 
						|
        for fn in (self.bc_path, self.bc_path2):
 | 
						|
            try:
 | 
						|
                os.unlink(fn)
 | 
						|
            except:
 | 
						|
                pass
 | 
						|
        self.assertTrue(compileall.compile_file(self.source_path,
 | 
						|
                                                force=False, quiet=True))
 | 
						|
        self.assertTrue(os.path.isfile(self.bc_path) and
 | 
						|
                        not os.path.isfile(self.bc_path2))
 | 
						|
        os.unlink(self.bc_path)
 | 
						|
        self.assertTrue(compileall.compile_dir(self.directory, force=False,
 | 
						|
                                               quiet=True))
 | 
						|
        self.assertTrue(os.path.isfile(self.bc_path) and
 | 
						|
                        os.path.isfile(self.bc_path2))
 | 
						|
        os.unlink(self.bc_path)
 | 
						|
        os.unlink(self.bc_path2)
 | 
						|
        # Test against bad files
 | 
						|
        self.add_bad_source_file()
 | 
						|
        self.assertFalse(compileall.compile_file(self.bad_source_path,
 | 
						|
                                                 force=False, quiet=2))
 | 
						|
        self.assertFalse(compileall.compile_dir(self.directory,
 | 
						|
                                                force=False, quiet=2))
 | 
						|
 | 
						|
    def test_compile_file_pathlike(self):
 | 
						|
        self.assertFalse(os.path.isfile(self.bc_path))
 | 
						|
        # we should also test the output
 | 
						|
        with support.captured_stdout() as stdout:
 | 
						|
            self.assertTrue(compileall.compile_file(pathlib.Path(self.source_path)))
 | 
						|
        self.assertRegex(stdout.getvalue(), r'Compiling ([^WindowsPath|PosixPath].*)')
 | 
						|
        self.assertTrue(os.path.isfile(self.bc_path))
 | 
						|
 | 
						|
    def test_compile_file_pathlike_ddir(self):
 | 
						|
        self.assertFalse(os.path.isfile(self.bc_path))
 | 
						|
        self.assertTrue(compileall.compile_file(pathlib.Path(self.source_path),
 | 
						|
                                                ddir=pathlib.Path('ddir_path'),
 | 
						|
                                                quiet=2))
 | 
						|
        self.assertTrue(os.path.isfile(self.bc_path))
 | 
						|
 | 
						|
    def test_compile_path(self):
 | 
						|
        with test.test_importlib.util.import_state(path=[self.directory]):
 | 
						|
            self.assertTrue(compileall.compile_path(quiet=2))
 | 
						|
 | 
						|
        with test.test_importlib.util.import_state(path=[self.directory]):
 | 
						|
            self.add_bad_source_file()
 | 
						|
            self.assertFalse(compileall.compile_path(skip_curdir=False,
 | 
						|
                                                     force=True, quiet=2))
 | 
						|
 | 
						|
    def test_no_pycache_in_non_package(self):
 | 
						|
        # Bug 8563 reported that __pycache__ directories got created by
 | 
						|
        # compile_file() for non-.py files.
 | 
						|
        data_dir = os.path.join(self.directory, 'data')
 | 
						|
        data_file = os.path.join(data_dir, 'file')
 | 
						|
        os.mkdir(data_dir)
 | 
						|
        # touch data/file
 | 
						|
        with open(data_file, 'wb'):
 | 
						|
            pass
 | 
						|
        compileall.compile_file(data_file)
 | 
						|
        self.assertFalse(os.path.exists(os.path.join(data_dir, '__pycache__')))
 | 
						|
 | 
						|
    def test_optimize(self):
 | 
						|
        # make sure compiling with different optimization settings than the
 | 
						|
        # interpreter's creates the correct file names
 | 
						|
        optimize, opt = (1, 1) if __debug__ else (0, '')
 | 
						|
        compileall.compile_dir(self.directory, quiet=True, optimize=optimize)
 | 
						|
        cached = importlib.util.cache_from_source(self.source_path,
 | 
						|
                                                  optimization=opt)
 | 
						|
        self.assertTrue(os.path.isfile(cached))
 | 
						|
        cached2 = importlib.util.cache_from_source(self.source_path2,
 | 
						|
                                                   optimization=opt)
 | 
						|
        self.assertTrue(os.path.isfile(cached2))
 | 
						|
        cached3 = importlib.util.cache_from_source(self.source_path3,
 | 
						|
                                                   optimization=opt)
 | 
						|
        self.assertTrue(os.path.isfile(cached3))
 | 
						|
 | 
						|
    def test_compile_dir_pathlike(self):
 | 
						|
        self.assertFalse(os.path.isfile(self.bc_path))
 | 
						|
        with support.captured_stdout() as stdout:
 | 
						|
            compileall.compile_dir(pathlib.Path(self.directory))
 | 
						|
        line = stdout.getvalue().splitlines()[0]
 | 
						|
        self.assertRegex(line, r'Listing ([^WindowsPath|PosixPath].*)')
 | 
						|
        self.assertTrue(os.path.isfile(self.bc_path))
 | 
						|
 | 
						|
    @skipUnless(_have_multiprocessing, "requires multiprocessing")
 | 
						|
    @mock.patch('concurrent.futures.ProcessPoolExecutor')
 | 
						|
    def test_compile_pool_called(self, pool_mock):
 | 
						|
        compileall.compile_dir(self.directory, quiet=True, workers=5)
 | 
						|
        self.assertTrue(pool_mock.called)
 | 
						|
 | 
						|
    def test_compile_workers_non_positive(self):
 | 
						|
        with self.assertRaisesRegex(ValueError,
 | 
						|
                                    "workers must be greater or equal to 0"):
 | 
						|
            compileall.compile_dir(self.directory, workers=-1)
 | 
						|
 | 
						|
    @skipUnless(_have_multiprocessing, "requires multiprocessing")
 | 
						|
    @mock.patch('concurrent.futures.ProcessPoolExecutor')
 | 
						|
    def test_compile_workers_cpu_count(self, pool_mock):
 | 
						|
        compileall.compile_dir(self.directory, quiet=True, workers=0)
 | 
						|
        self.assertEqual(pool_mock.call_args[1]['max_workers'], None)
 | 
						|
 | 
						|
    @skipUnless(_have_multiprocessing, "requires multiprocessing")
 | 
						|
    @mock.patch('concurrent.futures.ProcessPoolExecutor')
 | 
						|
    @mock.patch('compileall.compile_file')
 | 
						|
    def test_compile_one_worker(self, compile_file_mock, pool_mock):
 | 
						|
        compileall.compile_dir(self.directory, quiet=True)
 | 
						|
        self.assertFalse(pool_mock.called)
 | 
						|
        self.assertTrue(compile_file_mock.called)
 | 
						|
 | 
						|
    @mock.patch('concurrent.futures.ProcessPoolExecutor', new=None)
 | 
						|
    @mock.patch('compileall.compile_file')
 | 
						|
    def test_compile_missing_multiprocessing(self, compile_file_mock):
 | 
						|
        compileall.compile_dir(self.directory, quiet=True, workers=5)
 | 
						|
        self.assertTrue(compile_file_mock.called)
 | 
						|
 | 
						|
    def test_compile_dir_maxlevels(self):
 | 
						|
        # Test the actual impact of maxlevels parameter
 | 
						|
        depth = 3
 | 
						|
        path = self.directory
 | 
						|
        for i in range(1, depth + 1):
 | 
						|
            path = os.path.join(path, f"dir_{i}")
 | 
						|
            source = os.path.join(path, 'script.py')
 | 
						|
            os.mkdir(path)
 | 
						|
            shutil.copyfile(self.source_path, source)
 | 
						|
        pyc_filename = importlib.util.cache_from_source(source)
 | 
						|
 | 
						|
        compileall.compile_dir(self.directory, quiet=True, maxlevels=depth - 1)
 | 
						|
        self.assertFalse(os.path.isfile(pyc_filename))
 | 
						|
 | 
						|
        compileall.compile_dir(self.directory, quiet=True, maxlevels=depth)
 | 
						|
        self.assertTrue(os.path.isfile(pyc_filename))
 | 
						|
 | 
						|
    def _test_ddir_only(self, *, ddir, parallel=True):
 | 
						|
        """Recursive compile_dir ddir must contain package paths; bpo39769."""
 | 
						|
        fullpath = ["test", "foo"]
 | 
						|
        path = self.directory
 | 
						|
        mods = []
 | 
						|
        for subdir in fullpath:
 | 
						|
            path = os.path.join(path, subdir)
 | 
						|
            os.mkdir(path)
 | 
						|
            script_helper.make_script(path, "__init__", "")
 | 
						|
            mods.append(script_helper.make_script(path, "mod",
 | 
						|
                                                  "def fn(): 1/0\nfn()\n"))
 | 
						|
        compileall.compile_dir(
 | 
						|
                self.directory, quiet=True, ddir=ddir,
 | 
						|
                workers=2 if parallel else 1)
 | 
						|
        self.assertTrue(mods)
 | 
						|
        for mod in mods:
 | 
						|
            self.assertTrue(mod.startswith(self.directory), mod)
 | 
						|
            modcode = importlib.util.cache_from_source(mod)
 | 
						|
            modpath = mod[len(self.directory+os.sep):]
 | 
						|
            _, _, err = script_helper.assert_python_failure(modcode)
 | 
						|
            expected_in = os.path.join(ddir, modpath)
 | 
						|
            mod_code_obj = test.test_importlib.util.get_code_from_pyc(modcode)
 | 
						|
            self.assertEqual(mod_code_obj.co_filename, expected_in)
 | 
						|
            self.assertIn(f'"{expected_in}"', os.fsdecode(err))
 | 
						|
 | 
						|
    def test_ddir_only_one_worker(self):
 | 
						|
        """Recursive compile_dir ddir= contains package paths; bpo39769."""
 | 
						|
        return self._test_ddir_only(ddir="<a prefix>", parallel=False)
 | 
						|
 | 
						|
    def test_ddir_multiple_workers(self):
 | 
						|
        """Recursive compile_dir ddir= contains package paths; bpo39769."""
 | 
						|
        return self._test_ddir_only(ddir="<a prefix>", parallel=True)
 | 
						|
 | 
						|
    def test_ddir_empty_only_one_worker(self):
 | 
						|
        """Recursive compile_dir ddir='' contains package paths; bpo39769."""
 | 
						|
        return self._test_ddir_only(ddir="", parallel=False)
 | 
						|
 | 
						|
    def test_ddir_empty_multiple_workers(self):
 | 
						|
        """Recursive compile_dir ddir='' contains package paths; bpo39769."""
 | 
						|
        return self._test_ddir_only(ddir="", parallel=True)
 | 
						|
 | 
						|
    def test_strip_only(self):
 | 
						|
        fullpath = ["test", "build", "real", "path"]
 | 
						|
        path = os.path.join(self.directory, *fullpath)
 | 
						|
        os.makedirs(path)
 | 
						|
        script = script_helper.make_script(path, "test", "1 / 0")
 | 
						|
        bc = importlib.util.cache_from_source(script)
 | 
						|
        stripdir = os.path.join(self.directory, *fullpath[:2])
 | 
						|
        compileall.compile_dir(path, quiet=True, stripdir=stripdir)
 | 
						|
        rc, out, err = script_helper.assert_python_failure(bc)
 | 
						|
        expected_in = os.path.join(*fullpath[2:])
 | 
						|
        self.assertIn(
 | 
						|
            expected_in,
 | 
						|
            str(err, encoding=sys.getdefaultencoding())
 | 
						|
        )
 | 
						|
        self.assertNotIn(
 | 
						|
            stripdir,
 | 
						|
            str(err, encoding=sys.getdefaultencoding())
 | 
						|
        )
 | 
						|
 | 
						|
    def test_prepend_only(self):
 | 
						|
        fullpath = ["test", "build", "real", "path"]
 | 
						|
        path = os.path.join(self.directory, *fullpath)
 | 
						|
        os.makedirs(path)
 | 
						|
        script = script_helper.make_script(path, "test", "1 / 0")
 | 
						|
        bc = importlib.util.cache_from_source(script)
 | 
						|
        prependdir = "/foo"
 | 
						|
        compileall.compile_dir(path, quiet=True, prependdir=prependdir)
 | 
						|
        rc, out, err = script_helper.assert_python_failure(bc)
 | 
						|
        expected_in = os.path.join(prependdir, self.directory, *fullpath)
 | 
						|
        self.assertIn(
 | 
						|
            expected_in,
 | 
						|
            str(err, encoding=sys.getdefaultencoding())
 | 
						|
        )
 | 
						|
 | 
						|
    def test_strip_and_prepend(self):
 | 
						|
        fullpath = ["test", "build", "real", "path"]
 | 
						|
        path = os.path.join(self.directory, *fullpath)
 | 
						|
        os.makedirs(path)
 | 
						|
        script = script_helper.make_script(path, "test", "1 / 0")
 | 
						|
        bc = importlib.util.cache_from_source(script)
 | 
						|
        stripdir = os.path.join(self.directory, *fullpath[:2])
 | 
						|
        prependdir = "/foo"
 | 
						|
        compileall.compile_dir(path, quiet=True,
 | 
						|
                               stripdir=stripdir, prependdir=prependdir)
 | 
						|
        rc, out, err = script_helper.assert_python_failure(bc)
 | 
						|
        expected_in = os.path.join(prependdir, *fullpath[2:])
 | 
						|
        self.assertIn(
 | 
						|
            expected_in,
 | 
						|
            str(err, encoding=sys.getdefaultencoding())
 | 
						|
        )
 | 
						|
        self.assertNotIn(
 | 
						|
            stripdir,
 | 
						|
            str(err, encoding=sys.getdefaultencoding())
 | 
						|
        )
 | 
						|
 | 
						|
    def test_strip_prepend_and_ddir(self):
 | 
						|
        fullpath = ["test", "build", "real", "path", "ddir"]
 | 
						|
        path = os.path.join(self.directory, *fullpath)
 | 
						|
        os.makedirs(path)
 | 
						|
        script_helper.make_script(path, "test", "1 / 0")
 | 
						|
        with self.assertRaises(ValueError):
 | 
						|
            compileall.compile_dir(path, quiet=True, ddir="/bar",
 | 
						|
                                   stripdir="/foo", prependdir="/bar")
 | 
						|
 | 
						|
    def test_multiple_optimization_levels(self):
 | 
						|
        script = script_helper.make_script(self.directory,
 | 
						|
                                           "test_optimization",
 | 
						|
                                           "a = 0")
 | 
						|
        bc = []
 | 
						|
        for opt_level in "", 1, 2, 3:
 | 
						|
            bc.append(importlib.util.cache_from_source(script,
 | 
						|
                                                       optimization=opt_level))
 | 
						|
        test_combinations = [[0, 1], [1, 2], [0, 2], [0, 1, 2]]
 | 
						|
        for opt_combination in test_combinations:
 | 
						|
            compileall.compile_file(script, quiet=True,
 | 
						|
                                    optimize=opt_combination)
 | 
						|
            for opt_level in opt_combination:
 | 
						|
                self.assertTrue(os.path.isfile(bc[opt_level]))
 | 
						|
                try:
 | 
						|
                    os.unlink(bc[opt_level])
 | 
						|
                except Exception:
 | 
						|
                    pass
 | 
						|
 | 
						|
    @os_helper.skip_unless_symlink
 | 
						|
    def test_ignore_symlink_destination(self):
 | 
						|
        # Create folders for allowed files, symlinks and prohibited area
 | 
						|
        allowed_path = os.path.join(self.directory, "test", "dir", "allowed")
 | 
						|
        symlinks_path = os.path.join(self.directory, "test", "dir", "symlinks")
 | 
						|
        prohibited_path = os.path.join(self.directory, "test", "dir", "prohibited")
 | 
						|
        os.makedirs(allowed_path)
 | 
						|
        os.makedirs(symlinks_path)
 | 
						|
        os.makedirs(prohibited_path)
 | 
						|
 | 
						|
        # Create scripts and symlinks and remember their byte-compiled versions
 | 
						|
        allowed_script = script_helper.make_script(allowed_path, "test_allowed", "a = 0")
 | 
						|
        prohibited_script = script_helper.make_script(prohibited_path, "test_prohibited", "a = 0")
 | 
						|
        allowed_symlink = os.path.join(symlinks_path, "test_allowed.py")
 | 
						|
        prohibited_symlink = os.path.join(symlinks_path, "test_prohibited.py")
 | 
						|
        os.symlink(allowed_script, allowed_symlink)
 | 
						|
        os.symlink(prohibited_script, prohibited_symlink)
 | 
						|
        allowed_bc = importlib.util.cache_from_source(allowed_symlink)
 | 
						|
        prohibited_bc = importlib.util.cache_from_source(prohibited_symlink)
 | 
						|
 | 
						|
        compileall.compile_dir(symlinks_path, quiet=True, limit_sl_dest=allowed_path)
 | 
						|
 | 
						|
        self.assertTrue(os.path.isfile(allowed_bc))
 | 
						|
        self.assertFalse(os.path.isfile(prohibited_bc))
 | 
						|
 | 
						|
 | 
						|
class CompileallTestsWithSourceEpoch(CompileallTestsBase,
 | 
						|
                                     unittest.TestCase,
 | 
						|
                                     metaclass=SourceDateEpochTestMeta,
 | 
						|
                                     source_date_epoch=True):
 | 
						|
    pass
 | 
						|
 | 
						|
 | 
						|
class CompileallTestsWithoutSourceEpoch(CompileallTestsBase,
 | 
						|
                                        unittest.TestCase,
 | 
						|
                                        metaclass=SourceDateEpochTestMeta,
 | 
						|
                                        source_date_epoch=False):
 | 
						|
    pass
 | 
						|
 | 
						|
 | 
						|
class EncodingTest(unittest.TestCase):
 | 
						|
    """Issue 6716: compileall should escape source code when printing errors
 | 
						|
    to stdout."""
 | 
						|
 | 
						|
    def setUp(self):
 | 
						|
        self.directory = tempfile.mkdtemp()
 | 
						|
        self.source_path = os.path.join(self.directory, '_test.py')
 | 
						|
        with open(self.source_path, 'w', encoding='utf-8') as file:
 | 
						|
            file.write('# -*- coding: utf-8 -*-\n')
 | 
						|
            file.write('print u"\u20ac"\n')
 | 
						|
 | 
						|
    def tearDown(self):
 | 
						|
        shutil.rmtree(self.directory)
 | 
						|
 | 
						|
    def test_error(self):
 | 
						|
        try:
 | 
						|
            orig_stdout = sys.stdout
 | 
						|
            sys.stdout = io.TextIOWrapper(io.BytesIO(),encoding='ascii')
 | 
						|
            compileall.compile_dir(self.directory)
 | 
						|
        finally:
 | 
						|
            sys.stdout = orig_stdout
 | 
						|
 | 
						|
 | 
						|
class CommandLineTestsBase:
 | 
						|
    """Test compileall's CLI."""
 | 
						|
 | 
						|
    @classmethod
 | 
						|
    def setUpClass(cls):
 | 
						|
        for path in filter(os.path.isdir, sys.path):
 | 
						|
            directory_created = False
 | 
						|
            directory = pathlib.Path(path) / '__pycache__'
 | 
						|
            path = directory / 'test.try'
 | 
						|
            try:
 | 
						|
                if not directory.is_dir():
 | 
						|
                    directory.mkdir()
 | 
						|
                    directory_created = True
 | 
						|
                path.write_text('# for test_compileall', encoding="utf-8")
 | 
						|
            except OSError:
 | 
						|
                sys_path_writable = False
 | 
						|
                break
 | 
						|
            finally:
 | 
						|
                os_helper.unlink(str(path))
 | 
						|
                if directory_created:
 | 
						|
                    directory.rmdir()
 | 
						|
        else:
 | 
						|
            sys_path_writable = True
 | 
						|
        cls._sys_path_writable = sys_path_writable
 | 
						|
 | 
						|
    def _skip_if_sys_path_not_writable(self):
 | 
						|
        if not self._sys_path_writable:
 | 
						|
            raise unittest.SkipTest('not all entries on sys.path are writable')
 | 
						|
 | 
						|
    def _get_run_args(self, args):
 | 
						|
        return [*support.optim_args_from_interpreter_flags(),
 | 
						|
                '-S', '-m', 'compileall',
 | 
						|
                *args]
 | 
						|
 | 
						|
    def assertRunOK(self, *args, **env_vars):
 | 
						|
        rc, out, err = script_helper.assert_python_ok(
 | 
						|
                         *self._get_run_args(args), **env_vars,
 | 
						|
                         PYTHONIOENCODING='utf-8')
 | 
						|
        self.assertEqual(b'', err)
 | 
						|
        return out
 | 
						|
 | 
						|
    def assertRunNotOK(self, *args, **env_vars):
 | 
						|
        rc, out, err = script_helper.assert_python_failure(
 | 
						|
                        *self._get_run_args(args), **env_vars,
 | 
						|
                        PYTHONIOENCODING='utf-8')
 | 
						|
        return rc, out, err
 | 
						|
 | 
						|
    def assertCompiled(self, fn):
 | 
						|
        path = importlib.util.cache_from_source(fn)
 | 
						|
        self.assertTrue(os.path.exists(path))
 | 
						|
 | 
						|
    def assertNotCompiled(self, fn):
 | 
						|
        path = importlib.util.cache_from_source(fn)
 | 
						|
        self.assertFalse(os.path.exists(path))
 | 
						|
 | 
						|
    def setUp(self):
 | 
						|
        self.directory = tempfile.mkdtemp()
 | 
						|
        self.addCleanup(os_helper.rmtree, self.directory)
 | 
						|
        self.pkgdir = os.path.join(self.directory, 'foo')
 | 
						|
        os.mkdir(self.pkgdir)
 | 
						|
        self.pkgdir_cachedir = os.path.join(self.pkgdir, '__pycache__')
 | 
						|
        # Create the __init__.py and a package module.
 | 
						|
        self.initfn = script_helper.make_script(self.pkgdir, '__init__', '')
 | 
						|
        self.barfn = script_helper.make_script(self.pkgdir, 'bar', '')
 | 
						|
 | 
						|
    def test_no_args_compiles_path(self):
 | 
						|
        # Note that -l is implied for the no args case.
 | 
						|
        self._skip_if_sys_path_not_writable()
 | 
						|
        bazfn = script_helper.make_script(self.directory, 'baz', '')
 | 
						|
        self.assertRunOK(PYTHONPATH=self.directory)
 | 
						|
        self.assertCompiled(bazfn)
 | 
						|
        self.assertNotCompiled(self.initfn)
 | 
						|
        self.assertNotCompiled(self.barfn)
 | 
						|
 | 
						|
    @without_source_date_epoch  # timestamp invalidation test
 | 
						|
    def test_no_args_respects_force_flag(self):
 | 
						|
        self._skip_if_sys_path_not_writable()
 | 
						|
        bazfn = script_helper.make_script(self.directory, 'baz', '')
 | 
						|
        self.assertRunOK(PYTHONPATH=self.directory)
 | 
						|
        pycpath = importlib.util.cache_from_source(bazfn)
 | 
						|
        # Set atime/mtime backward to avoid file timestamp resolution issues
 | 
						|
        os.utime(pycpath, (time.time()-60,)*2)
 | 
						|
        mtime = os.stat(pycpath).st_mtime
 | 
						|
        # Without force, no recompilation
 | 
						|
        self.assertRunOK(PYTHONPATH=self.directory)
 | 
						|
        mtime2 = os.stat(pycpath).st_mtime
 | 
						|
        self.assertEqual(mtime, mtime2)
 | 
						|
        # Now force it.
 | 
						|
        self.assertRunOK('-f', PYTHONPATH=self.directory)
 | 
						|
        mtime2 = os.stat(pycpath).st_mtime
 | 
						|
        self.assertNotEqual(mtime, mtime2)
 | 
						|
 | 
						|
    def test_no_args_respects_quiet_flag(self):
 | 
						|
        self._skip_if_sys_path_not_writable()
 | 
						|
        script_helper.make_script(self.directory, 'baz', '')
 | 
						|
        noisy = self.assertRunOK(PYTHONPATH=self.directory)
 | 
						|
        self.assertIn(b'Listing ', noisy)
 | 
						|
        quiet = self.assertRunOK('-q', PYTHONPATH=self.directory)
 | 
						|
        self.assertNotIn(b'Listing ', quiet)
 | 
						|
 | 
						|
    # Ensure that the default behavior of compileall's CLI is to create
 | 
						|
    # PEP 3147/PEP 488 pyc files.
 | 
						|
    for name, ext, switch in [
 | 
						|
        ('normal', 'pyc', []),
 | 
						|
        ('optimize', 'opt-1.pyc', ['-O']),
 | 
						|
        ('doubleoptimize', 'opt-2.pyc', ['-OO']),
 | 
						|
    ]:
 | 
						|
        def f(self, ext=ext, switch=switch):
 | 
						|
            script_helper.assert_python_ok(*(switch +
 | 
						|
                ['-m', 'compileall', '-q', self.pkgdir]))
 | 
						|
            # Verify the __pycache__ directory contents.
 | 
						|
            self.assertTrue(os.path.exists(self.pkgdir_cachedir))
 | 
						|
            expected = sorted(base.format(sys.implementation.cache_tag, ext)
 | 
						|
                              for base in ('__init__.{}.{}', 'bar.{}.{}'))
 | 
						|
            self.assertEqual(sorted(os.listdir(self.pkgdir_cachedir)), expected)
 | 
						|
            # Make sure there are no .pyc files in the source directory.
 | 
						|
            self.assertFalse([fn for fn in os.listdir(self.pkgdir)
 | 
						|
                              if fn.endswith(ext)])
 | 
						|
        locals()['test_pep3147_paths_' + name] = f
 | 
						|
 | 
						|
    def test_legacy_paths(self):
 | 
						|
        # Ensure that with the proper switch, compileall leaves legacy
 | 
						|
        # pyc files, and no __pycache__ directory.
 | 
						|
        self.assertRunOK('-b', '-q', self.pkgdir)
 | 
						|
        # Verify the __pycache__ directory contents.
 | 
						|
        self.assertFalse(os.path.exists(self.pkgdir_cachedir))
 | 
						|
        expected = sorted(['__init__.py', '__init__.pyc', 'bar.py',
 | 
						|
                           'bar.pyc'])
 | 
						|
        self.assertEqual(sorted(os.listdir(self.pkgdir)), expected)
 | 
						|
 | 
						|
    def test_multiple_runs(self):
 | 
						|
        # Bug 8527 reported that multiple calls produced empty
 | 
						|
        # __pycache__/__pycache__ directories.
 | 
						|
        self.assertRunOK('-q', self.pkgdir)
 | 
						|
        # Verify the __pycache__ directory contents.
 | 
						|
        self.assertTrue(os.path.exists(self.pkgdir_cachedir))
 | 
						|
        cachecachedir = os.path.join(self.pkgdir_cachedir, '__pycache__')
 | 
						|
        self.assertFalse(os.path.exists(cachecachedir))
 | 
						|
        # Call compileall again.
 | 
						|
        self.assertRunOK('-q', self.pkgdir)
 | 
						|
        self.assertTrue(os.path.exists(self.pkgdir_cachedir))
 | 
						|
        self.assertFalse(os.path.exists(cachecachedir))
 | 
						|
 | 
						|
    @without_source_date_epoch  # timestamp invalidation test
 | 
						|
    def test_force(self):
 | 
						|
        self.assertRunOK('-q', self.pkgdir)
 | 
						|
        pycpath = importlib.util.cache_from_source(self.barfn)
 | 
						|
        # set atime/mtime backward to avoid file timestamp resolution issues
 | 
						|
        os.utime(pycpath, (time.time()-60,)*2)
 | 
						|
        mtime = os.stat(pycpath).st_mtime
 | 
						|
        # without force, no recompilation
 | 
						|
        self.assertRunOK('-q', self.pkgdir)
 | 
						|
        mtime2 = os.stat(pycpath).st_mtime
 | 
						|
        self.assertEqual(mtime, mtime2)
 | 
						|
        # now force it.
 | 
						|
        self.assertRunOK('-q', '-f', self.pkgdir)
 | 
						|
        mtime2 = os.stat(pycpath).st_mtime
 | 
						|
        self.assertNotEqual(mtime, mtime2)
 | 
						|
 | 
						|
    def test_recursion_control(self):
 | 
						|
        subpackage = os.path.join(self.pkgdir, 'spam')
 | 
						|
        os.mkdir(subpackage)
 | 
						|
        subinitfn = script_helper.make_script(subpackage, '__init__', '')
 | 
						|
        hamfn = script_helper.make_script(subpackage, 'ham', '')
 | 
						|
        self.assertRunOK('-q', '-l', self.pkgdir)
 | 
						|
        self.assertNotCompiled(subinitfn)
 | 
						|
        self.assertFalse(os.path.exists(os.path.join(subpackage, '__pycache__')))
 | 
						|
        self.assertRunOK('-q', self.pkgdir)
 | 
						|
        self.assertCompiled(subinitfn)
 | 
						|
        self.assertCompiled(hamfn)
 | 
						|
 | 
						|
    def test_recursion_limit(self):
 | 
						|
        subpackage = os.path.join(self.pkgdir, 'spam')
 | 
						|
        subpackage2 = os.path.join(subpackage, 'ham')
 | 
						|
        subpackage3 = os.path.join(subpackage2, 'eggs')
 | 
						|
        for pkg in (subpackage, subpackage2, subpackage3):
 | 
						|
            script_helper.make_pkg(pkg)
 | 
						|
 | 
						|
        subinitfn = os.path.join(subpackage, '__init__.py')
 | 
						|
        hamfn = script_helper.make_script(subpackage, 'ham', '')
 | 
						|
        spamfn = script_helper.make_script(subpackage2, 'spam', '')
 | 
						|
        eggfn = script_helper.make_script(subpackage3, 'egg', '')
 | 
						|
 | 
						|
        self.assertRunOK('-q', '-r 0', self.pkgdir)
 | 
						|
        self.assertNotCompiled(subinitfn)
 | 
						|
        self.assertFalse(
 | 
						|
            os.path.exists(os.path.join(subpackage, '__pycache__')))
 | 
						|
 | 
						|
        self.assertRunOK('-q', '-r 1', self.pkgdir)
 | 
						|
        self.assertCompiled(subinitfn)
 | 
						|
        self.assertCompiled(hamfn)
 | 
						|
        self.assertNotCompiled(spamfn)
 | 
						|
 | 
						|
        self.assertRunOK('-q', '-r 2', self.pkgdir)
 | 
						|
        self.assertCompiled(subinitfn)
 | 
						|
        self.assertCompiled(hamfn)
 | 
						|
        self.assertCompiled(spamfn)
 | 
						|
        self.assertNotCompiled(eggfn)
 | 
						|
 | 
						|
        self.assertRunOK('-q', '-r 5', self.pkgdir)
 | 
						|
        self.assertCompiled(subinitfn)
 | 
						|
        self.assertCompiled(hamfn)
 | 
						|
        self.assertCompiled(spamfn)
 | 
						|
        self.assertCompiled(eggfn)
 | 
						|
 | 
						|
    @os_helper.skip_unless_symlink
 | 
						|
    def test_symlink_loop(self):
 | 
						|
        # Currently, compileall ignores symlinks to directories.
 | 
						|
        # If that limitation is ever lifted, it should protect against
 | 
						|
        # recursion in symlink loops.
 | 
						|
        pkg = os.path.join(self.pkgdir, 'spam')
 | 
						|
        script_helper.make_pkg(pkg)
 | 
						|
        os.symlink('.', os.path.join(pkg, 'evil'))
 | 
						|
        os.symlink('.', os.path.join(pkg, 'evil2'))
 | 
						|
        self.assertRunOK('-q', self.pkgdir)
 | 
						|
        self.assertCompiled(os.path.join(
 | 
						|
            self.pkgdir, 'spam', 'evil', 'evil2', '__init__.py'
 | 
						|
        ))
 | 
						|
 | 
						|
    def test_quiet(self):
 | 
						|
        noisy = self.assertRunOK(self.pkgdir)
 | 
						|
        quiet = self.assertRunOK('-q', self.pkgdir)
 | 
						|
        self.assertNotEqual(b'', noisy)
 | 
						|
        self.assertEqual(b'', quiet)
 | 
						|
 | 
						|
    def test_silent(self):
 | 
						|
        script_helper.make_script(self.pkgdir, 'crunchyfrog', 'bad(syntax')
 | 
						|
        _, quiet, _ = self.assertRunNotOK('-q', self.pkgdir)
 | 
						|
        _, silent, _ = self.assertRunNotOK('-qq', self.pkgdir)
 | 
						|
        self.assertNotEqual(b'', quiet)
 | 
						|
        self.assertEqual(b'', silent)
 | 
						|
 | 
						|
    def test_regexp(self):
 | 
						|
        self.assertRunOK('-q', '-x', r'ba[^\\/]*$', self.pkgdir)
 | 
						|
        self.assertNotCompiled(self.barfn)
 | 
						|
        self.assertCompiled(self.initfn)
 | 
						|
 | 
						|
    def test_multiple_dirs(self):
 | 
						|
        pkgdir2 = os.path.join(self.directory, 'foo2')
 | 
						|
        os.mkdir(pkgdir2)
 | 
						|
        init2fn = script_helper.make_script(pkgdir2, '__init__', '')
 | 
						|
        bar2fn = script_helper.make_script(pkgdir2, 'bar2', '')
 | 
						|
        self.assertRunOK('-q', self.pkgdir, pkgdir2)
 | 
						|
        self.assertCompiled(self.initfn)
 | 
						|
        self.assertCompiled(self.barfn)
 | 
						|
        self.assertCompiled(init2fn)
 | 
						|
        self.assertCompiled(bar2fn)
 | 
						|
 | 
						|
    def test_d_compile_error(self):
 | 
						|
        script_helper.make_script(self.pkgdir, 'crunchyfrog', 'bad(syntax')
 | 
						|
        rc, out, err = self.assertRunNotOK('-q', '-d', 'dinsdale', self.pkgdir)
 | 
						|
        self.assertRegex(out, b'File "dinsdale')
 | 
						|
 | 
						|
    def test_d_runtime_error(self):
 | 
						|
        bazfn = script_helper.make_script(self.pkgdir, 'baz', 'raise Exception')
 | 
						|
        self.assertRunOK('-q', '-d', 'dinsdale', self.pkgdir)
 | 
						|
        fn = script_helper.make_script(self.pkgdir, 'bing', 'import baz')
 | 
						|
        pyc = importlib.util.cache_from_source(bazfn)
 | 
						|
        os.rename(pyc, os.path.join(self.pkgdir, 'baz.pyc'))
 | 
						|
        os.remove(bazfn)
 | 
						|
        rc, out, err = script_helper.assert_python_failure(fn, __isolated=False)
 | 
						|
        self.assertRegex(err, b'File "dinsdale')
 | 
						|
 | 
						|
    def test_include_bad_file(self):
 | 
						|
        rc, out, err = self.assertRunNotOK(
 | 
						|
            '-i', os.path.join(self.directory, 'nosuchfile'), self.pkgdir)
 | 
						|
        self.assertRegex(out, b'rror.*nosuchfile')
 | 
						|
        self.assertNotRegex(err, b'Traceback')
 | 
						|
        self.assertFalse(os.path.exists(importlib.util.cache_from_source(
 | 
						|
                                            self.pkgdir_cachedir)))
 | 
						|
 | 
						|
    def test_include_file_with_arg(self):
 | 
						|
        f1 = script_helper.make_script(self.pkgdir, 'f1', '')
 | 
						|
        f2 = script_helper.make_script(self.pkgdir, 'f2', '')
 | 
						|
        f3 = script_helper.make_script(self.pkgdir, 'f3', '')
 | 
						|
        f4 = script_helper.make_script(self.pkgdir, 'f4', '')
 | 
						|
        with open(os.path.join(self.directory, 'l1'), 'w', encoding="utf-8") as l1:
 | 
						|
            l1.write(os.path.join(self.pkgdir, 'f1.py')+os.linesep)
 | 
						|
            l1.write(os.path.join(self.pkgdir, 'f2.py')+os.linesep)
 | 
						|
        self.assertRunOK('-i', os.path.join(self.directory, 'l1'), f4)
 | 
						|
        self.assertCompiled(f1)
 | 
						|
        self.assertCompiled(f2)
 | 
						|
        self.assertNotCompiled(f3)
 | 
						|
        self.assertCompiled(f4)
 | 
						|
 | 
						|
    def test_include_file_no_arg(self):
 | 
						|
        f1 = script_helper.make_script(self.pkgdir, 'f1', '')
 | 
						|
        f2 = script_helper.make_script(self.pkgdir, 'f2', '')
 | 
						|
        f3 = script_helper.make_script(self.pkgdir, 'f3', '')
 | 
						|
        f4 = script_helper.make_script(self.pkgdir, 'f4', '')
 | 
						|
        with open(os.path.join(self.directory, 'l1'), 'w', encoding="utf-8") as l1:
 | 
						|
            l1.write(os.path.join(self.pkgdir, 'f2.py')+os.linesep)
 | 
						|
        self.assertRunOK('-i', os.path.join(self.directory, 'l1'))
 | 
						|
        self.assertNotCompiled(f1)
 | 
						|
        self.assertCompiled(f2)
 | 
						|
        self.assertNotCompiled(f3)
 | 
						|
        self.assertNotCompiled(f4)
 | 
						|
 | 
						|
    def test_include_on_stdin(self):
 | 
						|
        f1 = script_helper.make_script(self.pkgdir, 'f1', '')
 | 
						|
        f2 = script_helper.make_script(self.pkgdir, 'f2', '')
 | 
						|
        f3 = script_helper.make_script(self.pkgdir, 'f3', '')
 | 
						|
        f4 = script_helper.make_script(self.pkgdir, 'f4', '')
 | 
						|
        p = script_helper.spawn_python(*(self._get_run_args(()) + ['-i', '-']))
 | 
						|
        p.stdin.write((f3+os.linesep).encode('ascii'))
 | 
						|
        script_helper.kill_python(p)
 | 
						|
        self.assertNotCompiled(f1)
 | 
						|
        self.assertNotCompiled(f2)
 | 
						|
        self.assertCompiled(f3)
 | 
						|
        self.assertNotCompiled(f4)
 | 
						|
 | 
						|
    def test_compiles_as_much_as_possible(self):
 | 
						|
        bingfn = script_helper.make_script(self.pkgdir, 'bing', 'syntax(error')
 | 
						|
        rc, out, err = self.assertRunNotOK('nosuchfile', self.initfn,
 | 
						|
                                           bingfn, self.barfn)
 | 
						|
        self.assertRegex(out, b'rror')
 | 
						|
        self.assertNotCompiled(bingfn)
 | 
						|
        self.assertCompiled(self.initfn)
 | 
						|
        self.assertCompiled(self.barfn)
 | 
						|
 | 
						|
    def test_invalid_arg_produces_message(self):
 | 
						|
        out = self.assertRunOK('badfilename')
 | 
						|
        self.assertRegex(out, b"Can't list 'badfilename'")
 | 
						|
 | 
						|
    def test_pyc_invalidation_mode(self):
 | 
						|
        script_helper.make_script(self.pkgdir, 'f1', '')
 | 
						|
        pyc = importlib.util.cache_from_source(
 | 
						|
            os.path.join(self.pkgdir, 'f1.py'))
 | 
						|
        self.assertRunOK('--invalidation-mode=checked-hash', self.pkgdir)
 | 
						|
        with open(pyc, 'rb') as fp:
 | 
						|
            data = fp.read()
 | 
						|
        self.assertEqual(int.from_bytes(data[4:8], 'little'), 0b11)
 | 
						|
        self.assertRunOK('--invalidation-mode=unchecked-hash', self.pkgdir)
 | 
						|
        with open(pyc, 'rb') as fp:
 | 
						|
            data = fp.read()
 | 
						|
        self.assertEqual(int.from_bytes(data[4:8], 'little'), 0b01)
 | 
						|
 | 
						|
    @skipUnless(_have_multiprocessing, "requires multiprocessing")
 | 
						|
    def test_workers(self):
 | 
						|
        bar2fn = script_helper.make_script(self.directory, 'bar2', '')
 | 
						|
        files = []
 | 
						|
        for suffix in range(5):
 | 
						|
            pkgdir = os.path.join(self.directory, 'foo{}'.format(suffix))
 | 
						|
            os.mkdir(pkgdir)
 | 
						|
            fn = script_helper.make_script(pkgdir, '__init__', '')
 | 
						|
            files.append(script_helper.make_script(pkgdir, 'bar2', ''))
 | 
						|
 | 
						|
        self.assertRunOK(self.directory, '-j', '0')
 | 
						|
        self.assertCompiled(bar2fn)
 | 
						|
        for file in files:
 | 
						|
            self.assertCompiled(file)
 | 
						|
 | 
						|
    @mock.patch('compileall.compile_dir')
 | 
						|
    def test_workers_available_cores(self, compile_dir):
 | 
						|
        with mock.patch("sys.argv",
 | 
						|
                        new=[sys.executable, self.directory, "-j0"]):
 | 
						|
            compileall.main()
 | 
						|
            self.assertTrue(compile_dir.called)
 | 
						|
            self.assertEqual(compile_dir.call_args[-1]['workers'], 0)
 | 
						|
 | 
						|
    def test_strip_and_prepend(self):
 | 
						|
        fullpath = ["test", "build", "real", "path"]
 | 
						|
        path = os.path.join(self.directory, *fullpath)
 | 
						|
        os.makedirs(path)
 | 
						|
        script = script_helper.make_script(path, "test", "1 / 0")
 | 
						|
        bc = importlib.util.cache_from_source(script)
 | 
						|
        stripdir = os.path.join(self.directory, *fullpath[:2])
 | 
						|
        prependdir = "/foo"
 | 
						|
        self.assertRunOK("-s", stripdir, "-p", prependdir, path)
 | 
						|
        rc, out, err = script_helper.assert_python_failure(bc)
 | 
						|
        expected_in = os.path.join(prependdir, *fullpath[2:])
 | 
						|
        self.assertIn(
 | 
						|
            expected_in,
 | 
						|
            str(err, encoding=sys.getdefaultencoding())
 | 
						|
        )
 | 
						|
        self.assertNotIn(
 | 
						|
            stripdir,
 | 
						|
            str(err, encoding=sys.getdefaultencoding())
 | 
						|
        )
 | 
						|
 | 
						|
    def test_multiple_optimization_levels(self):
 | 
						|
        path = os.path.join(self.directory, "optimizations")
 | 
						|
        os.makedirs(path)
 | 
						|
        script = script_helper.make_script(path,
 | 
						|
                                           "test_optimization",
 | 
						|
                                           "a = 0")
 | 
						|
        bc = []
 | 
						|
        for opt_level in "", 1, 2, 3:
 | 
						|
            bc.append(importlib.util.cache_from_source(script,
 | 
						|
                                                       optimization=opt_level))
 | 
						|
        test_combinations = [["0", "1"],
 | 
						|
                             ["1", "2"],
 | 
						|
                             ["0", "2"],
 | 
						|
                             ["0", "1", "2"]]
 | 
						|
        for opt_combination in test_combinations:
 | 
						|
            self.assertRunOK(path, *("-o" + str(n) for n in opt_combination))
 | 
						|
            for opt_level in opt_combination:
 | 
						|
                self.assertTrue(os.path.isfile(bc[int(opt_level)]))
 | 
						|
                try:
 | 
						|
                    os.unlink(bc[opt_level])
 | 
						|
                except Exception:
 | 
						|
                    pass
 | 
						|
 | 
						|
    @os_helper.skip_unless_symlink
 | 
						|
    def test_ignore_symlink_destination(self):
 | 
						|
        # Create folders for allowed files, symlinks and prohibited area
 | 
						|
        allowed_path = os.path.join(self.directory, "test", "dir", "allowed")
 | 
						|
        symlinks_path = os.path.join(self.directory, "test", "dir", "symlinks")
 | 
						|
        prohibited_path = os.path.join(self.directory, "test", "dir", "prohibited")
 | 
						|
        os.makedirs(allowed_path)
 | 
						|
        os.makedirs(symlinks_path)
 | 
						|
        os.makedirs(prohibited_path)
 | 
						|
 | 
						|
        # Create scripts and symlinks and remember their byte-compiled versions
 | 
						|
        allowed_script = script_helper.make_script(allowed_path, "test_allowed", "a = 0")
 | 
						|
        prohibited_script = script_helper.make_script(prohibited_path, "test_prohibited", "a = 0")
 | 
						|
        allowed_symlink = os.path.join(symlinks_path, "test_allowed.py")
 | 
						|
        prohibited_symlink = os.path.join(symlinks_path, "test_prohibited.py")
 | 
						|
        os.symlink(allowed_script, allowed_symlink)
 | 
						|
        os.symlink(prohibited_script, prohibited_symlink)
 | 
						|
        allowed_bc = importlib.util.cache_from_source(allowed_symlink)
 | 
						|
        prohibited_bc = importlib.util.cache_from_source(prohibited_symlink)
 | 
						|
 | 
						|
        self.assertRunOK(symlinks_path, "-e", allowed_path)
 | 
						|
 | 
						|
        self.assertTrue(os.path.isfile(allowed_bc))
 | 
						|
        self.assertFalse(os.path.isfile(prohibited_bc))
 | 
						|
 | 
						|
    def test_hardlink_bad_args(self):
 | 
						|
        # Bad arguments combination, hardlink deduplication make sense
 | 
						|
        # only for more than one optimization level
 | 
						|
        self.assertRunNotOK(self.directory, "-o 1", "--hardlink-dupes")
 | 
						|
 | 
						|
    def test_hardlink(self):
 | 
						|
        # 'a = 0' code produces the same bytecode for the 3 optimization
 | 
						|
        # levels. All three .pyc files must have the same inode (hardlinks).
 | 
						|
        #
 | 
						|
        # If deduplication is disabled, all pyc files must have different
 | 
						|
        # inodes.
 | 
						|
        for dedup in (True, False):
 | 
						|
            with tempfile.TemporaryDirectory() as path:
 | 
						|
                with self.subTest(dedup=dedup):
 | 
						|
                    script = script_helper.make_script(path, "script", "a = 0")
 | 
						|
                    pycs = get_pycs(script)
 | 
						|
 | 
						|
                    args = ["-q", "-o 0", "-o 1", "-o 2"]
 | 
						|
                    if dedup:
 | 
						|
                        args.append("--hardlink-dupes")
 | 
						|
                    self.assertRunOK(path, *args)
 | 
						|
 | 
						|
                    self.assertEqual(is_hardlink(pycs[0], pycs[1]), dedup)
 | 
						|
                    self.assertEqual(is_hardlink(pycs[1], pycs[2]), dedup)
 | 
						|
                    self.assertEqual(is_hardlink(pycs[0], pycs[2]), dedup)
 | 
						|
 | 
						|
 | 
						|
class CommandLineTestsWithSourceEpoch(CommandLineTestsBase,
 | 
						|
                                       unittest.TestCase,
 | 
						|
                                       metaclass=SourceDateEpochTestMeta,
 | 
						|
                                       source_date_epoch=True):
 | 
						|
    pass
 | 
						|
 | 
						|
 | 
						|
class CommandLineTestsNoSourceEpoch(CommandLineTestsBase,
 | 
						|
                                     unittest.TestCase,
 | 
						|
                                     metaclass=SourceDateEpochTestMeta,
 | 
						|
                                     source_date_epoch=False):
 | 
						|
    pass
 | 
						|
 | 
						|
 | 
						|
 | 
						|
class HardlinkDedupTestsBase:
 | 
						|
    # Test hardlink_dupes parameter of compileall.compile_dir()
 | 
						|
 | 
						|
    def setUp(self):
 | 
						|
        self.path = None
 | 
						|
 | 
						|
    @contextlib.contextmanager
 | 
						|
    def temporary_directory(self):
 | 
						|
        with tempfile.TemporaryDirectory() as path:
 | 
						|
            self.path = path
 | 
						|
            yield path
 | 
						|
            self.path = None
 | 
						|
 | 
						|
    def make_script(self, code, name="script"):
 | 
						|
        return script_helper.make_script(self.path, name, code)
 | 
						|
 | 
						|
    def compile_dir(self, *, dedup=True, optimize=(0, 1, 2), force=False):
 | 
						|
        compileall.compile_dir(self.path, quiet=True, optimize=optimize,
 | 
						|
                               hardlink_dupes=dedup, force=force)
 | 
						|
 | 
						|
    def test_bad_args(self):
 | 
						|
        # Bad arguments combination, hardlink deduplication make sense
 | 
						|
        # only for more than one optimization level
 | 
						|
        with self.temporary_directory():
 | 
						|
            self.make_script("pass")
 | 
						|
            with self.assertRaises(ValueError):
 | 
						|
                compileall.compile_dir(self.path, quiet=True, optimize=0,
 | 
						|
                                       hardlink_dupes=True)
 | 
						|
            with self.assertRaises(ValueError):
 | 
						|
                # same optimization level specified twice:
 | 
						|
                # compile_dir() removes duplicates
 | 
						|
                compileall.compile_dir(self.path, quiet=True, optimize=[0, 0],
 | 
						|
                                       hardlink_dupes=True)
 | 
						|
 | 
						|
    def create_code(self, docstring=False, assertion=False):
 | 
						|
        lines = []
 | 
						|
        if docstring:
 | 
						|
            lines.append("'module docstring'")
 | 
						|
        lines.append('x = 1')
 | 
						|
        if assertion:
 | 
						|
            lines.append("assert x == 1")
 | 
						|
        return '\n'.join(lines)
 | 
						|
 | 
						|
    def iter_codes(self):
 | 
						|
        for docstring in (False, True):
 | 
						|
            for assertion in (False, True):
 | 
						|
                code = self.create_code(docstring=docstring, assertion=assertion)
 | 
						|
                yield (code, docstring, assertion)
 | 
						|
 | 
						|
    def test_disabled(self):
 | 
						|
        # Deduplication disabled, no hardlinks
 | 
						|
        for code, docstring, assertion in self.iter_codes():
 | 
						|
            with self.subTest(docstring=docstring, assertion=assertion):
 | 
						|
                with self.temporary_directory():
 | 
						|
                    script = self.make_script(code)
 | 
						|
                    pycs = get_pycs(script)
 | 
						|
                    self.compile_dir(dedup=False)
 | 
						|
                    self.assertFalse(is_hardlink(pycs[0], pycs[1]))
 | 
						|
                    self.assertFalse(is_hardlink(pycs[0], pycs[2]))
 | 
						|
                    self.assertFalse(is_hardlink(pycs[1], pycs[2]))
 | 
						|
 | 
						|
    def check_hardlinks(self, script, docstring=False, assertion=False):
 | 
						|
        pycs = get_pycs(script)
 | 
						|
        self.assertEqual(is_hardlink(pycs[0], pycs[1]),
 | 
						|
                         not assertion)
 | 
						|
        self.assertEqual(is_hardlink(pycs[0], pycs[2]),
 | 
						|
                         not assertion and not docstring)
 | 
						|
        self.assertEqual(is_hardlink(pycs[1], pycs[2]),
 | 
						|
                         not docstring)
 | 
						|
 | 
						|
    def test_hardlink(self):
 | 
						|
        # Test deduplication on all combinations
 | 
						|
        for code, docstring, assertion in self.iter_codes():
 | 
						|
            with self.subTest(docstring=docstring, assertion=assertion):
 | 
						|
                with self.temporary_directory():
 | 
						|
                    script = self.make_script(code)
 | 
						|
                    self.compile_dir()
 | 
						|
                    self.check_hardlinks(script, docstring, assertion)
 | 
						|
 | 
						|
    def test_only_two_levels(self):
 | 
						|
        # Don't build the 3 optimization levels, but only 2
 | 
						|
        for opts in ((0, 1), (1, 2), (0, 2)):
 | 
						|
            with self.subTest(opts=opts):
 | 
						|
                with self.temporary_directory():
 | 
						|
                    # code with no dostring and no assertion:
 | 
						|
                    # same bytecode for all optimization levels
 | 
						|
                    script = self.make_script(self.create_code())
 | 
						|
                    self.compile_dir(optimize=opts)
 | 
						|
                    pyc1 = get_pyc(script, opts[0])
 | 
						|
                    pyc2 = get_pyc(script, opts[1])
 | 
						|
                    self.assertTrue(is_hardlink(pyc1, pyc2))
 | 
						|
 | 
						|
    def test_duplicated_levels(self):
 | 
						|
        # compile_dir() must not fail if optimize contains duplicated
 | 
						|
        # optimization levels and/or if optimization levels are not sorted.
 | 
						|
        with self.temporary_directory():
 | 
						|
            # code with no dostring and no assertion:
 | 
						|
            # same bytecode for all optimization levels
 | 
						|
            script = self.make_script(self.create_code())
 | 
						|
            self.compile_dir(optimize=[1, 0, 1, 0])
 | 
						|
            pyc1 = get_pyc(script, 0)
 | 
						|
            pyc2 = get_pyc(script, 1)
 | 
						|
            self.assertTrue(is_hardlink(pyc1, pyc2))
 | 
						|
 | 
						|
    def test_recompilation(self):
 | 
						|
        # Test compile_dir() when pyc files already exists and the script
 | 
						|
        # content changed
 | 
						|
        with self.temporary_directory():
 | 
						|
            script = self.make_script("a = 0")
 | 
						|
            self.compile_dir()
 | 
						|
            # All three levels have the same inode
 | 
						|
            self.check_hardlinks(script)
 | 
						|
 | 
						|
            pycs = get_pycs(script)
 | 
						|
            inode = os.stat(pycs[0]).st_ino
 | 
						|
 | 
						|
            # Change of the module content
 | 
						|
            script = self.make_script("print(0)")
 | 
						|
 | 
						|
            # Recompilation without -o 1
 | 
						|
            self.compile_dir(optimize=[0, 2], force=True)
 | 
						|
 | 
						|
            # opt-1.pyc should have the same inode as before and others should not
 | 
						|
            self.assertEqual(inode, os.stat(pycs[1]).st_ino)
 | 
						|
            self.assertTrue(is_hardlink(pycs[0], pycs[2]))
 | 
						|
            self.assertNotEqual(inode, os.stat(pycs[2]).st_ino)
 | 
						|
            # opt-1.pyc and opt-2.pyc have different content
 | 
						|
            self.assertFalse(filecmp.cmp(pycs[1], pycs[2], shallow=True))
 | 
						|
 | 
						|
    def test_import(self):
 | 
						|
        # Test that import updates a single pyc file when pyc files already
 | 
						|
        # exists and the script content changed
 | 
						|
        with self.temporary_directory():
 | 
						|
            script = self.make_script(self.create_code(), name="module")
 | 
						|
            self.compile_dir()
 | 
						|
            # All three levels have the same inode
 | 
						|
            self.check_hardlinks(script)
 | 
						|
 | 
						|
            pycs = get_pycs(script)
 | 
						|
            inode = os.stat(pycs[0]).st_ino
 | 
						|
 | 
						|
            # Change of the module content
 | 
						|
            script = self.make_script("print(0)", name="module")
 | 
						|
 | 
						|
            # Import the module in Python with -O (optimization level 1)
 | 
						|
            script_helper.assert_python_ok(
 | 
						|
                "-O", "-c", "import module", __isolated=False, PYTHONPATH=self.path
 | 
						|
            )
 | 
						|
 | 
						|
            # Only opt-1.pyc is changed
 | 
						|
            self.assertEqual(inode, os.stat(pycs[0]).st_ino)
 | 
						|
            self.assertEqual(inode, os.stat(pycs[2]).st_ino)
 | 
						|
            self.assertFalse(is_hardlink(pycs[1], pycs[2]))
 | 
						|
            # opt-1.pyc and opt-2.pyc have different content
 | 
						|
            self.assertFalse(filecmp.cmp(pycs[1], pycs[2], shallow=True))
 | 
						|
 | 
						|
 | 
						|
class HardlinkDedupTestsWithSourceEpoch(HardlinkDedupTestsBase,
 | 
						|
                                        unittest.TestCase,
 | 
						|
                                        metaclass=SourceDateEpochTestMeta,
 | 
						|
                                        source_date_epoch=True):
 | 
						|
    pass
 | 
						|
 | 
						|
 | 
						|
class HardlinkDedupTestsNoSourceEpoch(HardlinkDedupTestsBase,
 | 
						|
                                      unittest.TestCase,
 | 
						|
                                      metaclass=SourceDateEpochTestMeta,
 | 
						|
                                      source_date_epoch=False):
 | 
						|
    pass
 | 
						|
 | 
						|
 | 
						|
if __name__ == "__main__":
 | 
						|
    unittest.main()
 |