mirror of
https://github.com/clinton-hall/nzbToMedia.git
synced 2025-08-21 05:43:16 -07:00
Move common libs to libs/common
This commit is contained in:
parent
8dbb1a2451
commit
1f4bd41bcc
1612 changed files with 962 additions and 10 deletions
923
libs/common/pbr/tests/test_packaging.py
Normal file
923
libs/common/pbr/tests/test_packaging.py
Normal file
|
@ -0,0 +1,923 @@
|
|||
# Copyright (c) 2013 New Dream Network, LLC (DreamHost)
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||
# implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
#
|
||||
# Copyright (C) 2013 Association of Universities for Research in Astronomy
|
||||
# (AURA)
|
||||
#
|
||||
# Redistribution and use in source and binary forms, with or without
|
||||
# modification, are permitted provided that the following conditions are met:
|
||||
#
|
||||
# 1. Redistributions of source code must retain the above copyright
|
||||
# notice, this list of conditions and the following disclaimer.
|
||||
#
|
||||
# 2. Redistributions in binary form must reproduce the above
|
||||
# copyright notice, this list of conditions and the following
|
||||
# disclaimer in the documentation and/or other materials provided
|
||||
# with the distribution.
|
||||
#
|
||||
# 3. The name of AURA and its representatives may not be used to
|
||||
# endorse or promote products derived from this software without
|
||||
# specific prior written permission.
|
||||
#
|
||||
# THIS SOFTWARE IS PROVIDED BY AURA ``AS IS'' AND ANY EXPRESS OR IMPLIED
|
||||
# WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
|
||||
# MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
||||
# DISCLAIMED. IN NO EVENT SHALL AURA BE LIABLE FOR ANY DIRECT, INDIRECT,
|
||||
# INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
|
||||
# BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
|
||||
|
||||
import email
|
||||
import email.errors
|
||||
import imp
|
||||
import os
|
||||
import re
|
||||
import sysconfig
|
||||
import tempfile
|
||||
import textwrap
|
||||
|
||||
import fixtures
|
||||
import mock
|
||||
import pkg_resources
|
||||
import six
|
||||
import testscenarios
|
||||
import testtools
|
||||
from testtools import matchers
|
||||
import virtualenv
|
||||
from wheel import wheelfile
|
||||
|
||||
from pbr import git
|
||||
from pbr import packaging
|
||||
from pbr.tests import base
|
||||
|
||||
|
||||
PBR_ROOT = os.path.abspath(os.path.join(__file__, '..', '..', '..'))
|
||||
|
||||
|
||||
class TestRepo(fixtures.Fixture):
|
||||
"""A git repo for testing with.
|
||||
|
||||
Use of TempHomeDir with this fixture is strongly recommended as due to the
|
||||
lack of config --local in older gits, it will write to the users global
|
||||
configuration without TempHomeDir.
|
||||
"""
|
||||
|
||||
def __init__(self, basedir):
|
||||
super(TestRepo, self).__init__()
|
||||
self._basedir = basedir
|
||||
|
||||
def setUp(self):
|
||||
super(TestRepo, self).setUp()
|
||||
base._run_cmd(['git', 'init', '.'], self._basedir)
|
||||
base._config_git()
|
||||
base._run_cmd(['git', 'add', '.'], self._basedir)
|
||||
|
||||
def commit(self, message_content='test commit'):
|
||||
files = len(os.listdir(self._basedir))
|
||||
path = self._basedir + '/%d' % files
|
||||
open(path, 'wt').close()
|
||||
base._run_cmd(['git', 'add', path], self._basedir)
|
||||
base._run_cmd(['git', 'commit', '-m', message_content], self._basedir)
|
||||
|
||||
def uncommit(self):
|
||||
base._run_cmd(['git', 'reset', '--hard', 'HEAD^'], self._basedir)
|
||||
|
||||
def tag(self, version):
|
||||
base._run_cmd(
|
||||
['git', 'tag', '-sm', 'test tag', version], self._basedir)
|
||||
|
||||
|
||||
class GPGKeyFixture(fixtures.Fixture):
|
||||
"""Creates a GPG key for testing.
|
||||
|
||||
It's recommended that this be used in concert with a unique home
|
||||
directory.
|
||||
"""
|
||||
|
||||
def setUp(self):
|
||||
super(GPGKeyFixture, self).setUp()
|
||||
tempdir = self.useFixture(fixtures.TempDir())
|
||||
gnupg_version_re = re.compile('^gpg\s.*\s([\d+])\.([\d+])\.([\d+])')
|
||||
gnupg_version = base._run_cmd(['gpg', '--version'], tempdir.path)
|
||||
for line in gnupg_version[0].split('\n'):
|
||||
gnupg_version = gnupg_version_re.match(line)
|
||||
if gnupg_version:
|
||||
gnupg_version = (int(gnupg_version.group(1)),
|
||||
int(gnupg_version.group(2)),
|
||||
int(gnupg_version.group(3)))
|
||||
break
|
||||
else:
|
||||
if gnupg_version is None:
|
||||
gnupg_version = (0, 0, 0)
|
||||
config_file = tempdir.path + '/key-config'
|
||||
f = open(config_file, 'wt')
|
||||
try:
|
||||
if gnupg_version[0] == 2 and gnupg_version[1] >= 1:
|
||||
f.write("""
|
||||
%no-protection
|
||||
%transient-key
|
||||
""")
|
||||
f.write("""
|
||||
%no-ask-passphrase
|
||||
Key-Type: RSA
|
||||
Name-Real: Example Key
|
||||
Name-Comment: N/A
|
||||
Name-Email: example@example.com
|
||||
Expire-Date: 2d
|
||||
Preferences: (setpref)
|
||||
%commit
|
||||
""")
|
||||
finally:
|
||||
f.close()
|
||||
# Note that --quick-random (--debug-quick-random in GnuPG 2.x)
|
||||
# does not have a corresponding preferences file setting and
|
||||
# must be passed explicitly on the command line instead
|
||||
if gnupg_version[0] == 1:
|
||||
gnupg_random = '--quick-random'
|
||||
elif gnupg_version[0] >= 2:
|
||||
gnupg_random = '--debug-quick-random'
|
||||
else:
|
||||
gnupg_random = ''
|
||||
base._run_cmd(
|
||||
['gpg', '--gen-key', '--batch', gnupg_random, config_file],
|
||||
tempdir.path)
|
||||
|
||||
|
||||
class Venv(fixtures.Fixture):
|
||||
"""Create a virtual environment for testing with.
|
||||
|
||||
:attr path: The path to the environment root.
|
||||
:attr python: The path to the python binary in the environment.
|
||||
"""
|
||||
|
||||
def __init__(self, reason, modules=(), pip_cmd=None):
|
||||
"""Create a Venv fixture.
|
||||
|
||||
:param reason: A human readable string to bake into the venv
|
||||
file path to aid diagnostics in the case of failures.
|
||||
:param modules: A list of modules to install, defaults to latest
|
||||
pip, wheel, and the working copy of PBR.
|
||||
:attr pip_cmd: A list to override the default pip_cmd passed to
|
||||
python for installing base packages.
|
||||
"""
|
||||
self._reason = reason
|
||||
if modules == ():
|
||||
pbr = 'file://%s#egg=pbr' % PBR_ROOT
|
||||
modules = ['pip', 'wheel', pbr]
|
||||
self.modules = modules
|
||||
if pip_cmd is None:
|
||||
self.pip_cmd = ['-m', 'pip', 'install']
|
||||
else:
|
||||
self.pip_cmd = pip_cmd
|
||||
|
||||
def _setUp(self):
|
||||
path = self.useFixture(fixtures.TempDir()).path
|
||||
virtualenv.create_environment(path, clear=True)
|
||||
python = os.path.join(path, 'bin', 'python')
|
||||
command = [python] + self.pip_cmd + ['-U']
|
||||
if self.modules and len(self.modules) > 0:
|
||||
command.extend(self.modules)
|
||||
self.useFixture(base.CapturedSubprocess(
|
||||
'mkvenv-' + self._reason, command))
|
||||
self.addCleanup(delattr, self, 'path')
|
||||
self.addCleanup(delattr, self, 'python')
|
||||
self.path = path
|
||||
self.python = python
|
||||
return path, python
|
||||
|
||||
|
||||
class CreatePackages(fixtures.Fixture):
|
||||
"""Creates packages from dict with defaults
|
||||
|
||||
:param package_dirs: A dict of package name to directory strings
|
||||
{'pkg_a': '/tmp/path/to/tmp/pkg_a', 'pkg_b': '/tmp/path/to/tmp/pkg_b'}
|
||||
"""
|
||||
|
||||
defaults = {
|
||||
'setup.py': textwrap.dedent(six.u("""\
|
||||
#!/usr/bin/env python
|
||||
import setuptools
|
||||
setuptools.setup(
|
||||
setup_requires=['pbr'],
|
||||
pbr=True,
|
||||
)
|
||||
""")),
|
||||
'setup.cfg': textwrap.dedent(six.u("""\
|
||||
[metadata]
|
||||
name = {pkg_name}
|
||||
"""))
|
||||
}
|
||||
|
||||
def __init__(self, packages):
|
||||
"""Creates packages from dict with defaults
|
||||
|
||||
:param packages: a dict where the keys are the package name and a
|
||||
value that is a second dict that may be empty, containing keys of
|
||||
filenames and a string value of the contents.
|
||||
{'package-a': {'requirements.txt': 'string', 'setup.cfg': 'string'}
|
||||
"""
|
||||
self.packages = packages
|
||||
|
||||
def _writeFile(self, directory, file_name, contents):
|
||||
path = os.path.abspath(os.path.join(directory, file_name))
|
||||
path_dir = os.path.dirname(path)
|
||||
if not os.path.exists(path_dir):
|
||||
if path_dir.startswith(directory):
|
||||
os.makedirs(path_dir)
|
||||
else:
|
||||
raise ValueError
|
||||
with open(path, 'wt') as f:
|
||||
f.write(contents)
|
||||
|
||||
def _setUp(self):
|
||||
tmpdir = self.useFixture(fixtures.TempDir()).path
|
||||
package_dirs = {}
|
||||
for pkg_name in self.packages:
|
||||
pkg_path = os.path.join(tmpdir, pkg_name)
|
||||
package_dirs[pkg_name] = pkg_path
|
||||
os.mkdir(pkg_path)
|
||||
for cf in ['setup.py', 'setup.cfg']:
|
||||
if cf in self.packages[pkg_name]:
|
||||
contents = self.packages[pkg_name].pop(cf)
|
||||
else:
|
||||
contents = self.defaults[cf].format(pkg_name=pkg_name)
|
||||
self._writeFile(pkg_path, cf, contents)
|
||||
|
||||
for cf in self.packages[pkg_name]:
|
||||
self._writeFile(pkg_path, cf, self.packages[pkg_name][cf])
|
||||
self.useFixture(TestRepo(pkg_path)).commit()
|
||||
self.addCleanup(delattr, self, 'package_dirs')
|
||||
self.package_dirs = package_dirs
|
||||
return package_dirs
|
||||
|
||||
|
||||
class TestPackagingInGitRepoWithCommit(base.BaseTestCase):
|
||||
|
||||
scenarios = [
|
||||
('preversioned', dict(preversioned=True)),
|
||||
('postversioned', dict(preversioned=False)),
|
||||
]
|
||||
|
||||
def setUp(self):
|
||||
super(TestPackagingInGitRepoWithCommit, self).setUp()
|
||||
self.repo = self.useFixture(TestRepo(self.package_dir))
|
||||
self.repo.commit()
|
||||
|
||||
def test_authors(self):
|
||||
self.run_setup('sdist', allow_fail=False)
|
||||
# One commit, something should be in the authors list
|
||||
with open(os.path.join(self.package_dir, 'AUTHORS'), 'r') as f:
|
||||
body = f.read()
|
||||
self.assertNotEqual(body, '')
|
||||
|
||||
def test_changelog(self):
|
||||
self.run_setup('sdist', allow_fail=False)
|
||||
with open(os.path.join(self.package_dir, 'ChangeLog'), 'r') as f:
|
||||
body = f.read()
|
||||
# One commit, something should be in the ChangeLog list
|
||||
self.assertNotEqual(body, '')
|
||||
|
||||
def test_changelog_handles_astrisk(self):
|
||||
self.repo.commit(message_content="Allow *.openstack.org to work")
|
||||
self.run_setup('sdist', allow_fail=False)
|
||||
with open(os.path.join(self.package_dir, 'ChangeLog'), 'r') as f:
|
||||
body = f.read()
|
||||
self.assertIn('\*', body)
|
||||
|
||||
def test_changelog_handles_dead_links_in_commit(self):
|
||||
self.repo.commit(message_content="See os_ for to_do about qemu_.")
|
||||
self.run_setup('sdist', allow_fail=False)
|
||||
with open(os.path.join(self.package_dir, 'ChangeLog'), 'r') as f:
|
||||
body = f.read()
|
||||
self.assertIn('os\_', body)
|
||||
self.assertIn('to\_do', body)
|
||||
self.assertIn('qemu\_', body)
|
||||
|
||||
def test_changelog_handles_backticks(self):
|
||||
self.repo.commit(message_content="Allow `openstack.org` to `work")
|
||||
self.run_setup('sdist', allow_fail=False)
|
||||
with open(os.path.join(self.package_dir, 'ChangeLog'), 'r') as f:
|
||||
body = f.read()
|
||||
self.assertIn('\`', body)
|
||||
|
||||
def test_manifest_exclude_honoured(self):
|
||||
self.run_setup('sdist', allow_fail=False)
|
||||
with open(os.path.join(
|
||||
self.package_dir,
|
||||
'pbr_testpackage.egg-info/SOURCES.txt'), 'r') as f:
|
||||
body = f.read()
|
||||
self.assertThat(
|
||||
body, matchers.Not(matchers.Contains('pbr_testpackage/extra.py')))
|
||||
self.assertThat(body, matchers.Contains('pbr_testpackage/__init__.py'))
|
||||
|
||||
def test_install_writes_changelog(self):
|
||||
stdout, _, _ = self.run_setup(
|
||||
'install', '--root', self.temp_dir + 'installed',
|
||||
allow_fail=False)
|
||||
self.expectThat(stdout, matchers.Contains('Generating ChangeLog'))
|
||||
|
||||
|
||||
class TestExtrafileInstallation(base.BaseTestCase):
|
||||
def test_install_glob(self):
|
||||
stdout, _, _ = self.run_setup(
|
||||
'install', '--root', self.temp_dir + 'installed',
|
||||
allow_fail=False)
|
||||
self.expectThat(
|
||||
stdout, matchers.Contains('copying data_files/a.txt'))
|
||||
self.expectThat(
|
||||
stdout, matchers.Contains('copying data_files/b.txt'))
|
||||
|
||||
|
||||
class TestPackagingInGitRepoWithoutCommit(base.BaseTestCase):
|
||||
|
||||
def setUp(self):
|
||||
super(TestPackagingInGitRepoWithoutCommit, self).setUp()
|
||||
self.useFixture(TestRepo(self.package_dir))
|
||||
self.run_setup('sdist', allow_fail=False)
|
||||
|
||||
def test_authors(self):
|
||||
# No commits, no authors in list
|
||||
with open(os.path.join(self.package_dir, 'AUTHORS'), 'r') as f:
|
||||
body = f.read()
|
||||
self.assertEqual('\n', body)
|
||||
|
||||
def test_changelog(self):
|
||||
# No commits, nothing should be in the ChangeLog list
|
||||
with open(os.path.join(self.package_dir, 'ChangeLog'), 'r') as f:
|
||||
body = f.read()
|
||||
self.assertEqual('CHANGES\n=======\n\n', body)
|
||||
|
||||
|
||||
class TestPackagingWheels(base.BaseTestCase):
|
||||
|
||||
def setUp(self):
|
||||
super(TestPackagingWheels, self).setUp()
|
||||
self.useFixture(TestRepo(self.package_dir))
|
||||
# Build the wheel
|
||||
self.run_setup('bdist_wheel', allow_fail=False)
|
||||
# Slowly construct the path to the generated whl
|
||||
dist_dir = os.path.join(self.package_dir, 'dist')
|
||||
relative_wheel_filename = os.listdir(dist_dir)[0]
|
||||
absolute_wheel_filename = os.path.join(
|
||||
dist_dir, relative_wheel_filename)
|
||||
wheel_file = wheelfile.WheelFile(absolute_wheel_filename)
|
||||
wheel_name = wheel_file.parsed_filename.group('namever')
|
||||
# Create a directory path to unpack the wheel to
|
||||
self.extracted_wheel_dir = os.path.join(dist_dir, wheel_name)
|
||||
# Extract the wheel contents to the directory we just created
|
||||
wheel_file.extractall(self.extracted_wheel_dir)
|
||||
wheel_file.close()
|
||||
|
||||
def test_data_directory_has_wsgi_scripts(self):
|
||||
# Build the path to the scripts directory
|
||||
scripts_dir = os.path.join(
|
||||
self.extracted_wheel_dir, 'pbr_testpackage-0.0.data/scripts')
|
||||
self.assertTrue(os.path.exists(scripts_dir))
|
||||
scripts = os.listdir(scripts_dir)
|
||||
|
||||
self.assertIn('pbr_test_wsgi', scripts)
|
||||
self.assertIn('pbr_test_wsgi_with_class', scripts)
|
||||
self.assertNotIn('pbr_test_cmd', scripts)
|
||||
self.assertNotIn('pbr_test_cmd_with_class', scripts)
|
||||
|
||||
def test_generates_c_extensions(self):
|
||||
built_package_dir = os.path.join(
|
||||
self.extracted_wheel_dir, 'pbr_testpackage')
|
||||
static_object_filename = 'testext.so'
|
||||
soabi = get_soabi()
|
||||
if soabi:
|
||||
static_object_filename = 'testext.{0}.so'.format(soabi)
|
||||
static_object_path = os.path.join(
|
||||
built_package_dir, static_object_filename)
|
||||
|
||||
self.assertTrue(os.path.exists(built_package_dir))
|
||||
self.assertTrue(os.path.exists(static_object_path))
|
||||
|
||||
|
||||
class TestPackagingHelpers(testtools.TestCase):
|
||||
|
||||
def test_generate_script(self):
|
||||
group = 'console_scripts'
|
||||
entry_point = pkg_resources.EntryPoint(
|
||||
name='test-ep',
|
||||
module_name='pbr.packaging',
|
||||
attrs=('LocalInstallScripts',))
|
||||
header = '#!/usr/bin/env fake-header\n'
|
||||
template = ('%(group)s %(module_name)s %(import_target)s '
|
||||
'%(invoke_target)s')
|
||||
|
||||
generated_script = packaging.generate_script(
|
||||
group, entry_point, header, template)
|
||||
|
||||
expected_script = (
|
||||
'#!/usr/bin/env fake-header\nconsole_scripts pbr.packaging '
|
||||
'LocalInstallScripts LocalInstallScripts'
|
||||
)
|
||||
self.assertEqual(expected_script, generated_script)
|
||||
|
||||
def test_generate_script_validates_expectations(self):
|
||||
group = 'console_scripts'
|
||||
entry_point = pkg_resources.EntryPoint(
|
||||
name='test-ep',
|
||||
module_name='pbr.packaging')
|
||||
header = '#!/usr/bin/env fake-header\n'
|
||||
template = ('%(group)s %(module_name)s %(import_target)s '
|
||||
'%(invoke_target)s')
|
||||
self.assertRaises(
|
||||
ValueError, packaging.generate_script, group, entry_point, header,
|
||||
template)
|
||||
|
||||
entry_point = pkg_resources.EntryPoint(
|
||||
name='test-ep',
|
||||
module_name='pbr.packaging',
|
||||
attrs=('attr1', 'attr2', 'attr3'))
|
||||
self.assertRaises(
|
||||
ValueError, packaging.generate_script, group, entry_point, header,
|
||||
template)
|
||||
|
||||
|
||||
class TestPackagingInPlainDirectory(base.BaseTestCase):
|
||||
|
||||
def setUp(self):
|
||||
super(TestPackagingInPlainDirectory, self).setUp()
|
||||
|
||||
def test_authors(self):
|
||||
self.run_setup('sdist', allow_fail=False)
|
||||
# Not a git repo, no AUTHORS file created
|
||||
filename = os.path.join(self.package_dir, 'AUTHORS')
|
||||
self.assertFalse(os.path.exists(filename))
|
||||
|
||||
def test_changelog(self):
|
||||
self.run_setup('sdist', allow_fail=False)
|
||||
# Not a git repo, no ChangeLog created
|
||||
filename = os.path.join(self.package_dir, 'ChangeLog')
|
||||
self.assertFalse(os.path.exists(filename))
|
||||
|
||||
def test_install_no_ChangeLog(self):
|
||||
stdout, _, _ = self.run_setup(
|
||||
'install', '--root', self.temp_dir + 'installed',
|
||||
allow_fail=False)
|
||||
self.expectThat(
|
||||
stdout, matchers.Not(matchers.Contains('Generating ChangeLog')))
|
||||
|
||||
|
||||
class TestPresenceOfGit(base.BaseTestCase):
|
||||
|
||||
def testGitIsInstalled(self):
|
||||
with mock.patch.object(git,
|
||||
'_run_shell_command') as _command:
|
||||
_command.return_value = 'git version 1.8.4.1'
|
||||
self.assertEqual(True, git._git_is_installed())
|
||||
|
||||
def testGitIsNotInstalled(self):
|
||||
with mock.patch.object(git,
|
||||
'_run_shell_command') as _command:
|
||||
_command.side_effect = OSError
|
||||
self.assertEqual(False, git._git_is_installed())
|
||||
|
||||
|
||||
class ParseRequirementsTest(base.BaseTestCase):
|
||||
|
||||
def test_empty_requirements(self):
|
||||
actual = packaging.parse_requirements([])
|
||||
self.assertEqual([], actual)
|
||||
|
||||
def test_default_requirements(self):
|
||||
"""Ensure default files used if no files provided."""
|
||||
tempdir = tempfile.mkdtemp()
|
||||
requirements = os.path.join(tempdir, 'requirements.txt')
|
||||
with open(requirements, 'w') as f:
|
||||
f.write('pbr')
|
||||
# the defaults are relative to where pbr is called from so we need to
|
||||
# override them. This is OK, however, as we want to validate that
|
||||
# defaults are used - not what those defaults are
|
||||
with mock.patch.object(packaging, 'REQUIREMENTS_FILES', (
|
||||
requirements,)):
|
||||
result = packaging.parse_requirements()
|
||||
self.assertEqual(['pbr'], result)
|
||||
|
||||
def test_override_with_env(self):
|
||||
"""Ensure environment variable used if no files provided."""
|
||||
_, tmp_file = tempfile.mkstemp(prefix='openstack', suffix='.setup')
|
||||
with open(tmp_file, 'w') as fh:
|
||||
fh.write("foo\nbar")
|
||||
self.useFixture(
|
||||
fixtures.EnvironmentVariable('PBR_REQUIREMENTS_FILES', tmp_file))
|
||||
self.assertEqual(['foo', 'bar'],
|
||||
packaging.parse_requirements())
|
||||
|
||||
def test_override_with_env_multiple_files(self):
|
||||
_, tmp_file = tempfile.mkstemp(prefix='openstack', suffix='.setup')
|
||||
with open(tmp_file, 'w') as fh:
|
||||
fh.write("foo\nbar")
|
||||
self.useFixture(
|
||||
fixtures.EnvironmentVariable('PBR_REQUIREMENTS_FILES',
|
||||
"no-such-file," + tmp_file))
|
||||
self.assertEqual(['foo', 'bar'],
|
||||
packaging.parse_requirements())
|
||||
|
||||
def test_index_present(self):
|
||||
tempdir = tempfile.mkdtemp()
|
||||
requirements = os.path.join(tempdir, 'requirements.txt')
|
||||
with open(requirements, 'w') as f:
|
||||
f.write('-i https://myindex.local')
|
||||
f.write(' --index-url https://myindex.local')
|
||||
f.write(' --extra-index-url https://myindex.local')
|
||||
result = packaging.parse_requirements([requirements])
|
||||
self.assertEqual([], result)
|
||||
|
||||
def test_nested_requirements(self):
|
||||
tempdir = tempfile.mkdtemp()
|
||||
requirements = os.path.join(tempdir, 'requirements.txt')
|
||||
nested = os.path.join(tempdir, 'nested.txt')
|
||||
with open(requirements, 'w') as f:
|
||||
f.write('-r ' + nested)
|
||||
with open(nested, 'w') as f:
|
||||
f.write('pbr')
|
||||
result = packaging.parse_requirements([requirements])
|
||||
self.assertEqual(['pbr'], result)
|
||||
|
||||
|
||||
class ParseRequirementsTestScenarios(base.BaseTestCase):
|
||||
|
||||
versioned_scenarios = [
|
||||
('non-versioned', {'versioned': False, 'expected': ['bar']}),
|
||||
('versioned', {'versioned': True, 'expected': ['bar>=1.2.3']})
|
||||
]
|
||||
|
||||
subdirectory_scenarios = [
|
||||
('non-subdirectory', {'has_subdirectory': False}),
|
||||
('has-subdirectory', {'has_subdirectory': True})
|
||||
]
|
||||
|
||||
scenarios = [
|
||||
('normal', {'url': "foo\nbar", 'expected': ['foo', 'bar']}),
|
||||
('normal_with_comments', {
|
||||
'url': "# this is a comment\nfoo\n# and another one\nbar",
|
||||
'expected': ['foo', 'bar']}),
|
||||
('removes_index_lines', {'url': '-f foobar', 'expected': []}),
|
||||
]
|
||||
|
||||
scenarios = scenarios + testscenarios.multiply_scenarios([
|
||||
('ssh_egg_url', {'url': 'git+ssh://foo.com/zipball#egg=bar'}),
|
||||
('git_https_egg_url', {'url': 'git+https://foo.com/zipball#egg=bar'}),
|
||||
('http_egg_url', {'url': 'https://foo.com/zipball#egg=bar'}),
|
||||
], versioned_scenarios, subdirectory_scenarios)
|
||||
|
||||
scenarios = scenarios + testscenarios.multiply_scenarios(
|
||||
[
|
||||
('git_egg_url',
|
||||
{'url': 'git://foo.com/zipball#egg=bar', 'name': 'bar'})
|
||||
], [
|
||||
('non-editable', {'editable': False}),
|
||||
('editable', {'editable': True}),
|
||||
],
|
||||
versioned_scenarios, subdirectory_scenarios)
|
||||
|
||||
def test_parse_requirements(self):
|
||||
tmp_file = tempfile.NamedTemporaryFile()
|
||||
req_string = self.url
|
||||
if hasattr(self, 'editable') and self.editable:
|
||||
req_string = ("-e %s" % req_string)
|
||||
if hasattr(self, 'versioned') and self.versioned:
|
||||
req_string = ("%s-1.2.3" % req_string)
|
||||
if hasattr(self, 'has_subdirectory') and self.has_subdirectory:
|
||||
req_string = ("%s&subdirectory=baz" % req_string)
|
||||
with open(tmp_file.name, 'w') as fh:
|
||||
fh.write(req_string)
|
||||
self.assertEqual(self.expected,
|
||||
packaging.parse_requirements([tmp_file.name]))
|
||||
|
||||
|
||||
class ParseDependencyLinksTest(base.BaseTestCase):
|
||||
|
||||
def setUp(self):
|
||||
super(ParseDependencyLinksTest, self).setUp()
|
||||
_, self.tmp_file = tempfile.mkstemp(prefix="openstack",
|
||||
suffix=".setup")
|
||||
|
||||
def test_parse_dependency_normal(self):
|
||||
with open(self.tmp_file, "w") as fh:
|
||||
fh.write("http://test.com\n")
|
||||
self.assertEqual(
|
||||
["http://test.com"],
|
||||
packaging.parse_dependency_links([self.tmp_file]))
|
||||
|
||||
def test_parse_dependency_with_git_egg_url(self):
|
||||
with open(self.tmp_file, "w") as fh:
|
||||
fh.write("-e git://foo.com/zipball#egg=bar")
|
||||
self.assertEqual(
|
||||
["git://foo.com/zipball#egg=bar"],
|
||||
packaging.parse_dependency_links([self.tmp_file]))
|
||||
|
||||
|
||||
class TestVersions(base.BaseTestCase):
|
||||
|
||||
scenarios = [
|
||||
('preversioned', dict(preversioned=True)),
|
||||
('postversioned', dict(preversioned=False)),
|
||||
]
|
||||
|
||||
def setUp(self):
|
||||
super(TestVersions, self).setUp()
|
||||
self.repo = self.useFixture(TestRepo(self.package_dir))
|
||||
self.useFixture(GPGKeyFixture())
|
||||
self.useFixture(base.DiveDir(self.package_dir))
|
||||
|
||||
def test_email_parsing_errors_are_handled(self):
|
||||
mocked_open = mock.mock_open()
|
||||
with mock.patch('pbr.packaging.open', mocked_open):
|
||||
with mock.patch('email.message_from_file') as message_from_file:
|
||||
message_from_file.side_effect = [
|
||||
email.errors.MessageError('Test'),
|
||||
{'Name': 'pbr_testpackage'}]
|
||||
version = packaging._get_version_from_pkg_metadata(
|
||||
'pbr_testpackage')
|
||||
|
||||
self.assertTrue(message_from_file.called)
|
||||
self.assertIsNone(version)
|
||||
|
||||
def test_capitalized_headers(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
self.repo.commit('Sem-Ver: api-break')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('2.0.0.dev1'))
|
||||
|
||||
def test_capitalized_headers_partial(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
self.repo.commit('Sem-ver: api-break')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('2.0.0.dev1'))
|
||||
|
||||
def test_tagged_version_has_tag_version(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
version = packaging._get_version_from_git('1.2.3')
|
||||
self.assertEqual('1.2.3', version)
|
||||
|
||||
def test_non_canonical_tagged_version_bump(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.4')
|
||||
self.repo.commit('Sem-Ver: api-break')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('2.0.0.dev1'))
|
||||
|
||||
def test_untagged_version_has_dev_version_postversion(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
self.repo.commit()
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('1.2.4.dev1'))
|
||||
|
||||
def test_untagged_pre_release_has_pre_dev_version_postversion(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3.0a1')
|
||||
self.repo.commit()
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('1.2.3.0a2.dev1'))
|
||||
|
||||
def test_untagged_version_minor_bump(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
self.repo.commit('sem-ver: deprecation')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('1.3.0.dev1'))
|
||||
|
||||
def test_untagged_version_major_bump(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
self.repo.commit('sem-ver: api-break')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('2.0.0.dev1'))
|
||||
|
||||
def test_untagged_version_has_dev_version_preversion(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
self.repo.commit()
|
||||
version = packaging._get_version_from_git('1.2.5')
|
||||
self.assertThat(version, matchers.StartsWith('1.2.5.dev1'))
|
||||
|
||||
def test_untagged_version_after_pre_has_dev_version_preversion(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3.0a1')
|
||||
self.repo.commit()
|
||||
version = packaging._get_version_from_git('1.2.5')
|
||||
self.assertThat(version, matchers.StartsWith('1.2.5.dev1'))
|
||||
|
||||
def test_untagged_version_after_rc_has_dev_version_preversion(self):
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3.0a1')
|
||||
self.repo.commit()
|
||||
version = packaging._get_version_from_git('1.2.3')
|
||||
self.assertThat(version, matchers.StartsWith('1.2.3.0a2.dev1'))
|
||||
|
||||
def test_preversion_too_low_simple(self):
|
||||
# That is, the target version is either already released or not high
|
||||
# enough for the semver requirements given api breaks etc.
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
self.repo.commit()
|
||||
# Note that we can't target 1.2.3 anymore - with 1.2.3 released we
|
||||
# need to be working on 1.2.4.
|
||||
err = self.assertRaises(
|
||||
ValueError, packaging._get_version_from_git, '1.2.3')
|
||||
self.assertThat(err.args[0], matchers.StartsWith('git history'))
|
||||
|
||||
def test_preversion_too_low_semver_headers(self):
|
||||
# That is, the target version is either already released or not high
|
||||
# enough for the semver requirements given api breaks etc.
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
self.repo.commit('sem-ver: feature')
|
||||
# Note that we can't target 1.2.4, the feature header means we need
|
||||
# to be working on 1.3.0 or above.
|
||||
err = self.assertRaises(
|
||||
ValueError, packaging._get_version_from_git, '1.2.4')
|
||||
self.assertThat(err.args[0], matchers.StartsWith('git history'))
|
||||
|
||||
def test_get_kwargs_corner_cases(self):
|
||||
# No tags:
|
||||
git_dir = self.repo._basedir + '/.git'
|
||||
get_kwargs = lambda tag: packaging._get_increment_kwargs(git_dir, tag)
|
||||
|
||||
def _check_combinations(tag):
|
||||
self.repo.commit()
|
||||
self.assertEqual(dict(), get_kwargs(tag))
|
||||
self.repo.commit('sem-ver: bugfix')
|
||||
self.assertEqual(dict(), get_kwargs(tag))
|
||||
self.repo.commit('sem-ver: feature')
|
||||
self.assertEqual(dict(minor=True), get_kwargs(tag))
|
||||
self.repo.uncommit()
|
||||
self.repo.commit('sem-ver: deprecation')
|
||||
self.assertEqual(dict(minor=True), get_kwargs(tag))
|
||||
self.repo.uncommit()
|
||||
self.repo.commit('sem-ver: api-break')
|
||||
self.assertEqual(dict(major=True), get_kwargs(tag))
|
||||
self.repo.commit('sem-ver: deprecation')
|
||||
self.assertEqual(dict(major=True, minor=True), get_kwargs(tag))
|
||||
_check_combinations('')
|
||||
self.repo.tag('1.2.3')
|
||||
_check_combinations('1.2.3')
|
||||
|
||||
def test_invalid_tag_ignored(self):
|
||||
# Fix for bug 1356784 - we treated any tag as a version, not just those
|
||||
# that are valid versions.
|
||||
self.repo.commit()
|
||||
self.repo.tag('1')
|
||||
self.repo.commit()
|
||||
# when the tree is tagged and its wrong:
|
||||
self.repo.tag('badver')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('1.0.1.dev1'))
|
||||
# When the tree isn't tagged, we also fall through.
|
||||
self.repo.commit()
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('1.0.1.dev2'))
|
||||
# We don't fall through x.y versions
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2')
|
||||
self.repo.commit()
|
||||
self.repo.tag('badver2')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('1.2.1.dev1'))
|
||||
# Or x.y.z versions
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
self.repo.commit()
|
||||
self.repo.tag('badver3')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('1.2.4.dev1'))
|
||||
# Or alpha/beta/pre versions
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.4.0a1')
|
||||
self.repo.commit()
|
||||
self.repo.tag('badver4')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('1.2.4.0a2.dev1'))
|
||||
# Non-release related tags are ignored.
|
||||
self.repo.commit()
|
||||
self.repo.tag('2')
|
||||
self.repo.commit()
|
||||
self.repo.tag('non-release-tag/2014.12.16-1')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertThat(version, matchers.StartsWith('2.0.1.dev1'))
|
||||
|
||||
def test_valid_tag_honoured(self):
|
||||
# Fix for bug 1370608 - we converted any target into a 'dev version'
|
||||
# even if there was a distance of 0 - indicating that we were on the
|
||||
# tag itself.
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.3.0.0a1')
|
||||
version = packaging._get_version_from_git()
|
||||
self.assertEqual('1.3.0.0a1', version)
|
||||
|
||||
def test_skip_write_git_changelog(self):
|
||||
# Fix for bug 1467440
|
||||
self.repo.commit()
|
||||
self.repo.tag('1.2.3')
|
||||
os.environ['SKIP_WRITE_GIT_CHANGELOG'] = '1'
|
||||
version = packaging._get_version_from_git('1.2.3')
|
||||
self.assertEqual('1.2.3', version)
|
||||
|
||||
def tearDown(self):
|
||||
super(TestVersions, self).tearDown()
|
||||
os.environ.pop('SKIP_WRITE_GIT_CHANGELOG', None)
|
||||
|
||||
|
||||
class TestRequirementParsing(base.BaseTestCase):
|
||||
|
||||
def test_requirement_parsing(self):
|
||||
pkgs = {
|
||||
'test_reqparse':
|
||||
{
|
||||
'requirements.txt': textwrap.dedent("""\
|
||||
bar
|
||||
quux<1.0; python_version=='2.6'
|
||||
requests-aws>=0.1.4 # BSD License (3 clause)
|
||||
Routes>=1.12.3,!=2.0,!=2.1;python_version=='2.7'
|
||||
requests-kerberos>=0.6;python_version=='2.7' # MIT
|
||||
"""),
|
||||
'setup.cfg': textwrap.dedent("""\
|
||||
[metadata]
|
||||
name = test_reqparse
|
||||
|
||||
[extras]
|
||||
test =
|
||||
foo
|
||||
baz>3.2 :python_version=='2.7' # MIT
|
||||
bar>3.3 :python_version=='2.7' # MIT # Apache
|
||||
""")},
|
||||
}
|
||||
pkg_dirs = self.useFixture(CreatePackages(pkgs)).package_dirs
|
||||
pkg_dir = pkg_dirs['test_reqparse']
|
||||
# pkg_resources.split_sections uses None as the title of an
|
||||
# anonymous section instead of the empty string. Weird.
|
||||
expected_requirements = {
|
||||
None: ['bar', 'requests-aws>=0.1.4'],
|
||||
":(python_version=='2.6')": ['quux<1.0'],
|
||||
":(python_version=='2.7')": ['Routes!=2.0,!=2.1,>=1.12.3',
|
||||
'requests-kerberos>=0.6'],
|
||||
'test': ['foo'],
|
||||
"test:(python_version=='2.7')": ['baz>3.2', 'bar>3.3']
|
||||
}
|
||||
venv = self.useFixture(Venv('reqParse'))
|
||||
bin_python = venv.python
|
||||
# Two things are tested by this
|
||||
# 1) pbr properly parses markers from requiremnts.txt and setup.cfg
|
||||
# 2) bdist_wheel causes pbr to not evaluate markers
|
||||
self._run_cmd(bin_python, ('setup.py', 'bdist_wheel'),
|
||||
allow_fail=False, cwd=pkg_dir)
|
||||
egg_info = os.path.join(pkg_dir, 'test_reqparse.egg-info')
|
||||
|
||||
requires_txt = os.path.join(egg_info, 'requires.txt')
|
||||
with open(requires_txt, 'rt') as requires:
|
||||
generated_requirements = dict(
|
||||
pkg_resources.split_sections(requires))
|
||||
|
||||
# NOTE(dhellmann): We have to spell out the comparison because
|
||||
# the rendering for version specifiers in a range is not
|
||||
# consistent across versions of setuptools.
|
||||
|
||||
for section, expected in expected_requirements.items():
|
||||
exp_parsed = [
|
||||
pkg_resources.Requirement.parse(s)
|
||||
for s in expected
|
||||
]
|
||||
gen_parsed = [
|
||||
pkg_resources.Requirement.parse(s)
|
||||
for s in generated_requirements[section]
|
||||
]
|
||||
self.assertEqual(exp_parsed, gen_parsed)
|
||||
|
||||
|
||||
def get_soabi():
|
||||
soabi = None
|
||||
try:
|
||||
soabi = sysconfig.get_config_var('SOABI')
|
||||
arch = sysconfig.get_config_var('MULTIARCH')
|
||||
except IOError:
|
||||
pass
|
||||
if soabi and arch and 'pypy' in sysconfig.get_scheme_names():
|
||||
soabi = '%s-%s' % (soabi, arch)
|
||||
if soabi is None and 'pypy' in sysconfig.get_scheme_names():
|
||||
# NOTE(sigmavirus24): PyPy only added support for the SOABI config var
|
||||
# to sysconfig in 2015. That was well after 2.2.1 was published in the
|
||||
# Ubuntu 14.04 archive.
|
||||
for suffix, _, _ in imp.get_suffixes():
|
||||
if suffix.startswith('.pypy') and suffix.endswith('.so'):
|
||||
soabi = suffix.split('.')[1]
|
||||
break
|
||||
return soabi
|
Loading…
Add table
Add a link
Reference in a new issue