mirror of https://github.com/sgoudham/Enso-Bot.git
You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1057 lines
38 KiB
Python
1057 lines
38 KiB
Python
5 years ago
|
# -*- coding: utf-8 -*-
|
||
|
#
|
||
|
# Copyright (C) 2012 The Python Software Foundation.
|
||
|
# See LICENSE.txt and CONTRIBUTORS.txt.
|
||
|
#
|
||
|
"""Implementation of the Metadata for Python packages PEPs.
|
||
|
|
||
|
Supports all metadata formats (1.0, 1.1, 1.2, 1.3/2.1 and withdrawn 2.0).
|
||
|
"""
|
||
|
from __future__ import unicode_literals
|
||
|
|
||
|
import codecs
|
||
|
from email import message_from_file
|
||
|
import json
|
||
|
import logging
|
||
|
import re
|
||
|
|
||
|
|
||
|
from . import DistlibException, __version__
|
||
|
from .compat import StringIO, string_types, text_type
|
||
|
from .markers import interpret
|
||
|
from .util import extract_by_key, get_extras
|
||
|
from .version import get_scheme, PEP440_VERSION_RE
|
||
|
|
||
|
logger = logging.getLogger(__name__)
|
||
|
|
||
|
|
||
|
class MetadataMissingError(DistlibException):
|
||
|
"""A required metadata is missing"""
|
||
|
|
||
|
|
||
|
class MetadataConflictError(DistlibException):
|
||
|
"""Attempt to read or write metadata fields that are conflictual."""
|
||
|
|
||
|
|
||
|
class MetadataUnrecognizedVersionError(DistlibException):
|
||
|
"""Unknown metadata version number."""
|
||
|
|
||
|
|
||
|
class MetadataInvalidError(DistlibException):
|
||
|
"""A metadata value is invalid"""
|
||
|
|
||
|
# public API of this module
|
||
|
__all__ = ['Metadata', 'PKG_INFO_ENCODING', 'PKG_INFO_PREFERRED_VERSION']
|
||
|
|
||
|
# Encoding used for the PKG-INFO files
|
||
|
PKG_INFO_ENCODING = 'utf-8'
|
||
|
|
||
|
# preferred version. Hopefully will be changed
|
||
|
# to 1.2 once PEP 345 is supported everywhere
|
||
|
PKG_INFO_PREFERRED_VERSION = '1.1'
|
||
|
|
||
|
_LINE_PREFIX_1_2 = re.compile('\n \\|')
|
||
|
_LINE_PREFIX_PRE_1_2 = re.compile('\n ')
|
||
|
_241_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
|
||
|
'Summary', 'Description',
|
||
|
'Keywords', 'Home-page', 'Author', 'Author-email',
|
||
|
'License')
|
||
|
|
||
|
_314_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
|
||
|
'Supported-Platform', 'Summary', 'Description',
|
||
|
'Keywords', 'Home-page', 'Author', 'Author-email',
|
||
|
'License', 'Classifier', 'Download-URL', 'Obsoletes',
|
||
|
'Provides', 'Requires')
|
||
|
|
||
|
_314_MARKERS = ('Obsoletes', 'Provides', 'Requires', 'Classifier',
|
||
|
'Download-URL')
|
||
|
|
||
|
_345_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
|
||
|
'Supported-Platform', 'Summary', 'Description',
|
||
|
'Keywords', 'Home-page', 'Author', 'Author-email',
|
||
|
'Maintainer', 'Maintainer-email', 'License',
|
||
|
'Classifier', 'Download-URL', 'Obsoletes-Dist',
|
||
|
'Project-URL', 'Provides-Dist', 'Requires-Dist',
|
||
|
'Requires-Python', 'Requires-External')
|
||
|
|
||
|
_345_MARKERS = ('Provides-Dist', 'Requires-Dist', 'Requires-Python',
|
||
|
'Obsoletes-Dist', 'Requires-External', 'Maintainer',
|
||
|
'Maintainer-email', 'Project-URL')
|
||
|
|
||
|
_426_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
|
||
|
'Supported-Platform', 'Summary', 'Description',
|
||
|
'Keywords', 'Home-page', 'Author', 'Author-email',
|
||
|
'Maintainer', 'Maintainer-email', 'License',
|
||
|
'Classifier', 'Download-URL', 'Obsoletes-Dist',
|
||
|
'Project-URL', 'Provides-Dist', 'Requires-Dist',
|
||
|
'Requires-Python', 'Requires-External', 'Private-Version',
|
||
|
'Obsoleted-By', 'Setup-Requires-Dist', 'Extension',
|
||
|
'Provides-Extra')
|
||
|
|
||
|
_426_MARKERS = ('Private-Version', 'Provides-Extra', 'Obsoleted-By',
|
||
|
'Setup-Requires-Dist', 'Extension')
|
||
|
|
||
|
# See issue #106: Sometimes 'Requires' and 'Provides' occur wrongly in
|
||
|
# the metadata. Include them in the tuple literal below to allow them
|
||
|
# (for now).
|
||
|
_566_FIELDS = _426_FIELDS + ('Description-Content-Type',
|
||
|
'Requires', 'Provides')
|
||
|
|
||
|
_566_MARKERS = ('Description-Content-Type',)
|
||
|
|
||
|
_ALL_FIELDS = set()
|
||
|
_ALL_FIELDS.update(_241_FIELDS)
|
||
|
_ALL_FIELDS.update(_314_FIELDS)
|
||
|
_ALL_FIELDS.update(_345_FIELDS)
|
||
|
_ALL_FIELDS.update(_426_FIELDS)
|
||
|
_ALL_FIELDS.update(_566_FIELDS)
|
||
|
|
||
|
EXTRA_RE = re.compile(r'''extra\s*==\s*("([^"]+)"|'([^']+)')''')
|
||
|
|
||
|
|
||
|
def _version2fieldlist(version):
|
||
|
if version == '1.0':
|
||
|
return _241_FIELDS
|
||
|
elif version == '1.1':
|
||
|
return _314_FIELDS
|
||
|
elif version == '1.2':
|
||
|
return _345_FIELDS
|
||
|
elif version in ('1.3', '2.1'):
|
||
|
return _345_FIELDS + _566_FIELDS
|
||
|
elif version == '2.0':
|
||
|
return _426_FIELDS
|
||
|
raise MetadataUnrecognizedVersionError(version)
|
||
|
|
||
|
|
||
|
def _best_version(fields):
|
||
|
"""Detect the best version depending on the fields used."""
|
||
|
def _has_marker(keys, markers):
|
||
|
for marker in markers:
|
||
|
if marker in keys:
|
||
|
return True
|
||
|
return False
|
||
|
|
||
|
keys = []
|
||
|
for key, value in fields.items():
|
||
|
if value in ([], 'UNKNOWN', None):
|
||
|
continue
|
||
|
keys.append(key)
|
||
|
|
||
|
possible_versions = ['1.0', '1.1', '1.2', '1.3', '2.0', '2.1']
|
||
|
|
||
|
# first let's try to see if a field is not part of one of the version
|
||
|
for key in keys:
|
||
|
if key not in _241_FIELDS and '1.0' in possible_versions:
|
||
|
possible_versions.remove('1.0')
|
||
|
logger.debug('Removed 1.0 due to %s', key)
|
||
|
if key not in _314_FIELDS and '1.1' in possible_versions:
|
||
|
possible_versions.remove('1.1')
|
||
|
logger.debug('Removed 1.1 due to %s', key)
|
||
|
if key not in _345_FIELDS and '1.2' in possible_versions:
|
||
|
possible_versions.remove('1.2')
|
||
|
logger.debug('Removed 1.2 due to %s', key)
|
||
|
if key not in _566_FIELDS and '1.3' in possible_versions:
|
||
|
possible_versions.remove('1.3')
|
||
|
logger.debug('Removed 1.3 due to %s', key)
|
||
|
if key not in _566_FIELDS and '2.1' in possible_versions:
|
||
|
if key != 'Description': # In 2.1, description allowed after headers
|
||
|
possible_versions.remove('2.1')
|
||
|
logger.debug('Removed 2.1 due to %s', key)
|
||
|
if key not in _426_FIELDS and '2.0' in possible_versions:
|
||
|
possible_versions.remove('2.0')
|
||
|
logger.debug('Removed 2.0 due to %s', key)
|
||
|
|
||
|
# possible_version contains qualified versions
|
||
|
if len(possible_versions) == 1:
|
||
|
return possible_versions[0] # found !
|
||
|
elif len(possible_versions) == 0:
|
||
|
logger.debug('Out of options - unknown metadata set: %s', fields)
|
||
|
raise MetadataConflictError('Unknown metadata set')
|
||
|
|
||
|
# let's see if one unique marker is found
|
||
|
is_1_1 = '1.1' in possible_versions and _has_marker(keys, _314_MARKERS)
|
||
|
is_1_2 = '1.2' in possible_versions and _has_marker(keys, _345_MARKERS)
|
||
|
is_2_1 = '2.1' in possible_versions and _has_marker(keys, _566_MARKERS)
|
||
|
is_2_0 = '2.0' in possible_versions and _has_marker(keys, _426_MARKERS)
|
||
|
if int(is_1_1) + int(is_1_2) + int(is_2_1) + int(is_2_0) > 1:
|
||
|
raise MetadataConflictError('You used incompatible 1.1/1.2/2.0/2.1 fields')
|
||
|
|
||
|
# we have the choice, 1.0, or 1.2, or 2.0
|
||
|
# - 1.0 has a broken Summary field but works with all tools
|
||
|
# - 1.1 is to avoid
|
||
|
# - 1.2 fixes Summary but has little adoption
|
||
|
# - 2.0 adds more features and is very new
|
||
|
if not is_1_1 and not is_1_2 and not is_2_1 and not is_2_0:
|
||
|
# we couldn't find any specific marker
|
||
|
if PKG_INFO_PREFERRED_VERSION in possible_versions:
|
||
|
return PKG_INFO_PREFERRED_VERSION
|
||
|
if is_1_1:
|
||
|
return '1.1'
|
||
|
if is_1_2:
|
||
|
return '1.2'
|
||
|
if is_2_1:
|
||
|
return '2.1'
|
||
|
|
||
|
return '2.0'
|
||
|
|
||
|
# This follows the rules about transforming keys as described in
|
||
|
# https://www.python.org/dev/peps/pep-0566/#id17
|
||
|
_ATTR2FIELD = {
|
||
|
name.lower().replace("-", "_"): name for name in _ALL_FIELDS
|
||
|
}
|
||
|
_FIELD2ATTR = {field: attr for attr, field in _ATTR2FIELD.items()}
|
||
|
|
||
|
_PREDICATE_FIELDS = ('Requires-Dist', 'Obsoletes-Dist', 'Provides-Dist')
|
||
|
_VERSIONS_FIELDS = ('Requires-Python',)
|
||
|
_VERSION_FIELDS = ('Version',)
|
||
|
_LISTFIELDS = ('Platform', 'Classifier', 'Obsoletes',
|
||
|
'Requires', 'Provides', 'Obsoletes-Dist',
|
||
|
'Provides-Dist', 'Requires-Dist', 'Requires-External',
|
||
|
'Project-URL', 'Supported-Platform', 'Setup-Requires-Dist',
|
||
|
'Provides-Extra', 'Extension')
|
||
|
_LISTTUPLEFIELDS = ('Project-URL',)
|
||
|
|
||
|
_ELEMENTSFIELD = ('Keywords',)
|
||
|
|
||
|
_UNICODEFIELDS = ('Author', 'Maintainer', 'Summary', 'Description')
|
||
|
|
||
|
_MISSING = object()
|
||
|
|
||
|
_FILESAFE = re.compile('[^A-Za-z0-9.]+')
|
||
|
|
||
|
|
||
|
def _get_name_and_version(name, version, for_filename=False):
|
||
|
"""Return the distribution name with version.
|
||
|
|
||
|
If for_filename is true, return a filename-escaped form."""
|
||
|
if for_filename:
|
||
|
# For both name and version any runs of non-alphanumeric or '.'
|
||
|
# characters are replaced with a single '-'. Additionally any
|
||
|
# spaces in the version string become '.'
|
||
|
name = _FILESAFE.sub('-', name)
|
||
|
version = _FILESAFE.sub('-', version.replace(' ', '.'))
|
||
|
return '%s-%s' % (name, version)
|
||
|
|
||
|
|
||
|
class LegacyMetadata(object):
|
||
|
"""The legacy metadata of a release.
|
||
|
|
||
|
Supports versions 1.0, 1.1, 1.2, 2.0 and 1.3/2.1 (auto-detected). You can
|
||
|
instantiate the class with one of these arguments (or none):
|
||
|
- *path*, the path to a metadata file
|
||
|
- *fileobj* give a file-like object with metadata as content
|
||
|
- *mapping* is a dict-like object
|
||
|
- *scheme* is a version scheme name
|
||
|
"""
|
||
|
# TODO document the mapping API and UNKNOWN default key
|
||
|
|
||
|
def __init__(self, path=None, fileobj=None, mapping=None,
|
||
|
scheme='default'):
|
||
|
if [path, fileobj, mapping].count(None) < 2:
|
||
|
raise TypeError('path, fileobj and mapping are exclusive')
|
||
|
self._fields = {}
|
||
|
self.requires_files = []
|
||
|
self._dependencies = None
|
||
|
self.scheme = scheme
|
||
|
if path is not None:
|
||
|
self.read(path)
|
||
|
elif fileobj is not None:
|
||
|
self.read_file(fileobj)
|
||
|
elif mapping is not None:
|
||
|
self.update(mapping)
|
||
|
self.set_metadata_version()
|
||
|
|
||
|
def set_metadata_version(self):
|
||
|
self._fields['Metadata-Version'] = _best_version(self._fields)
|
||
|
|
||
|
def _write_field(self, fileobj, name, value):
|
||
|
fileobj.write('%s: %s\n' % (name, value))
|
||
|
|
||
|
def __getitem__(self, name):
|
||
|
return self.get(name)
|
||
|
|
||
|
def __setitem__(self, name, value):
|
||
|
return self.set(name, value)
|
||
|
|
||
|
def __delitem__(self, name):
|
||
|
field_name = self._convert_name(name)
|
||
|
try:
|
||
|
del self._fields[field_name]
|
||
|
except KeyError:
|
||
|
raise KeyError(name)
|
||
|
|
||
|
def __contains__(self, name):
|
||
|
return (name in self._fields or
|
||
|
self._convert_name(name) in self._fields)
|
||
|
|
||
|
def _convert_name(self, name):
|
||
|
if name in _ALL_FIELDS:
|
||
|
return name
|
||
|
name = name.replace('-', '_').lower()
|
||
|
return _ATTR2FIELD.get(name, name)
|
||
|
|
||
|
def _default_value(self, name):
|
||
|
if name in _LISTFIELDS or name in _ELEMENTSFIELD:
|
||
|
return []
|
||
|
return 'UNKNOWN'
|
||
|
|
||
|
def _remove_line_prefix(self, value):
|
||
|
if self.metadata_version in ('1.0', '1.1'):
|
||
|
return _LINE_PREFIX_PRE_1_2.sub('\n', value)
|
||
|
else:
|
||
|
return _LINE_PREFIX_1_2.sub('\n', value)
|
||
|
|
||
|
def __getattr__(self, name):
|
||
|
if name in _ATTR2FIELD:
|
||
|
return self[name]
|
||
|
raise AttributeError(name)
|
||
|
|
||
|
#
|
||
|
# Public API
|
||
|
#
|
||
|
|
||
|
# dependencies = property(_get_dependencies, _set_dependencies)
|
||
|
|
||
|
def get_fullname(self, filesafe=False):
|
||
|
"""Return the distribution name with version.
|
||
|
|
||
|
If filesafe is true, return a filename-escaped form."""
|
||
|
return _get_name_and_version(self['Name'], self['Version'], filesafe)
|
||
|
|
||
|
def is_field(self, name):
|
||
|
"""return True if name is a valid metadata key"""
|
||
|
name = self._convert_name(name)
|
||
|
return name in _ALL_FIELDS
|
||
|
|
||
|
def is_multi_field(self, name):
|
||
|
name = self._convert_name(name)
|
||
|
return name in _LISTFIELDS
|
||
|
|
||
|
def read(self, filepath):
|
||
|
"""Read the metadata values from a file path."""
|
||
|
fp = codecs.open(filepath, 'r', encoding='utf-8')
|
||
|
try:
|
||
|
self.read_file(fp)
|
||
|
finally:
|
||
|
fp.close()
|
||
|
|
||
|
def read_file(self, fileob):
|
||
|
"""Read the metadata values from a file object."""
|
||
|
msg = message_from_file(fileob)
|
||
|
self._fields['Metadata-Version'] = msg['metadata-version']
|
||
|
|
||
|
# When reading, get all the fields we can
|
||
|
for field in _ALL_FIELDS:
|
||
|
if field not in msg:
|
||
|
continue
|
||
|
if field in _LISTFIELDS:
|
||
|
# we can have multiple lines
|
||
|
values = msg.get_all(field)
|
||
|
if field in _LISTTUPLEFIELDS and values is not None:
|
||
|
values = [tuple(value.split(',')) for value in values]
|
||
|
self.set(field, values)
|
||
|
else:
|
||
|
# single line
|
||
|
value = msg[field]
|
||
|
if value is not None and value != 'UNKNOWN':
|
||
|
self.set(field, value)
|
||
|
|
||
|
# PEP 566 specifies that the body be used for the description, if
|
||
|
# available
|
||
|
body = msg.get_payload()
|
||
|
self["Description"] = body if body else self["Description"]
|
||
|
# logger.debug('Attempting to set metadata for %s', self)
|
||
|
# self.set_metadata_version()
|
||
|
|
||
|
def write(self, filepath, skip_unknown=False):
|
||
|
"""Write the metadata fields to filepath."""
|
||
|
fp = codecs.open(filepath, 'w', encoding='utf-8')
|
||
|
try:
|
||
|
self.write_file(fp, skip_unknown)
|
||
|
finally:
|
||
|
fp.close()
|
||
|
|
||
|
def write_file(self, fileobject, skip_unknown=False):
|
||
|
"""Write the PKG-INFO format data to a file object."""
|
||
|
self.set_metadata_version()
|
||
|
|
||
|
for field in _version2fieldlist(self['Metadata-Version']):
|
||
|
values = self.get(field)
|
||
|
if skip_unknown and values in ('UNKNOWN', [], ['UNKNOWN']):
|
||
|
continue
|
||
|
if field in _ELEMENTSFIELD:
|
||
|
self._write_field(fileobject, field, ','.join(values))
|
||
|
continue
|
||
|
if field not in _LISTFIELDS:
|
||
|
if field == 'Description':
|
||
|
if self.metadata_version in ('1.0', '1.1'):
|
||
|
values = values.replace('\n', '\n ')
|
||
|
else:
|
||
|
values = values.replace('\n', '\n |')
|
||
|
values = [values]
|
||
|
|
||
|
if field in _LISTTUPLEFIELDS:
|
||
|
values = [','.join(value) for value in values]
|
||
|
|
||
|
for value in values:
|
||
|
self._write_field(fileobject, field, value)
|
||
|
|
||
|
def update(self, other=None, **kwargs):
|
||
|
"""Set metadata values from the given iterable `other` and kwargs.
|
||
|
|
||
|
Behavior is like `dict.update`: If `other` has a ``keys`` method,
|
||
|
they are looped over and ``self[key]`` is assigned ``other[key]``.
|
||
|
Else, ``other`` is an iterable of ``(key, value)`` iterables.
|
||
|
|
||
|
Keys that don't match a metadata field or that have an empty value are
|
||
|
dropped.
|
||
|
"""
|
||
|
def _set(key, value):
|
||
|
if key in _ATTR2FIELD and value:
|
||
|
self.set(self._convert_name(key), value)
|
||
|
|
||
|
if not other:
|
||
|
# other is None or empty container
|
||
|
pass
|
||
|
elif hasattr(other, 'keys'):
|
||
|
for k in other.keys():
|
||
|
_set(k, other[k])
|
||
|
else:
|
||
|
for k, v in other:
|
||
|
_set(k, v)
|
||
|
|
||
|
if kwargs:
|
||
|
for k, v in kwargs.items():
|
||
|
_set(k, v)
|
||
|
|
||
|
def set(self, name, value):
|
||
|
"""Control then set a metadata field."""
|
||
|
name = self._convert_name(name)
|
||
|
|
||
|
if ((name in _ELEMENTSFIELD or name == 'Platform') and
|
||
|
not isinstance(value, (list, tuple))):
|
||
|
if isinstance(value, string_types):
|
||
|
value = [v.strip() for v in value.split(',')]
|
||
|
else:
|
||
|
value = []
|
||
|
elif (name in _LISTFIELDS and
|
||
|
not isinstance(value, (list, tuple))):
|
||
|
if isinstance(value, string_types):
|
||
|
value = [value]
|
||
|
else:
|
||
|
value = []
|
||
|
|
||
|
if logger.isEnabledFor(logging.WARNING):
|
||
|
project_name = self['Name']
|
||
|
|
||
|
scheme = get_scheme(self.scheme)
|
||
|
if name in _PREDICATE_FIELDS and value is not None:
|
||
|
for v in value:
|
||
|
# check that the values are valid
|
||
|
if not scheme.is_valid_matcher(v.split(';')[0]):
|
||
|
logger.warning(
|
||
|
"'%s': '%s' is not valid (field '%s')",
|
||
|
project_name, v, name)
|
||
|
# FIXME this rejects UNKNOWN, is that right?
|
||
|
elif name in _VERSIONS_FIELDS and value is not None:
|
||
|
if not scheme.is_valid_constraint_list(value):
|
||
|
logger.warning("'%s': '%s' is not a valid version (field '%s')",
|
||
|
project_name, value, name)
|
||
|
elif name in _VERSION_FIELDS and value is not None:
|
||
|
if not scheme.is_valid_version(value):
|
||
|
logger.warning("'%s': '%s' is not a valid version (field '%s')",
|
||
|
project_name, value, name)
|
||
|
|
||
|
if name in _UNICODEFIELDS:
|
||
|
if name == 'Description':
|
||
|
value = self._remove_line_prefix(value)
|
||
|
|
||
|
self._fields[name] = value
|
||
|
|
||
|
def get(self, name, default=_MISSING):
|
||
|
"""Get a metadata field."""
|
||
|
name = self._convert_name(name)
|
||
|
if name not in self._fields:
|
||
|
if default is _MISSING:
|
||
|
default = self._default_value(name)
|
||
|
return default
|
||
|
if name in _UNICODEFIELDS:
|
||
|
value = self._fields[name]
|
||
|
return value
|
||
|
elif name in _LISTFIELDS:
|
||
|
value = self._fields[name]
|
||
|
if value is None:
|
||
|
return []
|
||
|
res = []
|
||
|
for val in value:
|
||
|
if name not in _LISTTUPLEFIELDS:
|
||
|
res.append(val)
|
||
|
else:
|
||
|
# That's for Project-URL
|
||
|
res.append((val[0], val[1]))
|
||
|
return res
|
||
|
|
||
|
elif name in _ELEMENTSFIELD:
|
||
|
value = self._fields[name]
|
||
|
if isinstance(value, string_types):
|
||
|
return value.split(',')
|
||
|
return self._fields[name]
|
||
|
|
||
|
def check(self, strict=False):
|
||
|
"""Check if the metadata is compliant. If strict is True then raise if
|
||
|
no Name or Version are provided"""
|
||
|
self.set_metadata_version()
|
||
|
|
||
|
# XXX should check the versions (if the file was loaded)
|
||
|
missing, warnings = [], []
|
||
|
|
||
|
for attr in ('Name', 'Version'): # required by PEP 345
|
||
|
if attr not in self:
|
||
|
missing.append(attr)
|
||
|
|
||
|
if strict and missing != []:
|
||
|
msg = 'missing required metadata: %s' % ', '.join(missing)
|
||
|
raise MetadataMissingError(msg)
|
||
|
|
||
|
for attr in ('Home-page', 'Author'):
|
||
|
if attr not in self:
|
||
|
missing.append(attr)
|
||
|
|
||
|
# checking metadata 1.2 (XXX needs to check 1.1, 1.0)
|
||
|
if self['Metadata-Version'] != '1.2':
|
||
|
return missing, warnings
|
||
|
|
||
|
scheme = get_scheme(self.scheme)
|
||
|
|
||
|
def are_valid_constraints(value):
|
||
|
for v in value:
|
||
|
if not scheme.is_valid_matcher(v.split(';')[0]):
|
||
|
return False
|
||
|
return True
|
||
|
|
||
|
for fields, controller in ((_PREDICATE_FIELDS, are_valid_constraints),
|
||
|
(_VERSIONS_FIELDS,
|
||
|
scheme.is_valid_constraint_list),
|
||
|
(_VERSION_FIELDS,
|
||
|
scheme.is_valid_version)):
|
||
|
for field in fields:
|
||
|
value = self.get(field, None)
|
||
|
if value is not None and not controller(value):
|
||
|
warnings.append("Wrong value for '%s': %s" % (field, value))
|
||
|
|
||
|
return missing, warnings
|
||
|
|
||
|
def todict(self, skip_missing=False):
|
||
|
"""Return fields as a dict.
|
||
|
|
||
|
Field names will be converted to use the underscore-lowercase style
|
||
|
instead of hyphen-mixed case (i.e. home_page instead of Home-page).
|
||
|
This is as per https://www.python.org/dev/peps/pep-0566/#id17.
|
||
|
"""
|
||
|
self.set_metadata_version()
|
||
|
|
||
|
fields = _version2fieldlist(self['Metadata-Version'])
|
||
|
|
||
|
data = {}
|
||
|
|
||
|
for field_name in fields:
|
||
|
if not skip_missing or field_name in self._fields:
|
||
|
key = _FIELD2ATTR[field_name]
|
||
|
if key != 'project_url':
|
||
|
data[key] = self[field_name]
|
||
|
else:
|
||
|
data[key] = [','.join(u) for u in self[field_name]]
|
||
|
|
||
|
return data
|
||
|
|
||
|
def add_requirements(self, requirements):
|
||
|
if self['Metadata-Version'] == '1.1':
|
||
|
# we can't have 1.1 metadata *and* Setuptools requires
|
||
|
for field in ('Obsoletes', 'Requires', 'Provides'):
|
||
|
if field in self:
|
||
|
del self[field]
|
||
|
self['Requires-Dist'] += requirements
|
||
|
|
||
|
# Mapping API
|
||
|
# TODO could add iter* variants
|
||
|
|
||
|
def keys(self):
|
||
|
return list(_version2fieldlist(self['Metadata-Version']))
|
||
|
|
||
|
def __iter__(self):
|
||
|
for key in self.keys():
|
||
|
yield key
|
||
|
|
||
|
def values(self):
|
||
|
return [self[key] for key in self.keys()]
|
||
|
|
||
|
def items(self):
|
||
|
return [(key, self[key]) for key in self.keys()]
|
||
|
|
||
|
def __repr__(self):
|
||
|
return '<%s %s %s>' % (self.__class__.__name__, self.name,
|
||
|
self.version)
|
||
|
|
||
|
|
||
|
METADATA_FILENAME = 'pydist.json'
|
||
|
WHEEL_METADATA_FILENAME = 'metadata.json'
|
||
|
LEGACY_METADATA_FILENAME = 'METADATA'
|
||
|
|
||
|
|
||
|
class Metadata(object):
|
||
|
"""
|
||
|
The metadata of a release. This implementation uses 2.0 (JSON)
|
||
|
metadata where possible. If not possible, it wraps a LegacyMetadata
|
||
|
instance which handles the key-value metadata format.
|
||
|
"""
|
||
|
|
||
|
METADATA_VERSION_MATCHER = re.compile(r'^\d+(\.\d+)*$')
|
||
|
|
||
|
NAME_MATCHER = re.compile('^[0-9A-Z]([0-9A-Z_.-]*[0-9A-Z])?$', re.I)
|
||
|
|
||
|
VERSION_MATCHER = PEP440_VERSION_RE
|
||
|
|
||
|
SUMMARY_MATCHER = re.compile('.{1,2047}')
|
||
|
|
||
|
METADATA_VERSION = '2.0'
|
||
|
|
||
|
GENERATOR = 'distlib (%s)' % __version__
|
||
|
|
||
|
MANDATORY_KEYS = {
|
||
|
'name': (),
|
||
|
'version': (),
|
||
|
'summary': ('legacy',),
|
||
|
}
|
||
|
|
||
|
INDEX_KEYS = ('name version license summary description author '
|
||
|
'author_email keywords platform home_page classifiers '
|
||
|
'download_url')
|
||
|
|
||
|
DEPENDENCY_KEYS = ('extras run_requires test_requires build_requires '
|
||
|
'dev_requires provides meta_requires obsoleted_by '
|
||
|
'supports_environments')
|
||
|
|
||
|
SYNTAX_VALIDATORS = {
|
||
|
'metadata_version': (METADATA_VERSION_MATCHER, ()),
|
||
|
'name': (NAME_MATCHER, ('legacy',)),
|
||
|
'version': (VERSION_MATCHER, ('legacy',)),
|
||
|
'summary': (SUMMARY_MATCHER, ('legacy',)),
|
||
|
}
|
||
|
|
||
|
__slots__ = ('_legacy', '_data', 'scheme')
|
||
|
|
||
|
def __init__(self, path=None, fileobj=None, mapping=None,
|
||
|
scheme='default'):
|
||
|
if [path, fileobj, mapping].count(None) < 2:
|
||
|
raise TypeError('path, fileobj and mapping are exclusive')
|
||
|
self._legacy = None
|
||
|
self._data = None
|
||
|
self.scheme = scheme
|
||
|
#import pdb; pdb.set_trace()
|
||
|
if mapping is not None:
|
||
|
try:
|
||
|
self._validate_mapping(mapping, scheme)
|
||
|
self._data = mapping
|
||
|
except MetadataUnrecognizedVersionError:
|
||
|
self._legacy = LegacyMetadata(mapping=mapping, scheme=scheme)
|
||
|
self.validate()
|
||
|
else:
|
||
|
data = None
|
||
|
if path:
|
||
|
with open(path, 'rb') as f:
|
||
|
data = f.read()
|
||
|
elif fileobj:
|
||
|
data = fileobj.read()
|
||
|
if data is None:
|
||
|
# Initialised with no args - to be added
|
||
|
self._data = {
|
||
|
'metadata_version': self.METADATA_VERSION,
|
||
|
'generator': self.GENERATOR,
|
||
|
}
|
||
|
else:
|
||
|
if not isinstance(data, text_type):
|
||
|
data = data.decode('utf-8')
|
||
|
try:
|
||
|
self._data = json.loads(data)
|
||
|
self._validate_mapping(self._data, scheme)
|
||
|
except ValueError:
|
||
|
# Note: MetadataUnrecognizedVersionError does not
|
||
|
# inherit from ValueError (it's a DistlibException,
|
||
|
# which should not inherit from ValueError).
|
||
|
# The ValueError comes from the json.load - if that
|
||
|
# succeeds and we get a validation error, we want
|
||
|
# that to propagate
|
||
|
self._legacy = LegacyMetadata(fileobj=StringIO(data),
|
||
|
scheme=scheme)
|
||
|
self.validate()
|
||
|
|
||
|
common_keys = set(('name', 'version', 'license', 'keywords', 'summary'))
|
||
|
|
||
|
none_list = (None, list)
|
||
|
none_dict = (None, dict)
|
||
|
|
||
|
mapped_keys = {
|
||
|
'run_requires': ('Requires-Dist', list),
|
||
|
'build_requires': ('Setup-Requires-Dist', list),
|
||
|
'dev_requires': none_list,
|
||
|
'test_requires': none_list,
|
||
|
'meta_requires': none_list,
|
||
|
'extras': ('Provides-Extra', list),
|
||
|
'modules': none_list,
|
||
|
'namespaces': none_list,
|
||
|
'exports': none_dict,
|
||
|
'commands': none_dict,
|
||
|
'classifiers': ('Classifier', list),
|
||
|
'source_url': ('Download-URL', None),
|
||
|
'metadata_version': ('Metadata-Version', None),
|
||
|
}
|
||
|
|
||
|
del none_list, none_dict
|
||
|
|
||
|
def __getattribute__(self, key):
|
||
|
common = object.__getattribute__(self, 'common_keys')
|
||
|
mapped = object.__getattribute__(self, 'mapped_keys')
|
||
|
if key in mapped:
|
||
|
lk, maker = mapped[key]
|
||
|
if self._legacy:
|
||
|
if lk is None:
|
||
|
result = None if maker is None else maker()
|
||
|
else:
|
||
|
result = self._legacy.get(lk)
|
||
|
else:
|
||
|
value = None if maker is None else maker()
|
||
|
if key not in ('commands', 'exports', 'modules', 'namespaces',
|
||
|
'classifiers'):
|
||
|
result = self._data.get(key, value)
|
||
|
else:
|
||
|
# special cases for PEP 459
|
||
|
sentinel = object()
|
||
|
result = sentinel
|
||
|
d = self._data.get('extensions')
|
||
|
if d:
|
||
|
if key == 'commands':
|
||
|
result = d.get('python.commands', value)
|
||
|
elif key == 'classifiers':
|
||
|
d = d.get('python.details')
|
||
|
if d:
|
||
|
result = d.get(key, value)
|
||
|
else:
|
||
|
d = d.get('python.exports')
|
||
|
if not d:
|
||
|
d = self._data.get('python.exports')
|
||
|
if d:
|
||
|
result = d.get(key, value)
|
||
|
if result is sentinel:
|
||
|
result = value
|
||
|
elif key not in common:
|
||
|
result = object.__getattribute__(self, key)
|
||
|
elif self._legacy:
|
||
|
result = self._legacy.get(key)
|
||
|
else:
|
||
|
result = self._data.get(key)
|
||
|
return result
|
||
|
|
||
|
def _validate_value(self, key, value, scheme=None):
|
||
|
if key in self.SYNTAX_VALIDATORS:
|
||
|
pattern, exclusions = self.SYNTAX_VALIDATORS[key]
|
||
|
if (scheme or self.scheme) not in exclusions:
|
||
|
m = pattern.match(value)
|
||
|
if not m:
|
||
|
raise MetadataInvalidError("'%s' is an invalid value for "
|
||
|
"the '%s' property" % (value,
|
||
|
key))
|
||
|
|
||
|
def __setattr__(self, key, value):
|
||
|
self._validate_value(key, value)
|
||
|
common = object.__getattribute__(self, 'common_keys')
|
||
|
mapped = object.__getattribute__(self, 'mapped_keys')
|
||
|
if key in mapped:
|
||
|
lk, _ = mapped[key]
|
||
|
if self._legacy:
|
||
|
if lk is None:
|
||
|
raise NotImplementedError
|
||
|
self._legacy[lk] = value
|
||
|
elif key not in ('commands', 'exports', 'modules', 'namespaces',
|
||
|
'classifiers'):
|
||
|
self._data[key] = value
|
||
|
else:
|
||
|
# special cases for PEP 459
|
||
|
d = self._data.setdefault('extensions', {})
|
||
|
if key == 'commands':
|
||
|
d['python.commands'] = value
|
||
|
elif key == 'classifiers':
|
||
|
d = d.setdefault('python.details', {})
|
||
|
d[key] = value
|
||
|
else:
|
||
|
d = d.setdefault('python.exports', {})
|
||
|
d[key] = value
|
||
|
elif key not in common:
|
||
|
object.__setattr__(self, key, value)
|
||
|
else:
|
||
|
if key == 'keywords':
|
||
|
if isinstance(value, string_types):
|
||
|
value = value.strip()
|
||
|
if value:
|
||
|
value = value.split()
|
||
|
else:
|
||
|
value = []
|
||
|
if self._legacy:
|
||
|
self._legacy[key] = value
|
||
|
else:
|
||
|
self._data[key] = value
|
||
|
|
||
|
@property
|
||
|
def name_and_version(self):
|
||
|
return _get_name_and_version(self.name, self.version, True)
|
||
|
|
||
|
@property
|
||
|
def provides(self):
|
||
|
if self._legacy:
|
||
|
result = self._legacy['Provides-Dist']
|
||
|
else:
|
||
|
result = self._data.setdefault('provides', [])
|
||
|
s = '%s (%s)' % (self.name, self.version)
|
||
|
if s not in result:
|
||
|
result.append(s)
|
||
|
return result
|
||
|
|
||
|
@provides.setter
|
||
|
def provides(self, value):
|
||
|
if self._legacy:
|
||
|
self._legacy['Provides-Dist'] = value
|
||
|
else:
|
||
|
self._data['provides'] = value
|
||
|
|
||
|
def get_requirements(self, reqts, extras=None, env=None):
|
||
|
"""
|
||
|
Base method to get dependencies, given a set of extras
|
||
|
to satisfy and an optional environment context.
|
||
|
:param reqts: A list of sometimes-wanted dependencies,
|
||
|
perhaps dependent on extras and environment.
|
||
|
:param extras: A list of optional components being requested.
|
||
|
:param env: An optional environment for marker evaluation.
|
||
|
"""
|
||
|
if self._legacy:
|
||
|
result = reqts
|
||
|
else:
|
||
|
result = []
|
||
|
extras = get_extras(extras or [], self.extras)
|
||
|
for d in reqts:
|
||
|
if 'extra' not in d and 'environment' not in d:
|
||
|
# unconditional
|
||
|
include = True
|
||
|
else:
|
||
|
if 'extra' not in d:
|
||
|
# Not extra-dependent - only environment-dependent
|
||
|
include = True
|
||
|
else:
|
||
|
include = d.get('extra') in extras
|
||
|
if include:
|
||
|
# Not excluded because of extras, check environment
|
||
|
marker = d.get('environment')
|
||
|
if marker:
|
||
|
include = interpret(marker, env)
|
||
|
if include:
|
||
|
result.extend(d['requires'])
|
||
|
for key in ('build', 'dev', 'test'):
|
||
|
e = ':%s:' % key
|
||
|
if e in extras:
|
||
|
extras.remove(e)
|
||
|
# A recursive call, but it should terminate since 'test'
|
||
|
# has been removed from the extras
|
||
|
reqts = self._data.get('%s_requires' % key, [])
|
||
|
result.extend(self.get_requirements(reqts, extras=extras,
|
||
|
env=env))
|
||
|
return result
|
||
|
|
||
|
@property
|
||
|
def dictionary(self):
|
||
|
if self._legacy:
|
||
|
return self._from_legacy()
|
||
|
return self._data
|
||
|
|
||
|
@property
|
||
|
def dependencies(self):
|
||
|
if self._legacy:
|
||
|
raise NotImplementedError
|
||
|
else:
|
||
|
return extract_by_key(self._data, self.DEPENDENCY_KEYS)
|
||
|
|
||
|
@dependencies.setter
|
||
|
def dependencies(self, value):
|
||
|
if self._legacy:
|
||
|
raise NotImplementedError
|
||
|
else:
|
||
|
self._data.update(value)
|
||
|
|
||
|
def _validate_mapping(self, mapping, scheme):
|
||
|
if mapping.get('metadata_version') != self.METADATA_VERSION:
|
||
|
raise MetadataUnrecognizedVersionError()
|
||
|
missing = []
|
||
|
for key, exclusions in self.MANDATORY_KEYS.items():
|
||
|
if key not in mapping:
|
||
|
if scheme not in exclusions:
|
||
|
missing.append(key)
|
||
|
if missing:
|
||
|
msg = 'Missing metadata items: %s' % ', '.join(missing)
|
||
|
raise MetadataMissingError(msg)
|
||
|
for k, v in mapping.items():
|
||
|
self._validate_value(k, v, scheme)
|
||
|
|
||
|
def validate(self):
|
||
|
if self._legacy:
|
||
|
missing, warnings = self._legacy.check(True)
|
||
|
if missing or warnings:
|
||
|
logger.warning('Metadata: missing: %s, warnings: %s',
|
||
|
missing, warnings)
|
||
|
else:
|
||
|
self._validate_mapping(self._data, self.scheme)
|
||
|
|
||
|
def todict(self):
|
||
|
if self._legacy:
|
||
|
return self._legacy.todict(True)
|
||
|
else:
|
||
|
result = extract_by_key(self._data, self.INDEX_KEYS)
|
||
|
return result
|
||
|
|
||
|
def _from_legacy(self):
|
||
|
assert self._legacy and not self._data
|
||
|
result = {
|
||
|
'metadata_version': self.METADATA_VERSION,
|
||
|
'generator': self.GENERATOR,
|
||
|
}
|
||
|
lmd = self._legacy.todict(True) # skip missing ones
|
||
|
for k in ('name', 'version', 'license', 'summary', 'description',
|
||
|
'classifier'):
|
||
|
if k in lmd:
|
||
|
if k == 'classifier':
|
||
|
nk = 'classifiers'
|
||
|
else:
|
||
|
nk = k
|
||
|
result[nk] = lmd[k]
|
||
|
kw = lmd.get('Keywords', [])
|
||
|
if kw == ['']:
|
||
|
kw = []
|
||
|
result['keywords'] = kw
|
||
|
keys = (('requires_dist', 'run_requires'),
|
||
|
('setup_requires_dist', 'build_requires'))
|
||
|
for ok, nk in keys:
|
||
|
if ok in lmd and lmd[ok]:
|
||
|
result[nk] = [{'requires': lmd[ok]}]
|
||
|
result['provides'] = self.provides
|
||
|
author = {}
|
||
|
maintainer = {}
|
||
|
return result
|
||
|
|
||
|
LEGACY_MAPPING = {
|
||
|
'name': 'Name',
|
||
|
'version': 'Version',
|
||
|
('extensions', 'python.details', 'license'): 'License',
|
||
|
'summary': 'Summary',
|
||
|
'description': 'Description',
|
||
|
('extensions', 'python.project', 'project_urls', 'Home'): 'Home-page',
|
||
|
('extensions', 'python.project', 'contacts', 0, 'name'): 'Author',
|
||
|
('extensions', 'python.project', 'contacts', 0, 'email'): 'Author-email',
|
||
|
'source_url': 'Download-URL',
|
||
|
('extensions', 'python.details', 'classifiers'): 'Classifier',
|
||
|
}
|
||
|
|
||
|
def _to_legacy(self):
|
||
|
def process_entries(entries):
|
||
|
reqts = set()
|
||
|
for e in entries:
|
||
|
extra = e.get('extra')
|
||
|
env = e.get('environment')
|
||
|
rlist = e['requires']
|
||
|
for r in rlist:
|
||
|
if not env and not extra:
|
||
|
reqts.add(r)
|
||
|
else:
|
||
|
marker = ''
|
||
|
if extra:
|
||
|
marker = 'extra == "%s"' % extra
|
||
|
if env:
|
||
|
if marker:
|
||
|
marker = '(%s) and %s' % (env, marker)
|
||
|
else:
|
||
|
marker = env
|
||
|
reqts.add(';'.join((r, marker)))
|
||
|
return reqts
|
||
|
|
||
|
assert self._data and not self._legacy
|
||
|
result = LegacyMetadata()
|
||
|
nmd = self._data
|
||
|
# import pdb; pdb.set_trace()
|
||
|
for nk, ok in self.LEGACY_MAPPING.items():
|
||
|
if not isinstance(nk, tuple):
|
||
|
if nk in nmd:
|
||
|
result[ok] = nmd[nk]
|
||
|
else:
|
||
|
d = nmd
|
||
|
found = True
|
||
|
for k in nk:
|
||
|
try:
|
||
|
d = d[k]
|
||
|
except (KeyError, IndexError):
|
||
|
found = False
|
||
|
break
|
||
|
if found:
|
||
|
result[ok] = d
|
||
|
r1 = process_entries(self.run_requires + self.meta_requires)
|
||
|
r2 = process_entries(self.build_requires + self.dev_requires)
|
||
|
if self.extras:
|
||
|
result['Provides-Extra'] = sorted(self.extras)
|
||
|
result['Requires-Dist'] = sorted(r1)
|
||
|
result['Setup-Requires-Dist'] = sorted(r2)
|
||
|
# TODO: any other fields wanted
|
||
|
return result
|
||
|
|
||
|
def write(self, path=None, fileobj=None, legacy=False, skip_unknown=True):
|
||
|
if [path, fileobj].count(None) != 1:
|
||
|
raise ValueError('Exactly one of path and fileobj is needed')
|
||
|
self.validate()
|
||
|
if legacy:
|
||
|
if self._legacy:
|
||
|
legacy_md = self._legacy
|
||
|
else:
|
||
|
legacy_md = self._to_legacy()
|
||
|
if path:
|
||
|
legacy_md.write(path, skip_unknown=skip_unknown)
|
||
|
else:
|
||
|
legacy_md.write_file(fileobj, skip_unknown=skip_unknown)
|
||
|
else:
|
||
|
if self._legacy:
|
||
|
d = self._from_legacy()
|
||
|
else:
|
||
|
d = self._data
|
||
|
if fileobj:
|
||
|
json.dump(d, fileobj, ensure_ascii=True, indent=2,
|
||
|
sort_keys=True)
|
||
|
else:
|
||
|
with codecs.open(path, 'w', 'utf-8') as f:
|
||
|
json.dump(d, f, ensure_ascii=True, indent=2,
|
||
|
sort_keys=True)
|
||
|
|
||
|
def add_requirements(self, requirements):
|
||
|
if self._legacy:
|
||
|
self._legacy.add_requirements(requirements)
|
||
|
else:
|
||
|
run_requires = self._data.setdefault('run_requires', [])
|
||
|
always = None
|
||
|
for entry in run_requires:
|
||
|
if 'environment' not in entry and 'extra' not in entry:
|
||
|
always = entry
|
||
|
break
|
||
|
if always is None:
|
||
|
always = { 'requires': requirements }
|
||
|
run_requires.insert(0, always)
|
||
|
else:
|
||
|
rset = set(always['requires']) | set(requirements)
|
||
|
always['requires'] = sorted(rset)
|
||
|
|
||
|
def __repr__(self):
|
||
|
name = self.name or '(no name)'
|
||
|
version = self.version or 'no version'
|
||
|
return '<%s %s %s (%s)>' % (self.__class__.__name__,
|
||
|
self.metadata_version, name, version)
|