Spaces:
Runtime error
Runtime error
""" | |
babel.util | |
~~~~~~~~~~ | |
Various utility classes and functions. | |
:copyright: (c) 2013-2023 by the Babel Team. | |
:license: BSD, see LICENSE for more details. | |
""" | |
from __future__ import annotations | |
import codecs | |
import collections | |
import datetime | |
import os | |
import re | |
import textwrap | |
from collections.abc import Generator, Iterable | |
from typing import IO, Any, TypeVar | |
from babel import dates, localtime | |
missing = object() | |
_T = TypeVar("_T") | |
def distinct(iterable: Iterable[_T]) -> Generator[_T, None, None]: | |
"""Yield all items in an iterable collection that are distinct. | |
Unlike when using sets for a similar effect, the original ordering of the | |
items in the collection is preserved by this function. | |
>>> print(list(distinct([1, 2, 1, 3, 4, 4]))) | |
[1, 2, 3, 4] | |
>>> print(list(distinct('foobar'))) | |
['f', 'o', 'b', 'a', 'r'] | |
:param iterable: the iterable collection providing the data | |
""" | |
seen = set() | |
for item in iter(iterable): | |
if item not in seen: | |
yield item | |
seen.add(item) | |
# Regexp to match python magic encoding line | |
PYTHON_MAGIC_COMMENT_re = re.compile( | |
br'[ \t\f]* \# .* coding[=:][ \t]*([-\w.]+)', re.VERBOSE) | |
def parse_encoding(fp: IO[bytes]) -> str | None: | |
"""Deduce the encoding of a source file from magic comment. | |
It does this in the same way as the `Python interpreter`__ | |
.. __: https://docs.python.org/3.4/reference/lexical_analysis.html#encoding-declarations | |
The ``fp`` argument should be a seekable file object. | |
(From Jeff Dairiki) | |
""" | |
pos = fp.tell() | |
fp.seek(0) | |
try: | |
line1 = fp.readline() | |
has_bom = line1.startswith(codecs.BOM_UTF8) | |
if has_bom: | |
line1 = line1[len(codecs.BOM_UTF8):] | |
m = PYTHON_MAGIC_COMMENT_re.match(line1) | |
if not m: | |
try: | |
import ast | |
ast.parse(line1.decode('latin-1')) | |
except (ImportError, SyntaxError, UnicodeEncodeError): | |
# Either it's a real syntax error, in which case the source is | |
# not valid python source, or line2 is a continuation of line1, | |
# in which case we don't want to scan line2 for a magic | |
# comment. | |
pass | |
else: | |
line2 = fp.readline() | |
m = PYTHON_MAGIC_COMMENT_re.match(line2) | |
if has_bom: | |
if m: | |
magic_comment_encoding = m.group(1).decode('latin-1') | |
if magic_comment_encoding != 'utf-8': | |
raise SyntaxError(f"encoding problem: {magic_comment_encoding} with BOM") | |
return 'utf-8' | |
elif m: | |
return m.group(1).decode('latin-1') | |
else: | |
return None | |
finally: | |
fp.seek(pos) | |
PYTHON_FUTURE_IMPORT_re = re.compile( | |
r'from\s+__future__\s+import\s+\(*(.+)\)*') | |
def parse_future_flags(fp: IO[bytes], encoding: str = 'latin-1') -> int: | |
"""Parse the compiler flags by :mod:`__future__` from the given Python | |
code. | |
""" | |
import __future__ | |
pos = fp.tell() | |
fp.seek(0) | |
flags = 0 | |
try: | |
body = fp.read().decode(encoding) | |
# Fix up the source to be (hopefully) parsable by regexpen. | |
# This will likely do untoward things if the source code itself is broken. | |
# (1) Fix `import (\n...` to be `import (...`. | |
body = re.sub(r'import\s*\([\r\n]+', 'import (', body) | |
# (2) Join line-ending commas with the next line. | |
body = re.sub(r',\s*[\r\n]+', ', ', body) | |
# (3) Remove backslash line continuations. | |
body = re.sub(r'\\\s*[\r\n]+', ' ', body) | |
for m in PYTHON_FUTURE_IMPORT_re.finditer(body): | |
names = [x.strip().strip('()') for x in m.group(1).split(',')] | |
for name in names: | |
feature = getattr(__future__, name, None) | |
if feature: | |
flags |= feature.compiler_flag | |
finally: | |
fp.seek(pos) | |
return flags | |
def pathmatch(pattern: str, filename: str) -> bool: | |
"""Extended pathname pattern matching. | |
This function is similar to what is provided by the ``fnmatch`` module in | |
the Python standard library, but: | |
* can match complete (relative or absolute) path names, and not just file | |
names, and | |
* also supports a convenience pattern ("**") to match files at any | |
directory level. | |
Examples: | |
>>> pathmatch('**.py', 'bar.py') | |
True | |
>>> pathmatch('**.py', 'foo/bar/baz.py') | |
True | |
>>> pathmatch('**.py', 'templates/index.html') | |
False | |
>>> pathmatch('./foo/**.py', 'foo/bar/baz.py') | |
True | |
>>> pathmatch('./foo/**.py', 'bar/baz.py') | |
False | |
>>> pathmatch('^foo/**.py', 'foo/bar/baz.py') | |
True | |
>>> pathmatch('^foo/**.py', 'bar/baz.py') | |
False | |
>>> pathmatch('**/templates/*.html', 'templates/index.html') | |
True | |
>>> pathmatch('**/templates/*.html', 'templates/foo/bar.html') | |
False | |
:param pattern: the glob pattern | |
:param filename: the path name of the file to match against | |
""" | |
symbols = { | |
'?': '[^/]', | |
'?/': '[^/]/', | |
'*': '[^/]+', | |
'*/': '[^/]+/', | |
'**/': '(?:.+/)*?', | |
'**': '(?:.+/)*?[^/]+', | |
} | |
if pattern.startswith('^'): | |
buf = ['^'] | |
pattern = pattern[1:] | |
elif pattern.startswith('./'): | |
buf = ['^'] | |
pattern = pattern[2:] | |
else: | |
buf = [] | |
for idx, part in enumerate(re.split('([?*]+/?)', pattern)): | |
if idx % 2: | |
buf.append(symbols[part]) | |
elif part: | |
buf.append(re.escape(part)) | |
match = re.match(f"{''.join(buf)}$", filename.replace(os.sep, "/")) | |
return match is not None | |
class TextWrapper(textwrap.TextWrapper): | |
wordsep_re = re.compile( | |
r'(\s+|' # any whitespace | |
r'(?<=[\w\!\"\'\&\.\,\?])-{2,}(?=\w))' # em-dash | |
) | |
def wraptext(text: str, width: int = 70, initial_indent: str = '', subsequent_indent: str = '') -> list[str]: | |
"""Simple wrapper around the ``textwrap.wrap`` function in the standard | |
library. This version does not wrap lines on hyphens in words. | |
:param text: the text to wrap | |
:param width: the maximum line width | |
:param initial_indent: string that will be prepended to the first line of | |
wrapped output | |
:param subsequent_indent: string that will be prepended to all lines save | |
the first of wrapped output | |
""" | |
wrapper = TextWrapper(width=width, initial_indent=initial_indent, | |
subsequent_indent=subsequent_indent, | |
break_long_words=False) | |
return wrapper.wrap(text) | |
# TODO (Babel 3.x): Remove this re-export | |
odict = collections.OrderedDict | |
class FixedOffsetTimezone(datetime.tzinfo): | |
"""Fixed offset in minutes east from UTC.""" | |
def __init__(self, offset: float, name: str | None = None) -> None: | |
self._offset = datetime.timedelta(minutes=offset) | |
if name is None: | |
name = 'Etc/GMT%+d' % offset | |
self.zone = name | |
def __str__(self) -> str: | |
return self.zone | |
def __repr__(self) -> str: | |
return f'<FixedOffset "{self.zone}" {self._offset}>' | |
def utcoffset(self, dt: datetime.datetime) -> datetime.timedelta: | |
return self._offset | |
def tzname(self, dt: datetime.datetime) -> str: | |
return self.zone | |
def dst(self, dt: datetime.datetime) -> datetime.timedelta: | |
return ZERO | |
# Export the localtime functionality here because that's | |
# where it was in the past. | |
# TODO(3.0): remove these aliases | |
UTC = dates.UTC | |
LOCALTZ = dates.LOCALTZ | |
get_localzone = localtime.get_localzone | |
STDOFFSET = localtime.STDOFFSET | |
DSTOFFSET = localtime.DSTOFFSET | |
DSTDIFF = localtime.DSTDIFF | |
ZERO = localtime.ZERO | |
def _cmp(a: Any, b: Any): | |
return (a > b) - (a < b) | |