mirror of
https://github.com/gryf/ebook-converter.git
synced 2026-04-19 04:33:34 +02:00
Fixed flake8 issues to several modules
This commit is contained in:
@@ -18,7 +18,7 @@ try:
|
||||
_author_pat = re.compile(tweaks['authors_split_regex'])
|
||||
except Exception:
|
||||
prints('Author split regexp:', tweaks['authors_split_regex'],
|
||||
'is invalid, using default')
|
||||
'is invalid, using default')
|
||||
_author_pat = re.compile(r'(?i),?\s+(and|with)\s+')
|
||||
|
||||
|
||||
@@ -76,7 +76,8 @@ def author_to_author_sort(author, method=None):
|
||||
if method == 'copy':
|
||||
return author
|
||||
|
||||
prefixes = {force_unicode(y).lower() for y in tweaks['author_name_prefixes']}
|
||||
prefixes = {force_unicode(y).lower()
|
||||
for y in tweaks['author_name_prefixes']}
|
||||
prefixes |= {y+'.' for y in prefixes}
|
||||
while True:
|
||||
if not tokens:
|
||||
@@ -87,7 +88,8 @@ def author_to_author_sort(author, method=None):
|
||||
else:
|
||||
break
|
||||
|
||||
suffixes = {force_unicode(y).lower() for y in tweaks['author_name_suffixes']}
|
||||
suffixes = {force_unicode(y).lower()
|
||||
for y in tweaks['author_name_suffixes']}
|
||||
suffixes |= {y+'.' for y in suffixes}
|
||||
|
||||
suffix = ''
|
||||
@@ -144,7 +146,7 @@ def get_title_sort_pat(lang=None):
|
||||
except:
|
||||
ans = frozenset((r'A\s+', r'The\s+', r'An\s+'))
|
||||
ans = '|'.join(ans)
|
||||
ans = '^(%s)'%ans
|
||||
ans = '^(%s)' % ans
|
||||
try:
|
||||
ans = re.compile(ans, re.IGNORECASE)
|
||||
except:
|
||||
@@ -154,7 +156,7 @@ def get_title_sort_pat(lang=None):
|
||||
|
||||
|
||||
_ignore_starts = '\'"'+''.join(chr(x) for x in
|
||||
list(range(0x2018, 0x201e))+[0x2032, 0x2033])
|
||||
list(range(0x2018, 0x201e))+[0x2032, 0x2033])
|
||||
|
||||
|
||||
def title_sort(title, order=None, lang=None):
|
||||
|
||||
@@ -12,8 +12,7 @@ from lxml import etree
|
||||
from ebook_converter.utils.date import parse_only_date
|
||||
from ebook_converter.utils.img import save_cover_data_to
|
||||
from ebook_converter.utils.imghdr import identify
|
||||
from ebook_converter import guess_type, guess_all_extensions, prints, \
|
||||
force_unicode
|
||||
from ebook_converter import guess_all_extensions, prints, force_unicode
|
||||
from ebook_converter.ebooks.metadata import MetaInformation, check_isbn
|
||||
from ebook_converter.ebooks.chardet import xml_to_unicode
|
||||
from ebook_converter.polyglot.binary import as_base64_unicode
|
||||
|
||||
@@ -10,11 +10,11 @@ import mimetypes
|
||||
import os
|
||||
import re
|
||||
import sys
|
||||
import textwrap
|
||||
import traceback
|
||||
import unittest
|
||||
import urllib.parse
|
||||
import uuid
|
||||
import traceback
|
||||
import textwrap
|
||||
|
||||
from lxml import etree
|
||||
from lxml.builder import ElementMaker
|
||||
@@ -32,7 +32,7 @@ from ebook_converter.ebooks.metadata import string_to_authors, \
|
||||
from ebook_converter.ebooks.metadata.book.base import Metadata
|
||||
from ebook_converter.utils.date import parse_date, isoformat
|
||||
from ebook_converter.utils.localization import get_lang, canonicalize_lang
|
||||
from ebook_converter import prints, guess_type
|
||||
from ebook_converter import prints
|
||||
from ebook_converter.utils.cleantext import clean_ascii_chars, clean_xml_chars
|
||||
from ebook_converter.utils.config import tweaks
|
||||
from ebook_converter.polyglot.urllib import unquote
|
||||
@@ -1807,8 +1807,7 @@ def test_m2o():
|
||||
class OPFTest(unittest.TestCase):
|
||||
|
||||
def setUp(self):
|
||||
self.stream = io.BytesIO(
|
||||
b'''\
|
||||
self.stream = io.BytesIO(b'''\
|
||||
<?xml version="1.0" encoding="UTF-8"?>
|
||||
<package version="2.0" xmlns="http://www.idpf.org/2007/opf" >
|
||||
<metadata xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:opf="http://www.idpf.org/2007/opf">
|
||||
@@ -1827,8 +1826,7 @@ b'''\
|
||||
<item id="1" href="a%20%7E%20b" media-type="text/txt" />
|
||||
</manifest>
|
||||
</package>
|
||||
'''
|
||||
)
|
||||
''')
|
||||
self.opf = OPF(self.stream, os.getcwd())
|
||||
|
||||
def testReading(self, opf=None):
|
||||
|
||||
Reference in New Issue
Block a user