From mboxrd@z Thu Jan 1 00:00:00 1970 From: Adolf Belka To: development@lists.ipfire.org Subject: [PATCH] client175: Convert python2 modules in source tarball to python3 Date: Thu, 05 Aug 2021 15:14:30 +0200 Message-ID: <20210805131430.2771006-1-adolf.belka@ipfire.org> MIME-Version: 1.0 Content-Type: multipart/mixed; boundary="===============0758129431786205935==" List-Id: --===============0758129431786205935== Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: quoted-printable - Patch created to convert all python modules to python3 compatibility that n= eed it. 2to3 converter used for this. - Start initscript changed to use python3 - No change required in rootfile - Execution of patch added to lfs file - Tested in vm machine. WUI page showed the same as with the python version. scan of directory for mp3 (.flac) files was successful. Could not test act= ual audio playing capability as my vm testbed does not have any audio setup at this = time. I believe that the purpose of client175 is to provide the WUI page and for = that my testing seemed to show everything working as expected. Tested-by: Adolf Belka Signed-off-by: Adolf Belka --- lfs/client175 | 1 + src/initscripts/packages/client175 | 2 +- ...vert-to-python3-using-2to3-converter.patch | 4923 +++++++++++++++++ 3 files changed, 4925 insertions(+), 1 deletion(-) create mode 100644 src/patches/client175_0.7-convert-to-python3-using-2to3-c= onverter.patch diff --git a/lfs/client175 b/lfs/client175 index e69e8e705..f2d23dda9 100644 --- a/lfs/client175 +++ b/lfs/client175 @@ -78,6 +78,7 @@ $(subst %,%_MD5,$(objects)) : $(TARGET) : $(patsubst %,$(DIR_DL)/%,$(objects)) @$(PREBUILD) @rm -rf $(DIR_APP) && cd $(DIR_SRC) && tar Jxf $(DIR_DL)/$(DL_FILE) + cd $(DIR_APP) && patch -Np1 < $(DIR_SRC)/src/patches/client175_0.7-convert-= to-python3-using-2to3-converter.patch=09 @rm -rf /srv/client175 mkdir -pv /srv/client175 =20 diff --git a/src/initscripts/packages/client175 b/src/initscripts/packages/cl= ient175 index 9f279b7fb..1dde39800 100644 --- a/src/initscripts/packages/client175 +++ b/src/initscripts/packages/client175 @@ -26,7 +26,7 @@ case "$1" in fi fi boot_mesg "Starting Client175 MPD WebIF..." - sudo -u nobody python /srv/client175/server.py > /var/log/client175 2>&1 & + sudo -u nobody python3 /srv/client175/server.py > /var/log/client175 2>&1 & echo $! > /var/run/client175.pid evaluate_retval ;; diff --git a/src/patches/client175_0.7-convert-to-python3-using-2to3-converte= r.patch b/src/patches/client175_0.7-convert-to-python3-using-2to3-converter.p= atch new file mode 100644 index 000000000..61cd8d9ca --- /dev/null +++ b/src/patches/client175_0.7-convert-to-python3-using-2to3-converter.patch @@ -0,0 +1,4923 @@ +diff -Naur client175_0.7-original/BeautifulSoup.py client175_0.7/BeautifulSo= up.py +--- client175_0.7-original/BeautifulSoup.py 2010-05-14 12:57:39.000000000 +0= 200 ++++ client175_0.7/BeautifulSoup.py 2021-08-03 14:39:30.213509172 +0200 +@@ -76,7 +76,7 @@ + SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE, DAMMIT. +=20 + """ +-from __future__ import generators ++ +=20 + __author__ =3D "Leonard Richardson (leonardr(a)segfault.org)" + __version__ =3D "3.0.8.1" +@@ -85,12 +85,12 @@ +=20 + from sgmllib import SGMLParser, SGMLParseError + import codecs +-import markupbase ++import _markupbase + import types + import re + import sgmllib + try: +- from htmlentitydefs import name2codepoint ++ from html.entities import name2codepoint + except ImportError: + name2codepoint =3D {} + try: +@@ -100,7 +100,7 @@ +=20 + #These hacks make Beautiful Soup able to parse XML with namespaces + sgmllib.tagfind =3D re.compile('[a-zA-Z][-_.:a-zA-Z0-9]*') +-markupbase._declname_match =3D re.compile(r'[a-zA-Z][-_.:a-zA-Z0-9]*\s*').m= atch ++_markupbase._declname_match =3D re.compile(r'[a-zA-Z][-_.:a-zA-Z0-9]*\s*').= match +=20 + DEFAULT_OUTPUT_ENCODING =3D "utf-8" +=20 +@@ -162,7 +162,7 @@ + #this element (and any children) hadn't been parsed. Connect + #the two. + lastChild =3D self._lastRecursiveChild() +- nextElement =3D lastChild.next ++ nextElement =3D lastChild.__next__ +=20 + if self.previous: + self.previous.next =3D nextElement +@@ -187,7 +187,7 @@ + return lastChild +=20 + def insert(self, position, newChild): +- if isinstance(newChild, basestring) \ ++ if isinstance(newChild, str) \ + and not isinstance(newChild, NavigableString): + newChild =3D NavigableString(newChild) +=20 +@@ -241,7 +241,7 @@ + newChild.nextSibling.previousSibling =3D newChild + newChildsLastElement.next =3D nextChild +=20 +- if newChildsLastElement.next: ++ if newChildsLastElement.__next__: + newChildsLastElement.next.previous =3D newChildsLastElement + self.contents.insert(position, newChild) +=20 +@@ -342,7 +342,7 @@ + return [element for element in generator() + if isinstance(element, Tag)] + # findAll*('tag-name') +- elif isinstance(name, basestring): ++ elif isinstance(name, str): + return [element for element in generator() + if isinstance(element, Tag) and + element.name =3D=3D name] +@@ -355,7 +355,7 @@ + g =3D generator() + while True: + try: +- i =3D g.next() ++ i =3D next(g) + except StopIteration: + break + if i: +@@ -371,7 +371,7 @@ + def nextGenerator(self): + i =3D self + while i is not None: +- i =3D i.next ++ i =3D i.__next__ + yield i +=20 + def nextSiblingGenerator(self): +@@ -406,22 +406,22 @@ + def toEncoding(self, s, encoding=3DNone): + """Encodes an object to a string in some encoding, or to Unicode. + .""" +- if isinstance(s, unicode): ++ if isinstance(s, str): + if encoding: + s =3D s.encode(encoding) + elif isinstance(s, str): + if encoding: + s =3D s.encode(encoding) + else: +- s =3D unicode(s) ++ s =3D str(s) + else: + if encoding: + s =3D self.toEncoding(str(s), encoding) + else: +- s =3D unicode(s) ++ s =3D str(s) + return s +=20 +-class NavigableString(unicode, PageElement): ++class NavigableString(str, PageElement): +=20 + def __new__(cls, value): + """Create a new NavigableString. +@@ -431,9 +431,9 @@ + passed in to the superclass's __new__ or the superclass won't know + how to handle non-ASCII characters. + """ +- if isinstance(value, unicode): +- return unicode.__new__(cls, value) +- return unicode.__new__(cls, value, DEFAULT_OUTPUT_ENCODING) ++ if isinstance(value, str): ++ return str.__new__(cls, value) ++ return str.__new__(cls, value, DEFAULT_OUTPUT_ENCODING) +=20 + def __getnewargs__(self): + return (NavigableString.__str__(self),) +@@ -445,7 +445,7 @@ + if attr =3D=3D 'string': + return self + else: +- raise AttributeError, "'%s' object has no attribute '%s'" % (se= lf.__class__.__name__, attr) ++ raise AttributeError("'%s' object has no attribute '%s'" % (sel= f.__class__.__name__, attr)) +=20 + def __unicode__(self): + return str(self).decode(DEFAULT_OUTPUT_ENCODING) +@@ -483,7 +483,7 @@ + def _invert(h): + "Cheap function to invert a hash." + i =3D {} +- for k,v in h.items(): ++ for k,v in list(h.items()): + i[v] =3D k + return i +=20 +@@ -502,23 +502,23 @@ + escaped.""" + x =3D match.group(1) + if self.convertHTMLEntities and x in name2codepoint: +- return unichr(name2codepoint[x]) ++ return chr(name2codepoint[x]) + elif x in self.XML_ENTITIES_TO_SPECIAL_CHARS: + if self.convertXMLEntities: + return self.XML_ENTITIES_TO_SPECIAL_CHARS[x] + else: +- return u'&%s;' % x ++ return '&%s;' % x + elif len(x) > 0 and x[0] =3D=3D '#': + # Handle numeric entities + if len(x) > 1 and x[1] =3D=3D 'x': +- return unichr(int(x[2:], 16)) ++ return chr(int(x[2:], 16)) + else: +- return unichr(int(x[1:])) ++ return chr(int(x[1:])) +=20 + elif self.escapeUnrecognizedEntities: +- return u'&%s;' % x ++ return '&%s;' % x + else: +- return u'&%s;' % x ++ return '&%s;' % x +=20 + def __init__(self, parser, name, attrs=3DNone, parent=3DNone, + previous=3DNone): +@@ -541,11 +541,11 @@ + self.escapeUnrecognizedEntities =3D parser.escapeUnrecognizedEntiti= es +=20 + # Convert any HTML, XML, or numeric entities in the attribute value= s. +- convert =3D lambda(k, val): (k, ++ convert =3D lambda k_val: (k_val[0], + re.sub("&(#\d+|#x[0-9a-fA-F]+|\w+);", + self._convertEntities, +- val)) +- self.attrs =3D map(convert, self.attrs) ++ k_val[1])) ++ self.attrs =3D list(map(convert, self.attrs)) +=20 + def getString(self): + if (len(self.contents) =3D=3D 1 +@@ -559,16 +559,16 @@ +=20 + string =3D property(getString, setString) +=20 +- def getText(self, separator=3Du""): ++ def getText(self, separator=3D""): + if not len(self.contents): +- return u"" +- stopNode =3D self._lastRecursiveChild().next ++ return "" ++ stopNode =3D self._lastRecursiveChild().__next__ + strings =3D [] + current =3D self.contents[0] + while current is not stopNode: + if isinstance(current, NavigableString): + strings.append(current.strip()) +- current =3D current.next ++ current =3D current.__next__ + return separator.join(strings) +=20 + text =3D property(getText) +@@ -591,7 +591,7 @@ + raise ValueError("Tag.index: element not in tag") +=20 + def has_key(self, key): +- return self._getAttrMap().has_key(key) ++ return key in self._getAttrMap() +=20 + def __getitem__(self, key): + """tag[key] returns the value of the 'key' attribute for the tag, +@@ -609,7 +609,7 @@ + def __contains__(self, x): + return x in self.contents +=20 +- def __nonzero__(self): ++ def __bool__(self): + "A tag is non-None even if it has no contents." + return True +=20 +@@ -635,14 +635,14 @@ + #We don't break because bad HTML can define the same + #attribute multiple times. + self._getAttrMap() +- if self.attrMap.has_key(key): ++ if key in self.attrMap: + del self.attrMap[key] +=20 + def __call__(self, *args, **kwargs): + """Calling a tag like a function is the same as calling its + findAll() method. Eg. tag('a') returns a list of all the A tags + found within this tag.""" +- return apply(self.findAll, args, kwargs) ++ return self.findAll(*args, **kwargs) +=20 + def __getattr__(self, tag): + #print "Getattr %s.%s" % (self.__class__, tag) +@@ -650,7 +650,7 @@ + return self.find(tag[:-3]) + elif tag.find('__') !=3D 0: + return self.find(tag) +- raise AttributeError, "'%s' object has no attribute '%s'" % (self._= _class__, tag) ++ raise AttributeError("'%s' object has no attribute '%s'" % (self.__= class__, tag)) +=20 + def __eq__(self, other): + """Returns true iff this tag has the same name, the same attributes, +@@ -703,7 +703,7 @@ + if self.attrs: + for key, val in self.attrs: + fmt =3D '%s=3D"%s"' +- if isinstance(val, basestring): ++ if isinstance(val, str): + if self.containsSubstitutions and '%SOUP-ENCODING%' in = val: + val =3D self.substituteEncoding(val, encoding) +=20 +@@ -780,7 +780,7 @@ + return + current =3D self.contents[0] + while current is not None: +- next =3D current.next ++ next =3D current.__next__ + if isinstance(current, Tag): + del current.contents[:] + current.parent =3D None +@@ -873,11 +873,11 @@ + def recursiveChildGenerator(self): + if not len(self.contents): + raise StopIteration +- stopNode =3D self._lastRecursiveChild().next ++ stopNode =3D self._lastRecursiveChild().__next__ + current =3D self.contents[0] + while current is not stopNode: + yield current +- current =3D current.next ++ current =3D current.__next__ +=20 +=20 + # Next, a couple classes to represent queries and their results. +@@ -887,7 +887,7 @@ +=20 + def __init__(self, name=3DNone, attrs=3D{}, text=3DNone, **kwargs): + self.name =3D name +- if isinstance(attrs, basestring): ++ if isinstance(attrs, str): + kwargs['class'] =3D _match_css_class(attrs) + attrs =3D None + if kwargs: +@@ -923,7 +923,7 @@ + else: + match =3D True + markupAttrMap =3D None +- for attr, matchAgainst in self.attrs.items(): ++ for attr, matchAgainst in list(self.attrs.items()): + if not markupAttrMap: + if hasattr(markupAttrs, 'get'): + markupAttrMap =3D markupAttrs +@@ -961,12 +961,12 @@ + found =3D self.searchTag(markup) + # If it's text, make sure the text matches. + elif isinstance(markup, NavigableString) or \ +- isinstance(markup, basestring): ++ isinstance(markup, str): + if self._matches(markup, self.text): + found =3D markup + else: +- raise Exception, "I don't know how to match against a %s" \ +- % markup.__class__ ++ raise Exception("I don't know how to match against a %s" \ ++ % markup.__class__) + return found +=20 + def _matches(self, markup, matchAgainst): +@@ -981,8 +981,8 @@ + #other ways of matching match the tag name as a string. + if isinstance(markup, Tag): + markup =3D markup.name +- if markup and not isinstance(markup, basestring): +- markup =3D unicode(markup) ++ if markup and not isinstance(markup, str): ++ markup =3D str(markup) + #Now we know that chunk is either a string, or None. + if hasattr(matchAgainst, 'match'): + # It's a regexp object. +@@ -990,10 +990,10 @@ + elif hasattr(matchAgainst, '__iter__'): # list-like + result =3D markup in matchAgainst + elif hasattr(matchAgainst, 'items'): +- result =3D markup.has_key(matchAgainst) +- elif matchAgainst and isinstance(markup, basestring): +- if isinstance(markup, unicode): +- matchAgainst =3D unicode(matchAgainst) ++ result =3D matchAgainst in markup ++ elif matchAgainst and isinstance(markup, str): ++ if isinstance(markup, str): ++ matchAgainst =3D str(matchAgainst) + else: + matchAgainst =3D str(matchAgainst) +=20 +@@ -1018,7 +1018,7 @@ + for portion in args: + if hasattr(portion, 'items'): + #It's a map. Merge it. +- for k,v in portion.items(): ++ for k,v in list(portion.items()): + built[k] =3D v + elif hasattr(portion, '__iter__'): # is a list + #It's a list. Map each item to the default. +@@ -1061,7 +1061,7 @@ + lambda x: '') + ] +=20 +- ROOT_TAG_NAME =3D u'[document]' ++ ROOT_TAG_NAME =3D '[document]' +=20 + HTML_ENTITIES =3D "html" + XML_ENTITIES =3D "xml" +@@ -1157,14 +1157,14 @@ + def _feed(self, inDocumentEncoding=3DNone, isHTML=3DFalse): + # Convert the document to Unicode. + markup =3D self.markup +- if isinstance(markup, unicode): ++ if isinstance(markup, str): + if not hasattr(self, 'originalEncoding'): + self.originalEncoding =3D None + else: + dammit =3D UnicodeDammit\ + (markup, [self.fromEncoding, inDocumentEncoding], + smartQuotesTo=3Dself.smartQuotesTo, isHTML=3DisHTML) +- markup =3D dammit.unicode ++ markup =3D dammit.str + self.originalEncoding =3D dammit.originalEncoding + self.declaredHTMLEncoding =3D dammit.declaredHTMLEncoding + if markup: +@@ -1203,8 +1203,8 @@ + def isSelfClosingTag(self, name): + """Returns true iff the given string is the name of a + self-closing tag according to this parser.""" +- return self.SELF_CLOSING_TAGS.has_key(name) \ +- or self.instanceSelfClosingTags.has_key(name) ++ return name in self.SELF_CLOSING_TAGS \ ++ or name in self.instanceSelfClosingTags +=20 + def reset(self): + Tag.__init__(self, self, self.ROOT_TAG_NAME) +@@ -1233,7 +1233,7 @@ +=20 + def endData(self, containerClass=3DNavigableString): + if self.currentData: +- currentData =3D u''.join(self.currentData) ++ currentData =3D ''.join(self.currentData) + if (currentData.translate(self.STRIP_ASCII_SPACES) =3D=3D '' and + not set([tag.name for tag in self.tagStack]).intersection( + self.PRESERVE_WHITESPACE_TAGS)): +@@ -1296,7 +1296,7 @@ +=20 + nestingResetTriggers =3D self.NESTABLE_TAGS.get(name) + isNestable =3D nestingResetTriggers !=3D None +- isResetNesting =3D self.RESET_NESTING_TAGS.has_key(name) ++ isResetNesting =3D name in self.RESET_NESTING_TAGS + popTo =3D None + inclusive =3D True + for i in range(len(self.tagStack)-1, 0, -1): +@@ -1309,7 +1309,7 @@ + if (nestingResetTriggers is not None + and p.name in nestingResetTriggers) \ + or (nestingResetTriggers is None and isResetNesting +- and self.RESET_NESTING_TAGS.has_key(p.name)): ++ and p.name in self.RESET_NESTING_TAGS): +=20 + #If we encounter one of the nesting reset triggers + #peculiar to this tag, or we encounter another tag +@@ -1380,7 +1380,7 @@ + object, possibly one with a %SOUP-ENCODING% slot into which an + encoding will be plugged later.""" + if text[:3] =3D=3D "xml": +- text =3D u"xml version=3D'1.0' encoding=3D'%SOUP-ENCODING%'" ++ text =3D "xml version=3D'1.0' encoding=3D'%SOUP-ENCODING%'" + self._toStringSubclass(text, ProcessingInstruction) +=20 + def handle_comment(self, text): +@@ -1390,7 +1390,7 @@ + def handle_charref(self, ref): + "Handle character references as data." + if self.convertEntities: +- data =3D unichr(int(ref)) ++ data =3D chr(int(ref)) + else: + data =3D '&#%s;' % ref + self.handle_data(data) +@@ -1402,7 +1402,7 @@ + data =3D None + if self.convertHTMLEntities: + try: +- data =3D unichr(name2codepoint[ref]) ++ data =3D chr(name2codepoint[ref]) + except KeyError: + pass +=20 +@@ -1511,7 +1511,7 @@ + BeautifulStoneSoup before writing your own subclass.""" +=20 + def __init__(self, *args, **kwargs): +- if not kwargs.has_key('smartQuotesTo'): ++ if 'smartQuotesTo' not in kwargs: + kwargs['smartQuotesTo'] =3D self.HTML_ENTITIES + kwargs['isHTML'] =3D True + BeautifulStoneSoup.__init__(self, *args, **kwargs) +@@ -1694,7 +1694,7 @@ + parent._getAttrMap() + if (isinstance(tag, Tag) and len(tag.contents) =3D=3D 1 and + isinstance(tag.contents[0], NavigableString) and +- not parent.attrMap.has_key(tag.name)): ++ tag.name not in parent.attrMap): + parent[tag.name] =3D tag.contents[0] + BeautifulStoneSoup.popTag(self) +=20 +@@ -1768,9 +1768,9 @@ + self._detectEncoding(markup, isHTML) + self.smartQuotesTo =3D smartQuotesTo + self.triedEncodings =3D [] +- if markup =3D=3D '' or isinstance(markup, unicode): ++ if markup =3D=3D '' or isinstance(markup, str): + self.originalEncoding =3D None +- self.unicode =3D unicode(markup) ++ self.str =3D str(markup) + return +=20 + u =3D None +@@ -1783,7 +1783,7 @@ + if u: break +=20 + # If no luck and we have auto-detection library, try that: +- if not u and chardet and not isinstance(self.markup, unicode): ++ if not u and chardet and not isinstance(self.markup, str): + u =3D self._convertFrom(chardet.detect(self.markup)['encoding']) +=20 + # As a last resort, try utf-8 and windows-1252: +@@ -1792,7 +1792,7 @@ + u =3D self._convertFrom(proposed_encoding) + if u: break +=20 +- self.unicode =3D u ++ self.str =3D u + if not u: self.originalEncoding =3D None +=20 + def _subMSChar(self, orig): +@@ -1819,7 +1819,7 @@ + "iso-8859-1", + "iso-8859-2"): + markup =3D re.compile("([\x80-\x9f])").sub \ +- (lambda(x): self._subMSChar(x.group(1)), ++ (lambda x: self._subMSChar(x.group(1)), + markup) +=20 + try: +@@ -1827,7 +1827,7 @@ + u =3D self._toUnicode(markup, proposed) + self.markup =3D u + self.originalEncoding =3D proposed +- except Exception, e: ++ except Exception as e: + # print "That didn't work!" + # print e + return None +@@ -1856,7 +1856,7 @@ + elif data[:4] =3D=3D '\xff\xfe\x00\x00': + encoding =3D 'utf-32le' + data =3D data[4:] +- newdata =3D unicode(data, encoding) ++ newdata =3D str(data, encoding) + return newdata +=20 + def _detectEncoding(self, xml_data, isHTML=3DFalse): +@@ -1869,41 +1869,41 @@ + elif xml_data[:4] =3D=3D '\x00\x3c\x00\x3f': + # UTF-16BE + sniffed_xml_encoding =3D 'utf-16be' +- xml_data =3D unicode(xml_data, 'utf-16be').encode('utf-8') ++ xml_data =3D str(xml_data, 'utf-16be').encode('utf-8') + elif (len(xml_data) >=3D 4) and (xml_data[:2] =3D=3D '\xfe\xff'= ) \ + and (xml_data[2:4] !=3D '\x00\x00'): + # UTF-16BE with BOM + sniffed_xml_encoding =3D 'utf-16be' +- xml_data =3D unicode(xml_data[2:], 'utf-16be').encode('utf-= 8') ++ xml_data =3D str(xml_data[2:], 'utf-16be').encode('utf-8') + elif xml_data[:4] =3D=3D '\x3c\x00\x3f\x00': + # UTF-16LE + sniffed_xml_encoding =3D 'utf-16le' +- xml_data =3D unicode(xml_data, 'utf-16le').encode('utf-8') ++ xml_data =3D str(xml_data, 'utf-16le').encode('utf-8') + elif (len(xml_data) >=3D 4) and (xml_data[:2] =3D=3D '\xff\xfe'= ) and \ + (xml_data[2:4] !=3D '\x00\x00'): + # UTF-16LE with BOM + sniffed_xml_encoding =3D 'utf-16le' +- xml_data =3D unicode(xml_data[2:], 'utf-16le').encode('utf-= 8') ++ xml_data =3D str(xml_data[2:], 'utf-16le').encode('utf-8') + elif xml_data[:4] =3D=3D '\x00\x00\x00\x3c': + # UTF-32BE + sniffed_xml_encoding =3D 'utf-32be' +- xml_data =3D unicode(xml_data, 'utf-32be').encode('utf-8') ++ xml_data =3D str(xml_data, 'utf-32be').encode('utf-8') + elif xml_data[:4] =3D=3D '\x3c\x00\x00\x00': + # UTF-32LE + sniffed_xml_encoding =3D 'utf-32le' +- xml_data =3D unicode(xml_data, 'utf-32le').encode('utf-8') ++ xml_data =3D str(xml_data, 'utf-32le').encode('utf-8') + elif xml_data[:4] =3D=3D '\x00\x00\xfe\xff': + # UTF-32BE with BOM + sniffed_xml_encoding =3D 'utf-32be' +- xml_data =3D unicode(xml_data[4:], 'utf-32be').encode('utf-= 8') ++ xml_data =3D str(xml_data[4:], 'utf-32be').encode('utf-8') + elif xml_data[:4] =3D=3D '\xff\xfe\x00\x00': + # UTF-32LE with BOM + sniffed_xml_encoding =3D 'utf-32le' +- xml_data =3D unicode(xml_data[4:], 'utf-32le').encode('utf-= 8') ++ xml_data =3D str(xml_data[4:], 'utf-32le').encode('utf-8') + elif xml_data[:3] =3D=3D '\xef\xbb\xbf': + # UTF-8 with BOM + sniffed_xml_encoding =3D 'utf-8' +- xml_data =3D unicode(xml_data[3:], 'utf-8').encode('utf-8') ++ xml_data =3D str(xml_data[3:], 'utf-8').encode('utf-8') + else: + sniffed_xml_encoding =3D 'ascii' + pass +@@ -1966,7 +1966,7 @@ + 250,251,252,253,254,255) + import string + c.EBCDIC_TO_ASCII_MAP =3D string.maketrans( \ +- ''.join(map(chr, range(256))), ''.join(map(chr, emap))) ++ ''.join(map(chr, list(range(256)))), ''.join(map(chr, emap))) + return s.translate(c.EBCDIC_TO_ASCII_MAP) +=20 + MS_CHARS =3D { '\x80' : ('euro', '20AC'), +@@ -2009,4 +2009,4 @@ + if __name__ =3D=3D '__main__': + import sys + soup =3D BeautifulSoup(sys.stdin) +- print soup.prettify() ++ print(soup.prettify()) +diff -Naur client175_0.7-original/cherrypy/cherryd client175_0.7/cherrypy/ch= erryd +--- client175_0.7-original/cherrypy/cherryd 2010-04-20 13:10:10.000000000 +0= 200 ++++ client175_0.7/cherrypy/cherryd 2021-08-03 15:37:40.098963967 +0200 +@@ -12,7 +12,7 @@ + """Subscribe all engine plugins and start the engine.""" + sys.path =3D [''] + sys.path + for i in imports or []: +- exec "import %s" % i ++ exec("import %s" % i) + =20 + for c in configfiles or []: + cherrypy.config.update(c) +diff -Naur client175_0.7-original/cherrypy/_cpcgifs.py client175_0.7/cherryp= y/_cpcgifs.py +--- client175_0.7-original/cherrypy/_cpcgifs.py 2010-04-20 13:10:10.00000000= 0 +0200 ++++ client175_0.7/cherrypy/_cpcgifs.py 2021-08-03 14:41:19.199896214 +0200 +@@ -6,7 +6,7 @@ + def __init__(self, *args, **kwds): + try: + cgi.FieldStorage.__init__(self, *args, **kwds) +- except ValueError, ex: ++ except ValueError as ex: + if str(ex) =3D=3D 'Maximum content length exceeded': + raise cherrypy.HTTPError(status=3D413) + else: +diff -Naur client175_0.7-original/cherrypy/_cpchecker.py client175_0.7/cherr= ypy/_cpchecker.py +--- client175_0.7-original/cherrypy/_cpchecker.py 2010-04-20 13:10:10.000000= 000 +0200 ++++ client175_0.7/cherrypy/_cpchecker.py 2021-08-03 14:41:30.971721551 +0200 +@@ -47,7 +47,7 @@ + global_config_contained_paths =3D False + =20 + def check_skipped_app_config(self): +- for sn, app in cherrypy.tree.apps.iteritems(): ++ for sn, app in cherrypy.tree.apps.items(): + if not isinstance(app, cherrypy.Application): + continue + if not app.config: +@@ -64,7 +64,7 @@ + def check_static_paths(self): + # Use the dummy Request object in the main thread. + request =3D cherrypy.request +- for sn, app in cherrypy.tree.apps.iteritems(): ++ for sn, app in cherrypy.tree.apps.items(): + if not isinstance(app, cherrypy.Application): + continue + request.app =3D app +@@ -130,9 +130,9 @@ + =20 + def _compat(self, config): + """Process config and warn on each obsolete or deprecated entry.""" +- for section, conf in config.iteritems(): ++ for section, conf in config.items(): + if isinstance(conf, dict): +- for k, v in conf.iteritems(): ++ for k, v in conf.items(): + if k in self.obsolete: + warnings.warn("%r is obsolete. Use %r instead.\n" + "section: [%s]" % +@@ -152,7 +152,7 @@ + def check_compatibility(self): + """Process config and warn on each obsolete or deprecated entry.""" + self._compat(cherrypy.config) +- for sn, app in cherrypy.tree.apps.iteritems(): ++ for sn, app in cherrypy.tree.apps.items(): + if not isinstance(app, cherrypy.Application): + continue + self._compat(app.config) +@@ -164,16 +164,16 @@ + =20 + def _known_ns(self, app): + ns =3D ["wsgi"] +- ns.extend(app.toolboxes.keys()) +- ns.extend(app.namespaces.keys()) +- ns.extend(app.request_class.namespaces.keys()) +- ns.extend(cherrypy.config.namespaces.keys()) ++ ns.extend(list(app.toolboxes.keys())) ++ ns.extend(list(app.namespaces.keys())) ++ ns.extend(list(app.request_class.namespaces.keys())) ++ ns.extend(list(cherrypy.config.namespaces.keys())) + ns +=3D self.extra_config_namespaces + =20 +- for section, conf in app.config.iteritems(): ++ for section, conf in app.config.items(): + is_path_section =3D section.startswith("/") + if is_path_section and isinstance(conf, dict): +- for k, v in conf.iteritems(): ++ for k, v in conf.items(): + atoms =3D k.split(".") + if len(atoms) > 1: + if atoms[0] not in ns: +@@ -197,7 +197,7 @@ + =20 + def check_config_namespaces(self): + """Process config and warn on each unknown config namespace.""" +- for sn, app in cherrypy.tree.apps.iteritems(): ++ for sn, app in cherrypy.tree.apps.items(): + if not isinstance(app, cherrypy.Application): + continue + self._known_ns(app) +@@ -210,8 +210,8 @@ + known_config_types =3D {} + =20 + def _populate_known_types(self): +- import __builtin__ +- builtins =3D [x for x in vars(__builtin__).values() ++ import builtins ++ builtins =3D [x for x in list(vars(__builtin__).values()) + if type(x) is type(str)] + =20 + def traverse(obj, namespace): +@@ -230,9 +230,9 @@ + msg =3D ("The config entry %r in section %r is of type %r, " + "which does not match the expected type %r.") + =20 +- for section, conf in config.iteritems(): ++ for section, conf in config.items(): + if isinstance(conf, dict): +- for k, v in conf.iteritems(): ++ for k, v in conf.items(): + if v is not None: + expected_type =3D self.known_config_types.get(k, No= ne) + vtype =3D type(v) +@@ -251,7 +251,7 @@ + def check_config_types(self): + """Assert that config values are of the same type as default values= .""" + self._known_types(cherrypy.config) +- for sn, app in cherrypy.tree.apps.iteritems(): ++ for sn, app in cherrypy.tree.apps.items(): + if not isinstance(app, cherrypy.Application): + continue + self._known_types(app.config) +@@ -261,7 +261,7 @@ + =20 + def check_localhost(self): + """Warn if any socket_host is 'localhost'. See #711.""" +- for k, v in cherrypy.config.iteritems(): ++ for k, v in cherrypy.config.items(): + if k =3D=3D 'server.socket_host' and v =3D=3D 'localhost': + warnings.warn("The use of 'localhost' as a socket host can " + "cause problems on newer systems, since 'localhost' can= " +diff -Naur client175_0.7-original/cherrypy/_cpconfig.py client175_0.7/cherry= py/_cpconfig.py +--- client175_0.7-original/cherrypy/_cpconfig.py 2010-04-20 13:10:10.0000000= 00 +0200 ++++ client175_0.7/cherrypy/_cpconfig.py 2021-08-03 14:41:39.459596077 +0200 +@@ -93,7 +93,7 @@ + style) context manager. + """ +=20 +-import ConfigParser ++import configparser + try: + set + except NameError: +@@ -138,7 +138,7 @@ +=20 + def as_dict(config): + """Return a dict from 'config' whether it is a dict, file, or filename.= """ +- if isinstance(config, basestring): ++ if isinstance(config, str): + config =3D _Parser().dict_from_file(config) + elif hasattr(config, 'read'): + config =3D _Parser().dict_from_file(config) +@@ -150,11 +150,11 @@ + If the given config is a filename, it will be appended to + the list of files to monitor for "autoreload" changes. + """ +- if isinstance(other, basestring): ++ if isinstance(other, str): + cherrypy.engine.autoreload.files.add(other) + =20 + # Load other into base +- for section, value_map in as_dict(other).iteritems(): ++ for section, value_map in as_dict(other).items(): + base.setdefault(section, {}).update(value_map) +=20 +=20 +@@ -196,14 +196,14 @@ + # with handler as callable: + # for k, v in ns_confs.get(ns, {}).iteritems(): + # callable(k, v) +- for ns, handler in self.iteritems(): ++ for ns, handler in self.items(): + exit =3D getattr(handler, "__exit__", None) + if exit: + callable =3D handler.__enter__() + no_exc =3D True + try: + try: +- for k, v in ns_confs.get(ns, {}).iteritems(): ++ for k, v in ns_confs.get(ns, {}).items(): + callable(k, v) + except: + # The exceptional case is handled here +@@ -218,7 +218,7 @@ + if no_exc and exit: + exit(None, None, None) + else: +- for k, v in ns_confs.get(ns, {}).iteritems(): ++ for k, v in ns_confs.get(ns, {}).items(): + handler(k, v) + =20 + def __repr__(self): +@@ -257,7 +257,7 @@ + =20 + def update(self, config): + """Update self from a dict, file or filename.""" +- if isinstance(config, basestring): ++ if isinstance(config, str): + # Filename + cherrypy.engine.autoreload.files.add(config) + config =3D _Parser().dict_from_file(config) +@@ -333,7 +333,7 @@ + Config.namespaces["tree"] =3D _tree_namespace_handler +=20 +=20 +-class _Parser(ConfigParser.ConfigParser): ++class _Parser(configparser.ConfigParser): + """Sub-class of ConfigParser that keeps the case of options and that ra= ises + an exception if the file cannot be read. + """ +@@ -342,7 +342,7 @@ + return optionstr + =20 + def read(self, filenames): +- if isinstance(filenames, basestring): ++ if isinstance(filenames, str): + filenames =3D [filenames] + for filename in filenames: + # try: +@@ -367,7 +367,7 @@ + value =3D self.get(section, option, raw, vars) + try: + value =3D unrepr(value) +- except Exception, x: ++ except Exception as x: + msg =3D ("Config error in section: %r, option: %r, " + "value: %r. Config values must be valid Python."= % + (section, option, value)) +diff -Naur client175_0.7-original/cherrypy/_cpdispatch.py client175_0.7/cher= rypy/_cpdispatch.py +--- client175_0.7-original/cherrypy/_cpdispatch.py 2010-04-20 13:10:10.00000= 0000 +0200 ++++ client175_0.7/cherrypy/_cpdispatch.py 2021-08-03 14:41:48.627460479 +0200 +@@ -23,7 +23,7 @@ + def __call__(self): + try: + return self.callable(*self.args, **self.kwargs) +- except TypeError, x: ++ except TypeError as x: + test_callable_spec(self.callable, self.args, self.kwargs) + raise +=20 +@@ -62,7 +62,7 @@ + except IndexError: + vararg_usage +=3D 1 +=20 +- for key in callable_kwargs.keys(): ++ for key in list(callable_kwargs.keys()): + try: + arg_usage[key] +=3D 1 + except KeyError: +@@ -76,7 +76,7 @@ +=20 + missing_args =3D [] + multiple_args =3D [] +- for key, usage in arg_usage.iteritems(): ++ for key, usage in arg_usage.items(): + if usage =3D=3D 0: + missing_args.append(key) + elif usage > 1: +@@ -258,7 +258,7 @@ + =20 + # Try successive objects (reverse order) + num_candidates =3D len(object_trail) - 1 +- for i in xrange(num_candidates, -1, -1): ++ for i in range(num_candidates, -1, -1): + =20 + name, candidate, nodeconf, curpath =3D object_trail[i] + if candidate is None: +diff -Naur client175_0.7-original/cherrypy/_cperror.py client175_0.7/cherryp= y/_cperror.py +--- client175_0.7-original/cherrypy/_cperror.py 2010-04-20 13:10:10.00000000= 0 +0200 ++++ client175_0.7/cherrypy/_cperror.py 2021-08-03 14:41:56.939337474 +0200 +@@ -3,7 +3,7 @@ + from cgi import escape as _escape + from sys import exc_info as _exc_info + from traceback import format_exception as _format_exception +-from urlparse import urljoin as _urljoin ++from urllib.parse import urljoin as _urljoin + from cherrypy.lib import http as _http +=20 +=20 +@@ -57,7 +57,7 @@ + import cherrypy + request =3D cherrypy.request + =20 +- if isinstance(urls, basestring): ++ if isinstance(urls, str): + urls =3D [urls] + =20 + abs_urls =3D [] +@@ -161,7 +161,7 @@ + for key in ["Accept-Ranges", "Age", "ETag", "Location", "Retry-After", + "Vary", "Content-Encoding", "Content-Length", "Expires", + "Content-Location", "Content-MD5", "Last-Modified"]: +- if respheaders.has_key(key): ++ if key in respheaders: + del respheaders[key] + =20 + if status !=3D 416: +@@ -171,7 +171,7 @@ + # specifies the current length of the selected resource. + # A response with status code 206 (Partial Content) MUST NOT + # include a Content-Range field with a byte-range- resp-spec of "*". +- if respheaders.has_key("Content-Range"): ++ if "Content-Range" in respheaders: + del respheaders["Content-Range"] +=20 +=20 +@@ -277,7 +277,7 @@ + =20 + try: + code, reason, message =3D _http.valid_status(status) +- except ValueError, x: ++ except ValueError as x: + raise cherrypy.HTTPError(500, x.args[0]) + =20 + # We can't use setdefault here, because some +@@ -291,7 +291,7 @@ + if kwargs.get('version') is None: + kwargs['version'] =3D cherrypy.__version__ + =20 +- for k, v in kwargs.iteritems(): ++ for k, v in kwargs.items(): + if v is None: + kwargs[k] =3D "" + else: +diff -Naur client175_0.7-original/cherrypy/_cplogging.py client175_0.7/cherr= ypy/_cplogging.py +--- client175_0.7-original/cherrypy/_cplogging.py 2010-04-20 13:10:10.000000= 000 +0200 ++++ client175_0.7/cherrypy/_cplogging.py 2021-08-03 14:42:04.739222052 +0200 +@@ -88,8 +88,8 @@ + 'f': inheaders.get('Referer', ''), + 'a': inheaders.get('User-Agent', ''), + } +- for k, v in atoms.items(): +- if isinstance(v, unicode): ++ for k, v in list(atoms.items()): ++ if isinstance(v, str): + v =3D v.encode('utf8') + elif not isinstance(v, str): + v =3D str(v) +diff -Naur client175_0.7-original/cherrypy/_cpmodpy.py client175_0.7/cherryp= y/_cpmodpy.py +--- client175_0.7-original/cherrypy/_cpmodpy.py 2010-04-20 13:10:10.00000000= 0 +0200 ++++ client175_0.7/cherrypy/_cpmodpy.py 2021-08-03 14:42:11.807117510 +0200 +@@ -56,7 +56,7 @@ + """ +=20 + import logging +-import StringIO ++import io +=20 + import cherrypy + from cherrypy._cperror import format_exc, bare_error +@@ -183,7 +183,7 @@ + path =3D req.uri + qs =3D req.args or "" + reqproto =3D req.protocol +- headers =3D req.headers_in.items() ++ headers =3D list(req.headers_in.items()) + rfile =3D _ReadOnlyRequest(req) + prev =3D None + =20 +@@ -202,7 +202,7 @@ + try: + request.run(method, path, qs, reqproto, headers, rf= ile) + break +- except cherrypy.InternalRedirect, ir: ++ except cherrypy.InternalRedirect as ir: + app.release_serving() + prev =3D request + =20 +@@ -220,7 +220,7 @@ + method =3D "GET" + path =3D ir.path + qs =3D ir.query_string +- rfile =3D StringIO.StringIO() ++ rfile =3D io.StringIO() + =20 + send_response(req, response.status, response.header_list, + response.body, response.stream) +@@ -251,7 +251,7 @@ + req.flush() + =20 + # Set response body +- if isinstance(body, basestring): ++ if isinstance(body, str): + req.write(body) + else: + for seg in body: +diff -Naur client175_0.7-original/cherrypy/_cprequest.py client175_0.7/cherr= ypy/_cprequest.py +--- client175_0.7-original/cherrypy/_cprequest.py 2010-04-20 13:10:10.000000= 000 +0200 ++++ client175_0.7/cherrypy/_cprequest.py 2021-08-03 14:42:19.091009678 +0200 +@@ -1,5 +1,5 @@ +=20 +-import Cookie ++import http.cookies + import os + import sys + import time +@@ -11,11 +11,9 @@ + from cherrypy.lib import http, file_generator +=20 +=20 +-class Hook(object): ++class Hook(object, metaclass=3Dcherrypy._AttributeDocstrings): + """A callback and its metadata: failsafe, priority, and kwargs.""" + =20 +- __metaclass__ =3D cherrypy._AttributeDocstrings +- =20 + callback =3D None + callback__doc =3D """ + The bare callable that this Hook object is wrapping, which will +@@ -63,7 +61,7 @@ + % (cls.__module__, cls.__name__, self.callback, + self.failsafe, self.priority, + ", ".join(['%s=3D%r' % (k, v) +- for k, v in self.kwargs.iteritems()]))) ++ for k, v in self.kwargs.items()]))) +=20 +=20 + class HookMap(dict): +@@ -111,14 +109,14 @@ + newmap =3D self.__class__() + # We can't just use 'update' because we want copies of the + # mutable values (each is a list) as well. +- for k, v in self.iteritems(): ++ for k, v in self.items(): + newmap[k] =3D v[:] + return newmap + copy =3D __copy__ + =20 + def __repr__(self): + cls =3D self.__class__ +- return "%s.%s(points=3D%r)" % (cls.__module__, cls.__name__, self.k= eys()) ++ return "%s.%s(points=3D%r)" % (cls.__module__, cls.__name__, list(s= elf.keys())) +=20 +=20 + # Config namespace handlers +@@ -129,7 +127,7 @@ + # hookpoint per path (e.g. "hooks.before_handler.1"). + # Little-known fact you only get from reading source ;) + hookpoint =3D k.split(".", 1)[0] +- if isinstance(v, basestring): ++ if isinstance(v, str): + v =3D cherrypy.lib.attributes(v) + if not isinstance(v, Hook): + v =3D Hook(v) +@@ -156,7 +154,7 @@ + 'before_error_response', 'after_error_response'] +=20 +=20 +-class Request(object): ++class Request(object, metaclass=3Dcherrypy._AttributeDocstrings): + """An HTTP request. + =20 + This object represents the metadata of an HTTP request message; +@@ -169,8 +167,6 @@ + the given URL, and the execution plan for generating a response. + """ + =20 +- __metaclass__ =3D cherrypy._AttributeDocstrings +- =20 + prev =3D None + prev__doc =3D """ + The previous Request object (if any). This should be None +@@ -251,7 +247,7 @@ + values (decoded according to RFC 2047 if necessary). See also: + http.HeaderMap, http.HeaderElement.""" + =20 +- cookie =3D Cookie.SimpleCookie() ++ cookie =3D http.cookies.SimpleCookie() + cookie__doc =3D """See help(Cookie).""" + =20 + rfile =3D None +@@ -529,7 +525,7 @@ + self.header_list =3D list(headers) + self.rfile =3D rfile + self.headers =3D http.HeaderMap() +- self.cookie =3D Cookie.SimpleCookie() ++ self.cookie =3D http.cookies.SimpleCookie() + self.handler =3D None + =20 + # path_info should be the path from the +@@ -608,7 +604,7 @@ + self.stage =3D 'before_finalize' + self.hooks.run('before_finalize') + cherrypy.response.finalize() +- except (cherrypy.HTTPRedirect, cherrypy.HTTPError), inst: ++ except (cherrypy.HTTPRedirect, cherrypy.HTTPError) as inst: + inst.set_response() + self.stage =3D 'before_finalize (HTTPError)' + self.hooks.run('before_finalize') +@@ -648,7 +644,7 @@ + if name =3D=3D 'Cookie': + try: + self.cookie.load(value) +- except Cookie.CookieError: ++ except http.cookies.CookieError: + msg =3D "Illegal cookie name %s" % value.split('=3D')[0] + raise cherrypy.HTTPError(400, msg) + =20 +@@ -709,7 +705,7 @@ + # won't parse the request body for params if the client + # didn't provide a "Content-Type" header. + if 'Content-Type' not in self.headers: +- h =3D http.HeaderMap(self.headers.items()) ++ h =3D http.HeaderMap(list(self.headers.items())) + h['Content-Type'] =3D '' + else: + h =3D self.headers +@@ -720,7 +716,7 @@ + # FieldStorage only recognizes PO= ST. + environ=3D{'REQUEST_METHOD': "POS= T"}, + keep_blank_values=3D1) +- except Exception, e: ++ except Exception as e: + if e.__class__.__name__ =3D=3D 'MaxSizeExceeded': + # Post data is too big + raise cherrypy.HTTPError(413) +@@ -746,7 +742,7 @@ + self.error_response() + self.hooks.run("after_error_response") + cherrypy.response.finalize() +- except cherrypy.HTTPRedirect, inst: ++ except cherrypy.HTTPRedirect as inst: + inst.set_response() + cherrypy.response.finalize() +=20 +@@ -763,7 +759,7 @@ + =20 + def __set__(self, obj, value): + # Convert the given value to an iterable object. +- if isinstance(value, basestring): ++ if isinstance(value, str): + # strings get wrapped in a list because iterating over a single + # item list is much faster than iterating over every character + # in a long string. +@@ -779,7 +775,7 @@ + obj._body =3D value +=20 +=20 +-class Response(object): ++class Response(object, metaclass=3Dcherrypy._AttributeDocstrings): + """An HTTP Response, including status, headers, and body. + =20 + Application developers should use Response.headers (a dict) to +@@ -788,8 +784,6 @@ + (key, value) tuples. + """ + =20 +- __metaclass__ =3D cherrypy._AttributeDocstrings +- =20 + # Class attributes for dev-time introspection. + status =3D "" + status__doc =3D """The HTTP Status-Code and Reason-Phrase.""" +@@ -808,7 +802,7 @@ + values (decoded according to RFC 2047 if necessary). See also: + http.HeaderMap, http.HeaderElement.""" + =20 +- cookie =3D Cookie.SimpleCookie() ++ cookie =3D http.cookies.SimpleCookie() + cookie__doc =3D """See help(Cookie).""" + =20 + body =3D Body() +@@ -842,7 +836,7 @@ + "Server": "CherryPy/" + cherrypy.__version__, + "Date": http.HTTPDate(self.time), + }) +- self.cookie =3D Cookie.SimpleCookie() ++ self.cookie =3D http.cookies.SimpleCookie() + =20 + def collapse_body(self): + """Collapse self.body to a single string; replace it and return it.= """ +@@ -854,7 +848,7 @@ + """Transform headers (and cookies) into self.header_list. (Core)""" + try: + code, reason, _ =3D http.valid_status(self.status) +- except ValueError, x: ++ except ValueError as x: + raise cherrypy.HTTPError(500, x.args[0]) + =20 + self.status =3D "%s %s" % (code, reason) +diff -Naur client175_0.7-original/cherrypy/_cpserver.py client175_0.7/cherry= py/_cpserver.py +--- client175_0.7-original/cherrypy/_cpserver.py 2010-04-20 13:10:10.0000000= 00 +0200 ++++ client175_0.7/cherrypy/_cpserver.py 2021-08-03 14:42:27.582884114 +0200 +@@ -73,7 +73,7 @@ + if httpserver is None: + from cherrypy import _cpwsgi_server + httpserver =3D _cpwsgi_server.CPWSGIServer() +- if isinstance(httpserver, basestring): ++ if isinstance(httpserver, str): + httpserver =3D attributes(httpserver)() + =20 + if self.socket_file: +diff -Naur client175_0.7-original/cherrypy/_cptools.py client175_0.7/cherryp= y/_cptools.py +--- client175_0.7-original/cherrypy/_cptools.py 2010-04-20 13:10:10.00000000= 0 +0200 ++++ client175_0.7/cherrypy/_cptools.py 2021-08-03 14:43:38.721831788 +0200 +@@ -30,8 +30,8 @@ + # Use this instead of importing inspect for less mem overhead. + import types + if isinstance(func, types.MethodType): +- func =3D func.im_func +- co =3D func.func_code ++ func =3D func.__func__ ++ co =3D func.__code__ + return co.co_varnames[:co.co_argcount] +=20 +=20 +@@ -105,7 +105,7 @@ + f._cp_config =3D {} + subspace =3D self.namespace + "." + self._name + "." + f._cp_config[subspace + "on"] =3D True +- for k, v in kwargs.iteritems(): ++ for k, v in kwargs.items(): + f._cp_config[subspace + k] =3D v + return f + return tool_decorator +@@ -286,7 +286,7 @@ + sess.regenerate() + =20 + # Grab cookie-relevant tool args +- conf =3D dict([(k, v) for k, v in self._merged_args().iteritems() ++ conf =3D dict([(k, v) for k, v in self._merged_args().items() + if k in ('path', 'path_header', 'name', 'timeout', + 'domain', 'secure')]) + _sessions.set_response_cookie(**conf) +@@ -346,7 +346,7 @@ + # if a method is not found, an xmlrpclib.Fault should be return= ed + # raising an exception here will do that; see + # cherrypy.lib.xmlrpc.on_error +- raise Exception, 'method "%s" is not supported' % attr ++ raise Exception('method "%s" is not supported' % attr) + =20 + conf =3D cherrypy.request.toolmaps['tools'].get("xmlrpc", {}) + _xmlrpc.respond(body, +@@ -399,7 +399,7 @@ + cherrypy._cache =3D kwargs.pop("cache_class", _caching.MemoryCa= che)() + =20 + # Take all remaining kwargs and set them on the Cache object. +- for k, v in kwargs.iteritems(): ++ for k, v in kwargs.items(): + setattr(cherrypy._cache, k, v) + =20 + if _caching.get(invalid_methods=3Dinvalid_methods): +@@ -452,7 +452,7 @@ + """Run tool._setup() for each tool in our toolmap.""" + map =3D cherrypy.request.toolmaps.get(self.namespace) + if map: +- for name, settings in map.items(): ++ for name, settings in list(map.items()): + if settings.get("on", False): + tool =3D getattr(self, name) + tool._setup() +diff -Naur client175_0.7-original/cherrypy/_cptree.py client175_0.7/cherrypy= /_cptree.py +--- client175_0.7-original/cherrypy/_cptree.py 2010-04-20 13:10:10.000000000= +0200 ++++ client175_0.7/cherrypy/_cptree.py 2021-08-03 14:43:50.457658068 +0200 +@@ -6,7 +6,7 @@ + from cherrypy.lib import http as _http +=20 +=20 +-class Application(object): ++class Application(object, metaclass=3Dcherrypy._AttributeDocstrings): + """A CherryPy Application. + =20 + Servers and gateways should not instantiate Request objects directly. +@@ -16,8 +16,6 @@ + (WSGI application object) for itself. + """ + =20 +- __metaclass__ =3D cherrypy._AttributeDocstrings +- =20 + root =3D None + root__doc =3D """ + The top-most container of page handlers for this app. Handlers should +@@ -103,7 +101,7 @@ + req =3D self.request_class(local, remote, scheme, sproto) + req.app =3D self + =20 +- for name, toolbox in self.toolboxes.iteritems(): ++ for name, toolbox in self.toolboxes.items(): + req.namespaces[name] =3D toolbox + =20 + resp =3D self.response_class() +@@ -171,7 +169,7 @@ + if isinstance(root, Application): + app =3D root + if script_name !=3D "" and script_name !=3D app.script_name: +- raise ValueError, "Cannot specify a different script name a= nd pass an Application instance to cherrypy.mount" ++ raise ValueError("Cannot specify a different script name an= d pass an Application instance to cherrypy.mount") + script_name =3D app.script_name + else: + app =3D Application(root, script_name) +diff -Naur client175_0.7-original/cherrypy/_cpwsgi.py client175_0.7/cherrypy= /_cpwsgi.py +--- client175_0.7-original/cherrypy/_cpwsgi.py 2010-04-20 13:10:10.000000000= +0200 ++++ client175_0.7/cherrypy/_cpwsgi.py 2021-08-03 14:44:08.117396886 +0200 +@@ -1,6 +1,6 @@ + """WSGI interface (see PEP 333).""" +=20 +-import StringIO as _StringIO ++import io as _StringIO + import sys as _sys +=20 + import cherrypy as _cherrypy +@@ -82,7 +82,7 @@ + except self.throws: + self.close() + raise +- except _cherrypy.InternalRedirect, ir: ++ except _cherrypy.InternalRedirect as ir: + self.environ['cherrypy.previous_request'] =3D _cherrypy.serving= .request + self.close() + self.iredirect(ir.path, ir.query_string) +@@ -158,9 +158,9 @@ + def __iter__(self): + return self + =20 +- def next(self): ++ def __next__(self): + try: +- chunk =3D self.iter_response.next() ++ chunk =3D next(self.iter_response) + # WSGI requires all data to be of type "str". This coercion sho= uld + # not take any time at all if chunk is already of type "str". + # If it's unicode, it could be a big performance hit (x ~500). +@@ -170,7 +170,7 @@ + except self.throws: + self.close() + raise +- except _cherrypy.InternalRedirect, ir: ++ except _cherrypy.InternalRedirect as ir: + self.environ['cherrypy.previous_request'] =3D _cherrypy.serving= .request + self.close() + self.iredirect(ir.path, ir.query_string) +diff -Naur client175_0.7-original/cherrypy/__init__.py client175_0.7/cherryp= y/__init__.py +--- client175_0.7-original/cherrypy/__init__.py 2010-04-20 13:10:10.00000000= 0 +0200 ++++ client175_0.7/cherrypy/__init__.py 2021-08-03 14:44:38.280950778 +0200 +@@ -59,7 +59,7 @@ +=20 + __version__ =3D "3.1.2" +=20 +-from urlparse import urljoin as _urljoin ++from urllib.parse import urljoin as _urljoin +=20 +=20 + class _AttributeDocstrings(type): +@@ -126,7 +126,7 @@ + =20 + newdoc =3D [cls.__doc__ or ""] + =20 +- dctnames =3D dct.keys() ++ dctnames =3D list(dct.keys()) + dctnames.sort() + =20 + for name in dctnames: +@@ -254,7 +254,7 @@ + except ImportError: + from cherrypy._cpthreadinglocal import local as _local +=20 +-class _Serving(_local): ++class _Serving(_local, metaclass=3D_AttributeDocstrings): + """An interface for registering request and response objects. + =20 + Rather than have a separate "thread local" object for the request and +@@ -265,8 +265,6 @@ + thread-safe way. + """ + =20 +- __metaclass__ =3D _AttributeDocstrings +- =20 + request =3D _cprequest.Request(_http.Host("127.0.0.1", 80), + _http.Host("127.0.0.1", 1111)) + request__doc =3D """ +@@ -338,7 +336,7 @@ + child =3D getattr(serving, self.__attrname__) + return len(child) + =20 +- def __nonzero__(self): ++ def __bool__(self): + child =3D getattr(serving, self.__attrname__) + return bool(child) +=20 +@@ -410,7 +408,7 @@ + def expose_(func): + func.exposed =3D True + if alias is not None: +- if isinstance(alias, basestring): ++ if isinstance(alias, str): + parents[alias.replace(".", "_")] =3D func + else: + for a in alias: +diff -Naur client175_0.7-original/cherrypy/lib/auth.py client175_0.7/cherryp= y/lib/auth.py +--- client175_0.7-original/cherrypy/lib/auth.py 2010-04-20 13:10:10.00000000= 0 +0200 ++++ client175_0.7/cherrypy/lib/auth.py 2021-08-03 14:45:01.296612330 +0200 +@@ -19,7 +19,7 @@ + users =3D users() # expect it to return a dictionary +=20 + if not isinstance(users, dict): +- raise ValueError, "Authentication users must be a dicti= onary" ++ raise ValueError("Authentication users must be a dictio= nary") + =20 + # fetch the user password + password =3D users.get(ah["username"], None) +@@ -28,7 +28,7 @@ + password =3D users(ah["username"]) + else: + if not isinstance(users, dict): +- raise ValueError, "Authentication users must be a dictionar= y" ++ raise ValueError("Authentication users must be a dictionary= ") + =20 + # fetch the user password + password =3D users.get(ah["username"], None) +diff -Naur client175_0.7-original/cherrypy/lib/caching.py client175_0.7/cher= rypy/lib/caching.py +--- client175_0.7-original/cherrypy/lib/caching.py 2010-04-20 13:10:10.00000= 0000 +0200 ++++ client175_0.7/cherrypy/lib/caching.py 2021-08-03 14:45:08.464504417 +0200 +@@ -45,7 +45,7 @@ + # See tickets #99 and #180 for more information. + while time: + now =3D time.time() +- for expiration_time, objects in self.expirations.items(): ++ for expiration_time, objects in list(self.expirations.items()): + if expiration_time <=3D now: + for obj_size, obj_key in objects: + try: +@@ -161,7 +161,7 @@ + # this was put into the cached copy, and should have been + # resurrected just above (response.headers =3D cache_data[1]). + cptools.validate_since() +- except cherrypy.HTTPRedirect, x: ++ except cherrypy.HTTPRedirect as x: + if x.status =3D=3D 304: + cherrypy._cache.tot_non_modified +=3D 1 + raise +@@ -188,7 +188,7 @@ + cherrypy.response.headers.elements('Vary')] + if vary: + sel_headers =3D dict([(k, v) for k, v +- in cherrypy.request.headers.iteritems() ++ in cherrypy.request.headers.items() + if k in vary]) + else: + sel_headers =3D {} +diff -Naur client175_0.7-original/cherrypy/lib/covercp.py client175_0.7/cher= rypy/lib/covercp.py +--- client175_0.7-original/cherrypy/lib/covercp.py 2010-04-20 13:10:10.00000= 0000 +0200 ++++ client175_0.7/cherrypy/lib/covercp.py 2021-08-03 14:45:14.876409627 +0200 +@@ -21,14 +21,14 @@ + import re + import sys + import cgi +-import urllib ++import urllib.request, urllib.parse, urllib.error + import os, os.path + localFile =3D os.path.join(os.path.dirname(__file__), "coverage.cache") +=20 + try: +- import cStringIO as StringIO ++ import io as StringIO + except ImportError: +- import StringIO ++ import io +=20 + try: + from coverage import the_coverage as coverage +@@ -190,7 +190,7 @@ + def _show_branch(root, base, path, pct=3D0, showpct=3DFalse, exclude=3D""): + =20 + # Show the directory name and any of our children +- dirs =3D [k for k, v in root.iteritems() if v] ++ dirs =3D [k for k, v in root.items() if v] + dirs.sort() + for name in dirs: + newpath =3D os.path.join(path, name) +@@ -199,7 +199,7 @@ + relpath =3D newpath[len(base):] + yield "| " * relpath.count(os.sep) + yield "%s\n" % \ +- (newpath, urllib.quote_plus(exclude), name) ++ (newpath, urllib.parse.quote_plus(exclude), name) + =20 + for chunk in _show_branch(root[name], base, newpath, pct, showpct, = exclude): + yield chunk +@@ -207,7 +207,7 @@ + # Now list the files + if path.lower().startswith(base): + relpath =3D path[len(base):] +- files =3D [k for k, v in root.iteritems() if not v] ++ files =3D [k for k, v in root.items() if not v] + files.sort() + for name in files: + newpath =3D os.path.join(path, name) +@@ -257,7 +257,7 @@ + """Return covered module names as a nested dict.""" + tree =3D {} + coverage.get_ready() +- runs =3D coverage.cexecuted.keys() ++ runs =3D list(coverage.cexecuted.keys()) + if runs: + for path in runs: + if not _skip_file(path, exclude) and not os.path.isdir(path): +@@ -287,7 +287,7 @@ + for atom in atoms: + path +=3D atom + os.sep + yield ("%s %s" +- % (path, urllib.quote_plus(exclude), atom, os.sep)) ++ % (path, urllib.parse.quote_plus(exclude), atom, os.sep)) + yield "" + =20 + yield "
" +diff -Naur client175_0.7-original/cherrypy/lib/cptools.py client175_0.7/cher= rypy/lib/cptools.py +--- client175_0.7-original/cherrypy/lib/cptools.py 2010-04-20 13:10:10.00000= 0000 +0200 ++++ client175_0.7/cherrypy/lib/cptools.py 2021-08-03 14:45:22.384298594 +0200 +@@ -236,7 +236,7 @@ + if error_msg: + body =3D self.login_screen(from_page, username, error_msg) + cherrypy.response.body =3D body +- if cherrypy.response.headers.has_key("Content-Length"): ++ if "Content-Length" in cherrypy.response.headers: + # Delete Content-Length header so finalize() recalcs it. + del cherrypy.response.headers["Content-Length"] + return True +@@ -265,7 +265,7 @@ + sess[self.session_key] =3D username =3D self.anonymous() + if not username: + cherrypy.response.body =3D self.login_screen(cherrypy.url(qs=3D= request.query_string)) +- if cherrypy.response.headers.has_key("Content-Length"): ++ if "Content-Length" in cherrypy.response.headers: + # Delete Content-Length header so finalize() recalcs it. + del cherrypy.response.headers["Content-Length"] + return True +@@ -287,7 +287,7 @@ +=20 + def session_auth(**kwargs): + sa =3D SessionAuth() +- for k, v in kwargs.iteritems(): ++ for k, v in kwargs.items(): + setattr(sa, k, v) + return sa.run() + session_auth.__doc__ =3D """Session authentication hook. +@@ -314,7 +314,7 @@ + # Sort by the standard points if possible. + from cherrypy import _cprequest + points =3D _cprequest.hookpoints +- for k in cherrypy.request.hooks.keys(): ++ for k in list(cherrypy.request.hooks.keys()): + if k not in points: + points.append(k) + =20 +@@ -395,7 +395,7 @@ + """ + if not media: + return +- if isinstance(media, basestring): ++ if isinstance(media, str): + media =3D [media] + =20 + # Parse the Accept request header, and try to match one +diff -Naur client175_0.7-original/cherrypy/lib/encoding.py client175_0.7/che= rrypy/lib/encoding.py +--- client175_0.7-original/cherrypy/lib/encoding.py 2010-04-20 13:10:10.0000= 00000 +0200 ++++ client175_0.7/cherrypy/lib/encoding.py 2021-08-03 14:45:29.700190398 +02= 00 +@@ -33,7 +33,7 @@ +=20 + def decode_params(encoding): + decoded_params =3D {} +- for key, value in cherrypy.request.params.items(): ++ for key, value in list(cherrypy.request.params.items()): + if not hasattr(value, 'file'): + # Skip the value if it is an uploaded file + if isinstance(value, list): +@@ -73,7 +73,7 @@ + """ + def encoder(body): + for chunk in body: +- if isinstance(chunk, unicode): ++ if isinstance(chunk, str): + chunk =3D chunk.encode(encoding, errors) + yield chunk + cherrypy.response.body =3D encoder(cherrypy.response.body) +@@ -84,7 +84,7 @@ + try: + body =3D [] + for chunk in cherrypy.response.body: +- if isinstance(chunk, unicode): ++ if isinstance(chunk, str): + chunk =3D chunk.encode(encoding, errors) + body.append(chunk) + cherrypy.response.body =3D body +@@ -101,7 +101,7 @@ + else: + response.collapse_body() + encoder =3D encode_string +- if response.headers.has_key("Content-Length"): ++ if "Content-Length" in response.headers: + # Delete Content-Length header so finalize() recalcs it. + # Encoded strings may be of different lengths from their + # unicode equivalents, and even from each other. For example: +@@ -179,7 +179,7 @@ + yield '\037\213' # magic header + yield '\010' # compression method + yield '\0' +- yield struct.pack("=3D content_length: + # From rfc 2616 sec 14.16: + # "If the server receives a request (other than one +@@ -101,8 +101,8 @@ + self.params =3D params + =20 + def __unicode__(self): +- p =3D [";%s=3D%s" % (k, v) for k, v in self.params.iteritems()] +- return u"%s%s" % (self.value, "".join(p)) ++ p =3D [";%s=3D%s" % (k, v) for k, v in self.params.items()] ++ return "%s%s" % (self.value, "".join(p)) + =20 + def __str__(self): + return str(self.__unicode__()) +@@ -264,14 +264,14 @@ + pm =3D {'x': int(pm[0]), 'y': int(pm[1])} + else: + pm =3D cgi.parse_qs(query_string, keep_blank_values) +- for key, val in pm.items(): ++ for key, val in list(pm.items()): + if len(val) =3D=3D 1: + pm[key] =3D val[0] + return pm +=20 + def params_from_CGI_form(form): + params =3D {} +- for key in form.keys(): ++ for key in list(form.keys()): + value_list =3D form[key] + if isinstance(value_list, list): + params[key] =3D [] +@@ -315,7 +315,7 @@ + return dict.has_key(self, str(key).title()) + =20 + def update(self, E): +- for k in E.keys(): ++ for k in list(E.keys()): + self[str(k).title()] =3D E[k] + =20 + def fromkeys(cls, seq, value=3DNone): +@@ -357,8 +357,8 @@ + def output(self, protocol=3D(1, 1)): + """Transform self into a list of (name, value) tuples.""" + header_list =3D [] +- for key, v in self.iteritems(): +- if isinstance(v, unicode): ++ for key, v in self.items(): ++ if isinstance(v, str): + # HTTP/1.0 says, "Words of *TEXT may contain octets + # from character sets other than US-ASCII." and + # "Recipients of header field TEXT containing octets +diff -Naur client175_0.7-original/cherrypy/lib/__init__.py client175_0.7/che= rrypy/lib/__init__.py +--- client175_0.7-original/cherrypy/lib/__init__.py 2010-04-20 13:10:10.0000= 00000 +0200 ++++ client175_0.7/cherrypy/lib/__init__.py 2021-08-03 14:46:02.407667654 +02= 00 +@@ -18,7 +18,7 @@ + """Load a module and retrieve an attribute of that module.""" + =20 + # Parse out the path, module, and attribute +- last_dot =3D full_attribute_name.rfind(u".") ++ last_dot =3D full_attribute_name.rfind(".") + attr_name =3D full_attribute_name[last_dot + 1:] + mod_path =3D full_attribute_name[:last_dot] + =20 +@@ -52,7 +52,7 @@ + return expr[subs] + =20 + def build_CallFunc(self, o): +- children =3D map(self.build, o.getChildren()) ++ children =3D list(map(self.build, o.getChildren())) + callee =3D children.pop(0) + kwargs =3D children.pop() or {} + starargs =3D children.pop() or () +@@ -60,7 +60,7 @@ + return callee(*args, **kwargs) + =20 + def build_List(self, o): +- return map(self.build, o.getChildren()) ++ return list(map(self.build, o.getChildren())) + =20 + def build_Const(self, o): + return o.value +@@ -69,7 +69,7 @@ + d =3D {} + i =3D iter(map(self.build, o.getChildren())) + for el in i: +- d[el] =3D i.next() ++ d[el] =3D next(i) + return d + =20 + def build_Tuple(self, o): +@@ -91,7 +91,7 @@ + =20 + # See if the Name is in __builtin__. + try: +- import __builtin__ ++ import builtins + return getattr(__builtin__, o.name) + except AttributeError: + pass +@@ -99,7 +99,7 @@ + raise TypeError("unrepr could not resolve the name %s" % repr(o.nam= e)) + =20 + def build_Add(self, o): +- left, right =3D map(self.build, o.getChildren()) ++ left, right =3D list(map(self.build, o.getChildren())) + return left + right + =20 + def build_Getattr(self, o): +diff -Naur client175_0.7-original/cherrypy/lib/profiler.py client175_0.7/che= rrypy/lib/profiler.py +--- client175_0.7-original/cherrypy/lib/profiler.py 2010-04-20 13:10:10.0000= 00000 +0200 ++++ client175_0.7/cherrypy/lib/profiler.py 2021-08-03 14:46:29.643218230 +02= 00 +@@ -61,9 +61,9 @@ + import sys +=20 + try: +- import cStringIO as StringIO ++ import io as StringIO + except ImportError: +- import StringIO ++ import io +=20 +=20 + _count =3D 0 +@@ -94,7 +94,7 @@ + =20 + def stats(self, filename, sortby=3D'cumulative'): + """stats(index) -> output of print_stats() for the given profile.""" +- sio =3D StringIO.StringIO() ++ sio =3D io.StringIO() + if sys.version_info >=3D (2, 5): + s =3D pstats.Stats(os.path.join(self.path, filename), stream=3D= sio) + s.strip_dirs() +diff -Naur client175_0.7-original/cherrypy/lib/safemime.py client175_0.7/che= rrypy/lib/safemime.py +--- client175_0.7-original/cherrypy/lib/safemime.py 2010-04-20 13:10:10.0000= 00000 +0200 ++++ client175_0.7/cherrypy/lib/safemime.py 2021-08-03 14:46:40.883037698 +02= 00 +@@ -95,13 +95,13 @@ + def __iter__(self): + return self.rfile + =20 +- def next(self): ++ def __next__(self): + if self.clen: + # Return '' if we've read all the data. + if self.bytes_read >=3D self.clen: + return '' + =20 +- data =3D self.rfile.next() ++ data =3D next(self.rfile) + self.bytes_read +=3D len(data) + return data +=20 +diff -Naur client175_0.7-original/cherrypy/lib/sessions.py client175_0.7/che= rrypy/lib/sessions.py +--- client175_0.7-original/cherrypy/lib/sessions.py 2010-04-20 13:10:10.0000= 00000 +0200 ++++ client175_0.7/cherrypy/lib/sessions.py 2021-08-03 14:46:48.498916658 +02= 00 +@@ -9,7 +9,7 @@ + import datetime + import os + try: +- import cPickle as pickle ++ import pickle as pickle + except ImportError: + import pickle + import random +@@ -31,11 +31,9 @@ +=20 + missing =3D object() +=20 +-class Session(object): ++class Session(object, metaclass=3Dcherrypy._AttributeDocstrings): + """A CherryPy dict-like Session object (one per request).""" + =20 +- __metaclass__ =3D cherrypy._AttributeDocstrings +- =20 + _id =3D None + id_observers =3D None + id_observers__doc =3D "A list of callbacks to which to pass new id's." +@@ -72,7 +70,7 @@ + self.id_observers =3D [] + self._data =3D {} + =20 +- for k, v in kwargs.iteritems(): ++ for k, v in kwargs.items(): + setattr(self, k, v) + =20 + if id is None: +@@ -192,7 +190,7 @@ + def has_key(self, key): + """D.has_key(k) -> True if D has a key k, else False.""" + if not self.loaded: self.load() +- return self._data.has_key(key) ++ return key in self._data + =20 + def get(self, key, default=3DNone): + """D.get(k[,d]) -> D[k] if k in D, else d. d defaults to None.""" +@@ -217,17 +215,17 @@ + def keys(self): + """D.keys() -> list of D's keys.""" + if not self.loaded: self.load() +- return self._data.keys() ++ return list(self._data.keys()) + =20 + def items(self): + """D.items() -> list of D's (key, value) pairs, as 2-tuples.""" + if not self.loaded: self.load() +- return self._data.items() ++ return list(self._data.items()) + =20 + def values(self): + """D.values() -> list of D's values.""" + if not self.loaded: self.load() +- return self._data.values() ++ return list(self._data.values()) +=20 +=20 + class RamSession(Session): +@@ -239,7 +237,7 @@ + def clean_up(self): + """Clean up expired sessions.""" + now =3D datetime.datetime.now() +- for id, (data, expiration_time) in self.cache.items(): ++ for id, (data, expiration_time) in list(self.cache.items()): + if expiration_time < now: + try: + del self.cache[id] +@@ -302,7 +300,7 @@ + # The 'storage_path' arg is required for file-based sessions. + kwargs['storage_path'] =3D os.path.abspath(kwargs['storage_path']) + =20 +- for k, v in kwargs.iteritems(): ++ for k, v in kwargs.items(): + setattr(cls, k, v) + =20 + # Warn if any lock files exist at startup. +@@ -426,7 +424,7 @@ + This should only be called once per process; this will be done + automatically when using sessions.init (as the built-in Tool does). + """ +- for k, v in kwargs.iteritems(): ++ for k, v in kwargs.items(): + setattr(cls, k, v) + =20 + self.db =3D self.get_db() +@@ -502,7 +500,7 @@ + This should only be called once per process; this will be done + automatically when using sessions.init (as the built-in Tool does). + """ +- for k, v in kwargs.iteritems(): ++ for k, v in kwargs.items(): + setattr(cls, k, v) + =20 + import memcache +diff -Naur client175_0.7-original/cherrypy/lib/static.py client175_0.7/cherr= ypy/lib/static.py +--- client175_0.7-original/cherrypy/lib/static.py 2010-04-20 13:10:10.000000= 000 +0200 ++++ client175_0.7/cherrypy/lib/static.py 2021-08-03 14:46:56.870784689 +0200 +@@ -7,7 +7,7 @@ + import re + import stat + import time +-import urllib ++import urllib.request, urllib.parse, urllib.error +=20 + import cherrypy + from cherrypy.lib import cptools, http, file_generator_limited +@@ -99,7 +99,7 @@ + boundary =3D mimetools.choose_boundary() + ct =3D "multipart/byteranges; boundary=3D%s" % boundary + response.headers['Content-Type'] =3D ct +- if response.headers.has_key("Content-Length"): ++ if "Content-Length" in response.headers: + # Delete Content-Length header so finalize() recalcs it. + del response.headers["Content-Length"] + =20 +@@ -189,7 +189,7 @@ + section =3D "/" + section =3D section.rstrip(r"\/") + branch =3D cherrypy.request.path_info[len(section) + 1:] +- branch =3D urllib.unquote(branch.lstrip(r"\/")) ++ branch =3D urllib.parse.unquote(branch.lstrip(r"\/")) + =20 + # If branch is "", filename will end in a slash + filename =3D os.path.join(dir, branch) +diff -Naur client175_0.7-original/cherrypy/lib/tidy.py client175_0.7/cherryp= y/lib/tidy.py +--- client175_0.7-original/cherrypy/lib/tidy.py 2010-04-20 13:10:10.00000000= 0 +0200 ++++ client175_0.7/cherrypy/lib/tidy.py 2021-08-03 14:47:04.222669487 +0200 +@@ -2,7 +2,7 @@ +=20 + import cgi + import os +-import StringIO ++import io + import traceback +=20 + import cherrypy +@@ -79,7 +79,7 @@ + =20 + if new_errs: + response.body =3D wrong_content('
'.join(new_errs), orig_b= ody) +- if response.headers.has_key("Content-Length"): ++ if "Content-Length" in response.headers: + # Delete Content-Length header so finalize() recalcs it. + del response.headers["Content-Length"] + return +@@ -95,23 +95,23 @@ + enctag =3D '' % enco= ding + orig_body =3D enctag + orig_body + =20 +- f =3D StringIO.StringIO(orig_body) ++ f =3D io.StringIO(orig_body) + try: + tree =3D parse(f) + except: + # Wrong XML +- body_file =3D StringIO.StringIO() ++ body_file =3D io.StringIO() + traceback.print_exc(file =3D body_file) + body_file =3D '
'.join(body_file.getvalue()) + response.body =3D wrong_content(body_file, orig_body, "XML") +- if response.headers.has_key("Content-Length"): ++ if "Content-Length" in response.headers: + # Delete Content-Length header so finalize() recalcs it. + del response.headers["Content-Length"] + return + =20 + if use_output: + response.body =3D [output] +- if response.headers.has_key("Content-Length"): ++ if "Content-Length" in response.headers: + # Delete Content-Length header so finalize() recalcs it. + del response.headers["Content-Length"] +=20 +@@ -178,7 +178,7 @@ + =20 + if new_errs: + response.body =3D wrong_content('
'.join(new_errs), orig_b= ody) +- if response.headers.has_key("Content-Length"): ++ if "Content-Length" in response.headers: + # Delete Content-Length header so finalize() recalcs it. + del response.headers["Content-Length"] +=20 +diff -Naur client175_0.7-original/cherrypy/lib/wsgiapp.py client175_0.7/cher= rypy/lib/wsgiapp.py +--- client175_0.7-original/cherrypy/lib/wsgiapp.py 2010-04-20 13:10:10.00000= 0000 +0200 ++++ client175_0.7/cherrypy/lib/wsgiapp.py 2021-08-03 14:47:15.738490577 +0200 +@@ -43,7 +43,7 @@ + headers =3D request.headers + environ["CONTENT_TYPE"] =3D headers.get("Content-type", "") + environ["CONTENT_LENGTH"] =3D headers.get("Content-length", "") +- for (k, v) in headers.iteritems(): ++ for (k, v) in headers.items(): + envname =3D "HTTP_" + k.upper().replace("-","_") + environ[envname] =3D v + return environ +diff -Naur client175_0.7-original/cherrypy/lib/xmlrpc.py client175_0.7/cherr= ypy/lib/xmlrpc.py +--- client175_0.7-original/cherrypy/lib/xmlrpc.py 2010-04-20 13:10:10.000000= 000 +0200 ++++ client175_0.7/cherrypy/lib/xmlrpc.py 2021-08-03 14:47:22.378387654 +0200 +@@ -6,8 +6,8 @@ + def process_body(): + """Return (params, method) from request body.""" + try: +- import xmlrpclib +- return xmlrpclib.loads(cherrypy.request.body.read()) ++ import xmlrpc.client ++ return xmlrpc.client.loads(cherrypy.request.body.read()) + except Exception: + return ('ERROR PARAMS', ), 'ERRORMETHOD' +=20 +@@ -35,15 +35,15 @@ +=20 +=20 + def respond(body, encoding=3D'utf-8', allow_none=3D0): +- import xmlrpclib +- if not isinstance(body, xmlrpclib.Fault): ++ import xmlrpc.client ++ if not isinstance(body, xmlrpc.client.Fault): + body =3D (body,) +- _set_response(xmlrpclib.dumps(body, methodresponse=3D1, ++ _set_response(xmlrpc.client.dumps(body, methodresponse=3D1, + encoding=3Dencoding, + allow_none=3Dallow_none)) +=20 + def on_error(*args, **kwargs): + body =3D str(sys.exc_info()[1]) +- import xmlrpclib +- _set_response(xmlrpclib.dumps(xmlrpclib.Fault(1, body))) ++ import xmlrpc.client ++ _set_response(xmlrpc.client.dumps(xmlrpc.client.Fault(1, body))) +=20 +diff -Naur client175_0.7-original/cherrypy/process/plugins.py client175_0.7/= cherrypy/process/plugins.py +--- client175_0.7-original/cherrypy/process/plugins.py 2010-04-20 13:10:10.0= 00000000 +0200 ++++ client175_0.7/cherrypy/process/plugins.py 2021-08-03 14:48:07.409700439 = +0200 +@@ -49,7 +49,7 @@ + =20 + # Map from signal numbers to names + signals =3D {} +- for k, v in vars(_signal).items(): ++ for k, v in list(vars(_signal).items()): + if k.startswith('SIG') and not k.startswith('SIG_'): + signals[v] =3D k + del k, v +@@ -65,14 +65,14 @@ + self._previous_handlers =3D {} + =20 + def subscribe(self): +- for sig, func in self.handlers.iteritems(): ++ for sig, func in self.handlers.items(): + try: + self.set_handler(sig, func) + except ValueError: + pass + =20 + def unsubscribe(self): +- for signum, handler in self._previous_handlers.iteritems(): ++ for signum, handler in self._previous_handlers.items(): + signame =3D self.signals[signum] + =20 + if handler is None: +@@ -100,7 +100,7 @@ + If the given signal name or number is not available on the current + platform, ValueError is raised. + """ +- if isinstance(signal, basestring): ++ if isinstance(signal, str): + signum =3D getattr(_signal, signal, None) + if signum is None: + raise ValueError("No such signal: %r" % signal) +@@ -162,7 +162,7 @@ + self.bus.log("pwd module not available; ignoring uid.", + level=3D30) + val =3D None +- elif isinstance(val, basestring): ++ elif isinstance(val, str): + val =3D pwd.getpwnam(val)[2] + self._uid =3D val + uid =3D property(_get_uid, _set_uid, doc=3D"The uid under which to run.= ") +@@ -175,7 +175,7 @@ + self.bus.log("grp module not available; ignoring gid.", + level=3D30) + val =3D None +- elif isinstance(val, basestring): ++ elif isinstance(val, str): + val =3D grp.getgrnam(val)[2] + self._gid =3D val + gid =3D property(_get_gid, _set_gid, doc=3D"The gid under which to run.= ") +@@ -296,7 +296,7 @@ + # This is the first parent. Exit, now that we've forked. + self.bus.log('Forking once.') + os._exit(0) +- except OSError, exc: ++ except OSError as exc: + # Python raises OSError rather than returning negative numbers. + sys.exit("%s: fork #1 failed: (%d) %s\n" + % (sys.argv[0], exc.errno, exc.strerror)) +@@ -309,7 +309,7 @@ + if pid > 0: + self.bus.log('Forking twice.') + os._exit(0) # Exit second parent +- except OSError, exc: ++ except OSError as exc: + sys.exit("%s: fork #2 failed: (%d) %s\n" + % (sys.argv[0], exc.errno, exc.strerror)) + =20 +@@ -440,7 +440,7 @@ + def run(self): + """Reload the process if registered files have been modified.""" + sysfiles =3D set() +- for k, m in sys.modules.items(): ++ for k, m in list(sys.modules.items()): + if re.match(self.match, k): + if hasattr(m, '__loader__'): + if hasattr(m.__loader__, 'archive'): +@@ -522,7 +522,7 @@ + =20 + def stop(self): + """Release all threads and run all 'stop_thread' listeners.""" +- for thread_ident, i in self.threads.iteritems(): ++ for thread_ident, i in self.threads.items(): + self.bus.publish('stop_thread', i) + self.threads.clear() + graceful =3D stop +diff -Naur client175_0.7-original/cherrypy/process/servers.py client175_0.7/= cherrypy/process/servers.py +--- client175_0.7-original/cherrypy/process/servers.py 2010-04-20 13:10:10.0= 00000000 +0200 ++++ client175_0.7/cherrypy/process/servers.py 2021-08-03 14:48:16.165568167 = +0200 +@@ -71,11 +71,11 @@ + """ + try: + self.httpserver.start() +- except KeyboardInterrupt, exc: ++ except KeyboardInterrupt as exc: + self.bus.log(" hit: shutting down HTTP server") + self.interrupt =3D exc + self.bus.exit() +- except SystemExit, exc: ++ except SystemExit as exc: + self.bus.log("SystemExit raised: shutting down HTTP server") + self.interrupt =3D exc + self.bus.exit() +@@ -238,7 +238,7 @@ + if not host: + raise ValueError("Host values of '' or None are not allowed.") + =20 +- for trial in xrange(50): ++ for trial in range(50): + try: + # we are expecting a free port, so reduce the timeout + check_port(host, port, timeout=3D0.1) +@@ -255,7 +255,7 @@ + if not host: + raise ValueError("Host values of '' or None are not allowed.") + =20 +- for trial in xrange(50): ++ for trial in range(50): + try: + check_port(host, port) + except IOError: +diff -Naur client175_0.7-original/cherrypy/process/win32.py client175_0.7/ch= errypy/process/win32.py +--- client175_0.7-original/cherrypy/process/win32.py 2010-04-20 13:10:10.000= 000000 +0200 ++++ client175_0.7/cherrypy/process/win32.py 2021-08-03 14:48:22.801468164 +0= 200 +@@ -1,7 +1,7 @@ + """Windows service. Requires pywin32.""" +=20 + import os +-import thread ++import _thread + import win32api + import win32con + import win32event +@@ -84,7 +84,7 @@ + return self.events[state] + except KeyError: + event =3D win32event.CreateEvent(None, 0, 0, +- u"WSPBus %s Event (pid=3D%r)" % ++ "WSPBus %s Event (pid=3D%r)" % + (state.name, os.getpid())) + self.events[state] =3D event + return event +@@ -128,7 +128,7 @@ + =20 + def key_for(self, obj): + """For the given value, return its corresponding key.""" +- for key, val in self.iteritems(): ++ for key, val in self.items(): + if val is obj: + return key + raise ValueError("The given object could not be found: %r" % obj) +diff -Naur client175_0.7-original/cherrypy/process/wspbus.py client175_0.7/c= herrypy/process/wspbus.py +--- client175_0.7-original/cherrypy/process/wspbus.py 2010-04-20 13:10:10.00= 0000000 +0200 ++++ client175_0.7/cherrypy/process/wspbus.py 2021-08-03 14:48:29.181372091 += 0200 +@@ -147,7 +147,7 @@ + output.append(listener(*args, **kwargs)) + except KeyboardInterrupt: + raise +- except SystemExit, e: ++ except SystemExit as e: + # If we have previous errors ensure the exit code is non-ze= ro + if exc and e.code =3D=3D 0: + e.code =3D 1 +@@ -195,7 +195,7 @@ + except: + # Any stop/exit errors will be logged inside publish(). + pass +- raise e_info[0], e_info[1], e_info[2] ++ raise e_info[0](e_info[1]).with_traceback(e_info[2]) + =20 + def exit(self): + """Stop all services and prepare to exit the process.""" +diff -Naur client175_0.7-original/cherrypy/wsgiserver/__init__.py client175_= 0.7/cherrypy/wsgiserver/__init__.py +--- client175_0.7-original/cherrypy/wsgiserver/__init__.py 2010-04-20 13:10:= 10.000000000 +0200 ++++ client175_0.7/cherrypy/wsgiserver/__init__.py 2021-08-03 14:49:05.556826= 062 +0200 +@@ -79,24 +79,24 @@ +=20 + import base64 + import os +-import Queue ++import queue + import re + quoted_slash =3D re.compile("(?i)%2F") + import rfc822 + import socket + try: +- import cStringIO as StringIO ++ import io as StringIO + except ImportError: +- import StringIO ++ import io +=20 +-_fileobject_uses_str_type =3D isinstance(socket._fileobject(None)._rbuf, ba= sestring) ++_fileobject_uses_str_type =3D isinstance(socket._fileobject(None)._rbuf, st= r) +=20 + import sys + import threading + import time + import traceback +-from urllib import unquote +-from urlparse import urlparse ++from urllib.parse import unquote ++from urllib.parse import urlparse + import warnings +=20 + try: +@@ -117,7 +117,7 @@ + errno_names =3D dir(errno) + nums =3D [getattr(errno, k) for k in errnames if k in errno_names] + # de-dupe the list +- return dict.fromkeys(nums).keys() ++ return list(dict.fromkeys(nums).keys()) +=20 + socket_error_eintr =3D plat_specific_errors("EINTR", "WSAEINTR") +=20 +@@ -153,7 +153,7 @@ + =20 + def __init__(self, apps): + try: +- apps =3D apps.items() ++ apps =3D list(apps.items()) + except AttributeError: + pass + =20 +@@ -239,8 +239,8 @@ + def __iter__(self): + return self + =20 +- def next(self): +- data =3D self.rfile.next() ++ def __next__(self): ++ data =3D next(self.rfile) + self.bytes_read +=3D len(data) + self._check_length() + return data +@@ -401,7 +401,7 @@ + # then all the http headers + try: + self.read_headers() +- except ValueError, ex: ++ except ValueError as ex: + self.simple_response("400 Bad Request", repr(ex.args)) + return + =20 +@@ -500,7 +500,7 @@ + def decode_chunked(self): + """Decode the 'chunked' transfer coding.""" + cl =3D 0 +- data =3D StringIO.StringIO() ++ data =3D io.StringIO() + while True: + line =3D self.rfile.readline().strip().split(";", 1) + chunk_size =3D int(line.pop(0), 16) +@@ -592,7 +592,7 @@ + =20 + try: + self.wfile.sendall("".join(buf)) +- except socket.error, x: ++ except socket.error as x: + if x.args[0] not in socket_errors_to_ignore: + raise + =20 +@@ -609,7 +609,7 @@ + # exc_info tuple." + if self.sent_headers: + try: +- raise exc_info[0], exc_info[1], exc_info[2] ++ raise exc_info[0](exc_info[1]).with_traceback(exc_info[2]) + finally: + exc_info =3D None + =20 +@@ -728,7 +728,7 @@ + try: + bytes_sent =3D self.send(data) + data =3D data[bytes_sent:] +- except socket.error, e: ++ except socket.error as e: + if e.args[0] not in socket_errors_nonblocking: + raise +=20 +@@ -745,7 +745,7 @@ + while True: + try: + return self._sock.recv(size) +- except socket.error, e: ++ except socket.error as e: + if (e.args[0] not in socket_errors_nonblocking + and e.args[0] not in socket_error_eintr): + raise +@@ -762,7 +762,7 @@ + buf.seek(0, 2) # seek end + if size < 0: + # Read until EOF +- self._rbuf =3D StringIO.StringIO() # reset _rbuf. we cons= ume it via buf. ++ self._rbuf =3D io.StringIO() # reset _rbuf. we consume it= via buf. + while True: + data =3D self.recv(rbufsize) + if not data: +@@ -776,11 +776,11 @@ + # Already have size bytes in our buffer? Extract and r= eturn. + buf.seek(0) + rv =3D buf.read(size) +- self._rbuf =3D StringIO.StringIO() ++ self._rbuf =3D io.StringIO() + self._rbuf.write(buf.read()) + return rv +=20 +- self._rbuf =3D StringIO.StringIO() # reset _rbuf. we cons= ume it via buf. ++ self._rbuf =3D io.StringIO() # reset _rbuf. we consume it= via buf. + while True: + left =3D size - buf_len + # recv() will malloc the amount of memory given as its +@@ -818,7 +818,7 @@ + buf.seek(0) + bline =3D buf.readline(size) + if bline.endswith('\n') or len(bline) =3D=3D size: +- self._rbuf =3D StringIO.StringIO() ++ self._rbuf =3D io.StringIO() + self._rbuf.write(buf.read()) + return bline + del bline +@@ -828,7 +828,7 @@ + # Speed up unbuffered case + buf.seek(0) + buffers =3D [buf.read()] +- self._rbuf =3D StringIO.StringIO() # reset _rbuf. we = consume it via buf. ++ self._rbuf =3D io.StringIO() # reset _rbuf. we consum= e it via buf. + data =3D None + recv =3D self.recv + while data !=3D "\n": +@@ -839,7 +839,7 @@ + return "".join(buffers) +=20 + buf.seek(0, 2) # seek end +- self._rbuf =3D StringIO.StringIO() # reset _rbuf. we cons= ume it via buf. ++ self._rbuf =3D io.StringIO() # reset _rbuf. we consume it= via buf. + while True: + data =3D self.recv(self._rbufsize) + if not data: +@@ -860,10 +860,10 @@ + if buf_len >=3D size: + buf.seek(0) + rv =3D buf.read(size) +- self._rbuf =3D StringIO.StringIO() ++ self._rbuf =3D io.StringIO() + self._rbuf.write(buf.read()) + return rv +- self._rbuf =3D StringIO.StringIO() # reset _rbuf. we cons= ume it via buf. ++ self._rbuf =3D io.StringIO() # reset _rbuf. we consume it= via buf. + while True: + data =3D self.recv(self._rbufsize) + if not data: +@@ -906,7 +906,7 @@ + try: + bytes_sent =3D self.send(data) + data =3D data[bytes_sent:] +- except socket.error, e: ++ except socket.error as e: + if e.args[0] not in socket_errors_nonblocking: + raise +=20 +@@ -923,7 +923,7 @@ + while True: + try: + return self._sock.recv(size) +- except socket.error, e: ++ except socket.error as e: + if (e.args[0] not in socket_errors_nonblocking + and e.args[0] not in socket_error_eintr): + raise +@@ -1065,7 +1065,7 @@ + time.sleep(self.ssl_retry) + except SSL.WantWriteError: + time.sleep(self.ssl_retry) +- except SSL.SysCallError, e: ++ except SSL.SysCallError as e: + if is_reader and e.args =3D=3D (-1, 'Unexpected EOF'): + return "" + =20 +@@ -1073,7 +1073,7 @@ + if is_reader and errnum in socket_errors_to_ignore: + return "" + raise socket.error(errnum) +- except SSL.Error, e: ++ except SSL.Error as e: + if is_reader and e.args =3D=3D (-1, 'Unexpected EOF'): + return "" + =20 +@@ -1175,7 +1175,7 @@ + if req.close_connection: + return + =20 +- except socket.error, e: ++ except socket.error as e: + errnum =3D e.args[0] + if errnum =3D=3D 'timed out': + if req and not req.sent_headers: +@@ -1187,7 +1187,7 @@ + return + except (KeyboardInterrupt, SystemExit): + raise +- except FatalSSLAlert, e: ++ except FatalSSLAlert as e: + # Close the connection. + return + except NoSSLError: +@@ -1198,7 +1198,7 @@ + "The client sent a plain HTTP request, but " + "this server only speaks HTTPS on this port.") + self.linger =3D True +- except Exception, e: ++ except Exception as e: + if req and not req.sent_headers: + req.simple_response("500 Internal Server Error", format_exc= ()) + =20 +@@ -1272,7 +1272,7 @@ + finally: + conn.close() + self.conn =3D None +- except (KeyboardInterrupt, SystemExit), exc: ++ except (KeyboardInterrupt, SystemExit) as exc: + self.server.interrupt =3D exc +=20 +=20 +@@ -1288,12 +1288,12 @@ + self.min =3D min + self.max =3D max + self._threads =3D [] +- self._queue =3D Queue.Queue() ++ self._queue =3D queue.Queue() + self.get =3D self._queue.get + =20 + def start(self): + """Start the pool of threads.""" +- for i in xrange(self.min): ++ for i in range(self.min): + self._threads.append(WorkerThread(self.server)) + for worker in self._threads: + worker.setName("CP WSGIServer " + worker.getName()) +@@ -1314,7 +1314,7 @@ + =20 + def grow(self, amount): + """Spawn new worker threads (not above self.max).""" +- for i in xrange(amount): ++ for i in range(amount): + if self.max > 0 and len(self._threads) >=3D self.max: + break + worker =3D WorkerThread(self.server) +@@ -1332,7 +1332,7 @@ + amount -=3D 1 + =20 + if amount > 0: +- for i in xrange(min(amount, len(self._threads) - self.min)): ++ for i in range(min(amount, len(self._threads) - self.min)): + # Put a number of shutdown requests on the queue equal + # to 'amount'. Once each of those is processed by a worker, + # that worker will terminate and be culled from our list +@@ -1369,7 +1369,7 @@ + except (AssertionError, + # Ignore repeated Ctrl-C. + # See http://www.cherrypy.org/ticket/691. +- KeyboardInterrupt), exc1: ++ KeyboardInterrupt) as exc1: + pass +=20 +=20 +@@ -1392,13 +1392,13 @@ + 'sock_shutdown', 'get_peer_certificate', 'want_read', + 'want_write', 'set_connect_state', 'set_accept_state', + 'connect_ex', 'sendall', 'settimeout'): +- exec """def %s(self, *args): ++ exec("""def %s(self, *args): + self._lock.acquire() + try: + return self._ssl_conn.%s(*args) + finally: + self._lock.release() +-""" % (f, f) ++""" % (f, f)) +=20 +=20 + try: +@@ -1557,7 +1557,7 @@ + self._interrupt =3D None + =20 + # Select the appropriate socket +- if isinstance(self.bind_addr, basestring): ++ if isinstance(self.bind_addr, str): + # AF_UNIX socket + =20 + # So we can reuse the socket... +@@ -1565,7 +1565,7 @@ + except: pass + =20 + # So everyone can access the socket... +- try: os.chmod(self.bind_addr, 0777) ++ try: os.chmod(self.bind_addr, 0o777) + except: pass + =20 + info =3D [(socket.AF_UNIX, socket.SOCK_STREAM, 0, "", self.bind= _addr)] +@@ -1586,14 +1586,14 @@ + af, socktype, proto, canonname, sa =3D res + try: + self.bind(af, socktype, proto) +- except socket.error, msg: ++ except socket.error as msg: + if self.socket: + self.socket.close() + self.socket =3D None + continue + break + if not self.socket: +- raise socket.error, msg ++ raise socket.error(msg) + =20 + # Timeout so KeyboardInterrupt can be caught on Win32 + self.socket.settimeout(1) +@@ -1632,7 +1632,7 @@ + =20 + # If listening on the IPV6 any address ('::' =3D IN6ADDR_ANY), + # activate dual-stack. See http://www.cherrypy.org/ticket/871. +- if (not isinstance(self.bind_addr, basestring) ++ if (not isinstance(self.bind_addr, str) + and self.bind_addr[0] =3D=3D '::' and family =3D=3D socket.= AF_INET6): + try: + self.socket.setsockopt(socket.IPPROTO_IPV6, socket.IPV6= _V6ONLY, 0) +@@ -1664,7 +1664,7 @@ + environ["ACTUAL_SERVER_PROTOCOL"] =3D self.protocol + environ["SERVER_NAME"] =3D self.server_name + =20 +- if isinstance(self.bind_addr, basestring): ++ if isinstance(self.bind_addr, str): + # AF_UNIX. This isn't really allowed by WSGI, which doesn't + # address unix domain sockets. But it's better than nothing. + environ["SERVER_PORT"] =3D "" +@@ -1682,7 +1682,7 @@ + # notice keyboard interrupts on Win32, which don't interrupt + # accept() by default + return +- except socket.error, x: ++ except socket.error as x: + if x.args[0] in socket_error_eintr: + # I *think* this is right. EINTR should occur when a signal + # is received during the accept() call; all docs say retry +@@ -1715,11 +1715,11 @@ + =20 + sock =3D getattr(self, "socket", None) + if sock: +- if not isinstance(self.bind_addr, basestring): ++ if not isinstance(self.bind_addr, str): + # Touch our own socket to make accept() return immediately. + try: + host, port =3D sock.getsockname()[:2] +- except socket.error, x: ++ except socket.error as x: + if x.args[0] not in socket_errors_to_ignore: + raise + else: +diff -Naur client175_0.7-original/covers.py client175_0.7/covers.py +--- client175_0.7-original/covers.py 2011-04-02 03:51:24.000000000 +0200 ++++ client175_0.7/covers.py 2021-08-03 14:39:46.713265947 +0200 +@@ -23,7 +23,7 @@ + # Exaile (http://www.exaile.org/). +=20 +=20 +-import hashlib, re, urllib, os, time, shutil, threading ++import hashlib, re, urllib.request, urllib.parse, urllib.error, os, time, s= hutil, threading + from xml.etree import ElementTree as ET + from datetime import datetime, timedelta +=20 +@@ -76,7 +76,7 @@ +=20 + def _findMusicBrainz(self, vars): + self._delay('last_MB_lookup') +- data =3D urllib.urlopen(self.urlMB % vars).read() ++ data =3D urllib.request.urlopen(self.urlMB % vars).read() + m =3D self.regexMB.search(data) + if not m: + return False +@@ -85,7 +85,7 @@ + url =3D "http://images.amazon.com/images/P/%s.01.%sZZZZZZZ.jpg" + for sz in ['L', 'M']: + image =3D url % (asin, sz) +- h =3D urllib.urlopen(image) ++ h =3D urllib.request.urlopen(image) + data =3D h.read() + h.close() + if len(data) > 1000: +@@ -95,21 +95,21 @@ +=20 + def _findLastFM_album(self, vars): + self._delay('last_FM_lookup') +- data =3D urllib.urlopen(self.urlFM % vars).read() ++ data =3D urllib.request.urlopen(self.urlFM % vars).read() + x =3D ET.XML(data) + if len(x) =3D=3D 0: +- print 'LASTFM SEARCH: ERROR PARSING LASTFM DATA!' ++ print('LASTFM SEARCH: ERROR PARSING LASTFM DATA!') + return False +=20 + c =3D x.find('coverart') + if len(c) =3D=3D 0: +- print 'LASTFM SEARCH: NO COVERART NODE IN LASTFM DATA!' ++ print('LASTFM SEARCH: NO COVERART NODE IN LASTFM DATA!') + return False +=20 + for sz in ['large', 'medium', 'small']: + image =3D c.findtext(sz, '') + if image > '' and not image.lower().endswith('.gif'): +- h =3D urllib.urlopen(image) ++ h =3D urllib.request.urlopen(image) + data =3D h.read() + h.close() + if hashlib.sha1(data).hexdigest() !=3D "57b2c37343f711c94e8= 3a37bd91bc4d18d2ed9d5": +@@ -120,13 +120,13 @@ +=20 + def _findLastFM_artist(self, vars): + self._delay('last_FM_lookup') +- data =3D urllib.urlopen(self.urlFM_artist % vars['artist']).read() ++ data =3D urllib.request.urlopen(self.urlFM_artist % vars['artist'])= .read() + m =3D self.regexFM_artist.search(data) + if m: + image =3D m.group(1) + if image.lower().endswith('.gif'): + return False +- h =3D urllib.urlopen(image) ++ h =3D urllib.request.urlopen(image) + data =3D h.read() + h.close() + if hashlib.sha1(data).hexdigest() !=3D "57b2c37343f711c94e83a37= bd91bc4d18d2ed9d5": +@@ -147,8 +147,8 @@ + try: + shutil.copy2(coverpath, cover_destination) + except IOError: +- print "Could not save cover to: " + cover_destinati= on +- print "For best performance, please ensure that the= directory exists and is writable." ++ print("Could not save cover to: " + cover_destinati= on) ++ print("For best performance, please ensure that the= directory exists and is writable.") + h =3D open(coverpath, 'r') + data =3D h.read() + h.close() +@@ -175,8 +175,8 @@ + return covername, None +=20 + vars =3D { +- 'album': urllib.quote_plus(album.encode("utf-8")), +- 'artist': urllib.quote_plus(artist.encode("utf-8")) ++ 'album': urllib.parse.quote_plus(album.encode("utf-8")), ++ 'artist': urllib.parse.quote_plus(artist.encode("utf-8")) + } +=20 + for fn in lookups: +@@ -188,8 +188,8 @@ + h.write(data) + h.close() + except: +- print "Could not save cover to: " + coverpath +- print "For best performance, please ensure that the= directory exists and is writable." ++ print("Could not save cover to: " + coverpath) ++ print("For best performance, please ensure that the= directory exists and is writable.") + covername =3D "" + return covername, data + except: +diff -Naur client175_0.7-original/lyricwiki.py client175_0.7/lyricwiki.py +--- client175_0.7-original/lyricwiki.py 2010-11-20 18:43:24.000000000 +0100 ++++ client175_0.7/lyricwiki.py 2021-08-03 14:40:00.301064572 +0200 +@@ -18,7 +18,7 @@ + # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, + # MA 02110-1301, USA. +=20 +-import json, urllib, os, hashlib, time ++import json, urllib.request, urllib.parse, urllib.error, os, hashlib, time +=20 + def _download(args): + """ +@@ -31,9 +31,9 @@ + for key in args: + str_args[key] =3D args[key].encode("utf-8") + =20 +- args =3D urllib.urlencode(str_args) ++ args =3D urllib.parse.urlencode(str_args) + =20 +- return urllib.urlopen(base + args).read() ++ return urllib.request.urlopen(base + args).read() +=20 + def _get_page_titles(artist, title): + """ +diff -Naur client175_0.7-original/metadata/_base.py client175_0.7/metadata/_= base.py +--- client175_0.7-original/metadata/_base.py 2010-05-01 21:15:15.000000000 += 0200 ++++ client175_0.7/metadata/_base.py 2021-08-03 14:51:24.062763658 +0200 +@@ -66,7 +66,7 @@ + self.mutagen =3D None + self.load() + self._reverse_mapping =3D dict(( +- (v,k) for k,v in self.tag_mapping.iteritems() )) ++ (v,k) for k,v in self.tag_mapping.items() )) +=20 + def load(self): + """ +@@ -99,7 +99,7 @@ +=20 + def _get_keys(self): + keys =3D [] +- for k in self._get_raw().keys(): ++ for k in list(self._get_raw().keys()): + if k in self._reverse_mapping: + keys.append(self._reverse_mapping[k]) + else: +@@ -149,11 +149,11 @@ + if t =3D=3D None and tag in self.tag_mapping: + try: + t =3D self._get_tag(raw, self.tag_mapping[tag]) +- if type(t) in [str, unicode]: ++ if type(t) in [str, str]: + t =3D [t] + else: + try: +- t =3D [unicode(u) for u in list(t)] ++ t =3D [str(u) for u in list(t)] + except UnicodeDecodeError: + t =3D t + except (KeyError, TypeError): +@@ -161,10 +161,10 @@ + if t =3D=3D None and self.others: + try: + t =3D self._get_tag(raw, tag) +- if type(t) in [str, unicode]: ++ if type(t) in [str, str]: + t =3D [t] + else: +- t =3D [unicode(u) for u in list(t)] ++ t =3D [str(u) for u in list(t)] + except (KeyError, TypeError): + pass +=20 +@@ -207,7 +207,7 @@ + pass +=20 + # tags starting with __ are internal and should not be written +- for tag in tagdict.keys(): ++ for tag in list(tagdict.keys()): + if tag.startswith("__"): + try: + del tagdict[tag] +diff -Naur client175_0.7-original/metadata/_id3.py client175_0.7/metadata/_i= d3.py +--- client175_0.7-original/metadata/_id3.py 2010-05-04 23:50:41.000000000 +0= 200 ++++ client175_0.7/metadata/_id3.py 2021-08-03 14:51:42.866483930 +0200 +@@ -70,7 +70,7 @@ +=20 + def _get_tag(self, raw, t): + if not raw.tags: return [] +- if t not in self.tag_mapping.itervalues(): ++ if t not in iter(self.tag_mapping.values()): + t =3D "TXXX:" + t + field =3D raw.tags.getall(t) + if len(field) <=3D 0: +@@ -78,27 +78,27 @@ + ret =3D [] + if t =3D=3D 'TDRC' or t =3D=3D 'TDOR': # values are ID3TimeStamps + for value in field: +- ret.extend([unicode(x) for x in value.text]) ++ ret.extend([str(x) for x in value.text]) + elif t =3D=3D 'USLT': # Lyrics are stored in plain old strings + for value in field: +- ret.append(unicode(value.text)) ++ ret.append(str(value.text)) + elif t =3D=3D 'WOAR': # URLS are stored in url not text + for value in field: +- ret.extend([unicode(x.replace('\n','').replace('\r','')) \ ++ ret.extend([str(x.replace('\n','').replace('\r','')) \ + for x in value.url]) + elif t =3D=3D 'APIC': + ret =3D [x.data for x in field] + else: + for value in field: + try: +- ret.extend([unicode(x.replace('\n','').replace('\r','')= ) \ ++ ret.extend([str(x.replace('\n','').replace('\r','')) \ + for x in value.text]) + except: + pass + return ret +=20 + def _set_tag(self, raw, tag, data): +- if tag not in self.tag_mapping.itervalues(): ++ if tag not in iter(self.tag_mapping.values()): + tag =3D "TXXX:" + tag + if raw.tags is not None: + raw.tags.delall(tag) +@@ -107,7 +107,7 @@ + raw.tags.add(frame) +=20 + def _del_tag(self, raw, tag): +- if tag not in self.tag_mapping.itervalues(): ++ if tag not in iter(self.tag_mapping.values()): + tag =3D "TXXX:" + tag + if raw.tags is not None: + raw.tags.delall(tag) +diff -Naur client175_0.7-original/metadata/__init__.py client175_0.7/metadat= a/__init__.py +--- client175_0.7-original/metadata/__init__.py 2010-05-01 21:15:15.00000000= 0 +0200 ++++ client175_0.7/metadata/__init__.py 2021-08-03 14:50:50.391263894 +0200 +@@ -30,7 +30,7 @@ + import os +=20 + from metadata._base import BaseFormat, NotWritable, NotReadable +-import urlparse ++import urllib.parse +=20 + from metadata import (ape, asf, flac, mod, mp3, mp4, mpc, ogg, sid, speex, + tta, wav, wv) +diff -Naur client175_0.7-original/metadata/mp4.py client175_0.7/metadata/mp4= .py +--- client175_0.7-original/metadata/mp4.py 2010-05-01 21:15:15.000000000 +02= 00 ++++ client175_0.7/metadata/mp4.py 2021-08-03 14:52:08.706100110 +0200 +@@ -45,7 +45,7 @@ + writable =3D True +=20 + def _get_tag(self, f, name): +- if not f.has_key(name): ++ if name not in f: + return [] + elif name in ['trkn', 'disk']: + ret =3D [] +@@ -60,7 +60,7 @@ + try: + f[name] =3D [] + for val in value: +- tmp =3D map(int, val.split('/')) ++ tmp =3D list(map(int, val.split('/'))) + f[name].append(tuple(tmp)) + except TypeError: + pass +diff -Naur client175_0.7-original/mpd.py client175_0.7/mpd.py +--- client175_0.7-original/mpd.py 2010-08-27 00:38:39.000000000 +0200 ++++ client175_0.7/mpd.py 2021-08-03 14:40:15.808834774 +0200 +@@ -64,7 +64,7 @@ + =20 + =20 + def extend_database(item): +- keys =3D item.keys() ++ keys =3D list(item.keys()) + if 'file' in keys: + item =3D extend_file(item) + elif 'directory' in keys: +@@ -192,8 +192,8 @@ + self.lock.acquire() + try: + return self._execute(command, args, retval) +- except (ConnectionError, socket.error), e: +- print "%s\n reconnecting..." % e ++ except (ConnectionError, socket.error) as e: ++ print("%s\n reconnecting..." % e) + try: + self.disconnect() + except: +@@ -343,13 +343,13 @@ + raise StopIteration +=20 + def _fetch_songs(self): +- return map(extend_file, self._read_songs()) ++ return list(map(extend_file, self._read_songs())) +=20 + def _fetch_playlists(self): + return self._fetch_objects(["playlist"]) +=20 + def _fetch_database(self): +- return map(extend_database, self._read_objects(["file", "directory"= , "playlist"])) ++ return list(map(extend_database, self._read_objects(["file", "direc= tory", "playlist"]))) +=20 + def _fetch_outputs(self): + return self._fetch_objects(["outputid"]) +@@ -397,7 +397,7 @@ + try: + sock =3D socket.socket(af, socktype, proto) + sock.connect(sa) +- except socket.error, msg: ++ except socket.error as msg: + if sock: + sock.close() + sock =3D None +@@ -425,8 +425,8 @@ + self.password(_password) + self._TAGS =3D self.tagtypes() + self._TAGS.extend(['Pos', 'Time', 'Id']) +- self._TAGS_LOWER =3D map(str.lower, self._TAGS) +- self._TAGMAP =3D dict(zip(self._TAGS, self._TAGS_LOWER)) ++ self._TAGS_LOWER =3D list(map(str.lower, self._TAGS)) ++ self._TAGMAP =3D dict(list(zip(self._TAGS, self._TAGS_LOWER))) + except: + self.disconnect() + raise +diff -Naur client175_0.7-original/mutagen/apev2.py client175_0.7/mutagen/ape= v2.py +--- client175_0.7-original/mutagen/apev2.py 2010-05-15 00:42:14.000000000 +0= 200 ++++ client175_0.7/mutagen/apev2.py 2021-08-03 14:53:33.292843836 +0200 +@@ -33,7 +33,7 @@ + __all__ =3D ["APEv2", "APEv2File", "Open", "delete"] +=20 + import struct +-from cStringIO import StringIO ++from io import StringIO +=20 + def is_valid_apev2_key(key): + return (2 <=3D len(key) <=3D 255 and min(key) >=3D ' ' and max(key) <= =3D '~' and +@@ -44,11 +44,11 @@ + # 1: Item contains binary information + # 2: Item is a locator of external stored information [e.g. URL] + # 3: reserved" +-TEXT, BINARY, EXTERNAL =3D range(3) ++TEXT, BINARY, EXTERNAL =3D list(range(3)) +=20 +-HAS_HEADER =3D 1L << 31 +-HAS_NO_FOOTER =3D 1L << 30 +-IS_HEADER =3D 1L << 29 ++HAS_HEADER =3D 1 << 31 ++HAS_NO_FOOTER =3D 1 << 30 ++IS_HEADER =3D 1 << 29 +=20 + class error(IOError): pass + class APENoHeaderError(error, ValueError): pass +@@ -199,7 +199,7 @@ +=20 + def pprint(self): + """Return tag key=3Dvalue pairs in a human-readable format.""" +- items =3D self.items() ++ items =3D list(self.items()) + items.sort() + return "\n".join(["%s=3D%s" % (k, v.pprint()) for k, v in items]) +=20 +@@ -271,7 +271,7 @@ +=20 + if not isinstance(value, _APEValue): + # let's guess at the content if we're not already a value... +- if isinstance(value, unicode): ++ if isinstance(value, str): + # unicode? we've got to be text. + value =3D APEValue(utf8(value), TEXT) + elif isinstance(value, list): +@@ -289,7 +289,7 @@ + self.__dict[key.lower()] =3D value +=20 + def keys(self): +- return [self.__casemap.get(key, key) for key in self.__dict.keys()] ++ return [self.__casemap.get(key, key) for key in list(self.__dict.ke= ys())] +=20 + def save(self, filename=3DNone): + """Save changes to a file. +@@ -318,7 +318,7 @@ + # "APE tags items should be sorted ascending by size... This is + # not a MUST, but STRONGLY recommended. Actually the items should + # be sorted by importance/byte, but this is not feasible." +- tags =3D [v._internal(k) for k, v in self.items()] ++ tags =3D [v._internal(k) for k, v in list(self.items())] + tags.sort(lambda a, b: cmp(len(a), len(b))) + num_tags =3D len(tags) + tags =3D "".join(tags) +@@ -401,20 +401,20 @@ + strings (with a null seperating the values), or arrays of strings.""" +=20 + def __unicode__(self): +- return unicode(str(self), "utf-8") ++ return str(str(self), "utf-8") +=20 + def __iter__(self): + """Iterate over the strings of the value (not the characters)""" +- return iter(unicode(self).split("\0")) ++ return iter(str(self).split("\0")) +=20 + def __getitem__(self, index): +- return unicode(self).split("\0")[index] ++ return str(self).split("\0")[index] +=20 + def __len__(self): + return self.value.count("\0") + 1 +=20 + def __cmp__(self, other): +- return cmp(unicode(self), other) ++ return cmp(str(self), other) +=20 + def __setitem__(self, index, value): + values =3D list(self) +@@ -434,7 +434,7 @@ +=20 + External values are usually URI or IRI strings. + """ +- def pprint(self): return "[External] %s" % unicode(self) ++ def pprint(self): return "[External] %s" % str(self) +=20 + class APEv2File(FileType): + class _Info(object): +diff -Naur client175_0.7-original/mutagen/asf.py client175_0.7/mutagen/asf.py +--- client175_0.7-original/mutagen/asf.py 2010-05-15 00:42:14.000000000 +0200 ++++ client175_0.7/mutagen/asf.py 2021-08-03 14:53:40.512736644 +0200 +@@ -49,14 +49,14 @@ +=20 + """ + values =3D [value for (k, value) in self if k =3D=3D key] +- if not values: raise KeyError, key ++ if not values: raise KeyError(key) + else: return values +=20 + def __delitem__(self, key): + """Delete all values associated with the key.""" +- to_delete =3D filter(lambda x: x[0] =3D=3D key, self) +- if not to_delete: raise KeyError, key +- else: map(self.remove, to_delete) ++ to_delete =3D [x for x in self if x[0] =3D=3D key] ++ if not to_delete: raise KeyError(key) ++ else: list(map(self.remove, to_delete)) +=20 + def __contains__(self, key): + """Return true if the key has any values.""" +@@ -78,15 +78,15 @@ + except KeyError: pass + for value in values: + if key in _standard_attribute_names: +- value =3D unicode(value) ++ value =3D str(value) + elif not isinstance(value, ASFBaseAttribute): +- if isinstance(value, basestring): ++ if isinstance(value, str): + value =3D ASFUnicodeAttribute(value) + elif isinstance(value, bool): + value =3D ASFBoolAttribute(value) + elif isinstance(value, int): + value =3D ASFDWordAttribute(value) +- elif isinstance(value, long): ++ elif isinstance(value, int): + value =3D ASFQWordAttribute(value) + self.append((key, value)) +=20 +@@ -162,7 +162,7 @@ + return self.value +=20 + def __cmp__(self, other): +- return cmp(unicode(self), other) ++ return cmp(str(self), other) +=20 +=20 + class ASFByteArrayAttribute(ASFBaseAttribute): +@@ -294,7 +294,7 @@ + GUID =3D ASFGUIDAttribute.TYPE +=20 + def ASFValue(value, kind, **kwargs): +- for t, c in _attribute_types.items(): ++ for t, c in list(_attribute_types.items()): + if kind =3D=3D t: + return c(value=3Dvalue, **kwargs) + raise ValueError("Unknown value type") +@@ -362,12 +362,12 @@ + texts.append(None) + pos =3D end + title, author, copyright, desc, rating =3D texts +- for key, value in dict( ++ for key, value in list(dict( + Title=3Dtitle, + Author=3Dauthor, + Copyright=3Dcopyright, + Description=3Ddesc, +- Rating=3Drating).items(): ++ Rating=3Drating).items()): + if value is not None: + asf.tags[key] =3D value +=20 +@@ -378,8 +378,8 @@ + return value[0].encode("utf-16-le") + "\x00\x00" + else: + return "" +- texts =3D map(render_text, _standard_attribute_names) +- data =3D struct.pack(" u'\x7f': ++ if max(v) > '\x7f': + enc =3D 3 + id3.add(mutagen.id3.TXXX(encoding=3Denc, text=3Dvalue, desc= =3Ddesc)) + else: +@@ -182,7 +182,7 @@ +=20 + def __setitem__(self, key, value): + key =3D key.lower() +- if isinstance(value, basestring): ++ if isinstance(value, str): + value =3D [value] + func =3D dict_match(self.Set, key, self.SetFallback) + if func is not None: +@@ -200,7 +200,7 @@ +=20 + def keys(self): + keys =3D [] +- for key in self.Get.keys(): ++ for key in list(self.Get.keys()): + if key in self.List: + keys.extend(self.List[key](self.__id3, key)) + elif key in self: +@@ -332,7 +332,7 @@ + except KeyError: + raise EasyID3KeyError(key) + else: +- return [u"%+f dB" % frame.gain] ++ return ["%+f dB" % frame.gain] +=20 + def gain_set(id3, key, value): + if len(value) !=3D 1: +@@ -362,7 +362,7 @@ + except KeyError: + raise EasyID3KeyError(key) + else: +- return [u"%f" % frame.peak] ++ return ["%f" % frame.peak] +=20 + def peak_set(id3, key, value): + if len(value) !=3D 1: +@@ -423,7 +423,7 @@ + "TSOT": "titlesort", + "TSRC": "isrc", + "TSST": "discsubtitle", +- }.iteritems(): ++ }.items(): + EasyID3.RegisterTextKey(key, frameid) +=20 + EasyID3.RegisterKey("genre", genre_get, genre_set, genre_delete) +@@ -444,20 +444,20 @@ + # http://bugs.musicbrainz.org/ticket/1383 + # http://musicbrainz.org/doc/MusicBrainzTag + for desc, key in { +- u"MusicBrainz Artist Id": "musicbrainz_artistid", +- u"MusicBrainz Album Id": "musicbrainz_albumid", +- u"MusicBrainz Album Artist Id": "musicbrainz_albumartistid", +- u"MusicBrainz TRM Id": "musicbrainz_trmid", +- u"MusicIP PUID": "musicip_puid", +- u"MusicMagic Fingerprint": "musicip_fingerprint", +- u"MusicBrainz Album Status": "musicbrainz_albumstatus", +- u"MusicBrainz Album Type": "musicbrainz_albumtype", +- u"MusicBrainz Album Release Country": "releasecountry", +- u"MusicBrainz Disc Id": "musicbrainz_discid", +- u"ASIN": "asin", +- u"ALBUMARTISTSORT": "albumartistsort", +- u"BARCODE": "barcode", +- }.iteritems(): ++ "MusicBrainz Artist Id": "musicbrainz_artistid", ++ "MusicBrainz Album Id": "musicbrainz_albumid", ++ "MusicBrainz Album Artist Id": "musicbrainz_albumartistid", ++ "MusicBrainz TRM Id": "musicbrainz_trmid", ++ "MusicIP PUID": "musicip_puid", ++ "MusicMagic Fingerprint": "musicip_fingerprint", ++ "MusicBrainz Album Status": "musicbrainz_albumstatus", ++ "MusicBrainz Album Type": "musicbrainz_albumtype", ++ "MusicBrainz Album Release Country": "releasecountry", ++ "MusicBrainz Disc Id": "musicbrainz_discid", ++ "ASIN": "asin", ++ "ALBUMARTISTSORT": "albumartistsort", ++ "BARCODE": "barcode", ++ }.items(): + EasyID3.RegisterTXXXKey(key, desc) +=20 + class EasyID3FileType(ID3FileType): +diff -Naur client175_0.7-original/mutagen/easymp4.py client175_0.7/mutagen/e= asymp4.py +--- client175_0.7-original/mutagen/easymp4.py 2010-05-15 00:42:14.000000000 = +0200 ++++ client175_0.7/mutagen/easymp4.py 2021-08-03 14:54:50.039718986 +0200 +@@ -86,11 +86,11 @@ + """ +=20 + def getter(tags, key): +- return map(unicode, tags[atomid]) ++ return list(map(str, tags[atomid])) +=20 + def setter(tags, key, value): + clamp =3D lambda x: int(min(max(min_value, x), max_value)) +- tags[atomid] =3D map(clamp, map(int, value)) ++ tags[atomid] =3D list(map(clamp, list(map(int, value)))) +=20 + def deleter(tags, key): + del(tags[atomid]) +@@ -103,9 +103,9 @@ + ret =3D [] + for (track, total) in tags[atomid]: + if total: +- ret.append(u"%d/%d" % (track, total)) ++ ret.append("%d/%d" % (track, total)) + else: +- ret.append(unicode(track)) ++ ret.append(str(track)) + return ret +=20 + def setter(tags, key, value): +@@ -143,7 +143,7 @@ + return [s.decode("utf-8", "replace") for s in tags[atomid]] +=20 + def setter(tags, key, value): +- tags[atomid] =3D map(utf8, value) ++ tags[atomid] =3D list(map(utf8, value)) +=20 + def deleter(tags, key): + del(tags[atomid]) +@@ -161,7 +161,7 @@ +=20 + def __setitem__(self, key, value): + key =3D key.lower() +- if isinstance(value, basestring): ++ if isinstance(value, str): + value =3D [value] + func =3D dict_match(self.Set, key) + if func is not None: +@@ -179,7 +179,7 @@ +=20 + def keys(self): + keys =3D [] +- for key in self.Get.keys(): ++ for key in list(self.Get.keys()): + if key in self.List: + keys.extend(self.List[key](self.__mp4, key)) + elif key in self: +@@ -195,7 +195,7 @@ + strings.append("%s=3D%s" % (key, value)) + return "\n".join(strings) +=20 +-for atomid, key in { ++for atomid, key in list({ + '\xa9nam': 'title', + '\xa9alb': 'album', + '\xa9ART': 'artist', +@@ -211,10 +211,10 @@ + 'soar': 'artistsort', + 'sonm': 'titlesort', + 'soco': 'composersort', +- }.items(): ++ }.items()): + EasyMP4Tags.RegisterTextKey(key, atomid) +=20 +-for name, key in { ++for name, key in list({ + 'MusicBrainz Artist Id': 'musicbrainz_artistid', + 'MusicBrainz Track Id': 'musicbrainz_trackid', + 'MusicBrainz Album Id': 'musicbrainz_albumid', +@@ -223,18 +223,18 @@ + 'MusicBrainz Album Status': 'musicbrainz_albumstatus', + 'MusicBrainz Album Type': 'musicbrainz_albumtype', + 'MusicBrainz Release Country': 'releasecountry', +- }.items(): ++ }.items()): + EasyMP4Tags.RegisterFreeformKey(key, name) +=20 +-for name, key in { ++for name, key in list({ + "tmpo": "bpm", +- }.items(): ++ }.items()): + EasyMP4Tags.RegisterIntKey(key, name) +=20 +-for name, key in { ++for name, key in list({ + "trkn": "tracknumber", + "disk": "discnumber", +- }.items(): ++ }.items()): + EasyMP4Tags.RegisterIntPairKey(key, name) +=20 + class EasyMP4(MP4): +diff -Naur client175_0.7-original/mutagen/flac.py client175_0.7/mutagen/flac= .py +--- client175_0.7-original/mutagen/flac.py 2010-05-15 00:42:14.000000000 +02= 00 ++++ client175_0.7/mutagen/flac.py 2021-08-03 14:54:56.799621843 +0200 +@@ -22,11 +22,12 @@ + __all__ =3D ["FLAC", "Open", "delete"] +=20 + import struct +-from cStringIO import StringIO +-from _vorbis import VCommentDict ++from io import StringIO ++from ._vorbis import VCommentDict + from mutagen import FileType + from mutagen._util import insert_bytes + from mutagen.id3 import BitPaddedInt ++from functools import reduce +=20 + class error(IOError): pass + class FLACNoHeaderError(error): pass +@@ -35,7 +36,7 @@ + def to_int_be(string): + """Convert an arbitrarily-long string to a long using big-endian + byte order.""" +- return reduce(lambda a, b: (a << 8) + ord(b), string, 0L) ++ return reduce(lambda a, b: (a << 8) + ord(b), string, 0) +=20 + class MetadataBlock(object): + """A generic block of FLAC metadata. +@@ -79,8 +80,8 @@ +=20 + The overall size of the rendered blocks does not change, so + this adds several bytes of padding for each merged block.""" +- paddings =3D filter(lambda x: isinstance(x, Padding), blocks) +- map(blocks.remove, paddings) ++ paddings =3D [x for x in blocks if isinstance(x, Padding)] ++ list(map(blocks.remove, paddings)) + padding =3D Padding() + # total padding size is the sum of padding sizes plus 4 bytes + # per removed header. +@@ -137,7 +138,7 @@ + bps_tail =3D bps_total >> 36 + bps_head =3D (sample_channels_bps & 1) << 4 + self.bits_per_sample =3D int(bps_head + bps_tail + 1) +- self.total_samples =3D bps_total & 0xFFFFFFFFFL ++ self.total_samples =3D bps_total & 0xFFFFFFFFF + self.length =3D self.total_samples / float(self.sample_rate) +=20 + self.md5_signature =3D to_int_be(data.read(16)) +@@ -161,12 +162,12 @@ + byte +=3D (self.total_samples >> 32) & 0xF + f.write(chr(byte)) + # last 32 of sample count +- f.write(struct.pack(">I", self.total_samples & 0xFFFFFFFFL)) ++ f.write(struct.pack(">I", self.total_samples & 0xFFFFFFFF)) + # MD5 signature + sig =3D self.md5_signature + f.write(struct.pack( +- ">4I", (sig >> 96) & 0xFFFFFFFFL, (sig >> 64) & 0xFFFFFFFFL, +- (sig >> 32) & 0xFFFFFFFFL, sig & 0xFFFFFFFFL)) ++ ">4I", (sig >> 96) & 0xFFFFFFFF, (sig >> 64) & 0xFFFFFFFF, ++ (sig >> 32) & 0xFFFFFFFF, sig & 0xFFFFFFFF)) + return f.getvalue() +=20 + def pprint(self): +@@ -425,8 +426,8 @@ +=20 + def __init__(self, data=3DNone): + self.type =3D 0 +- self.mime =3D u'' +- self.desc =3D u'' ++ self.mime =3D '' ++ self.desc =3D '' + self.width =3D 0 + self.height =3D 0 + self.depth =3D 0 +@@ -601,11 +602,10 @@ +=20 + def clear_pictures(self): + """Delete all pictures from the file.""" +- self.metadata_blocks =3D filter(lambda b: b.code !=3D Picture.code, +- self.metadata_blocks) ++ self.metadata_blocks =3D [b for b in self.metadata_blocks if b.code= !=3D Picture.code] +=20 + def __get_pictures(self): +- return filter(lambda b: b.code =3D=3D Picture.code, self.metadata_b= locks) ++ return [b for b in self.metadata_blocks if b.code =3D=3D Picture.co= de] + pictures =3D property(__get_pictures, doc=3D"List of embedded pictures") +=20 + def save(self, filename=3DNone, deleteid3=3DFalse): +diff -Naur client175_0.7-original/mutagen/id3.py client175_0.7/mutagen/id3.py +--- client175_0.7-original/mutagen/id3.py 2010-05-15 00:42:14.000000000 +0200 ++++ client175_0.7/mutagen/id3.py 2021-08-03 14:55:05.431497226 +0200 +@@ -79,7 +79,7 @@ + raise ValueError('Requested bytes (%s) less than zero' % si= ze) + if size > self.__filesize: + raise EOFError('Requested %#x of %#x (%s)' %=20 +- (long(size), long(self.__filesize), self.filename)) ++ (int(size), int(self.__filesize), self.filename)) + except AttributeError: pass + data =3D self.__fileobj.read(size) + if len(data) !=3D size: raise EOFError +@@ -115,18 +115,18 @@ + self.size =3D 0 + raise ID3NoHeaderError("%s: too small (%d bytes)" %( + filename, self.__filesize)) +- except (ID3NoHeaderError, ID3UnsupportedVersionError), err: ++ except (ID3NoHeaderError, ID3UnsupportedVersionError) as err: + self.size =3D 0 + import sys + stack =3D sys.exc_info()[2] + try: self.__fileobj.seek(-128, 2) +- except EnvironmentError: raise err, None, stack ++ except EnvironmentError: raise err.with_traceback(stack) + else: + frames =3D ParseID3v1(self.__fileobj.read(128)) + if frames is not None: + self.version =3D (1, 1) +- map(self.add, frames.values()) +- else: raise err, None, stack ++ list(map(self.add, list(frames.values()))) ++ else: raise err.with_traceback(stack) + else: + frames =3D self.__known_frames + if frames is None: +@@ -159,14 +159,14 @@ + if key in self: return [self[key]] + else: + key =3D key + ":" +- return [v for s,v in self.items() if s.startswith(key)] ++ return [v for s,v in list(self.items()) if s.startswith(key)] +=20 + def delall(self, key): + """Delete all tags of a given kind; see getall.""" + if key in self: del(self[key]) + else: + key =3D key + ":" +- for k in filter(lambda s: s.startswith(key), self.keys()): ++ for k in [s for s in list(self.keys()) if s.startswith(key)]: + del(self[k]) +=20 + def setall(self, key, values): +@@ -184,7 +184,7 @@ + However, ID3 frames can have multiple keys: + POPM=3Duser(a)example.org=3D3 128/255 + """ +- return "\n".join(map(Frame.pprint, self.values())) ++ return "\n".join(map(Frame.pprint, list(self.values()))) +=20 + def loaded_frame(self, tag): + """Deprecated; use the add method.""" +@@ -343,9 +343,9 @@ +=20 + # Sort frames by 'importance' + order =3D ["TIT2", "TPE1", "TRCK", "TALB", "TPOS", "TDRC", "TCON"] +- order =3D dict(zip(order, range(len(order)))) ++ order =3D dict(list(zip(order, list(range(len(order)))))) + last =3D len(order) +- frames =3D self.items() ++ frames =3D list(self.items()) + frames.sort(lambda a, b: cmp(order.get(a[0][:4], last), + order.get(b[0][:4], last))) +=20 +@@ -355,7 +355,7 @@ + if not framedata: + try: + self.delete(filename) +- except EnvironmentError, err: ++ except EnvironmentError as err: + from errno import ENOENT + if err.errno !=3D ENOENT: raise + return +@@ -365,7 +365,7 @@ +=20 + if filename is None: filename =3D self.filename + try: f =3D open(filename, 'rb+') +- except IOError, err: ++ except IOError as err: + from errno import ENOENT + if err.errno !=3D ENOENT: raise + f =3D open(filename, 'ab') # create, then reopen +@@ -393,7 +393,7 @@ +=20 + try: + f.seek(-128, 2) +- except IOError, err: ++ except IOError as err: + from errno import EINVAL + if err.errno !=3D EINVAL: raise + f.seek(0, 2) # ensure read won't get "TAG" +@@ -540,7 +540,7 @@ + def __new__(cls, value, bits=3D7, bigendian=3DTrue): + "Strips 8-bits bits out of every byte" + mask =3D (1<<(bits))-1 +- if isinstance(value, (int, long)): ++ if isinstance(value, int): + bytes =3D [] + while value: + bytes.append(value & ((1< width: +- raise ValueError, 'Value too wide (%d bytes)' % len(bytes) ++ raise ValueError('Value too wide (%d bytes)' % len(bytes)) + else: bytes.extend([0] * (width-len(bytes))) + if bigendian: bytes.reverse() + return ''.join(map(chr, bytes)) +@@ -654,7 +654,7 @@ + def validate(self, frame, value): + if 0 <=3D value <=3D 3: return value + if value is None: return None +- raise ValueError, 'Invalid Encoding: %r' % value ++ raise ValueError('Invalid Encoding: %r' % value) +=20 + class StringSpec(Spec): + def __init__(self, name, length): +@@ -666,8 +666,8 @@ + else: return (str(value) + '\x00' * s.len)[:s.len] + def validate(s, frame, value): + if value is None: return None +- if isinstance(value, basestring) and len(value) =3D=3D s.len: retur= n value +- raise ValueError, 'Invalid StringSpec[%d] data: %r' % (s.len, value) ++ if isinstance(value, str) and len(value) =3D=3D s.len: return value ++ raise ValueError('Invalid StringSpec[%d] data: %r' % (s.len, value)) +=20 + class BinaryDataSpec(Spec): + def read(self, frame, data): return data, '' +@@ -696,14 +696,14 @@ + data, ret =3D data[0:offset], data[offset+2:]; break + except ValueError: pass +=20 +- if len(data) < len(term): return u'', ret ++ if len(data) < len(term): return '', ret + return data.decode(enc), ret +=20 + def write(self, frame, value): + enc, term =3D self._encodings[frame.encoding] + return value.encode(enc) + term +=20 +- def validate(self, frame, value): return unicode(value) ++ def validate(self, frame, value): return str(value) +=20 + class MultiSpec(Spec): + def __init__(self, name, *specs, **kw): +@@ -735,7 +735,7 @@ +=20 + def validate(self, frame, value): + if value is None: return [] +- if self.sep and isinstance(value, basestring): ++ if self.sep and isinstance(value, str): + value =3D value.split(self.sep) + if isinstance(value, list): + if len(self.specs) =3D=3D 1: +@@ -744,7 +744,7 @@ + return [=20 + [s.validate(frame, v) for (v,s) in zip(val, self.specs)] + for val in value ] +- raise ValueError, 'Invalid MultiSpec data: %r' % value ++ raise ValueError('Invalid MultiSpec data: %r' % value) +=20 + class EncodedNumericTextSpec(EncodedTextSpec): pass + class EncodedNumericPartTextSpec(EncodedTextSpec): pass +@@ -758,7 +758,7 @@ + def write(self, data, value): + return value.encode('latin1') + '\x00' +=20 +- def validate(self, frame, value): return unicode(value) ++ def validate(self, frame, value): return str(value) +=20 + class ID3TimeStamp(object): + """A time stamp in ID3v2 format. +@@ -782,9 +782,9 @@ + parts =3D [self.year, self.month, self.day, + self.hour, self.minute, self.second] + pieces =3D [] +- for i, part in enumerate(iter(iter(parts).next, None)): ++ for i, part in enumerate(iter(iter(parts).__next__, None)): + pieces.append(self.__formats[i]%part + self.__seps[i]) +- return u''.join(pieces)[:-1] ++ return ''.join(pieces)[:-1] +=20 + def set_text(self, text, splitre=3Dre.compile('[-T:/.]|\s+')): + year, month, day, hour, minute, second =3D \ +@@ -812,11 +812,11 @@ +=20 + def validate(self, frame, value): + try: return ID3TimeStamp(value) +- except TypeError: raise ValueError, "Invalid ID3TimeStamp: %r" % va= lue ++ except TypeError: raise ValueError("Invalid ID3TimeStamp: %r" % val= ue) +=20 + class ChannelSpec(ByteSpec): + (OTHER, MASTER, FRONTRIGHT, FRONTLEFT, BACKRIGHT, BACKLEFT, FRONTCENTRE, +- BACKCENTRE, SUBWOOFER) =3D range(9) ++ BACKCENTRE, SUBWOOFER) =3D list(range(9)) +=20 + class VolumeAdjustmentSpec(Spec): + def read(self, frame, data): +@@ -900,7 +900,7 @@ + freq /=3D 2.0 + adj /=3D 512.0 + adjustments[freq] =3D adj +- adjustments =3D adjustments.items() ++ adjustments =3D list(adjustments.items()) + adjustments.sort() + return adjustments, data +=20 +@@ -1033,21 +1033,21 @@ + data =3D data[4:] + if tflags & Frame.FLAG24_UNSYNCH or id3.f_unsynch: + try: data =3D unsynch.decode(data) +- except ValueError, err: ++ except ValueError as err: + if id3.PEDANTIC: +- raise ID3BadUnsynchData, '%s: %r' % (err, data) ++ raise ID3BadUnsynchData('%s: %r' % (err, data)) + if tflags & Frame.FLAG24_ENCRYPT: + raise ID3EncryptionUnsupportedError + if tflags & Frame.FLAG24_COMPRESS: + try: data =3D data.decode('zlib') +- except zlibError, err: ++ except zlibError as err: + # the initial mutagen that went out with QL 0.12 did not + # write the 4 bytes of uncompressed size. Compensate. + data =3D datalen_bytes + data + try: data =3D data.decode('zlib') +- except zlibError, err: ++ except zlibError as err: + if id3.PEDANTIC: +- raise ID3BadCompressedData, '%s: %r' % (err, da= ta) ++ raise ID3BadCompressedData('%s: %r' % (err, dat= a)) +=20 + elif (2,3,0) <=3D id3.version: + if tflags & Frame.FLAG23_COMPRESS: +@@ -1057,9 +1057,9 @@ + raise ID3EncryptionUnsupportedError + if tflags & Frame.FLAG23_COMPRESS: + try: data =3D data.decode('zlib') +- except zlibError, err: ++ except zlibError as err: + if id3.PEDANTIC: +- raise ID3BadCompressedData, '%s: %r' % (err, data) ++ raise ID3BadCompressedData('%s: %r' % (err, data)) +=20 + frame =3D cls() + frame._rawdata =3D data +@@ -1138,12 +1138,12 @@ + """ +=20 + _framespec =3D [ EncodingSpec('encoding'), +- MultiSpec('text', EncodedTextSpec('text'), sep=3Du'\u0000') ] ++ MultiSpec('text', EncodedTextSpec('text'), sep=3D'\u0000') ] + def __str__(self): return self.__unicode__().encode('utf-8') +- def __unicode__(self): return u'\u0000'.join(self.text) ++ def __unicode__(self): return '\u0000'.join(self.text) + def __eq__(self, other): + if isinstance(other, str): return str(self) =3D=3D other +- elif isinstance(other, unicode): return unicode(self) =3D=3D other ++ elif isinstance(other, str): return str(self) =3D=3D other + return self.text =3D=3D other + def __getitem__(self, item): return self.text[item] + def __iter__(self): return iter(self.text) +@@ -1160,7 +1160,7 @@ + """ +=20 + _framespec =3D [ EncodingSpec('encoding'), +- MultiSpec('text', EncodedNumericTextSpec('text'), sep=3Du'\u0000') ] ++ MultiSpec('text', EncodedNumericTextSpec('text'), sep=3D'\u0000') ] +=20 + def __pos__(self): + """Return the numerical value of the string.""" +@@ -1176,7 +1176,7 @@ + """ +=20 + _framespec =3D [ EncodingSpec('encoding'), +- MultiSpec('text', EncodedNumericPartTextSpec('text'), sep=3Du'\u000= 0') ] ++ MultiSpec('text', EncodedNumericPartTextSpec('text'), sep=3D'\u0000= ') ] + def __pos__(self): + return int(self.text[0].split("/")[0]) +=20 +@@ -1188,7 +1188,7 @@ + """ +=20 + _framespec =3D [ EncodingSpec('encoding'), +- MultiSpec('text', TimeStampSpec('stamp'), sep=3Du',') ] ++ MultiSpec('text', TimeStampSpec('stamp'), sep=3D',') ] + def __str__(self): return self.__unicode__().encode('utf-8') + def __unicode__(self): return ','.join([stamp.text for stamp in self.te= xt]) + def _pprint(self): +@@ -1235,9 +1235,9 @@ + for value in self.text: + if value.isdigit(): + try: genres.append(self.GENRES[int(value)]) +- except IndexError: genres.append(u"Unknown") +- elif value =3D=3D "CR": genres.append(u"Cover") +- elif value =3D=3D "RX": genres.append(u"Remix") ++ except IndexError: genres.append("Unknown") ++ elif value =3D=3D "CR": genres.append("Cover") ++ elif value =3D=3D "RX": genres.append("Remix") + elif value: + newgenres =3D [] + genreid, dummy, genrename =3D genre_re.match(value).groups() +@@ -1245,11 +1245,11 @@ + if genreid: + for gid in genreid[1:-1].split(")("): + if gid.isdigit() and int(gid) < len(self.GENRES): +- gid =3D unicode(self.GENRES[int(gid)]) ++ gid =3D str(self.GENRES[int(gid)]) + newgenres.append(gid) +- elif gid =3D=3D "CR": newgenres.append(u"Cover") +- elif gid =3D=3D "RX": newgenres.append(u"Remix") +- else: newgenres.append(u"Unknown") ++ elif gid =3D=3D "CR": newgenres.append("Cover") ++ elif gid =3D=3D "RX": newgenres.append("Remix") ++ else: newgenres.append("Unknown") +=20 + if genrename: + # "Unescaping" the first parenthesis +@@ -1261,8 +1261,8 @@ + return genres +=20 + def __set_genres(self, genres): +- if isinstance(genres, basestring): genres =3D [genres] +- self.text =3D map(self.__decode, genres) ++ if isinstance(genres, str): genres =3D [genres] ++ self.text =3D list(map(self.__decode, genres)) +=20 + def __decode(self, value): + if isinstance(value, str): +@@ -1333,7 +1333,7 @@ + the same). Many taggers use this frame to store freeform keys. + """ + _framespec =3D [ EncodingSpec('encoding'), EncodedTextSpec('desc'), +- MultiSpec('text', EncodedTextSpec('text'), sep=3Du'\u0000') ] ++ MultiSpec('text', EncodedTextSpec('text'), sep=3D'\u0000') ] + HashKey =3D property(lambda s: '%s:%s' % (s.FrameID, s.desc)) + def _pprint(self): return "%s=3D%s" % (self.desc, " / ".join(self.text)) +=20 +@@ -1448,7 +1448,7 @@ + """ + _framespec =3D [ EncodingSpec('encoding'), StringSpec('lang', 3), + EncodedTextSpec('desc'), +- MultiSpec('text', EncodedTextSpec('text'), sep=3Du'\u0000') ] ++ MultiSpec('text', EncodedTextSpec('text'), sep=3D'\u0000') ] + HashKey =3D property(lambda s: '%s:%s:%r' % (s.FrameID, s.desc, s.lang)) + def _pprint(self): return "%s=3D%r=3D%s" % ( + self.desc, self.lang, " / ".join(self.text)) +@@ -1545,7 +1545,7 @@ +=20 + def __eq__(self, other): return self.count =3D=3D other + def __pos__(self): return self.count +- def _pprint(self): return unicode(self.count) ++ def _pprint(self): return str(self.count) +=20 + class POPM(FrameOpt): + """Popularimeter. +@@ -1774,7 +1774,7 @@ + ASPIIndexSpec("Fi") ] + def __eq__(self, other): return self.Fi =3D=3D other +=20 +-Frames =3D dict([(k,v) for (k,v) in globals().items() ++Frames =3D dict([(k,v) for (k,v) in list(globals().items()) + if len(k)=3D=3D4 and isinstance(v, type) and issubclass(v, Frame)]) + """All supported ID3v2 frames, keyed by frame name.""" + del(k); del(v) +@@ -1867,7 +1867,7 @@ + _framespec =3D [ StringSpec('frameid', 3), Latin1TextSpec('url') ] + _optionalspec =3D [ BinaryDataSpec('data') ] +=20 +-Frames_2_2 =3D dict([(k,v) for (k,v) in globals().items() ++Frames_2_2 =3D dict([(k,v) for (k,v) in list(globals().items()) + if len(k)=3D=3D3 and isinstance(v, type) and issubclass(v, Frame)]) +=20 + # support open(filename) as interface +@@ -1886,8 +1886,8 @@ + if tag !=3D "TAG": return None + def fix(string): + return string.split("\x00")[0].strip().decode('latin1') +- title, artist, album, year, comment =3D map( +- fix, [title, artist, album, year, comment]) ++ title, artist, album, year, comment =3D list(map( ++ fix, [title, artist, album, year, comment])) +=20 + if title: frames["TIT2"] =3D TIT2(encoding=3D0, text=3Dtitle) + if artist: frames["TPE1"] =3D TPE1(encoding=3D0, text=3D[artist]) +@@ -1907,8 +1907,8 @@ +=20 + v1 =3D {} +=20 +- for v2id, name in {"TIT2": "title", "TPE1": "artist", +- "TALB": "album"}.items(): ++ for v2id, name in list({"TIT2": "title", "TPE1": "artist", ++ "TALB": "album"}.items()): + if v2id in id3: + text =3D id3[v2id].text[0].encode('latin1', 'replace')[:30] + else: +diff -Naur client175_0.7-original/mutagen/__init__.py client175_0.7/mutagen/= __init__.py +--- client175_0.7-original/mutagen/__init__.py 2010-05-15 00:42:14.000000000= +0200 ++++ client175_0.7/mutagen/__init__.py 2021-08-03 14:55:18.899303101 +0200 +@@ -82,7 +82,7 @@ +=20 + If the file has no tags at all, a KeyError is raised. + """ +- if self.tags is None: raise KeyError, key ++ if self.tags is None: raise KeyError(key) + else: return self.tags[key] +=20 + def __setitem__(self, key, value): +@@ -100,7 +100,7 @@ +=20 + If the file has no tags at all, a KeyError is raised. + """ +- if self.tags is None: raise KeyError, key ++ if self.tags is None: raise KeyError(key) + else: del(self.tags[key]) +=20 + def keys(self): +@@ -109,7 +109,7 @@ + If the file has no tags at all, an empty list is returned. + """ + if self.tags is None: return [] +- else: return self.tags.keys() ++ else: return list(self.tags.keys()) +=20 + def delete(self, filename=3DNone): + """Remove tags from a file.""" +@@ -210,7 +210,7 @@ + for Kind in options] + finally: + fileobj.close() +- results =3D zip(results, options) ++ results =3D list(zip(results, options)) + results.sort() + (score, name), Kind =3D results[-1] + if score > 0: return Kind(filename) +diff -Naur client175_0.7-original/mutagen/m4a.py client175_0.7/mutagen/m4a.py +--- client175_0.7-original/mutagen/m4a.py 2010-05-15 00:42:14.000000000 +0200 ++++ client175_0.7/mutagen/m4a.py 2021-08-03 14:55:39.954998104 +0200 +@@ -25,7 +25,7 @@ + import struct + import sys +=20 +-from cStringIO import StringIO ++from io import StringIO +=20 + from mutagen import FileType, Metadata + from mutagen._constants import GENRES +@@ -119,7 +119,7 @@ + if child.name =3D=3D remaining[0]: + return child[remaining[1:]] + else: +- raise KeyError, "%r not found" % remaining[0] ++ raise KeyError("%r not found" % remaining[0]) +=20 + def __repr__(self): + klass =3D self.__class__.__name__ +@@ -166,13 +166,13 @@ + 'names' may be a list of atoms (['moov', 'udta']) or a string + specifying the complete path ('moov.udta'). + """ +- if isinstance(names, basestring): ++ if isinstance(names, str): + names =3D names.split(".") + for child in self.atoms: + if child.name =3D=3D names[0]: + return child[names[1:]] + else: +- raise KeyError, "%s not found" % names[0] ++ raise KeyError("%s not found" % names[0]) +=20 + def __repr__(self): + return "\n".join([repr(child) for child in self.atoms]) +@@ -202,7 +202,7 @@ +=20 + def load(self, atoms, fileobj): + try: ilst =3D atoms["moov.udta.meta.ilst"] +- except KeyError, key: ++ except KeyError as key: + raise M4AMetadataError(key) + for atom in ilst.children: + fileobj.seek(atom.offset + 8) +@@ -210,14 +210,16 @@ + parse =3D self.__atoms.get(atom.name, (M4ATags.__parse_text,))[= 0] + parse(self, atom, data) +=20 +- def __key_sort((key1, v1), (key2, v2)): ++ def __key_sort(xxx_todo_changeme, xxx_todo_changeme1): + # iTunes always writes the tags in order of "relevance", try + # to copy it as closely as possible. ++ (key1, v1) =3D xxx_todo_changeme ++ (key2, v2) =3D xxx_todo_changeme1 + order =3D ["\xa9nam", "\xa9ART", "\xa9wrt", "\xa9alb", + "\xa9gen", "gnre", "trkn", "disk", + "\xa9day", "cpil", "tmpo", "\xa9too", + "----", "covr", "\xa9lyr"] +- order =3D dict(zip(order, range(len(order)))) ++ order =3D dict(list(zip(order, list(range(len(order)))))) + last =3D len(order) + # If there's no key-based way to distinguish, order by length. + # If there's still no way, go by string comparison on the +@@ -229,7 +231,7 @@ + def save(self, filename): + """Save the metadata to the given filename.""" + values =3D [] +- items =3D self.items() ++ items =3D list(self.items()) + items.sort(self.__key_sort) + for key, value in items: + render =3D self.__atoms.get( +@@ -411,7 +413,7 @@ +=20 + def pprint(self): + values =3D [] +- for key, value in self.iteritems(): ++ for key, value in self.items(): + key =3D key.decode('latin1') + try: values.append("%s=3D%s" % (key, value)) + except UnicodeDecodeError: +@@ -475,13 +477,13 @@ + try: + atoms =3D Atoms(fileobj) + try: self.info =3D M4AInfo(atoms, fileobj) +- except StandardError, err: +- raise M4AStreamInfoError, err, sys.exc_info()[2] ++ except Exception as err: ++ raise M4AStreamInfoError(err).with_traceback(sys.exc_info()= [2]) + try: self.tags =3D M4ATags(atoms, fileobj) + except M4AMetadataError: + self.tags =3D None +- except StandardError, err: +- raise M4AMetadataError, err, sys.exc_info()[2] ++ except Exception as err: ++ raise M4AMetadataError(err).with_traceback(sys.exc_info()[2= ]) + finally: + fileobj.close() +=20 +diff -Naur client175_0.7-original/mutagen/mp3.py client175_0.7/mutagen/mp3.py +--- client175_0.7-original/mutagen/mp3.py 2010-05-15 00:42:14.000000000 +0200 ++++ client175_0.7/mutagen/mp3.py 2021-08-03 14:55:59.774710002 +0200 +@@ -19,7 +19,7 @@ + class InvalidMPEGHeader(error, IOError): pass +=20 + # Mode values. +-STEREO, JOINTSTEREO, DUALCHANNEL, MONO =3D range(4) ++STEREO, JOINTSTEREO, DUALCHANNEL, MONO =3D list(range(4)) +=20 + class MPEGInfo(object): + """MPEG audio stream information +@@ -46,7 +46,7 @@ +=20 + # Map (version, layer) tuples to bitrates. + __BITRATE =3D { +- (1, 1): range(0, 480, 32), ++ (1, 1): list(range(0, 480, 32)), + (1, 2): [0, 32, 48, 56, 64, 80, 96, 112,128,160,192,224,256,320,384= ], + (1, 3): [0, 32, 40, 48, 56, 64, 80, 96, 112,128,160,192,224,256,320= ], + (2, 1): [0, 32, 48, 56, 64, 80, 96, 112,128,144,160,176,192,224,256= ], +@@ -95,7 +95,7 @@ + # and 90% through the file. + for i in [offset, 0.3 * size, 0.6 * size, 0.9 * size]: + try: self.__try(fileobj, int(i), size - offset) +- except error, e: pass ++ except error as e: pass + else: break + # If we can't find any two consecutive frames, try to find just + # one frame back at the original offset given. +diff -Naur client175_0.7-original/mutagen/mp4.py client175_0.7/mutagen/mp4.py +--- client175_0.7-original/mutagen/mp4.py 2010-05-15 00:42:14.000000000 +0200 ++++ client175_0.7/mutagen/mp4.py 2021-08-03 14:56:05.542625890 +0200 +@@ -121,7 +121,7 @@ + if child.name =3D=3D remaining[0]: + return child[remaining[1:]] + else: +- raise KeyError, "%r not found" % remaining[0] ++ raise KeyError("%r not found" % remaining[0]) +=20 + def __repr__(self): + klass =3D self.__class__.__name__ +@@ -168,13 +168,13 @@ + 'names' may be a list of atoms (['moov', 'udta']) or a string + specifying the complete path ('moov.udta'). + """ +- if isinstance(names, basestring): ++ if isinstance(names, str): + names =3D names.split(".") + for child in self.atoms: + if child.name =3D=3D names[0]: + return child[names[1:]] + else: +- raise KeyError, "%s not found" % names[0] ++ raise KeyError("%s not found" % names[0]) +=20 + def __repr__(self): + return "\n".join([repr(child) for child in self.atoms]) +@@ -242,7 +242,7 @@ +=20 + def load(self, atoms, fileobj): + try: ilst =3D atoms["moov.udta.meta.ilst"] +- except KeyError, key: ++ except KeyError as key: + raise MP4MetadataError(key) + for atom in ilst.children: + fileobj.seek(atom.offset + 8) +@@ -250,14 +250,16 @@ + info =3D self.__atoms.get(atom.name, (type(self).__parse_text, = None)) + info[0](self, atom, data, *info[2:]) +=20 +- def __key_sort((key1, v1), (key2, v2)): ++ def __key_sort(xxx_todo_changeme, xxx_todo_changeme1): + # iTunes always writes the tags in order of "relevance", try + # to copy it as closely as possible. ++ (key1, v1) =3D xxx_todo_changeme ++ (key2, v2) =3D xxx_todo_changeme1 + order =3D ["\xa9nam", "\xa9ART", "\xa9wrt", "\xa9alb", + "\xa9gen", "gnre", "trkn", "disk", + "\xa9day", "cpil", "pgap", "pcst", "tmpo", + "\xa9too", "----", "covr", "\xa9lyr"] +- order =3D dict(zip(order, range(len(order)))) ++ order =3D dict(list(zip(order, list(range(len(order)))))) + last =3D len(order) + # If there's no key-based way to distinguish, order by length. + # If there's still no way, go by string comparison on the +@@ -269,14 +271,14 @@ + def save(self, filename): + """Save the metadata to the given filename.""" + values =3D [] +- items =3D self.items() ++ items =3D list(self.items()) + items.sort(self.__key_sort) + for key, value in items: + info =3D self.__atoms.get(key[:4], (None, type(self).__render_t= ext)) + try: + values.append(info[1](self, key, value, *info[2:])) +- except (TypeError, ValueError), s: +- raise MP4MetadataValueError, s, sys.exc_info()[2] ++ except (TypeError, ValueError) as s: ++ raise MP4MetadataValueError(s).with_traceback(sys.exc_info(= )[2]) + data =3D Atom.render("ilst", "".join(values)) +=20 + # Find the old atoms. +@@ -440,7 +442,7 @@ + dummy, mean, name =3D key.split(":", 2) + mean =3D struct.pack(">I4sI", len(mean) + 12, "mean", 0) + mean + name =3D struct.pack(">I4sI", len(name) + 12, "name", 0) + name +- if isinstance(value, basestring): ++ if isinstance(value, str): + value =3D [value] + return Atom.render("----", mean + name + "".join([ + struct.pack(">I4s2I", len(data) + 16, "data", 1, 0) + data +@@ -492,7 +494,7 @@ + raise MP4MetadataValueError( + "tmpo must be a list of 16 bit integers") +=20 +- values =3D map(cdata.to_ushort_be, value) ++ values =3D list(map(cdata.to_ushort_be, value)) + return self.__render_data(key, 0x15, values) +=20 + def __parse_bool(self, atom, data): +@@ -531,10 +533,10 @@ + if value: + self[atom.name] =3D value + def __render_text(self, key, value, flags=3D1): +- if isinstance(value, basestring): ++ if isinstance(value, str): + value =3D [value] + return self.__render_data( +- key, flags, map(utf8, value)) ++ key, flags, list(map(utf8, value))) +=20 + def delete(self, filename): + self.clear() +@@ -556,13 +558,13 @@ +=20 + def pprint(self): + values =3D [] +- for key, value in self.iteritems(): ++ for key, value in self.items(): + key =3D key.decode('latin1') + if key =3D=3D "covr": + values.append("%s=3D%s" % (key, ", ".join( + ["[%d bytes of data]" % len(data) for data in value]))) + elif isinstance(value, list): +- values.append("%s=3D%s" % (key, " / ".join(map(unicode, val= ue)))) ++ values.append("%s=3D%s" % (key, " / ".join(map(str, value))= )) + else: + values.append("%s=3D%s" % (key, value)) + return "\n".join(values) +@@ -658,13 +660,13 @@ + try: + atoms =3D Atoms(fileobj) + try: self.info =3D MP4Info(atoms, fileobj) +- except StandardError, err: +- raise MP4StreamInfoError, err, sys.exc_info()[2] ++ except Exception as err: ++ raise MP4StreamInfoError(err).with_traceback(sys.exc_info()= [2]) + try: self.tags =3D self.MP4Tags(atoms, fileobj) + except MP4MetadataError: + self.tags =3D None +- except StandardError, err: +- raise MP4MetadataError, err, sys.exc_info()[2] ++ except Exception as err: ++ raise MP4MetadataError(err).with_traceback(sys.exc_info()[2= ]) + finally: + fileobj.close() +=20 +diff -Naur client175_0.7-original/mutagen/oggflac.py client175_0.7/mutagen/o= ggflac.py +--- client175_0.7-original/mutagen/oggflac.py 2010-05-15 00:42:14.000000000 = +0200 ++++ client175_0.7/mutagen/oggflac.py 2021-08-03 14:56:29.786272048 +0200 +@@ -21,7 +21,7 @@ +=20 + import struct +=20 +-from cStringIO import StringIO ++from io import StringIO +=20 + from mutagen.flac import StreamInfo, VCFLACDict + from mutagen.ogg import OggPage, OggFileType, error as OggError +diff -Naur client175_0.7-original/mutagen/ogg.py client175_0.7/mutagen/ogg.py +--- client175_0.7-original/mutagen/ogg.py 2010-05-15 00:42:14.000000000 +0200 ++++ client175_0.7/mutagen/ogg.py 2021-08-03 14:56:22.722375298 +0200 +@@ -20,7 +20,7 @@ + import sys + import zlib +=20 +-from cStringIO import StringIO ++from io import StringIO +=20 + from mutagen import FileType + from mutagen._util import cdata, insert_bytes, delete_bytes +@@ -57,7 +57,7 @@ +=20 + version =3D 0 + __type_flags =3D 0 +- position =3D 0L ++ position =3D 0 + serial =3D 0 + sequence =3D 0 + offset =3D None +@@ -103,8 +103,8 @@ + lacings.append(total) + self.complete =3D False +=20 +- self.packets =3D map(fileobj.read, lacings) +- if map(len, self.packets) !=3D lacings: ++ self.packets =3D list(map(fileobj.read, lacings)) ++ if list(map(len, self.packets)) !=3D lacings: + raise error("unable to read full data") +=20 + def __eq__(self, other): +@@ -301,7 +301,7 @@ + if page.packets[-1]: + page.complete =3D False + if len(page.packets) =3D=3D 1: +- page.position =3D -1L ++ page.position =3D -1 + else: + page.packets.pop(-1) + pages.append(page) +@@ -334,7 +334,7 @@ +=20 + # Number the new pages starting from the first old page. + first =3D old_pages[0].sequence +- for page, seq in zip(new_pages, range(first, first + len(new_pages)= )): ++ for page, seq in zip(new_pages, list(range(first, first + len(new_p= ages)))): + page.sequence =3D seq + page.serial =3D old_pages[0].serial +=20 +@@ -346,7 +346,7 @@ + new_pages[-1].last =3D old_pages[-1].last + new_pages[-1].complete =3D old_pages[-1].complete + if not new_pages[-1].complete and len(new_pages[-1].packets) =3D=3D= 1: +- new_pages[-1].position =3D -1L ++ new_pages[-1].position =3D -1 +=20 + new_data =3D "".join(map(klass.write, new_pages)) +=20 +@@ -454,10 +454,10 @@ + denom =3D self.info.fps + self.info.length =3D samples / float(denom) +=20 +- except error, e: +- raise self._Error, e, sys.exc_info()[2] ++ except error as e: ++ raise self._Error(e).with_traceback(sys.exc_info()[2]) + except EOFError: +- raise self._Error, "no appropriate stream found" ++ raise self._Error("no appropriate stream found") + finally: + fileobj.close() +=20 +@@ -473,10 +473,10 @@ + fileobj =3D file(filename, "rb+") + try: + try: self.tags._inject(fileobj) +- except error, e: +- raise self._Error, e, sys.exc_info()[2] ++ except error as e: ++ raise self._Error(e).with_traceback(sys.exc_info()[2]) + except EOFError: +- raise self._Error, "no appropriate stream found" ++ raise self._Error("no appropriate stream found") + finally: + fileobj.close() +=20 +@@ -490,9 +490,9 @@ + fileobj =3D file(filename, "rb+") + try: + try: self.tags._inject(fileobj) +- except error, e: +- raise self._Error, e, sys.exc_info()[2] ++ except error as e: ++ raise self._Error(e).with_traceback(sys.exc_info()[2]) + except EOFError: +- raise self._Error, "no appropriate stream found" ++ raise self._Error("no appropriate stream found") + finally: + fileobj.close() +diff -Naur client175_0.7-original/mutagen/_util.py client175_0.7/mutagen/_ut= il.py +--- client175_0.7-original/mutagen/_util.py 2010-05-15 00:42:14.000000000 +0= 200 ++++ client175_0.7/mutagen/_util.py 2021-08-03 14:57:18.085567569 +0200 +@@ -32,7 +32,7 @@ + """ +=20 + def __iter__(self): +- return iter(self.keys()) ++ return iter(list(self.keys())) +=20 + def has_key(self, key): + try: self[key] +@@ -40,18 +40,18 @@ + else: return True + __contains__ =3D has_key +=20 +- iterkeys =3D lambda self: iter(self.keys()) ++ iterkeys =3D lambda self: iter(list(self.keys())) +=20 + def values(self): +- return map(self.__getitem__, self.keys()) +- itervalues =3D lambda self: iter(self.values()) ++ return list(map(self.__getitem__, list(self.keys()))) ++ itervalues =3D lambda self: iter(list(self.values())) +=20 + def items(self): +- return zip(self.keys(), self.values()) +- iteritems =3D lambda s: iter(s.items()) ++ return list(zip(list(self.keys()), list(self.values()))) ++ iteritems =3D lambda s: iter(list(s.items())) +=20 + def clear(self): +- map(self.__delitem__, self.keys()) ++ list(map(self.__delitem__, list(self.keys()))) +=20 + def pop(self, key, *args): + if len(args) > 1: +@@ -65,7 +65,7 @@ +=20 + def popitem(self): + try: +- key =3D self.keys()[0] ++ key =3D list(self.keys())[0] + return key, self.pop(key) + except IndexError: raise KeyError("dictionary is empty") +=20 +@@ -74,7 +74,7 @@ + self.update(kwargs) + other =3D {} +=20 +- try: map(self.__setitem__, other.keys(), other.values()) ++ try: list(map(self.__setitem__, list(other.keys()), list(other.valu= es()))) + except AttributeError: + for key, value in other: + self[key] =3D value +@@ -90,14 +90,14 @@ + except KeyError: return default +=20 + def __repr__(self): +- return repr(dict(self.items())) ++ return repr(dict(list(self.items()))) +=20 + def __cmp__(self, other): + if other is None: return 1 +- else: return cmp(dict(self.items()), other) ++ else: return cmp(dict(list(self.items())), other) +=20 + def __len__(self): +- return len(self.keys()) ++ return len(list(self.keys())) +=20 + class DictProxy(DictMixin): + def __init__(self, *args, **kwargs): +@@ -114,7 +114,7 @@ + del(self.__dict[key]) +=20 + def keys(self): +- return self.__dict.keys() ++ return list(self.__dict.keys()) +=20 + class cdata(object): + """C character buffer to Python numeric type conversions.""" +@@ -300,7 +300,7 @@ + """Convert a basestring to a valid UTF-8 str.""" + if isinstance(data, str): + return data.decode("utf-8", "replace").encode("utf-8") +- elif isinstance(data, unicode): ++ elif isinstance(data, str): + return data.encode("utf-8") + else: raise TypeError("only unicode/str types can be converted to UTF-8= ") +=20 +@@ -308,7 +308,7 @@ + try: + return d[key] + except KeyError: +- for pattern, value in d.iteritems(): ++ for pattern, value in d.items(): + if fnmatchcase(key, pattern): + return value + return default +diff -Naur client175_0.7-original/mutagen/_vorbis.py client175_0.7/mutagen/_= vorbis.py +--- client175_0.7-original/mutagen/_vorbis.py 2010-05-15 00:42:14.000000000 = +0200 ++++ client175_0.7/mutagen/_vorbis.py 2021-08-03 14:57:26.321443147 +0200 +@@ -16,7 +16,7 @@ +=20 + import sys +=20 +-from cStringIO import StringIO ++from io import StringIO +=20 + import mutagen + from mutagen._util import DictMixin, cdata +@@ -54,7 +54,7 @@ + vendor -- the stream 'vendor' (i.e. writer); default 'Mutagen' + """ +=20 +- vendor =3D u"Mutagen " + mutagen.version_string ++ vendor =3D "Mutagen " + mutagen.version_string +=20 + def __init__(self, data=3DNone, *args, **kwargs): + # Collect the args to pass to load, this lets child classes +@@ -90,16 +90,16 @@ + except (OverflowError, MemoryError): + raise error("cannot read %d bytes, too large" % length) + try: tag, value =3D string.split('=3D', 1) +- except ValueError, err: ++ except ValueError as err: + if errors =3D=3D "ignore": + continue + elif errors =3D=3D "replace": +- tag, value =3D u"unknown%d" % i, string ++ tag, value =3D "unknown%d" % i, string + else: +- raise VorbisEncodingError, str(err), sys.exc_info()= [2] ++ raise VorbisEncodingError(str(err)).with_traceback(= sys.exc_info()[2]) + try: tag =3D tag.encode('ascii', errors) + except UnicodeEncodeError: +- raise VorbisEncodingError, "invalid tag name %r" % tag ++ raise VorbisEncodingError("invalid tag name %r" % tag) + else: + if is_valid_key(tag): self.append((tag, value)) + if framing and not ord(fileobj.read(1)) & 0x01: +@@ -115,7 +115,7 @@ + any invalid keys or values are found, a ValueError is raised. + """ +=20 +- if not isinstance(self.vendor, unicode): ++ if not isinstance(self.vendor, str): + try: self.vendor.decode('utf-8') + except UnicodeDecodeError: raise ValueError +=20 +@@ -123,7 +123,7 @@ + try: + if not is_valid_key(key): raise ValueError + except: raise ValueError("%r is not a valid key" % key) +- if not isinstance(value, unicode): ++ if not isinstance(value, str): + try: value.encode("utf-8") + except: raise ValueError("%r is not a valid value" % value) + else: return True +@@ -181,15 +181,15 @@ + """ + key =3D key.lower().encode('ascii') + values =3D [value for (k, value) in self if k.lower() =3D=3D key] +- if not values: raise KeyError, key ++ if not values: raise KeyError(key) + else: return values +=20 + def __delitem__(self, key): + """Delete all values associated with the key.""" + key =3D key.lower().encode('ascii') +- to_delete =3D filter(lambda x: x[0].lower() =3D=3D key, self) +- if not to_delete:raise KeyError, key +- else: map(self.remove, to_delete) ++ to_delete =3D [x for x in self if x[0].lower() =3D=3D key] ++ if not to_delete:raise KeyError(key) ++ else: list(map(self.remove, to_delete)) +=20 + def __contains__(self, key): + """Return true if the key has any values.""" +@@ -220,4 +220,4 @@ +=20 + def as_dict(self): + """Return a copy of the comment data in a real dict.""" +- return dict((key, self[key]) for key in self.keys()) ++ return dict((key, self[key]) for key in list(self.keys())) +diff -Naur client175_0.7-original/server.py client175_0.7/server.py +--- client175_0.7-original/server.py 2011-04-06 13:18:04.000000000 +0200 ++++ client175_0.7/server.py 2021-08-03 14:40:26.132681859 +0200 +@@ -20,7 +20,7 @@ + # MA 02110-1301, USA. +=20 +=20 +-import cherrypy, json, os, pwd, urllib, urllib2, sys ++import cherrypy, json, os, pwd, urllib.request, urllib.parse, urllib.error,= urllib.request, urllib.error, urllib.parse, sys + from BeautifulSoup import BeautifulSoup + from time import sleep + from datetime import datetime, timedelta +@@ -55,7 +55,7 @@ + PASSWORD =3D None + RUN_AS =3D pwd.getpwuid(os.getuid())[0] +=20 +-if os.environ.has_key("MPD_HOST"): ++if "MPD_HOST" in os.environ: + mpd_host =3D str(os.environ["MPD_HOST"]) + if "@" in mpd_host: + mpd_host =3D mpd_host.split("@") +@@ -64,7 +64,7 @@ + else: + HOST =3D mpd_host +=20 +-if os.environ.has_key("MPD_PORT"): ++if "MPD_PORT" in os.environ: + PORT =3D int(os.environ["MPD_PORT"]) +=20 + HOST =3D cherrypy.config.get('mpd_host', HOST) +@@ -124,7 +124,7 @@ +=20 + def add(self, *args, **kwargs): + if len(kwargs) > 0: +- args =3D list(args) + kwargs.values() ++ args =3D list(args) + list(kwargs.values()) + if len(args) =3D=3D 2: + if args[0] in ('file', 'directory'): + d =3D args[1] +@@ -146,7 +146,7 @@ + if ext in ['mp3', 'pgg', 'wav', 'flac', 'aac', 'mod', 'wma'= ]: + mpd.add(d) + else: +- sock =3D urllib2.urlopen(d) ++ sock =3D urllib.request.urlopen(d) + data =3D sock.read() + info =3D sock.info() + mime =3D info.gettype() +@@ -201,13 +201,13 @@ + """ +=20 + if len(kwargs) > 0: +- args =3D list(args) + kwargs.values() ++ args =3D list(args) + list(kwargs.values()) + try: + if len(args) =3D=3D 1: + args =3D args[0] +- print args ++ print(args) + result =3D mpd.execute(args) +- except MPDError, e: ++ except MPDError as e: + raise cherrypy.HTTPError(501, message=3Dstr(e)) + return json.dumps(result) + default.exposed =3D True +@@ -231,7 +231,7 @@ + return "WAV editing not supported." +=20 + tags =3D {} +- for tag, val in kwargs.items(): ++ for tag, val in list(kwargs.items()): + tag =3D tag.lower() + if tag =3D=3D 'track': + tags['tracknumber'] =3D val +@@ -239,7 +239,7 @@ + tags['discnumber'] =3D val + else: + tags[tag] =3D val +- print '%s[%s] =3D "%s"' % (id, tag, val) ++ print('%s[%s] =3D "%s"' % (id, tag, val)) +=20 + f =3D metadata.get_format(loc) + f.write_tags(tags) +@@ -249,7 +249,7 @@ + try: + mpd.update(id) + updating =3D True +- except MPDError, e: ++ except MPDError as e: + if str(e) =3D=3D "[54(a)0] {update} already updating": + sleep(0.01) + else: +@@ -327,7 +327,7 @@ + d =3D [] + skip =3D ('type', 'time', 'ptime', 'songs') + for item in data: +- for key, val in item.items(): ++ for key, val in list(item.items()): + if key not in skip: + if filter in str(val).lower(): + d.append(item) +@@ -444,7 +444,7 @@ + """ + try: + return mpd.raw(cmd) +- except MPDError, e: ++ except MPDError as e: + raise cherrypy.HTTPError(501, message=3Dstr(e)) + protocol.exposed =3D True +=20 +@@ -514,9 +514,9 @@ + return json.dumps(s) + n =3D 0 + while n < 50: +- if mpd.state.get('uptime', '') <> client_uptime: ++ if mpd.state.get('uptime', '') !=3D client_uptime: + return json.dumps(mpd.state) +- if mpd.state.get('updating_db', '') <> client_updating_db: ++ if mpd.state.get('updating_db', '') !=3D client_updating_db: + return json.dumps(mpd.state) + sleep(0.1) + n +=3D 1 +@@ -649,11 +649,11 @@ + if sport is None: + sport =3D "8080" +=20 +- print "" +- print "=3D" * 60 +- print "Server Ready." +- print "Client175 is available at: http://%s%s:%s" % (shost, sport, SER= VER_ROOT) +- print "=3D" * 60 +- print "" ++ print("") ++ print("=3D" * 60) ++ print("Server Ready.") ++ print("Client175 is available at: http://%s%s:%s" % (shost, sport, SER= VER_ROOT)) ++ print("=3D" * 60) ++ print("") +=20 + serve() --=20 2.32.0 --===============0758129431786205935==--