Source code for coherence.upnp.core.utils

# Licensed under the MIT license
# http://opensource.org/licenses/mit-license.php

import xml.etree.ElementTree as ET
# Copyright (C) 2006 Fluendo, S.A. (www.fluendo.com).
# Copyright 2006, Frank Scholz <coherence@beebits.net>
from urllib.parse import urlsplit, urlparse

from twisted.internet import reactor, defer, abstract
from twisted.python import failure
from twisted.web import client
from twisted.web import http, static
from twisted.web import proxy, resource, server

from coherence import SERVER_ID
from coherence import log

logger = log.get_logger('utils')

try:
    from twisted.protocols._c_urlarg import unquote
except ImportError:
    from urllib.parse import unquote

try:
    import netifaces

    have_netifaces = True
except ImportError:
    have_netifaces = False


[docs]def means_true(value): if isinstance(value, str): value = value.lower() return value in [True, 1, '1', 'true', 'yes', 'ok']
[docs]def generalise_boolean(value): """ standardize the different boolean incarnations transform anything that looks like a "True" into a '1', and everything else into a '0' """ if means_true(value): return '1' return '0'
generalize_boolean = generalise_boolean
[docs]def parse_xml(data, encoding="utf-8", dump_invalid_data=False): parser = ET.XMLParser() # my version of twisted.web returns page_infos as a dictionary in # the second item of the data list # :fixme: This must be handled where twisted.web is fetching the data if isinstance(data, (list, tuple)): data = data[0] try: data = data.encode(encoding) except UnicodeDecodeError: pass # Guess from who we're getting this? data = data.replace(b'\x00', b'') try: parser.feed(data) except Exception as error: if dump_invalid_data: print(error, repr(data)) parser.close() raise else: return ET.ElementTree(parser.close())
[docs]def parse_http_response(data): """ don't try to get the body, there are reponses without """ if isinstance(data, bytes): data = data.decode('utf-8') header = data.split('\r\n\r\n')[0] lines = header.split('\r\n') cmd = lines[0].split(' ') lines = [x.replace(': ', ':', 1) for x in lines[1:]] lines = [x for x in lines if len(x) > 0] headers = [x.split(':', 1) for x in lines] headers = dict([(x[0].lower(), x[1]) for x in headers]) return cmd, headers
[docs]def get_ip_address(ifname): """ determine the IP address by interface name http://aspn.activestate.com/ASPN/Cookbook/Python/Recipe/439094 (c) Paul Cannon Uses the Linux SIOCGIFADDR ioctl to find the IP address associated with a network interface, given the name of that interface, e.g. "eth0". The address is returned as a string containing a dotted quad. Updated to work on BSD. OpenBSD and OSX share the same value for SIOCGIFADDR, and its likely that other BSDs do too. Updated to work on Windows, using the optional Python module netifaces http://alastairs-place.net/netifaces/ Thx Lawrence for that patch! """ if have_netifaces: if ifname in netifaces.interfaces(): iface = netifaces.ifaddresses(ifname) ifaceadr = iface[netifaces.AF_INET] # we now have a list of address dictionaries, # there may be multiple addresses bound return ifaceadr[0]['addr'] import sys if sys.platform in ('win32', 'sunos5'): return '127.0.0.1' from os import uname import socket import fcntl import struct system_type = uname()[0] if system_type == "Linux": SIOCGIFADDR = 0x8915 else: SIOCGIFADDR = 0xc0206921 s = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) try: ip = socket.inet_ntoa(fcntl.ioctl( s.fileno(), SIOCGIFADDR, struct.pack(b'256s', ifname[:15].encode('ascii')) )[20:24]) except Exception: ip = '127.0.0.1' # print('ip is: {}'.format(ip)) return ip
[docs]def get_host_address(): """ try to get determine the interface used for the default route, as this is most likely the interface we should bind to (on a single homed host!) """ import sys if sys.platform == 'win32': if have_netifaces: interfaces = netifaces.interfaces() if len(interfaces): # on windows assume first interface is primary return get_ip_address(interfaces[0]) else: try: route_file = '/proc/net/route' route = open(route_file) if route: tmp = route.readline() # skip first line while tmp != '': tmp = route.readline() li = tmp.split('\t') if (len(li) > 2): if li[1] == '00000000': # default route... route.close() return get_ip_address(li[0]) except IOError as msg: """ fallback to parsing the output of netstat """ from twisted.internet import utils def result(r): from os import uname (osname, _, _, _, _) = uname() osname = osname.lower() lines = r.split('\n') for li in lines: li = li.strip(' \r\n') parts = [x.strip() for x in li.split(' ') if len(x) > 0] if parts[0] in ('0.0.0.0', 'default'): if osname[:6] == 'darwin': return get_ip_address(parts[5]) else: return get_ip_address(parts[-1]) return '127.0.0.1' def fail(f): return '127.0.0.1' d = utils.getProcessOutput('netstat', ['-rn']) d.addCallback(result) d.addErrback(fail) return d except Exception as msg: import traceback traceback.print_exc() """ return localhost if we haven't found anything """ return '127.0.0.1'
[docs]def de_chunk_payload(response): import io """ This method takes a chunked HTTP data object and unchunks it.""" newresponse = io.StringIO() # chunked encoding consists of a bunch of lines with # a length in hex followed by a data chunk and a CRLF pair. response = io.StringIO(response) def read_chunk_length(): line = response.readline() try: len = int(line.strip(), 16) except ValueError: len = 0 return len len = read_chunk_length() while (len > 0): newresponse.write(response.read(len)) line = response.readline() # after chunk and before next chunk length len = read_chunk_length() return newresponse.getvalue()
[docs]class Request(server.Request):
[docs] def process(self): "Process a request." # get site from channel self.site = self.channel.site # set various default headers self.setHeader(b'server', SERVER_ID.encode('ascii')) self.setHeader(b'date', http.datetimeToString()) self.setHeader(b'content-type', b"text/html") # Resource Identification url = self.path if isinstance(url, bytes): url = url.decode('utf-8') # remove trailing "/", if ever url = url.rstrip('/') scheme, netloc, path, query, fragment = urlsplit(url) clean_path = path[1:] self.prepath = [] if path == "": self.postpath = [] else: raw_p = list(map(unquote, clean_path.split('/'))) self.postpath = list(i.encode('ascii') for i in raw_p) try: def deferred_rendering(r): if isinstance(r, str): r = r.encode('ascii') self.render(r) resrc = self.site.getResourceFor(self) if resrc is None: self.setResponseCode( http.NOT_FOUND, "Error: No resource for path {}".format( path).encode('ascii')) self.finish() elif isinstance(resrc, defer.Deferred): resrc.addCallback(deferred_rendering) resrc.addErrback(self.processingFailed) else: if isinstance(resrc, str): resrc = resrc.encode('ascii') self.render(resrc) except Exception as e: logger.error('Error on render Request: {}'.format(e)) self.processingFailed(failure.Failure())
[docs]class Site(server.Site): noisy = False requestFactory = Request
[docs] def startFactory(self): pass
# http._logDateTimeStart()
[docs]class ProxyClient(proxy.ProxyClient, log.LogAble): def __init__(self, command, rest, version, headers, data, father): log.LogAble.__init__(self) # headers["connection"] = "close" self.send_data = 0 proxy.ProxyClient.__init__(self, command, rest, version, headers, data, father)
[docs] def handleStatus(self, version, code, message): if message: # Add a whitespace to message, this allows empty messages # transparently message = " %s" % (message,) if version == 'ICY': version = 'HTTP/1.1' proxy.ProxyClient.handleStatus(self, version, code, message)
[docs] def handleHeader(self, key, value): if not key.startswith('icy-'): proxy.ProxyClient.handleHeader(self, key, value)
[docs] def handleResponsePart(self, buffer): self.send_data += len(buffer) proxy.ProxyClient.handleResponsePart(self, buffer)
[docs]class ProxyClientFactory(proxy.ProxyClientFactory): # :fixme: Why here proxy.ProxyClient is used instad of our own # ProxyClent? Is out ProxyClient used at all? protocol = proxy.ProxyClient
[docs]class ReverseProxyResource(proxy.ReverseProxyResource): """ Resource that renders the results gotten from another server Put this resource in the tree to cause everything below it to be relayed to a different server. @ivar proxyClientFactoryClass: a proxy client factory class, used to create new connections. @type proxyClientFactoryClass: L{ClientFactory} @ivar reactor: the reactor used to create connections. @type reactor: object providing L{twisted.internet.interfaces.IReactorTCP} """ proxyClientFactoryClass = ProxyClientFactory
[docs] def __init__(self, host, port, path, reactor=reactor): """ @param host: the host of the web server to proxy. @type host: C{str} @param port: the port of the web server to proxy. @type port: C{port} @param path: the base path to fetch data from. Note that you shouldn't put any trailing slashes in it, it will be added automatically in request. For example, if you put B{/foo}, a request on B{/bar} will be proxied to B{/foo/bar}. @type path: C{str} """ resource.Resource.__init__(self) self.host = host self.port = port self.path = path self.qs = '' self.reactor = reactor
[docs] def getChild(self, path, request): return ReverseProxyResource( self.host, self.port, self.path + '/' + path)
[docs] def render(self, request): """ Render a request by forwarding it to the proxied server. """ # RFC 2616 tells us that we can omit the port if it's the default port, # but we have to provide it otherwise if self.port == 80: request.received_headers['host'] = self.host else: request.received_headers['host'] = "%s:%d" % (self.host, self.port) request.content.seek(0, 0) qs = urlparse(request.uri)[4] if qs == '': qs = self.qs if qs: rest = self.path + '?' + qs else: rest = self.path clientFactory = self.proxyClientFactoryClass( request.method, rest, request.clientproto, request.getAllHeaders(), request.content.read(), request) self.reactor.connectTCP(self.host, self.port, clientFactory) return server.NOT_DONE_YET
[docs] def resetTarget(self, host, port, path, qs=''): self.host = host self.port = port self.path = path self.qs = qs
[docs]class ReverseProxyUriResource(ReverseProxyResource): uri = None def __init__(self, uri, reactor=reactor): self.uri = uri _, host_port, path, params, _ = urlsplit(uri) if host_port.find(':') != -1: host, port = tuple(host_port.split(':')) port = int(port) else: host = host_port port = 80 if path == '': path = '/' if params == '': rest = path else: rest = '?'.join((path, params)) ReverseProxyResource.__init__(self, host, port, rest, reactor)
[docs] def resetUri(self, uri): self.uri = uri _, host_port, path, params, _ = urlsplit(uri) if host_port.find(':') != -1: host, port = tuple(host_port.split(':')) port = int(port) else: host = host_port port = 80 self.resetTarget(host, port, path, params)
# already on twisted.web since at least 8.0.0
[docs]class myHTTPPageGetter(client.HTTPPageGetter): followRedirect = True
[docs]class HeaderAwareHTTPClientFactory(client.HTTPClientFactory): protocol = myHTTPPageGetter noisy = False
[docs] def buildProtocol(self, addr): p = client.HTTPClientFactory.buildProtocol(self, addr) p.method = self.method p.followRedirect = self.followRedirect return p
[docs] def page(self, page): client.HTTPClientFactory.page(self, (page, self.response_headers))
# deprecated, do not use # already in twisted.web since at least 1.3.0 HeaderAwareHTTPDownloader = client.HTTPDownloader
[docs]def getPage(url, contextFactory=None, *args, **kwargs): """ Download a web page as a string. Download a page. Return a deferred, which will callback with a page (as a string) or errback with a description of the error. See HTTPClientFactory to see what extra args can be passed. """ # This function is like twisted.web.client.getPage, except it uses # our HeaderAwareHTTPClientFactory instead of HTTPClientFactory # and sets the user agent. url_bytes = url if not isinstance(url, bytes): url_bytes = url.encode('ascii') if args is None: args = [] if kwargs is None: kwargs = {} if 'headers' in kwargs and 'user-agent' in kwargs['headers']: kwargs['agent'] = kwargs['headers']['user-agent'] elif 'agent' not in kwargs: kwargs['agent'] = "Coherence PageGetter" new_kwargs = {} for k, v in kwargs.items(): if k == 'headers': new_kwargs[k] = {} for kh, vh in kwargs['headers'].items(): h_key = kh if isinstance(kh, bytes) else kh.encode('ascii') h_val = vh if isinstance(kh, bytes) else vh.encode('ascii') new_kwargs['headers'][h_key] = h_val else: new_kwargs[k] = v logger.info('getPage [url]: {} [type: {}]'.format(url, type(url))) logger.debug('\t->[args]: {} [type: {}]'.format(args, type(args))) logger.debug('\t->[kwargs]: {}'.format(kwargs)) logger.debug('\t->[new_kwargs]: {}]'.format(new_kwargs)) return client._makeGetterFactory( url_bytes, HeaderAwareHTTPClientFactory, contextFactory=contextFactory, *args, **new_kwargs).deferred
[docs]def downloadPage(url, file, contextFactory=None, *args, **kwargs): """Download a web page to a file. @param file: path to file on filesystem, or file-like object. See twisted.web.client.HTTPDownloader to see what extra args can be passed. """ url_bytes = url if not isinstance(url, bytes): url_bytes = url.encode('ascii') if 'headers' in kwargs and 'user-agent' in kwargs['headers']: kwargs['agent'] = kwargs['headers']['user-agent'] elif 'agent' not in kwargs: kwargs['agent'] = "Coherence PageGetter" new_kwargs = {} for k, v in kwargs.items(): if k == 'headers': new_kwargs[k] = {} for kh, vh in kwargs['headers'].items(): h_key = kh if isinstance(kh, bytes) else kh.encode('ascii') h_val = vh if isinstance(kh, bytes) else vh.encode('ascii') new_kwargs['headers'][h_key] = h_val else: new_kwargs[k] = v logger.info('downloadPage [url]: {} [type: {}]'.format(url, type(url))) logger.debug('\t->[args]: {} [type: {}]'.format(args, type(args))) logger.debug('\t->[kwargs]: {}'.format(kwargs)) logger.debug('\t->[new_kwargs]: {}]'.format(new_kwargs)) return client.downloadPage( url_bytes, file, contextFactory=contextFactory, *args, **new_kwargs)
# StaticFile used to be a patched version of static.File. The later # was fixed in TwistedWeb 8.2.0 and 9.0.0, while the patched variant # contained deprecated and removed code. StaticFile = static.File
[docs]class BufferFile(static.File): """ taken from twisted.web.static and modified accordingly to the patch by John-Mark Gurney http://resnet.uoregon.edu/~gurney_j/jmpc/dist/twisted.web.static.patch """ def __init__(self, path, target_size=0, *args): static.File.__init__(self, path, *args) self.target_size = target_size self.upnp_retry = None
[docs] def render(self, request): # print "" # print "BufferFile", request # FIXME detect when request is REALLY finished if request is None or request.finished: logger.info("No request to render!") return '' """You know what you doing.""" self.restat() if self.type is None: self.type, self.encoding = static.getTypeAndEncoding( self.basename(), self.contentTypes, self.contentEncodings, self.defaultType) if not self.exists(): return self.childNotFound.render(request) if self.isdir(): return self.redirect(request) # for content-length if (self.target_size > 0): fsize = size = int(self.target_size) else: fsize = size = int(self.getFileSize()) # print fsize if size == int(self.getFileSize()): request.setHeader(b'accept-ranges', b'bytes') if self.type: request.setHeader( b'content-type', self.type if isinstance( self.type, bytes) else self.type.encode('ascii')) if self.encoding: request.setHeader( b'content-encoding', self.encoding if isinstance( self.encoding, bytes) else self.encoding.encode('ascii')) try: f = self.openForReading() except IOError as e: import errno if e.errno == errno.EACCES: return resource.ForbiddenResource().render(request) else: raise if request.setLastModified(self.getmtime()) is http.CACHED: return '' trans = True range = request.getHeader('range') # print "StaticFile", range tsize = size if range is not None: # This is a request for partial data... bytesrange = range.split('=') assert bytesrange[0] == 'bytes', \ "Syntactically invalid http range header!" start, end = bytesrange[1].split('-', 1) if start: start = int(start) # Are we requesting something # beyond the current size of the file? if (start >= self.getFileSize()): # Retry later! logger.info(bytesrange) logger.info( "Requesting data beyond current scope -> " "postpone rendering!") self.upnp_retry = reactor.callLater( 1.0, self.render, request) return server.NOT_DONE_YET f.seek(start) if end: # print(":%s" % end) end = int(end) else: end = size - 1 else: lastbytes = int(end) if size < lastbytes: lastbytes = size start = size - lastbytes f.seek(start) fsize = lastbytes end = size - 1 size = end + 1 fsize = end - int(start) + 1 # start is the byte offset to begin, and end is the byte offset # to end.. fsize is size to send, tsize is the real size of # the file, and size is the byte position to stop sending. if fsize <= 0: request.setResponseCode(http.REQUESTED_RANGE_NOT_SATISFIABLE) fsize = tsize trans = False else: request.setResponseCode(http.PARTIAL_CONTENT) request.setHeader(b'content-range', ("bytes %s-%s/%s " % ( str(start), str(end), str(tsize))).encode('ascii')) # print "StaticFile", start, end, tsize request.setHeader('content-length', str(fsize)) if request.method == b'HEAD' or trans is False: # pretend we're a HEAD request, so content-length # won't be overwritten. request.method = b'HEAD' return '' # print "StaticFile out", request.headers, request.code # return data # size is the byte position to stop sending, not how many bytes to send BufferFileTransfer(f, size - f.tell(), request) # and make sure the connection doesn't get closed return server.NOT_DONE_YET
[docs]class BufferFileTransfer(object): """ A class to represent the transfer of a file over the network. """ request = None def __init__(self, file, size, request): self.file = file self.size = size self.request = request self.written = self.file.tell() request.registerProducer(self, 0)
[docs] def resumeProducing(self): # print "resumeProducing", self.request,self.size,self.written if not self.request: return data = self.file.read( min(abstract.FileDescriptor.bufferSize, self.size - self.written)) if data: self.written += len(data) # this .write will spin the reactor, calling .doWrite and then # .resumeProducing again, so be prepared for a re-entrant call self.request.write(data) if self.request and self.file.tell() == self.size: self.request.unregisterProducer() self.request.finish() self.request = None
[docs] def pauseProducing(self): pass
[docs] def stopProducing(self): # print "stopProducing",self.request self.request.unregisterProducer() self.file.close() self.request.finish() self.request = None
from datetime import datetime, tzinfo, timedelta import random
[docs]class _tz(tzinfo):
[docs] def utcoffset(self, dt): return self._offset
[docs] def tzname(self, dt): return self._name
[docs] def dst(self, dt): return timedelta(0)
[docs]class _CET(_tz): _offset = timedelta(minutes=60) _name = 'CET'
[docs]class _CEST(_tz): _offset = timedelta(minutes=120) _name = 'CEST'
_bdates = [datetime(1997, 2, 28, 17, 20, tzinfo=_CET()), # Sebastian Oliver datetime(1999, 9, 19, 4, 12, tzinfo=_CEST()), # Patrick Niklas datetime(2000, 9, 23, 4, 8, tzinfo=_CEST()), # Saskia Alexa datetime(2003, 7, 23, 1, 18, tzinfo=_CEST()), # Mara Sophie # you are the best! ]
[docs]def datefaker(): return random.choice(_bdates)
[docs]def cmp(x, y): """ Replacement for built-in function cmp that was removed in Python 3 Compare the two objects x and y and return an integer according to the outcome. The return value is negative if x < y, zero if x == y and strictly positive if x > y. """ return (x > y) - (x < y)