| desc
				 stringlengths 3 26.7k | decl
				 stringlengths 11 7.89k | bodies
				 stringlengths 8 553k | 
|---|---|---|
| 
	'__init__: This has to be present'
 | 
	def __init__(self):
 | 
	    self.a = 1
    def XXX22():
        'XXX22:    This    has    to    be    present'
        pass
 | 
| 
	'XXX22: This has to be present'
 | 
	def XXX11():
 | 
	    pass
 | 
| 
	'Remove recursive references to allow garbage
collector to collect this object.'
 | 
	def cleanup(self):
 | 
	    for dict in (self.rule2func, self.rules, self.rule2name):
        for i in dict.keys():
            dict[i] = None
    for i in dir(self):
        setattr(self, i, None)
 | 
| 
	'Disassemble a code object, returning a list of \'Token\'.
The main part of this procedure is modelled after
dis.disassemble().'
 | 
	def disassemble(self, co, classname=None, deob=0):
 | 
	    rv = []
    customize = {}
    Token = self.Token
    self.code = array('B', co.co_code)
    linestarts = list(dis.findlinestarts(co))
    varnames = list(co.co_varnames)
    if deob:
        linestarts = self.deobfuscate(co, linestarts, varnames)
    code = self.code
    n = len(code)
    self.prev = [0]
    for i in self.op_range(0, n):
        op = code[i]
        self.prev.append(i)
        if (op >= HAVE_ARGUMENT):
            self.prev.append(i)
            self.prev.append(i)
    self.lines = []
    linetuple = namedtuple('linetuple', ['l_no', 'next'])
    j = 0
    linestartoffsets = {a for (a, _) in linestarts}
    (prev_start_byte, prev_line_no) = linestarts[0]
    for (start_byte, line_no) in linestarts[1:]:
        while (j < start_byte):
            self.lines.append(linetuple(prev_line_no, start_byte))
            j += 1
        last_op = code[self.prev[start_byte]]
        (prev_start_byte, prev_line_no) = (start_byte, line_no)
    while (j < n):
        self.lines.append(linetuple(prev_line_no, n))
        j += 1
    if classname:
        classname = (('_' + classname.lstrip('_')) + '__')
        def unmangle(name):
            if (name.startswith(classname) and (name[(-2):] != '__')):
                return name[(len(classname) - 2):]
            return name
        free = [unmangle(name) for name in (co.co_cellvars + co.co_freevars)]
        names = [unmangle(name) for name in co.co_names]
        varnames = [unmangle(name) for name in varnames]
    else:
        free = (co.co_cellvars + co.co_freevars)
        names = co.co_names
    self.load_asserts = set()
    for i in self.op_range(0, n):
        if ((code[i] == PJIT) and (code[(i + 3)] == LOAD_GLOBAL)):
            if (names[(code[(i + 4)] + (256 * code[(i + 5)]))] == 'AssertionError'):
                self.load_asserts.add((i + 3))
    cf = self.find_jump_targets(code)
    last_stmt = self.next_stmt[0]
    i = self.next_stmt[last_stmt]
    replace = {}
    while (i < (n - 1)):
        if (self.lines[last_stmt].next > i):
            if (code[last_stmt] == PRINT_ITEM):
                if (code[i] == PRINT_ITEM):
                    replace[i] = 'PRINT_ITEM_CONT'
                elif (code[i] == PRINT_NEWLINE):
                    replace[i] = 'PRINT_NEWLINE_CONT'
        last_stmt = i
        i = self.next_stmt[i]
    imports = self.all_instr(0, n, (IMPORT_NAME, IMPORT_FROM, IMPORT_STAR))
    if (len(imports) > 1):
        last_import = imports[0]
        for i in imports[1:]:
            if (self.lines[last_import].next > i):
                if (code[last_import] == IMPORT_NAME == code[i]):
                    replace[i] = 'IMPORT_NAME_CONT'
            last_import = i
    extended_arg = 0
    for offset in self.op_range(0, n):
        if (offset in cf):
            k = 0
            for j in cf[offset]:
                rv.append(Token('COME_FROM', None, repr(j), offset=('%s_%d' % (offset, k))))
                k += 1
        op = code[offset]
        opname = dis.opname[op]
        oparg = None
        pattr = None
        if (op >= HAVE_ARGUMENT):
            oparg = ((code[(offset + 1)] + (code[(offset + 2)] * 256)) + extended_arg)
            extended_arg = 0
            if (op == dis.EXTENDED_ARG):
                extended_arg = (oparg * 65536L)
                continue
            if (op in dis.hasconst):
                const = co.co_consts[oparg]
                if (type(const) == types.CodeType):
                    oparg = const
                    if (const.co_name == '<lambda>'):
                        assert (opname == 'LOAD_CONST')
                        opname = 'LOAD_LAMBDA'
                    elif (const.co_name == '<genexpr>'):
                        opname = 'LOAD_GENEXPR'
                    elif (const.co_name == '<dictcomp>'):
                        opname = 'LOAD_DICTCOMP'
                    elif (const.co_name == '<setcomp>'):
                        opname = 'LOAD_SETCOMP'
                    pattr = (('<code_object    ' + const.co_name) + '>')
                else:
                    pattr = const
            elif (op in dis.hasname):
                pattr = names[oparg]
            elif (op in dis.hasjrel):
                pattr = repr(((offset + 3) + oparg))
            elif (op in dis.hasjabs):
                pattr = repr(oparg)
            elif (op in dis.haslocal):
                pattr = varnames[oparg]
            elif (op in dis.hascompare):
                pattr = dis.cmp_op[oparg]
            elif (op in dis.hasfree):
                pattr = free[oparg]
        if (op in (BUILD_LIST, BUILD_TUPLE, BUILD_SET, BUILD_SLICE, UNPACK_SEQUENCE, MAKE_FUNCTION, CALL_FUNCTION, MAKE_CLOSURE, CALL_FUNCTION_VAR, CALL_FUNCTION_KW, CALL_FUNCTION_VAR_KW, DUP_TOPX, RAISE_VARARGS)):
            if ((op == BUILD_TUPLE) and (code[self.prev[offset]] == LOAD_CLOSURE)):
                continue
            else:
                opname = ('%s_%d' % (opname, oparg))
                if (op != BUILD_SLICE):
                    customize[opname] = oparg
        elif (op == JA):
            target = self.get_target(offset)
            if (target < offset):
                if ((offset in self.stmts) and (code[(offset + 3)] not in (END_FINALLY, POP_BLOCK)) and (offset not in self.not_continue)):
                    opname = 'CONTINUE'
                else:
                    opname = 'JUMP_BACK'
        elif (op == LOAD_GLOBAL):
            if (offset in self.load_asserts):
                opname = 'LOAD_ASSERT'
        elif (op == RETURN_VALUE):
            if (offset in self.return_end_ifs):
                opname = 'RETURN_END_IF'
        if (offset not in replace):
            rv.append(Token(opname, oparg, pattr, offset, linestart=(offset in linestartoffsets)))
        else:
            rv.append(Token(replace[offset], oparg, pattr, offset, linestart=(offset in linestartoffsets)))
    if self.showasm:
        out = self.out
        for t in rv:
            print >>out, t
        print >>out
    return (rv, customize)
 | 
| 
	'Find the first <instr> in the block from start to end.
<instr> is any python bytecode instruction or a list of opcodes
If <instr> is an opcode with a target (like a jump), a target
destination can be specified which must match precisely if exact
is True, or if exact is False, the instruction which has a target
closest to <target> will be returned.
Return index to it or None if not found.'
 | 
	def first_instr(self, start, end, instr, target=None, exact=True):
 | 
	    code = self.code
    assert ((start >= 0) and (end <= len(code)))
    try:
        (None in instr)
    except:
        instr = [instr]
    pos = None
    distance = len(code)
    for i in self.op_range(start, end):
        op = code[i]
        if (op in instr):
            if (target is None):
                return i
            dest = self.get_target(i, op)
            if (dest == target):
                return i
            elif (not exact):
                _distance = abs((target - dest))
                if (_distance < distance):
                    distance = _distance
                    pos = i
    return pos
 | 
| 
	'Find the last <instr> in the block from start to end.
<instr> is any python bytecode instruction or a list of opcodes
If <instr> is an opcode with a target (like a jump), a target
destination can be specified which must match precisely if exact
is True, or if exact is False, the instruction which has a target
closest to <target> will be returned.
Return index to it or None if not found.'
 | 
	def last_instr(self, start, end, instr, target=None, exact=True):
 | 
	    code = self.code
    if (not ((start >= 0) and (end <= len(code)))):
        return None
    try:
        (None in instr)
    except:
        instr = [instr]
    pos = None
    distance = len(code)
    for i in self.op_range(start, end):
        op = code[i]
        if (op in instr):
            if (target is None):
                pos = i
            else:
                dest = self.get_target(i, op)
                if (dest == target):
                    distance = 0
                    pos = i
                elif (not exact):
                    _distance = abs((target - dest))
                    if (_distance <= distance):
                        distance = _distance
                        pos = i
    return pos
 | 
| 
	'Find all <instr> in the block from start to end.
<instr> is any python bytecode instruction or a list of opcodes
If <instr> is an opcode with a target (like a jump), a target
destination can be specified which must match precisely.
Return a list with indexes to them or [] if none found.'
 | 
	def all_instr(self, start, end, instr, target=None, include_beyond_target=False):
 | 
	    code = self.code
    assert ((start >= 0) and (end <= len(code)))
    try:
        (None in instr)
    except:
        instr = [instr]
    result = []
    for i in self.op_range(start, end):
        op = code[i]
        if (op in instr):
            if (target is None):
                result.append(i)
            else:
                t = self.get_target(i, op)
                if (include_beyond_target and (t >= target)):
                    result.append(i)
                elif (t == target):
                    result.append(i)
    return result
 | 
| 
	'Find all <instr> in the block from start to end.
<instr> is any python bytecode instruction or a list of opcodes
If <instr> is an opcode with a target (like a jump), a target
destination can be specified which must match precisely.
Return a list with indexes to them or [] if none found.'
 | 
	def rem_or(self, start, end, instr, target=None, include_beyond_target=False):
 | 
	    code = self.code
    assert ((start >= 0) and (end <= len(code)))
    try:
        (None in instr)
    except:
        instr = [instr]
    result = []
    for i in self.op_range(start, end):
        op = code[i]
        if (op in instr):
            if (target is None):
                result.append(i)
            else:
                t = self.get_target(i, op)
                if (include_beyond_target and (t >= target)):
                    result.append(i)
                elif (t == target):
                    result.append(i)
    pjits = self.all_instr(start, end, PJIT)
    filtered = []
    for pjit in pjits:
        tgt = (self.get_target(pjit) - 3)
        for i in result:
            if ((i <= pjit) or (i >= tgt)):
                filtered.append(i)
        result = filtered
        filtered = []
    return result
 | 
| 
	'Return the next jump that was generated by an except SomeException:
construct in a try...except...else clause or None if not found.'
 | 
	def next_except_jump(self, start):
 | 
	    if (self.code[start] == DUP_TOP):
        except_match = self.first_instr(start, len(self.code), POP_JUMP_IF_FALSE)
        if except_match:
            jmp = self.prev[self.get_target(except_match)]
            self.ignore_if.add(except_match)
            self.not_continue.add(jmp)
            return jmp
    count_END_FINALLY = 0
    count_SETUP_ = 0
    for i in self.op_range(start, len(self.code)):
        op = self.code[i]
        if (op == END_FINALLY):
            if (count_END_FINALLY == count_SETUP_):
                assert (self.code[self.prev[i]] in (JA, JF, RETURN_VALUE))
                self.not_continue.add(self.prev[i])
                return self.prev[i]
            count_END_FINALLY += 1
        elif (op in (SETUP_EXCEPT, SETUP_WITH, SETUP_FINALLY)):
            count_SETUP_ += 1
 | 
| 
	'Restrict pos to parent boundaries.'
 | 
	def restrict_to_parent(self, target, parent):
 | 
	    if (not (parent['start'] < target < parent['end'])):
        target = parent['end']
    return target
 | 
| 
	'Detect structures and their boundaries to fix optimizied jumps
in python2.3+'
 | 
	def detect_structure(self, pos, op=None):
 | 
	    code = self.code
    if (op is None):
        op = code[pos]
    parent = self.structs[0]
    start = parent['start']
    end = parent['end']
    for s in self.structs:
        _start = s['start']
        _end = s['end']
        if ((_start <= pos < _end) and ((_start >= start) and (_end <= end))):
            start = _start
            end = _end
            parent = s
    origStructCount = len(self.structs)
    if (op == SETUP_LOOP):
        start = (pos + 3)
        target = self.get_target(pos, op)
        end = self.restrict_to_parent(target, parent)
        if (target != end):
            self.fixed_jumps[pos] = end
        (line_no, next_line_byte) = self.lines[pos]
        jump_back = self.last_instr(start, end, JA, next_line_byte, False)
        if (jump_back and (jump_back != self.prev[end]) and (code[(jump_back + 3)] in (JA, JF))):
            if ((code[self.prev[end]] == RETURN_VALUE) or ((code[self.prev[end]] == POP_BLOCK) and (code[self.prev[self.prev[end]]] == RETURN_VALUE))):
                jump_back = None
        if (not jump_back):
            jump_back = (self.last_instr(start, end, RETURN_VALUE) + 1)
            if (not jump_back):
                return
            if (code[self.prev[next_line_byte]] not in (PJIF, PJIT)):
                loop_type = 'for'
            else:
                loop_type = 'while'
                self.ignore_if.add(self.prev[next_line_byte])
            target = next_line_byte
            end = (jump_back + 3)
        else:
            if (self.get_target(jump_back) >= next_line_byte):
                jump_back = self.last_instr(start, end, JA, start, False)
            if ((end > (jump_back + 4)) and (code[end] in (JF, JA))):
                if (code[(jump_back + 4)] in (JA, JF)):
                    if (self.get_target((jump_back + 4)) == self.get_target(end)):
                        self.fixed_jumps[pos] = (jump_back + 4)
                        end = (jump_back + 4)
            elif (target < pos):
                self.fixed_jumps[pos] = (jump_back + 4)
                end = (jump_back + 4)
            target = self.get_target(jump_back, JA)
            if (code[target] in (FOR_ITER, GET_ITER)):
                loop_type = 'for'
            else:
                loop_type = 'while'
                test = self.prev[next_line_byte]
                if (test == pos):
                    loop_type = 'while    1'
                else:
                    self.ignore_if.add(test)
                    test_target = self.get_target(test)
                    if (test_target > (jump_back + 3)):
                        jump_back = test_target
            self.not_continue.add(jump_back)
        self.loops.append(target)
        self.structs.append({'type': (loop_type + '-loop'), 'start': target, 'end': jump_back})
        if ((jump_back + 3) != end):
            self.structs.append({'type': (loop_type + '-else'), 'start': (jump_back + 3), 'end': end})
    elif (op == SETUP_EXCEPT):
        start = (pos + 3)
        target = self.get_target(pos, op)
        end = self.restrict_to_parent(target, parent)
        if (target != end):
            self.fixed_jumps[pos] = end
        self.structs.append({'type': 'try', 'start': start, 'end': (end - 4)})
        end_else = start_else = self.get_target(self.prev[end])
        i = end
        while (self.code[i] != END_FINALLY):
            jmp = self.next_except_jump(i)
            if (self.code[jmp] == RETURN_VALUE):
                self.structs.append({'type': 'except', 'start': i, 'end': (jmp + 1)})
                i = (jmp + 1)
            else:
                if (self.get_target(jmp) != start_else):
                    end_else = self.get_target(jmp)
                if (self.code[jmp] == JF):
                    self.fixed_jumps[jmp] = (-1)
                self.structs.append({'type': 'except', 'start': i, 'end': jmp})
                i = (jmp + 3)
        if (end_else != start_else):
            r_end_else = self.restrict_to_parent(end_else, parent)
            self.structs.append({'type': 'try-else', 'start': (i + 1), 'end': r_end_else})
            self.fixed_jumps[i] = r_end_else
        else:
            self.fixed_jumps[i] = (i + 1)
    elif (op in (PJIF, PJIT)):
        start = (pos + 3)
        target = self.get_target(pos, op)
        rtarget = self.restrict_to_parent(target, parent)
        pre = self.prev
        if ((target != rtarget) and (parent['type'] == 'and/or')):
            self.fixed_jumps[pos] = rtarget
            return
        if ((code[pre[target]] in (JUMP_IF_FALSE_OR_POP, JUMP_IF_TRUE_OR_POP, PJIF, PJIT)) and (target > pos)):
            self.fixed_jumps[pos] = pre[target]
            self.structs.append({'type': 'and/or', 'start': start, 'end': pre[target]})
            return
        if (op == PJIF):
            match = self.rem_or(start, self.next_stmt[pos], PJIF, target)
            match = self.remove_mid_line_ifs(match)
            if match:
                if ((code[pre[rtarget]] in (JF, JA)) and (pre[rtarget] not in self.stmts) and (self.restrict_to_parent(self.get_target(pre[rtarget]), parent) == rtarget)):
                    if ((code[pre[pre[rtarget]]] == JA) and self.remove_mid_line_ifs([pos]) and (target == self.get_target(pre[pre[rtarget]])) and ((pre[pre[rtarget]] not in self.stmts) or (self.get_target(pre[pre[rtarget]]) > pre[pre[rtarget]])) and (1 == len(self.remove_mid_line_ifs(self.rem_or(start, pre[pre[rtarget]], (PJIF, PJIT), target))))):
                        pass
                    elif ((code[pre[pre[rtarget]]] == RETURN_VALUE) and self.remove_mid_line_ifs([pos]) and (1 == len((set(self.remove_mid_line_ifs(self.rem_or(start, pre[pre[rtarget]], (PJIF, PJIT), target))) | set(self.remove_mid_line_ifs(self.rem_or(start, pre[pre[rtarget]], (PJIF, PJIT, JA), pre[rtarget], True))))))):
                        pass
                    else:
                        fix = None
                        jump_ifs = self.all_instr(start, self.next_stmt[pos], PJIF)
                        last_jump_good = True
                        for j in jump_ifs:
                            if (target == self.get_target(j)):
                                if ((self.lines[j].next == (j + 3)) and last_jump_good):
                                    fix = j
                                    break
                            else:
                                last_jump_good = False
                        self.fixed_jumps[pos] = (fix or match[(-1)])
                        return
                else:
                    self.fixed_jumps[pos] = match[(-1)]
                    return
        else:
            if ((pos + 3) in self.load_asserts):
                if (code[pre[rtarget]] == RAISE_VARARGS):
                    return
                self.load_asserts.remove((pos + 3))
            next = self.next_stmt[pos]
            if (pre[next] == pos):
                pass
            elif ((code[next] in (JF, JA)) and (target == self.get_target(next))):
                if (code[pre[next]] == PJIF):
                    if ((code[next] == JF) or (target != rtarget) or (code[pre[pre[rtarget]]] not in (JA, RETURN_VALUE))):
                        self.fixed_jumps[pos] = pre[next]
                        return
            elif ((code[next] == JA) and (code[target] in (JA, JF))):
                next_target = self.get_target(next)
                if (self.get_target(target) == next_target):
                    self.fixed_jumps[pos] = pre[next]
                    return
                elif ((code[next_target] in (JA, JF)) and (self.get_target(next_target) == self.get_target(target))):
                    self.fixed_jumps[pos] = pre[next]
                    return
        if (pos in self.ignore_if):
            return
        if ((code[pre[rtarget]] == JA) and (pre[rtarget] in self.stmts) and (pre[rtarget] != pos) and (pre[pre[rtarget]] != pos)):
            if ((code[rtarget] == JA) and (code[(rtarget + 3)] == POP_BLOCK)):
                if (code[pre[pre[rtarget]]] != JA):
                    pass
                elif (self.get_target(pre[pre[rtarget]]) != target):
                    pass
                else:
                    rtarget = pre[rtarget]
            else:
                rtarget = pre[rtarget]
        if (code[pre[rtarget]] in (JA, JF)):
            if_end = self.get_target(pre[rtarget])
            if ((if_end < pre[rtarget]) and (code[pre[if_end]] == SETUP_LOOP)):
                if (if_end > start):
                    return
            end = self.restrict_to_parent(if_end, parent)
            self.structs.append({'type': 'if-then', 'start': start, 'end': pre[rtarget]})
            self.not_continue.add(pre[rtarget])
            if (rtarget < end):
                self.structs.append({'type': 'if-else', 'start': rtarget, 'end': end})
        elif (code[pre[rtarget]] == RETURN_VALUE):
            self.structs.append({'type': 'if-then', 'start': start, 'end': rtarget})
            self.return_end_ifs.add(pre[rtarget])
    elif (op in (JUMP_IF_FALSE_OR_POP, JUMP_IF_TRUE_OR_POP)):
        target = self.get_target(pos, op)
        self.fixed_jumps[pos] = self.restrict_to_parent(target, parent)
 | 
| 
	'Detect all offsets in a byte code which are jump targets.
Return the list of offsets.
This procedure is modelled after dis.findlables(), but here
for each target the number of jumps are counted.'
 | 
	def find_jump_targets(self, code):
 | 
	    hasjrel = dis.hasjrel
    hasjabs = dis.hasjabs
    n = len(code)
    self.structs = [{'type': 'root', 'start': 0, 'end': (n - 1)}]
    self.loops = []
    self.fixed_jumps = {}
    self.ignore_if = set()
    self.build_stmt_indices()
    self.not_continue = set()
    self.return_end_ifs = set()
    targets = {}
    for i in self.op_range(0, n):
        op = code[i]
        self.detect_structure(i, op)
        if (op >= HAVE_ARGUMENT):
            label = self.fixed_jumps.get(i)
            oparg = (code[(i + 1)] + (code[(i + 2)] * 256))
            if (label is None):
                if ((op in hasjrel) and (op != FOR_ITER)):
                    label = ((i + 3) + oparg)
                elif (op in hasjabs):
                    if (op in (JUMP_IF_FALSE_OR_POP, JUMP_IF_TRUE_OR_POP)):
                        if (oparg > i):
                            label = oparg
            if ((label is not None) and (label != (-1))):
                targets[label] = (targets.get(label, []) + [i])
        elif ((op == END_FINALLY) and (i in self.fixed_jumps)):
            label = self.fixed_jumps[i]
            targets[label] = (targets.get(label, []) + [i])
    return targets
 | 
| 
	'exec_stmt ::= expr exprlist DUP_TOP EXEC_STMT
exec_stmt ::= expr exprlist EXEC_STMT'
 | 
	def n_exec_stmt(self, node):
 | 
	    self.write(self.indent, 'exec    ')
    self.preorder(node[0])
    if (node[1][0] != NONE):
        sep = '    in    '
        for subnode in node[1]:
            self.write(sep)
            sep = ',    '
            self.preorder(subnode)
    self.print_()
    self.prune()
 | 
| 
	'prettyprint a mapexpr
\'mapexpr\' is something like k = {\'a\': 1, \'b\': 42 }"'
 | 
	def n_mapexpr(self, node):
 | 
	    p = self.prec
    self.prec = 100
    assert (node[(-1)] == 'kvlist')
    node = node[(-1)]
    self.indentMore(INDENT_PER_LEVEL)
    line_seperator = (',\n' + self.indent)
    sep = INDENT_PER_LEVEL[:(-1)]
    self.write('{')
    for kv in node:
        assert (kv in ('kv', 'kv2', 'kv3'))
        if (kv == 'kv'):
            name = self.traverse(kv[(-2)], indent='')
            value = self.traverse(kv[1], indent=(self.indent + ((len(name) + 2) * '    ')))
        elif (kv == 'kv2'):
            name = self.traverse(kv[1], indent='')
            value = self.traverse(kv[(-3)], indent=(self.indent + ((len(name) + 2) * '    ')))
        elif (kv == 'kv3'):
            name = self.traverse(kv[(-2)], indent='')
            value = self.traverse(kv[0], indent=(self.indent + ((len(name) + 2) * '    ')))
        self.write(sep, name, ':    ', value)
        sep = line_seperator
    self.write('}')
    self.indentLess(INDENT_PER_LEVEL)
    self.prec = p
    self.prune()
 | 
| 
	'prettyprint a list or tuple'
 | 
	def n_build_list(self, node):
 | 
	    p = self.prec
    self.prec = 100
    lastnode = node.pop()
    lastnodetype = lastnode.type
    if lastnodetype.startswith('BUILD_LIST'):
        self.write('[')
        endchar = ']'
    elif lastnodetype.startswith('BUILD_TUPLE'):
        self.write('(')
        endchar = ')'
    elif lastnodetype.startswith('BUILD_SET'):
        self.write('{')
        endchar = '}'
    elif lastnodetype.startswith('ROT_TWO'):
        self.write('(')
        endchar = ')'
    else:
        raise 'Internal    Error:    n_build_list    expects    list    or    tuple'
    flat_elems = []
    for elem in node:
        if (elem == 'expr1024'):
            for subelem in elem:
                for subsubelem in subelem:
                    flat_elems.append(subsubelem)
        elif (elem == 'expr32'):
            for subelem in elem:
                flat_elems.append(subelem)
        else:
            flat_elems.append(elem)
    self.indentMore(INDENT_PER_LEVEL)
    if (lastnode.attr > 3):
        line_separator = (',\n' + self.indent)
    else:
        line_separator = ',    '
    sep = INDENT_PER_LEVEL[:(-1)]
    for elem in flat_elems:
        if (elem == 'ROT_THREE'):
            continue
        assert (elem == 'expr')
        value = self.traverse(elem)
        self.write(sep, value)
        sep = line_separator
    if ((lastnode.attr == 1) and lastnodetype.startswith('BUILD_TUPLE')):
        self.write(',')
    self.write(endchar)
    self.indentLess(INDENT_PER_LEVEL)
    self.prec = p
    self.prune()
 | 
| 
	'Special handling for opcodes that take a variable number
of arguments -- we add a new entry for each in TABLE_R.'
 | 
	def customize(self, customize):
 | 
	    for (k, v) in customize.items():
        if TABLE_R.has_key(k):
            continue
        op = k[:k.rfind('_')]
        if (op == 'CALL_FUNCTION'):
            TABLE_R[k] = ('%c(%P)', 0, (1, (-1), ',    ', 100))
        elif (op in ('CALL_FUNCTION_VAR', 'CALL_FUNCTION_VAR_KW', 'CALL_FUNCTION_KW')):
            if (v == 0):
                str = '%c(%C'
                p2 = (0, 0, None)
            else:
                str = '%c(%C,    '
                p2 = (1, (-2), ',    ')
            if (op == 'CALL_FUNCTION_VAR'):
                str += '*%c)'
                entry = (str, 0, p2, (-2))
            elif (op == 'CALL_FUNCTION_KW'):
                str += '**%c)'
                entry = (str, 0, p2, (-2))
            else:
                str += '*%c,    **%c)'
                if p2[2]:
                    p2 = (1, (-3), ',    ')
                entry = (str, 0, p2, (-3), (-2))
            TABLE_R[k] = entry
 | 
| 
	'If the name of the formal parameter starts with dot,
it\'s a tuple parameter, like this:
#          def MyFunc(xx, (a,b,c), yy):
#                  print a, b*2, c*42
In byte-code, the whole tuple is assigned to parameter \'.1\' and
then the tuple gets unpacked to \'a\', \'b\' and \'c\'.
Since identifiers starting with a dot are illegal in Python,
we can search for the byte-code equivalent to \'(a,b,c) = .1\''
 | 
	def get_tuple_parameter(self, ast, name):
 | 
	    assert (ast == 'stmts')
    for i in range(len(ast)):
        assert (ast[i][0] == 'stmt')
        node = ast[i][0][0]
        if ((node == 'assign') and (node[0] == ASSIGN_TUPLE_PARAM(name))):
            del ast[i]
            assert (node[1] == 'designator')
            return (('(' + self.traverse(node[1])) + ')')
    raise ("Can't    find    tuple    parameter" % name)
 | 
| 
	'Dump function defintion, doc string, and function body.'
 | 
	def make_function(self, node, isLambda, nested=1):
 | 
	    def build_param(ast, name, default):
        'build    parameters:\n                                                                -    handle    defaults\n                                                                -    handle    format    tuple    parameters\n                                                '
        if name.startswith('.'):
            name = self.get_tuple_parameter(ast, name)
        if default:
            if self.showast:
                print '--', name
                print default
                print '--'
            result = ('%s    =    %s' % (name, self.traverse(default, indent='')))
            if (result[(-2):] == '=    '):
                result += 'None'
            return result
        else:
            return name
    defparams = node[:node[(-1)].attr]
    code = node[(-2)].attr
    assert (type(code) == CodeType)
    code = Code(code, self.scanner, self.currentclass)
    argc = code.co_argcount
    paramnames = list(code.co_varnames[:argc])
    paramnames.reverse()
    defparams.reverse()
    try:
        ast = self.build_ast(code._tokens, code._customize, isLambda=isLambda, noneInNames=('None' in code.co_names))
    except ParserError as p:
        self.write(str(p))
        self.ERROR = p
        return
    params = []
    for (name, default) in map((lambda a, b: (a, b)), paramnames, defparams):
        params.append(build_param(ast, name, default))
    params.reverse()
    if (4 & code.co_flags):
        params.append(('*%s' % code.co_varnames[argc]))
        argc += 1
    if (8 & code.co_flags):
        params.append(('**%s' % code.co_varnames[argc]))
        argc += 1
    indent = self.indent
    if isLambda:
        self.write('lambda    ', ',    '.join(params), ':    ')
    else:
        self.print_('(', ',    '.join(params), '):')
    if ((len(code.co_consts) > 0) and (code.co_consts[0] != None)):
        self.print_docstring(indent, code.co_consts[0])
    code._tokens = None
    assert (ast == 'stmts')
    all_globals = find_all_globals(ast, set())
    for g in ((all_globals & self.mod_globs) | find_globals(ast, set())):
        self.print_(self.indent, 'global    ', g)
    self.mod_globs -= all_globals
    rn = (('None' in code.co_names) and (not find_none(ast)))
    self.gen_source(ast, code._customize, isLambda=isLambda, returnNone=rn)
    code._tokens = None
    code._customize = None
 | 
| 
	'Dump class definition, doc string and class body.'
 | 
	def build_class(self, code):
 | 
	    assert (type(code) == CodeType)
    code = Code(code, self.scanner, self.currentclass)
    indent = self.indent
    ast = self.build_ast(code._tokens, code._customize)
    code._tokens = None
    assert (ast == 'stmts')
    if (ast[0][0] == NAME_MODULE):
        del ast[0]
    if (code.co_consts and (code.co_consts[0] != None) and (ast[0][0] == ASSIGN_DOC_STRING(code.co_consts[0]))):
        self.print_docstring(indent, code.co_consts[0])
        self.print_()
        del ast[0]
    if (ast[(-1)][0] == RETURN_LOCALS):
        del ast[(-1)]
    for g in find_globals(ast, set()):
        self.print_(indent, 'global    ', g)
    self.gen_source(ast, code._customize)
    code._tokens = None
    code._customize = None
 | 
| 
	'convert AST to source code'
 | 
	def gen_source(self, ast, customize, isLambda=0, returnNone=False):
 | 
	    rn = self.return_none
    self.return_none = returnNone
    if (len(ast) == 0):
        self.print_(self.indent, 'pass')
    else:
        self.customize(customize)
        if isLambda:
            self.write(self.traverse(ast, isLambda=isLambda))
        else:
            self.print_(self.traverse(ast, isLambda=isLambda))
    self.return_none = rn
 | 
| 
	'Places new Gmail notifications in the Notifier\'s queue.'
 | 
	def handleEmailNotifications(self, lastDate):
 | 
	    emails = Gmail.fetchUnreadEmails(self.profile, since=lastDate)
    if emails:
        lastDate = Gmail.getMostRecentDate(emails)
    def styleEmail(e):
        return ('New    email    from    %s.' % Gmail.getSender(e))
    for e in emails:
        self.q.put(styleEmail(e))
    return lastDate
 | 
| 
	'Returns a notification. Note that this function is consuming.'
 | 
	def getNotification(self):
 | 
	    try:
        notif = self.q.get(block=False)
        return notif
    except Queue.Empty:
        return None
 | 
| 
	'Return a list of notifications in chronological order.
Note that this function is consuming, so consecutive calls
will yield different results.'
 | 
	def getAllNotifications(self):
 | 
	    notifs = []
    notif = self.getNotification()
    while notif:
        notifs.append(notif)
        notif = self.getNotification()
    return notifs
 | 
| 
	'Calculates a revision from phrases by using the SHA1 hash function.
Arguments:
phrases -- a list of phrases
Returns:
A revision string for given phrases.'
 | 
	@classmethod
def phrases_to_revision(cls, phrases):
 | 
	    sorted_phrases = sorted(phrases)
    joined_phrases = '\n'.join(sorted_phrases)
    sha1 = hashlib.sha1()
    sha1.update(joined_phrases)
    return sha1.hexdigest()
 | 
| 
	'Initializes a new Vocabulary instance.
Optional Arguments:
name -- (optional) the name of the vocabulary (Default: \'default\')
path -- (optional) the path in which the vocabulary exists or will
be created (Default: \'.\')'
 | 
	def __init__(self, name='default', path='.'):
 | 
	    self.name = name
    self.path = os.path.abspath(os.path.join(path, self.PATH_PREFIX, name))
    self._logger = logging.getLogger(__name__)
 | 
| 
	'Returns:
The path of the the revision file as string'
 | 
	@property
def revision_file(self):
 | 
	    return os.path.join(self.path, 'revision')
 | 
| 
	'Checks if the vocabulary is compiled by checking if the revision file
is readable. This method should be overridden by subclasses to check
for class-specific additional files, too.
Returns:
True if the dictionary is compiled, else False'
 | 
	@abstractproperty
def is_compiled(self):
 | 
	    return os.access(self.revision_file, os.R_OK)
 | 
| 
	'Reads the compiled revision from the revision file.
Returns:
the revision of this vocabulary (i.e. the string
inside the revision file), or None if is_compiled
if False'
 | 
	@property
def compiled_revision(self):
 | 
	    if (not self.is_compiled):
        return None
    with open(self.revision_file, 'r') as f:
        revision = f.read().strip()
    self._logger.debug("compiled_revision    is    '%s'", revision)
    return revision
 | 
| 
	'Convenience method to check if this vocabulary exactly contains the
phrases passed to this method.
Arguments:
phrases -- a list of phrases
Returns:
True if phrases exactly matches the phrases inside this
vocabulary.'
 | 
	def matches_phrases(self, phrases):
 | 
	    return (self.compiled_revision == self.phrases_to_revision(phrases))
 | 
| 
	'Compiles this vocabulary. If the force argument is True, compilation
will be forced regardless of necessity (which means that the
preliminary check if the current revision already equals the
revision after compilation will be skipped).
This method is not meant to be overridden by subclasses - use the
_compile_vocabulary()-method instead.
Arguments:
phrases -- a list of phrases that this vocabulary will contain
force -- (optional) forces compilation (Default: False)
Returns:
The revision of the compiled vocabulary'
 | 
	def compile(self, phrases, force=False):
 | 
	    revision = self.phrases_to_revision(phrases)
    if ((not force) and (self.compiled_revision == revision)):
        self._logger.debug(('Compilation    not    neccessary,    compiled    ' + 'version    matches    phrases.'))
        return revision
    if (not os.path.exists(self.path)):
        self._logger.debug(("Vocabulary    dir    '%s'    does    not    exist,    " + 'creating...'), self.path)
        try:
            os.makedirs(self.path)
        except OSError:
            self._logger.error("Couldn't    create    vocabulary    dir    '%s'", self.path, exc_info=True)
            raise
    try:
        with open(self.revision_file, 'w') as f:
            f.write(revision)
    except (OSError, IOError):
        self._logger.error("Couldn't    write    revision    file    in    '%s'", self.revision_file, exc_info=True)
        raise
    else:
        self._logger.info('Starting    compilation...')
        try:
            self._compile_vocabulary(phrases)
        except Exception as e:
            self._logger.error(('Fatal    compilation    Error    occured,    ' + 'cleaning    up...'), exc_info=True)
            try:
                os.remove(self.revision_file)
            except OSError:
                pass
            raise e
        else:
            self._logger.info('Compilation    done.')
    return revision
 | 
| 
	'Checks if the vocabulary is compiled by checking if the revision
file is readable.
Returns:
True if this vocabulary has been compiled, else False'
 | 
	@property
def is_compiled(self):
 | 
	    return super(self.__class__, self).is_compiled
 | 
| 
	'Does nothing (because this is a dummy class for testing purposes).'
 | 
	def _compile_vocabulary(self, phrases):
 | 
	    pass
 | 
| 
	'Returns:
The path of the the pocketsphinx languagemodel file as string'
 | 
	@property
def languagemodel_file(self):
 | 
	    return os.path.join(self.path, 'languagemodel')
 | 
| 
	'Returns:
The path of the pocketsphinx dictionary file as string'
 | 
	@property
def dictionary_file(self):
 | 
	    return os.path.join(self.path, 'dictionary')
 | 
| 
	'Checks if the vocabulary is compiled by checking if the revision,
languagemodel and dictionary files are readable.
Returns:
True if this vocabulary has been compiled, else False'
 | 
	@property
def is_compiled(self):
 | 
	    return (super(self.__class__, self).is_compiled and os.access(self.languagemodel_file, os.R_OK) and os.access(self.dictionary_file, os.R_OK))
 | 
| 
	'Convenience property to use this Vocabulary with the __init__() method
of the pocketsphinx.Decoder class.
Returns:
A dict containing kwargs for the pocketsphinx.Decoder.__init__()
method.
Example:
decoder = pocketsphinx.Decoder(**vocab_instance.decoder_kwargs,
hmm=\'/path/to/hmm\')'
 | 
	@property
def decoder_kwargs(self):
 | 
	    return {'lm': self.languagemodel_file, 'dict': self.dictionary_file}
 | 
| 
	'Compiles the vocabulary to the Pocketsphinx format by creating a
languagemodel and a dictionary.
Arguments:
phrases -- a list of phrases that this vocabulary will contain'
 | 
	def _compile_vocabulary(self, phrases):
 | 
	    text = '    '.join([('<s>    %s    </s>' % phrase) for phrase in phrases])
    self._logger.debug('Compiling    languagemodel...')
    vocabulary = self._compile_languagemodel(text, self.languagemodel_file)
    self._logger.debug('Starting    dictionary...')
    self._compile_dictionary(vocabulary, self.dictionary_file)
 | 
| 
	'Compiles the languagemodel from a text.
Arguments:
text -- the text the languagemodel will be generated from
output_file -- the path of the file this languagemodel will
be written to
Returns:
A list of all unique words this vocabulary contains.'
 | 
	def _compile_languagemodel(self, text, output_file):
 | 
	    with tempfile.NamedTemporaryFile(suffix='.vocab', delete=False) as f:
        vocab_file = f.name
    self._logger.debug("Creating    vocab    file:    '%s'", vocab_file)
    cmuclmtk.text2vocab(text, vocab_file)
    self._logger.debug("Creating    languagemodel    file:    '%s'", output_file)
    cmuclmtk.text2lm(text, output_file, vocab_file=vocab_file)
    self._logger.debug(('Getting    words    from    vocab    file    and    removing    it    ' + 'afterwards...'))
    words = []
    with open(vocab_file, 'r') as f:
        for line in f:
            line = line.strip()
            if ((not line.startswith('#')) and (line not in ('<s>', '</s>'))):
                words.append(line)
    os.remove(vocab_file)
    return words
 | 
| 
	'Compiles the dictionary from a list of words.
Arguments:
words -- a list of all unique words this vocabulary contains
output_file -- the path of the file this dictionary will
be written to'
 | 
	def _compile_dictionary(self, words, output_file):
 | 
	    self._logger.debug('Getting    phonemes    for    %d    words...', len(words))
    g2pconverter = PhonetisaurusG2P(**PhonetisaurusG2P.get_config())
    phonemes = g2pconverter.translate(words)
    self._logger.debug("Creating    dict    file:    '%s'", output_file)
    with open(output_file, 'w') as f:
        for (word, pronounciations) in phonemes.items():
            for (i, pronounciation) in enumerate(pronounciations, start=1):
                if (i == 1):
                    line = ('%s DCTB %s\n' % (word, pronounciation))
                else:
                    line = ('%s(%d) DCTB %s\n' % (word, i, pronounciation))
                f.write(line)
 | 
| 
	'Returns:
The path of the the julius dfa file as string'
 | 
	@property
def dfa_file(self):
 | 
	    return os.path.join(self.path, 'dfa')
 | 
| 
	'Returns:
The path of the the julius dict file as string'
 | 
	@property
def dict_file(self):
 | 
	    return os.path.join(self.path, 'dict')
 | 
| 
	'Prepare the client and music variables'
 | 
	def __init__(self, server='localhost', port=6600):
 | 
	    self.server = server
    self.port = port
    self.client = mpd.MPDClient()
    self.client.timeout = None
    self.client.idletimeout = None
    self.client.connect(self.server, self.port)
    self.playlists = [x['playlist'] for x in self.client.listplaylists()]
    self.client.clear()
    for playlist in self.playlists:
        self.client.load(playlist)
    self.songs = []
    self.song_titles = []
    self.song_artists = []
    soup = self.client.playlist()
    for i in range(0, (len(soup) / 10)):
        index = (i * 10)
        id = soup[index].strip()
        title = soup[(index + 3)].strip().upper()
        artist = soup[(index + 2)].strip().upper()
        album = soup[(index + 4)].strip().upper()
        self.songs.append(Song(id, title, artist, album))
        self.song_titles.append(title)
        self.song_artists.append(artist)
 | 
| 
	'Plays the current song or accepts a song to play.
Arguments:
songs -- a list of song objects
playlist_name -- user-defined, something like "Love Song Playlist"'
 | 
	@reconnect
def play(self, songs=False, playlist_name=False):
 | 
	    if songs:
        self.client.clear()
        for song in songs:
            try:
                self.client.add(song.id)
            except:
                pass
    if playlist_name:
        self.client.clear()
        self.client.load(playlist_name)
    self.client.play()
 | 
| 
	'Returns the list of unique words that comprise song and artist titles'
 | 
	def get_soup(self):
 | 
	    soup = []
    for song in self.songs:
        song_words = song.title.split('    ')
        artist_words = song.artist.split('    ')
        soup.extend(song_words)
        soup.extend(artist_words)
    title_trans = ''.join(((chr(c) if (chr(c).isupper() or chr(c).islower()) else '_') for c in range(256)))
    soup = [x.decode('utf-8').encode('ascii', 'ignore').upper().translate(title_trans).replace('_', '') for x in soup]
    soup = [x for x in soup if (x != '')]
    return list(set(soup))
 | 
| 
	'Returns the list of unique words that comprise playlist names'
 | 
	def get_soup_playlist(self):
 | 
	    soup = []
    for name in self.playlists:
        soup.extend(name.split('    '))
    title_trans = ''.join(((chr(c) if (chr(c).isupper() or chr(c).islower()) else '_') for c in range(256)))
    soup = [x.decode('utf-8').encode('ascii', 'ignore').upper().translate(title_trans).replace('_', '') for x in soup]
    soup = [x for x in soup if (x != '')]
    return list(set(soup))
 | 
| 
	'Returns the list of PHRASES that comprise song and artist titles'
 | 
	def get_soup_separated(self):
 | 
	    title_soup = [song.title for song in self.songs]
    artist_soup = [song.artist for song in self.songs]
    soup = list(set((title_soup + artist_soup)))
    title_trans = ''.join(((chr(c) if (chr(c).isupper() or chr(c).islower()) else '_') for c in range(256)))
    soup = [x.decode('utf-8').encode('ascii', 'ignore').upper().translate(title_trans).replace('_', '    ') for x in soup]
    soup = [re.sub('    +', '    ', x) for x in soup if (x != '')]
    return soup
 | 
| 
	'Returns songs matching a query best as possible on either artist
field, etc'
 | 
	def fuzzy_songs(self, query):
 | 
	    query = query.upper()
    matched_song_titles = difflib.get_close_matches(query, self.song_titles)
    matched_song_artists = difflib.get_close_matches(query, self.song_artists)
    strict_priority_title = [x for x in matched_song_titles if (x == query)]
    strict_priority_artists = [x for x in matched_song_artists if (x == query)]
    if strict_priority_title:
        matched_song_titles = strict_priority_title
    if strict_priority_artists:
        matched_song_artists = strict_priority_artists
    matched_songs_bytitle = [song for song in self.songs if (song.title in matched_song_titles)]
    matched_songs_byartist = [song for song in self.songs if (song.artist in matched_song_artists)]
    matches = list(set((matched_songs_bytitle + matched_songs_byartist)))
    return matches
 | 
| 
	'returns playlist names that match query best as possible'
 | 
	def fuzzy_playlists(self, query):
 | 
	    query = query.upper()
    lookup = {n.upper(): n for n in self.playlists}
    results = [lookup[r] for r in difflib.get_close_matches(query, lookup)]
    return results
 | 
| 
	'Initiates the pocketsphinx instance.
Arguments:
speaker -- handles platform-independent audio output
passive_stt_engine -- performs STT while Jasper is in passive listen
mode
acive_stt_engine -- performs STT while Jasper is in active listen mode'
 | 
	def __init__(self, speaker, passive_stt_engine, active_stt_engine):
 | 
	    self._logger = logging.getLogger(__name__)
    self.speaker = speaker
    self.passive_stt_engine = passive_stt_engine
    self.active_stt_engine = active_stt_engine
    self._logger.info((('Initializing    PyAudio.    ALSA/Jack    error    messages    ' + 'that    pop    up    during    this    process    are    normal    and    ') + 'can    usually    be    safely    ignored.'))
    self._audio = pyaudio.PyAudio()
    self._logger.info('Initialization    of    PyAudio    completed.')
 | 
| 
	'Listens for PERSONA in everyday sound. Times out after LISTEN_TIME, so
needs to be restarted.'
 | 
	def passiveListen(self, PERSONA):
 | 
	    THRESHOLD_MULTIPLIER = 1.8
    RATE = 16000
    CHUNK = 1024
    THRESHOLD_TIME = 1
    LISTEN_TIME = 10
    stream = self._audio.open(format=pyaudio.paInt16, channels=1, rate=RATE, input=True, frames_per_buffer=CHUNK)
    frames = []
    lastN = [i for i in range(30)]
    for i in range(0, ((RATE / CHUNK) * THRESHOLD_TIME)):
        data = stream.read(CHUNK)
        frames.append(data)
        lastN.pop(0)
        lastN.append(self.getScore(data))
        average = (sum(lastN) / len(lastN))
    THRESHOLD = (average * THRESHOLD_MULTIPLIER)
    frames = []
    didDetect = False
    for i in range(0, ((RATE / CHUNK) * LISTEN_TIME)):
        data = stream.read(CHUNK)
        frames.append(data)
        score = self.getScore(data)
        if (score > THRESHOLD):
            didDetect = True
            break
    if (not didDetect):
        print 'No    disturbance    detected'
        stream.stop_stream()
        stream.close()
        return (None, None)
    frames = frames[(-20):]
    DELAY_MULTIPLIER = 1
    for i in range(0, ((RATE / CHUNK) * DELAY_MULTIPLIER)):
        data = stream.read(CHUNK)
        frames.append(data)
    stream.stop_stream()
    stream.close()
    with tempfile.NamedTemporaryFile(mode='w+b') as f:
        wav_fp = wave.open(f, 'wb')
        wav_fp.setnchannels(1)
        wav_fp.setsampwidth(pyaudio.get_sample_size(pyaudio.paInt16))
        wav_fp.setframerate(RATE)
        wav_fp.writeframes(''.join(frames))
        wav_fp.close()
        f.seek(0)
        transcribed = self.passive_stt_engine.transcribe(f)
    if any(((PERSONA in phrase) for phrase in transcribed)):
        return (THRESHOLD, PERSONA)
    return (False, transcribed)
 | 
| 
	'Records until a second of silence or times out after 12 seconds
Returns the first matching string or None'
 | 
	def activeListen(self, THRESHOLD=None, LISTEN=True, MUSIC=False):
 | 
	    options = self.activeListenToAllOptions(THRESHOLD, LISTEN, MUSIC)
    if options:
        return options[0]
 | 
| 
	'Records until a second of silence or times out after 12 seconds
Returns a list of the matching options or None'
 | 
	def activeListenToAllOptions(self, THRESHOLD=None, LISTEN=True, MUSIC=False):
 | 
	    RATE = 16000
    CHUNK = 1024
    LISTEN_TIME = 12
    if (THRESHOLD is None):
        THRESHOLD = self.fetchThreshold()
    self.speaker.play(jasperpath.data('audio', 'beep_hi.wav'))
    stream = self._audio.open(format=pyaudio.paInt16, channels=1, rate=RATE, input=True, frames_per_buffer=CHUNK)
    frames = []
    lastN = [(THRESHOLD * 1.2) for i in range(30)]
    for i in range(0, ((RATE / CHUNK) * LISTEN_TIME)):
        data = stream.read(CHUNK)
        frames.append(data)
        score = self.getScore(data)
        lastN.pop(0)
        lastN.append(score)
        average = (sum(lastN) / float(len(lastN)))
        if (average < (THRESHOLD * 0.8)):
            break
    self.speaker.play(jasperpath.data('audio', 'beep_lo.wav'))
    stream.stop_stream()
    stream.close()
    with tempfile.SpooledTemporaryFile(mode='w+b') as f:
        wav_fp = wave.open(f, 'wb')
        wav_fp.setnchannels(1)
        wav_fp.setsampwidth(pyaudio.get_sample_size(pyaudio.paInt16))
        wav_fp.setframerate(RATE)
        wav_fp.writeframes(''.join(frames))
        wav_fp.close()
        f.seek(0)
        return self.active_stt_engine.transcribe(f)
 | 
| 
	'Delegates user input to the handling function when activated.'
 | 
	def handleForever(self):
 | 
	    self._logger.info("Starting    to    handle    conversation    with    keyword    '%s'.", self.persona)
    while True:
        notifications = self.notifier.getAllNotifications()
        for notif in notifications:
            self._logger.info("Received    notification:    '%s'", str(notif))
        self._logger.debug("Started    listening    for    keyword    '%s'", self.persona)
        (threshold, transcribed) = self.mic.passiveListen(self.persona)
        self._logger.debug("Stopped    listening    for    keyword    '%s'", self.persona)
        if ((not transcribed) or (not threshold)):
            self._logger.info('Nothing    has    been    said    or    transcribed.')
            continue
        self._logger.info("Keyword    '%s'    has    been    said!", self.persona)
        self._logger.debug('Started    to    listen    actively    with    threshold:    %r', threshold)
        input = self.mic.activeListenToAllOptions(threshold)
        self._logger.debug('Stopped    to    listen    actively    with    threshold:    %r', threshold)
        if input:
            self.brain.query(input)
        else:
            self.mic.say('Pardon?')
 | 
| 
	'Instantiates a new Brain object, which cross-references user
input with a list of modules. Note that the order of brain.modules
matters, as the Brain will cease execution on the first module
that accepts a given input.
Arguments:
mic -- used to interact with the user (for both input and output)
profile -- contains information related to the user (e.g., phone
number)'
 | 
	def __init__(self, mic, profile):
 | 
	    self.mic = mic
    self.profile = profile
    self.modules = self.get_modules()
    self._logger = logging.getLogger(__name__)
 | 
| 
	'Dynamically loads all the modules in the modules folder and sorts
them by the PRIORITY key. If no PRIORITY is defined for a given
module, a priority of 0 is assumed.'
 | 
	@classmethod
def get_modules(cls):
 | 
	    logger = logging.getLogger(__name__)
    locations = [jasperpath.PLUGIN_PATH]
    logger.debug('Looking    for    modules    in:    %s', ',    '.join([("'%s'" % location) for location in locations]))
    modules = []
    for (finder, name, ispkg) in pkgutil.walk_packages(locations):
        try:
            loader = finder.find_module(name)
            mod = loader.load_module(name)
        except:
            logger.warning("Skipped    module    '%s'    due    to    an    error.", name, exc_info=True)
        else:
            if hasattr(mod, 'WORDS'):
                logger.debug("Found    module    '%s'    with    words:    %r", name, mod.WORDS)
                modules.append(mod)
            else:
                logger.warning(("Skipped    module    '%s'    because    it    misses    " + 'the    WORDS    constant.'), name)
    modules.sort(key=(lambda mod: (mod.PRIORITY if hasattr(mod, 'PRIORITY') else 0)), reverse=True)
    return modules
 | 
| 
	'Passes user input to the appropriate module, testing it against
each candidate module\'s isValid function.
Arguments:
text -- user input, typically speech, to be parsed by a module'
 | 
	def query(self, texts):
 | 
	    for module in self.modules:
        for text in texts:
            if module.isValid(text):
                self._logger.debug(("'%s'    is    a    valid    phrase    for    module    " + "'%s'"), text, module.__name__)
                try:
                    module.handle(text, self.mic, self.profile)
                except Exception:
                    self._logger.error('Failed    to    execute    module', exc_info=True)
                    self.mic.say(("I'm    sorry.    I    had    some    trouble    with    " + 'that    operation.    Please    try    again    later.'))
                else:
                    self._logger.debug(("Handling    of    phrase    '%s'    by    " + "module    '%s'    completed"), text, module.__name__)
                finally:
                    return
    self._logger.debug(('No    module    was    able    to    handle    any    of    these    ' + 'phrases:    %r'), texts)
 | 
| 
	'Initiates the pocketsphinx instance.
Arguments:
vocabulary -- a PocketsphinxVocabulary instance
hmm_dir -- the path of the Hidden Markov Model (HMM)'
 | 
	def __init__(self, vocabulary, hmm_dir=('/usr/local/share/' + 'pocketsphinx/model/hmm/en_US/hub4wsj_sc_8k')):
 | 
	    self._logger = logging.getLogger(__name__)
    try:
        import pocketsphinx as ps
    except:
        import pocketsphinx as ps
    with tempfile.NamedTemporaryFile(prefix='psdecoder_', suffix='.log', delete=False) as f:
        self._logfile = f.name
    self._logger.debug(('Initializing    PocketSphinx    Decoder    with    hmm_dir    ' + "'%s'"), hmm_dir)
    if (not os.path.exists(hmm_dir)):
        msg = (("hmm_dir    '%s'    does    not    exist!    Please    make    sure    that    you    " + 'have    set    the    correct    hmm_dir    in    your    profile.') % hmm_dir)
        self._logger.error(msg)
        raise RuntimeError(msg)
    missing_hmm_files = []
    for fname in ('mdef', 'feat.params', 'means', 'noisedict', 'transition_matrices', 'variances'):
        if (not os.path.exists(os.path.join(hmm_dir, fname))):
            missing_hmm_files.append(fname)
    mixweights = os.path.exists(os.path.join(hmm_dir, 'mixture_weights'))
    sendump = os.path.exists(os.path.join(hmm_dir, 'sendump'))
    if ((not mixweights) and (not sendump)):
        missing_hmm_files.append('mixture_weights    or    sendump')
    if missing_hmm_files:
        self._logger.warning((("hmm_dir    '%s'    is    missing    files:    %s.    Please    " + 'make    sure    that    you    have    set    the    correct    ') + 'hmm_dir    in    your    profile.'), hmm_dir, ',    '.join(missing_hmm_files))
    self._decoder = ps.Decoder(hmm=hmm_dir, logfn=self._logfile, **vocabulary.decoder_kwargs)
 | 
| 
	'Performs STT, transcribing an audio file and returning the result.
Arguments:
fp -- a file object containing audio data'
 | 
	def transcribe(self, fp):
 | 
	    fp.seek(44)
    data = fp.read()
    self._decoder.start_utt()
    self._decoder.process_raw(data, False, True)
    self._decoder.end_utt()
    result = self._decoder.get_hyp()
    with open(self._logfile, 'r+') as f:
        for line in f:
            self._logger.debug(line.strip())
        f.truncate()
    transcribed = [result[0]]
    self._logger.info('Transcribed:    %r', transcribed)
    return transcribed
 | 
| 
	'Arguments:
api_key - the public api key which allows access to Google APIs'
 | 
	def __init__(self, api_key=None, language='en-us'):
 | 
	    self._logger = logging.getLogger(__name__)
    self._request_url = None
    self._language = None
    self._api_key = None
    self._http = requests.Session()
    self.language = language
    self.api_key = api_key
 | 
| 
	'Performs STT via the Google Speech API, transcribing an audio file and
returning an English string.
Arguments:
audio_file_path -- the path to the .wav file to be transcribed'
 | 
	def transcribe(self, fp):
 | 
	    if (not self.api_key):
        self._logger.critical(('API    key    missing,    transcription    request    ' + 'aborted.'))
        return []
    elif (not self.language):
        self._logger.critical(('Language    info    missing,    transcription    ' + 'request    aborted.'))
        return []
    wav = wave.open(fp, 'rb')
    frame_rate = wav.getframerate()
    wav.close()
    data = fp.read()
    headers = {'content-type': ('audio/l16;    rate=%s' % frame_rate)}
    r = self._http.post(self.request_url, data=data, headers=headers)
    try:
        r.raise_for_status()
    except requests.exceptions.HTTPError:
        self._logger.critical('Request    failed    with    http    status    %d', r.status_code)
        if (r.status_code == requests.codes['forbidden']):
            self._logger.warning(('Status    403    is    probably    caused    by    an    ' + 'invalid    Google    API    key.'))
        return []
    r.encoding = 'utf-8'
    try:
        response = json.loads(list(r.text.strip().split('\n', 1))[(-1)])
        if (len(response['result']) == 0):
            raise ValueError('Nothing    has    been    transcribed.')
        results = [alt['transcript'] for alt in response['result'][0]['alternative']]
    except ValueError as e:
        self._logger.warning('Empty    response:    %s', e.args[0])
        results = []
    except (KeyError, IndexError):
        self._logger.warning('Cannot    parse    response.', exc_info=True)
        results = []
    else:
        results = tuple((result.upper() for result in results))
        self._logger.info('Transcribed:    %r', results)
    return results
 | 
| 
	'Does Jasper recognize his name (i.e., passive listen)?'
 | 
	def testTranscribeJasper(self):
 | 
	    with open(self.jasper_clip, mode='rb') as f:
        transcription = self.passive_stt_engine.transcribe(f)
    self.assertIn('JASPER', transcription)
 | 
| 
	'Does Jasper recognize \'time\' (i.e., active listen)?'
 | 
	def testTranscribe(self):
 | 
	    with open(self.time_clip, mode='rb') as f:
        transcription = self.active_stt_engine.transcribe(f)
    self.assertIn('TIME', transcription)
 | 
| 
	'Does Brain correctly log errors when raised by modules?'
 | 
	def testLog(self):
 | 
	    my_brain = TestBrain._emptyBrain()
    unclear = my_brain.modules[(-1)]
    with mock.patch.object(unclear, 'handle') as mocked_handle:
        with mock.patch.object(my_brain._logger, 'error') as mocked_log:
            mocked_handle.side_effect = KeyError('foo')
            my_brain.query('zzz    gibberish    zzz')
            self.assertTrue(mocked_log.called)
 | 
| 
	'Does Brain sort modules by priority?'
 | 
	def testSortByPriority(self):
 | 
	    my_brain = TestBrain._emptyBrain()
    priorities = filter((lambda m: hasattr(m, 'PRIORITY')), my_brain.modules)
    target = sorted(priorities, key=(lambda m: m.PRIORITY), reverse=True)
    self.assertEqual(target, priorities)
 | 
| 
	'Does Brain correctly send query to higher-priority module?'
 | 
	def testPriority(self):
 | 
	    my_brain = TestBrain._emptyBrain()
    hn_module = 'HN'
    hn = filter((lambda m: (m.__name__ == hn_module)), my_brain.modules)[0]
    with mock.patch.object(hn, 'handle') as mocked_handle:
        my_brain.query(['hacker    news'])
        self.assertTrue(mocked_handle.called)
 | 
| 
	'Generic method for spoofing conversation.
Arguments:
query -- The initial input to the server.
inputs -- Additional input, if conversation is extended.
Returns:
The server\'s responses, in a list.'
 | 
	def runConversation(self, query, inputs, module):
 | 
	    self.assertTrue(module.isValid(query))
    mic = test_mic.Mic(inputs)
    module.handle(query, mic, self.profile)
    return mic.outputs
 | 
| 
	'The lines below is a spider contract. For more info see:
http://doc.scrapy.org/en/latest/topics/contracts.html
@url http://www.dmoz.org/Computers/Programming/Languages/Python/Resources/
@scrapes name'
 | 
	def parse(self, response):
 | 
	    sites = response.css('#site-list-content    >    div.site-item    >    div.title-and-desc')
    items = []
    for site in sites:
        item = Website()
        item['name'] = site.css('a    >    div.site-title::text').extract_first().strip()
        item['url'] = site.xpath('a/@href').extract_first().strip()
        item['description'] = site.css('div.site-descr::text').extract_first().strip()
        items.append(item)
    return items
 | 
| 
	'Since the original function always creates the directory, to resolve
the issue a new function had to be created. It\'s a simple copy and
was reduced for this case.
More info at:
https://github.com/scrapy/scrapy/pull/2005'
 | 
	def _copytree(self, src, dst):
 | 
	    ignore = IGNORE
    names = os.listdir(src)
    ignored_names = ignore(src, names)
    if (not os.path.exists(dst)):
        os.makedirs(dst)
    for name in names:
        if (name in ignored_names):
            continue
        srcname = os.path.join(src, name)
        dstname = os.path.join(dst, name)
        if os.path.isdir(srcname):
            self._copytree(srcname, dstname)
        else:
            copy2(srcname, dstname)
    copystat(src, dst)
 | 
| 
	'Command syntax (preferably one-line). Do not include command name.'
 | 
	def syntax(self):
 | 
	    return ''
 | 
| 
	'A short description of the command'
 | 
	def short_desc(self):
 | 
	    return ''
 | 
| 
	'A long description of the command. Return short description when not
available. It cannot contain newlines, since contents will be formatted
by optparser which removes newlines and wraps text.'
 | 
	def long_desc(self):
 | 
	    return self.short_desc()
 | 
| 
	'An extensive help for the command. It will be shown when using the
"help" command. It can contain newlines, since not post-formatting will
be applied to its contents.'
 | 
	def help(self):
 | 
	    return self.long_desc()
 | 
| 
	'Populate option parse with options available for this command'
 | 
	def add_options(self, parser):
 | 
	    group = OptionGroup(parser, 'Global    Options')
    group.add_option('--logfile', metavar='FILE', help='log    file.    if    omitted    stderr    will    be    used')
    group.add_option('-L', '--loglevel', metavar='LEVEL', default=None, help=('log    level    (default:    %s)' % self.settings['LOG_LEVEL']))
    group.add_option('--nolog', action='store_true', help='disable    logging    completely')
    group.add_option('--profile', metavar='FILE', default=None, help='write    python    cProfile    stats    to    FILE')
    group.add_option('--pidfile', metavar='FILE', help='write    process    ID    to    FILE')
    group.add_option('-s', '--set', action='append', default=[], metavar='NAME=VALUE', help='set/override    setting    (may    be    repeated)')
    group.add_option('--pdb', action='store_true', help='enable    pdb    on    failure')
    parser.add_option_group(group)
 | 
| 
	'Entry point for running commands'
 | 
	def run(self, args, opts):
 | 
	    raise NotImplementedError
 | 
| 
	'Generate the spider module, based on the given template'
 | 
	def _genspider(self, module, name, domain, template_name, template_file):
 | 
	    tvars = {'project_name': self.settings.get('BOT_NAME'), 'ProjectName': string_camelcase(self.settings.get('BOT_NAME')), 'module': module, 'name': name, 'domain': domain, 'classname': ('%sSpider' % ''.join((s.capitalize() for s in module.split('_'))))}
    if self.settings.get('NEWSPIDER_MODULE'):
        spiders_module = import_module(self.settings['NEWSPIDER_MODULE'])
        spiders_dir = abspath(dirname(spiders_module.__file__))
    else:
        spiders_module = None
        spiders_dir = '.'
    spider_file = ('%s.py' % join(spiders_dir, module))
    shutil.copyfile(template_file, spider_file)
    render_templatefile(spider_file, **tvars)
    print(('Created    spider    %r    using    template    %r    ' % (name, template_name)), end=('' if spiders_module else '\n'))
    if spiders_module:
        print(('in    module:\n        %s.%s' % (spiders_module.__name__, module)))
 | 
| 
	'You can use this function to update the Scrapy objects that will be
available in the shell'
 | 
	def update_vars(self, vars):
 | 
	    pass
 | 
| 
	'Start the execution engine'
 | 
	@defer.inlineCallbacks
def start(self):
 | 
	    assert (not self.running), 'Engine    already    running'
    self.start_time = time()
    (yield self.signals.send_catch_log_deferred(signal=signals.engine_started))
    self.running = True
    self._closewait = defer.Deferred()
    (yield self._closewait)
 | 
| 
	'Stop the execution engine gracefully'
 | 
	def stop(self):
 | 
	    assert self.running, 'Engine    not    running'
    self.running = False
    dfd = self._close_all_spiders()
    return dfd.addBoth((lambda _: self._finish_stopping_engine()))
 | 
| 
	'Close the execution engine gracefully.
If it has already been started, stop it. In all cases, close all spiders
and the downloader.'
 | 
	def close(self):
 | 
	    if self.running:
        return self.stop()
    elif self.open_spiders:
        return self._close_all_spiders()
    else:
        return defer.succeed(self.downloader.close())
 | 
| 
	'Pause the execution engine'
 | 
	def pause(self):
 | 
	    self.paused = True
 | 
| 
	'Resume the execution engine'
 | 
	def unpause(self):
 | 
	    self.paused = False
 | 
| 
	'Does the engine have capacity to handle more spiders'
 | 
	def has_capacity(self):
 | 
	    return (not bool(self.slot))
 | 
| 
	'Called when a spider gets idle. This function is called when there
are no remaining pages to download or schedule. It can be called
multiple times. If some extension raises a DontCloseSpider exception
(in the spider_idle signal handler) the spider is not closed until the
next loop and this function is guaranteed to be called (at least) once
again for this spider.'
 | 
	def _spider_idle(self, spider):
 | 
	    res = self.signals.send_catch_log(signal=signals.spider_idle, spider=spider, dont_log=DontCloseSpider)
    if any(((isinstance(x, Failure) and isinstance(x.value, DontCloseSpider)) for (_, x) in res)):
        return
    if self.spider_is_idle(spider):
        self.close_spider(spider, reason='finished')
 | 
| 
	'Close (cancel) spider and clear all its outstanding requests'
 | 
	def close_spider(self, spider, reason='cancelled'):
 | 
	    slot = self.slot
    if slot.closing:
        return slot.closing
    logger.info('Closing    spider    (%(reason)s)', {'reason': reason}, extra={'spider': spider})
    dfd = slot.close()
    def log_failure(msg):
        def errback(failure):
            logger.error(msg, exc_info=failure_to_exc_info(failure), extra={'spider': spider})
        return errback
    dfd.addBoth((lambda _: self.downloader.close()))
    dfd.addErrback(log_failure('Downloader    close    failure'))
    dfd.addBoth((lambda _: self.scraper.close_spider(spider)))
    dfd.addErrback(log_failure('Scraper    close    failure'))
    dfd.addBoth((lambda _: slot.scheduler.close(reason)))
    dfd.addErrback(log_failure('Scheduler    close    failure'))
    dfd.addBoth((lambda _: self.signals.send_catch_log_deferred(signal=signals.spider_closed, spider=spider, reason=reason)))
    dfd.addErrback(log_failure('Error    while    sending    spider_close    signal'))
    dfd.addBoth((lambda _: self.crawler.stats.close_spider(spider, reason=reason)))
    dfd.addErrback(log_failure('Stats    close    failure'))
    dfd.addBoth((lambda _: logger.info('Spider    closed    (%(reason)s)', {'reason': reason}, extra={'spider': spider})))
    dfd.addBoth((lambda _: setattr(self, 'slot', None)))
    dfd.addErrback(log_failure('Error    while    unassigning    slot'))
    dfd.addBoth((lambda _: setattr(self, 'spider', None)))
    dfd.addErrback(log_failure('Error    while    unassigning    spider'))
    dfd.addBoth((lambda _: self._spider_closed_callback(spider)))
    return dfd
 | 
| 
	'Open the given spider for scraping and allocate resources for it'
 | 
	@defer.inlineCallbacks
def open_spider(self, spider):
 | 
	    self.slot = Slot()
    (yield self.itemproc.open_spider(spider))
 | 
| 
	'Close a spider being scraped and release its resources'
 | 
	def close_spider(self, spider):
 | 
	    slot = self.slot
    slot.closing = defer.Deferred()
    slot.closing.addCallback(self.itemproc.close_spider)
    self._check_if_closing(spider, slot)
    return slot.closing
 | 
| 
	'Return True if there isn\'t any more spiders to process'
 | 
	def is_idle(self):
 | 
	    return (not self.slot)
 | 
| 
	'Handle the downloaded response or failure through the spider
callback/errback'
 | 
	def _scrape(self, response, request, spider):
 | 
	    assert isinstance(response, (Response, Failure))
    dfd = self._scrape2(response, request, spider)
    dfd.addErrback(self.handle_spider_error, request, response, spider)
    dfd.addCallback(self.handle_spider_output, request, response, spider)
    return dfd
 | 
| 
	'Handle the different cases of request\'s result been a Response or a
Failure'
 | 
	def _scrape2(self, request_result, request, spider):
 | 
	    if (not isinstance(request_result, Failure)):
        return self.spidermw.scrape_response(self.call_spider, request_result, request, spider)
    else:
        dfd = self.call_spider(request_result, request, spider)
        return dfd.addErrback(self._log_download_errors, request_result, request, spider)
 | 
| 
	'Process each Request/Item (given in the output parameter) returned
from the given spider'
 | 
	def _process_spidermw_output(self, output, request, response, spider):
 | 
	    if isinstance(output, Request):
        self.crawler.engine.crawl(request=output, spider=spider)
    elif isinstance(output, (BaseItem, dict)):
        self.slot.itemproc_size += 1
        dfd = self.itemproc.process_item(output, spider)
        dfd.addBoth(self._itemproc_finished, output, response, spider)
        return dfd
    elif (output is None):
        pass
    else:
        typename = type(output).__name__
        logger.error('Spider    must    return    Request,    BaseItem,    dict    or    None,    got    %(typename)r    in    %(request)s', {'request': request, 'typename': typename}, extra={'spider': spider})
 | 
| 
	'Log and silence errors that come from the engine (typically download
errors that got propagated thru here)'
 | 
	def _log_download_errors(self, spider_failure, download_failure, request, spider):
 | 
	    if (isinstance(download_failure, Failure) and (not download_failure.check(IgnoreRequest))):
        if download_failure.frames:
            logger.error('Error    downloading    %(request)s', {'request': request}, exc_info=failure_to_exc_info(download_failure), extra={'spider': spider})
        else:
            errmsg = download_failure.getErrorMessage()
            if errmsg:
                logger.error('Error    downloading    %(request)s:    %(errmsg)s', {'request': request, 'errmsg': errmsg}, extra={'spider': spider})
    if (spider_failure is not download_failure):
        return spider_failure
 | 
| 
	'ItemProcessor finished for the given ``item`` and returned ``output``'
 | 
	def _itemproc_finished(self, output, item, response, spider):
 | 
	    self.slot.itemproc_size -= 1
    if isinstance(output, Failure):
        ex = output.value
        if isinstance(ex, DropItem):
            logkws = self.logformatter.dropped(item, ex, response, spider)
            logger.log(extra={'spider': spider}, *logformatter_adapter(logkws))
            return self.signals.send_catch_log_deferred(signal=signals.item_dropped, item=item, response=response, spider=spider, exception=output.value)
        else:
            logger.error('Error    processing    %(item)s', {'item': item}, exc_info=failure_to_exc_info(output), extra={'spider': spider})
    else:
        logkws = self.logformatter.scraped(output, response, spider)
        logger.log(extra={'spider': spider}, *logformatter_adapter(logkws))
        return self.signals.send_catch_log_deferred(signal=signals.item_scraped, item=output, response=response, spider=spider)
 | 
| 
	'Lazy-load the downloadhandler for a scheme
only on the first request for that scheme.'
 | 
	def _get_handler(self, scheme):
 | 
	    if (scheme in self._handlers):
        return self._handlers[scheme]
    if (scheme in self._notconfigured):
        return None
    if (scheme not in self._schemes):
        self._notconfigured[scheme] = 'no    handler    available    for    that    scheme'
        return None
    path = self._schemes[scheme]
    try:
        dhcls = load_object(path)
        dh = dhcls(self._crawler.settings)
    except NotConfigured as ex:
        self._notconfigured[scheme] = str(ex)
        return None
    except Exception as ex:
        logger.error('Loading    "%(clspath)s"    for    scheme    "%(scheme)s"', {'clspath': path, 'scheme': scheme}, exc_info=True, extra={'crawler': self._crawler})
        self._notconfigured[scheme] = str(ex)
        return None
    else:
        self._handlers[scheme] = dh
    return self._handlers[scheme]
 | 
| 
	'Return a deferred for the HTTP download'
 | 
	def download_request(self, request, spider):
 | 
	    agent = ScrapyAgent(contextFactory=self._contextFactory, pool=self._pool, maxsize=getattr(spider, 'download_maxsize', self._default_maxsize), warnsize=getattr(spider, 'download_warnsize', self._default_warnsize), fail_on_dataloss=self._fail_on_dataloss)
    return agent.download_request(request)
 | 
| 
	'Asks the proxy to open a tunnel.'
 | 
	def requestTunnel(self, protocol):
 | 
	    tunnelReq = tunnel_request_data(self._tunneledHost, self._tunneledPort, self._proxyAuthHeader)
    protocol.transport.write(tunnelReq)
    self._protocolDataReceived = protocol.dataReceived
    protocol.dataReceived = self.processProxyResponse
    self._protocol = protocol
    return protocol
 | 
| 
	'Processes the response from the proxy. If the tunnel is successfully
created, notifies the client that we are ready to send requests. If not
raises a TunnelError.'
 | 
	def processProxyResponse(self, rcvd_bytes):
 | 
	    self._connectBuffer += rcvd_bytes
    if ('\r\n\r\n' not in self._connectBuffer):
        return
    self._protocol.dataReceived = self._protocolDataReceived
    respm = TunnelingTCP4ClientEndpoint._responseMatcher.match(self._connectBuffer)
    if (respm and (int(respm.group('status')) == 200)):
        try:
            sslOptions = self._contextFactory.creatorForNetloc(self._tunneledHost, self._tunneledPort)
        except AttributeError:
            sslOptions = self._contextFactory
        self._protocol.transport.startTLS(sslOptions, self._protocolFactory)
        self._tunnelReadyDeferred.callback(self._protocol)
    else:
        if respm:
            extra = {'status': int(respm.group('status')), 'reason': respm.group('reason').strip()}
        else:
            extra = rcvd_bytes[:32]
        self._tunnelReadyDeferred.errback(TunnelError(('Could    not    open    CONNECT    tunnel    with    proxy    %s:%s    [%r]' % (self._host, self._port, extra))))
 | 
| 
	'Propagates the errback to the appropriate deferred.'
 | 
	def connectFailed(self, reason):
 | 
	    self._tunnelReadyDeferred.errback(reason)
 | 
| 
	'Return a deferred for the HTTP download'
 | 
	def download_request(self, request, spider):
 | 
	    factory = self.HTTPClientFactory(request)
    self._connect(factory)
    return factory.deferred
 | 
| 
	'Upload file to S3 storage'
 | 
	def persist_file(self, path, buf, info, meta=None, headers=None):
 | 
	    key_name = ('%s%s' % (self.prefix, path))
    buf.seek(0)
    if self.is_botocore:
        extra = self._headers_to_botocore_kwargs(self.HEADERS)
        if headers:
            extra.update(self._headers_to_botocore_kwargs(headers))
        return threads.deferToThread(self.s3_client.put_object, Bucket=self.bucket, Key=key_name, Body=buf, Metadata={k: str(v) for (k, v) in six.iteritems((meta or {}))}, ACL=self.POLICY, **extra)
    else:
        b = self._get_boto_bucket()
        k = b.new_key(key_name)
        if meta:
            for (metakey, metavalue) in six.iteritems(meta):
                k.set_metadata(metakey, str(metavalue))
        h = self.HEADERS.copy()
        if headers:
            h.update(headers)
        return threads.deferToThread(k.set_contents_from_string, buf.getvalue(), headers=h, policy=self.POLICY)
 | 
| 
	'Convert headers to botocore keyword agruments.'
 | 
	def _headers_to_botocore_kwargs(self, headers):
 | 
	    mapping = CaselessDict({'Content-Type': 'ContentType', 'Cache-Control': 'CacheControl', 'Content-Disposition': 'ContentDisposition', 'Content-Encoding': 'ContentEncoding', 'Content-Language': 'ContentLanguage', 'Content-Length': 'ContentLength', 'Content-MD5': 'ContentMD5', 'Expires': 'Expires', 'X-Amz-Grant-Full-Control': 'GrantFullControl', 'X-Amz-Grant-Read': 'GrantRead', 'X-Amz-Grant-Read-ACP': 'GrantReadACP', 'X-Amz-Grant-Write-ACP': 'GrantWriteACP'})
    extra = {}
    for (key, value) in six.iteritems(headers):
        try:
            kwarg = mapping[key]
        except KeyError:
            raise TypeError(('Header    "%s"    is    not    supported    by    botocore' % key))
        else:
            extra[kwarg] = value
    return extra
 | 
			Subsets and Splits
				
	
				
			
				
No community queries yet
The top public SQL queries from the community will appear here once available.
