|
|
|
import crochet
|
|
|
|
import datetime
|
|
|
|
import json
|
|
|
|
import pappyproxy
|
|
|
|
import pygments
|
|
|
|
import pprint
|
|
|
|
import shlex
|
|
|
|
import urllib
|
|
|
|
|
|
|
|
from pappyproxy.util import PappyException, utc2local, load_reqlist, print_table, print_request_rows, get_req_data_row
|
|
|
|
from pappyproxy.http import Request, repeatable_parse_qs
|
|
|
|
from twisted.internet import defer
|
|
|
|
from pappyproxy.plugin import async_main_context_ids
|
|
|
|
from pappyproxy.colors import Colors, Styles, verb_color, scode_color, path_formatter, host_color
|
|
|
|
from pygments.formatters import TerminalFormatter
|
|
|
|
from pygments.lexers.data import JsonLexer
|
|
|
|
from pygments.lexers.html import XmlLexer
|
|
|
|
|
|
|
|
###################
|
|
|
|
## Helper functions
|
|
|
|
|
|
|
|
def view_full_message(request, headers_only=False):
|
|
|
|
if headers_only:
|
|
|
|
print request.headers_section_pretty
|
|
|
|
else:
|
|
|
|
print request.full_message_pretty
|
|
|
|
|
|
|
|
def print_request_extended(request):
|
|
|
|
# Prints extended info for the request
|
|
|
|
title = "Request Info (reqid=%s)" % request.reqid
|
|
|
|
print Styles.TABLE_HEADER + title + Colors.ENDC
|
|
|
|
reqlen = len(request.body)
|
|
|
|
reqlen = '%d bytes' % reqlen
|
|
|
|
rsplen = 'No response'
|
|
|
|
|
|
|
|
mangle_str = 'Nothing mangled'
|
|
|
|
if request.unmangled:
|
|
|
|
mangle_str = 'Request'
|
|
|
|
|
|
|
|
if request.response:
|
|
|
|
response_code = str(request.response.response_code) + \
|
|
|
|
' ' + request.response.response_text
|
|
|
|
response_code = scode_color(response_code) + response_code + Colors.ENDC
|
|
|
|
rsplen = len(request.response.body)
|
|
|
|
rsplen = '%d bytes' % rsplen
|
|
|
|
|
|
|
|
if request.response.unmangled:
|
|
|
|
if mangle_str == 'Nothing mangled':
|
|
|
|
mangle_str = 'Response'
|
|
|
|
else:
|
|
|
|
mangle_str += ' and Response'
|
|
|
|
else:
|
|
|
|
response_code = ''
|
|
|
|
|
|
|
|
time_str = '--'
|
|
|
|
if request.time_start and request.time_end:
|
|
|
|
time_delt = request.time_end - request.time_start
|
|
|
|
time_str = "%.2f sec" % time_delt.total_seconds()
|
|
|
|
|
|
|
|
if request.is_ssl:
|
|
|
|
is_ssl = 'YES'
|
|
|
|
else:
|
|
|
|
is_ssl = 'NO'
|
|
|
|
|
|
|
|
if request.time_start:
|
|
|
|
dtobj = utc2local(request.time_start)
|
|
|
|
time_made_str = dtobj.strftime('%a, %b %d, %Y, %I:%M:%S %p')
|
|
|
|
else:
|
|
|
|
time_made_str = '--'
|
|
|
|
|
|
|
|
verb = verb_color(request.verb) + request.verb + Colors.ENDC
|
|
|
|
host = host_color(request.host) + request.host + Colors.ENDC
|
|
|
|
|
|
|
|
print_pairs = []
|
|
|
|
print_pairs.append(('Made on', time_made_str))
|
|
|
|
print_pairs.append(('ID', request.reqid))
|
|
|
|
print_pairs.append(('URL', request.url_color))
|
|
|
|
print_pairs.append(('Host', host))
|
|
|
|
print_pairs.append(('Path', path_formatter(request.full_path)))
|
|
|
|
print_pairs.append(('Verb', verb))
|
|
|
|
print_pairs.append(('Status Code', response_code))
|
|
|
|
print_pairs.append(('Request Length', reqlen))
|
|
|
|
print_pairs.append(('Response Length', rsplen))
|
|
|
|
if request.response and request.response.unmangled:
|
|
|
|
print_pairs.append(('Unmangled Response Length', len(request.response.unmangled.full_response)))
|
|
|
|
print_pairs.append(('Time', time_str))
|
|
|
|
print_pairs.append(('Port', request.port))
|
|
|
|
print_pairs.append(('SSL', is_ssl))
|
|
|
|
print_pairs.append(('Mangled', mangle_str))
|
|
|
|
print_pairs.append(('Tags', ', '.join(request.tags)))
|
|
|
|
if request.plugin_data:
|
|
|
|
print_pairs.append(('Plugin Data', request.plugin_data))
|
|
|
|
|
|
|
|
for k, v in print_pairs:
|
|
|
|
print Styles.KV_KEY+str(k)+': '+Styles.KV_VAL+str(v)
|
|
|
|
|
|
|
|
def print_tree(tree):
|
|
|
|
# Prints a tree. Takes in a sorted list of path tuples
|
|
|
|
_print_tree_helper(tree, 0, [])
|
|
|
|
|
|
|
|
def guess_pretty_print_fmt(msg):
|
|
|
|
if 'content-type' in msg.headers:
|
|
|
|
if 'json' in msg.headers['content-type']:
|
|
|
|
return 'json'
|
|
|
|
elif 'www-form' in msg.headers['content-type']:
|
|
|
|
return 'form'
|
|
|
|
elif 'application/xml' in msg.headers['content-type']:
|
|
|
|
return 'xml'
|
|
|
|
return 'text'
|
|
|
|
|
|
|
|
def pretty_print_body(fmt, body):
|
|
|
|
try:
|
|
|
|
if fmt.lower() == 'json':
|
|
|
|
d = json.loads(body.strip())
|
|
|
|
s = json.dumps(d, indent=4, sort_keys=True)
|
|
|
|
print pygments.highlight(s, JsonLexer(), TerminalFormatter())
|
|
|
|
elif fmt.lower() == 'form':
|
|
|
|
qs = repeatable_parse_qs(body)
|
|
|
|
for k, v in qs.all_pairs():
|
|
|
|
s = Colors.GREEN
|
|
|
|
s += '%s: ' % urllib.unquote(k)
|
|
|
|
s += Colors.ENDC
|
|
|
|
s += urllib.unquote(v)
|
|
|
|
print s
|
|
|
|
elif fmt.lower() == 'text':
|
|
|
|
print body
|
|
|
|
elif fmt.lower() == 'xml':
|
|
|
|
import xml.dom.minidom
|
|
|
|
xml = xml.dom.minidom.parseString(body)
|
|
|
|
print pygments.highlight(xml.toprettyxml(), XmlLexer(), TerminalFormatter())
|
|
|
|
else:
|
|
|
|
raise PappyException('"%s" is not a valid format' % fmt)
|
|
|
|
except PappyException as e:
|
|
|
|
raise e
|
|
|
|
except:
|
|
|
|
raise PappyException('Body could not be parsed as "%s"' % fmt)
|
|
|
|
|
|
|
|
def _get_tree_prefix(depth, print_bars, last):
|
|
|
|
if depth == 0:
|
|
|
|
return u''
|
|
|
|
else:
|
|
|
|
ret = u''
|
|
|
|
pb = print_bars + [True]
|
|
|
|
for i in range(depth):
|
|
|
|
if pb[i]:
|
|
|
|
ret += u'\u2502 '
|
|
|
|
else:
|
|
|
|
ret += u' '
|
|
|
|
if last:
|
|
|
|
ret += u'\u2514\u2500\u2500 '
|
|
|
|
else:
|
|
|
|
ret += u'\u251c\u2500\u2500 '
|
|
|
|
return ret
|
|
|
|
|
|
|
|
def _print_tree_helper(tree, depth, print_bars):
|
|
|
|
# Takes in a tree and prints it at the given depth
|
|
|
|
if tree == [] or tree == [()]:
|
|
|
|
return
|
|
|
|
while tree[0] == ():
|
|
|
|
tree = tree[1:]
|
|
|
|
if tree == [] or tree == [()]:
|
|
|
|
return
|
|
|
|
if len(tree) == 1 and len(tree[0]) == 1:
|
|
|
|
print _get_tree_prefix(depth, print_bars + [False], True) + tree[0][0]
|
|
|
|
return
|
|
|
|
|
|
|
|
curkey = tree[0][0]
|
|
|
|
subtree = []
|
|
|
|
for row in tree:
|
|
|
|
if row[0] != curkey:
|
|
|
|
if curkey == '':
|
|
|
|
curkey = '/'
|
|
|
|
print _get_tree_prefix(depth, print_bars, False) + curkey
|
|
|
|
if depth == 0:
|
|
|
|
_print_tree_helper(subtree, depth+1, print_bars + [False])
|
|
|
|
else:
|
|
|
|
_print_tree_helper(subtree, depth+1, print_bars + [True])
|
|
|
|
curkey = row[0]
|
|
|
|
subtree = []
|
|
|
|
subtree.append(row[1:])
|
|
|
|
if curkey == '':
|
|
|
|
curkey = '/'
|
|
|
|
print _get_tree_prefix(depth, print_bars, True) + curkey
|
|
|
|
_print_tree_helper(subtree, depth+1, print_bars + [False])
|
|
|
|
|
|
|
|
def print_params(req, params=None):
|
|
|
|
if not req.url_params.all_pairs() and not req.body:
|
|
|
|
print 'Request %s has no url or data parameters' % req.reqid
|
|
|
|
print ''
|
|
|
|
if req.url_params.all_pairs():
|
|
|
|
print Styles.TABLE_HEADER + "Url Params" + Colors.ENDC
|
|
|
|
for k, v in req.url_params.all_pairs():
|
|
|
|
if params is None or (params and k in params):
|
|
|
|
print Styles.KV_KEY+str(k)+': '+Styles.KV_VAL+str(v)
|
|
|
|
print ''
|
|
|
|
if req.body:
|
|
|
|
print Styles.TABLE_HEADER + "Body/POST Params" + Colors.ENDC
|
|
|
|
pretty_print_body(guess_pretty_print_fmt(req), req.body)
|
|
|
|
print ''
|
|
|
|
if req.cookies.all_pairs():
|
|
|
|
print Styles.TABLE_HEADER + "Cookies" + Colors.ENDC
|
|
|
|
for k, v in req.cookies.all_pairs():
|
|
|
|
if params is None or (params and k in params):
|
|
|
|
print Styles.KV_KEY+str(k)+': '+Styles.KV_VAL+str(v)
|
|
|
|
print ''
|
|
|
|
# multiform request when we support it
|
|
|
|
|
|
|
|
def add_param(found_params, kind, k, v, reqid):
|
|
|
|
if not k in found_params:
|
|
|
|
found_params[k] = {}
|
|
|
|
if kind in found_params[k]:
|
|
|
|
found_params[k][kind].append((reqid, v))
|
|
|
|
else:
|
|
|
|
found_params[k][kind] = [(reqid, v)]
|
|
|
|
|
|
|
|
def print_param_info(param_info):
|
|
|
|
for k, d in param_info.iteritems():
|
|
|
|
print Styles.TABLE_HEADER + k + Colors.ENDC
|
|
|
|
for param_type, valpairs in d.iteritems():
|
|
|
|
print param_type
|
|
|
|
value_ids = {}
|
|
|
|
for reqid, val in valpairs:
|
|
|
|
ids = value_ids.get(val, [])
|
|
|
|
ids.append(reqid)
|
|
|
|
value_ids[val] = ids
|
|
|
|
for val, ids in value_ids.iteritems():
|
|
|
|
if len(ids) <= 15:
|
|
|
|
idstr = ', '.join(ids)
|
|
|
|
else:
|
|
|
|
idstr = ', '.join(ids[:15]) + '...'
|
|
|
|
if val == '':
|
|
|
|
printstr = (Colors.RED + 'BLANK' + Colors.ENDC + 'x%d (%s)') % (len(ids), idstr)
|
|
|
|
else:
|
|
|
|
printstr = (Colors.GREEN + '%s' + Colors.ENDC + 'x%d (%s)') % (val, len(ids), idstr)
|
|
|
|
print printstr
|
|
|
|
print ''
|
|
|
|
|
|
|
|
####################
|
|
|
|
## Command functions
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def list_reqs(line):
|
|
|
|
"""
|
|
|
|
List the most recent in-context requests. By default shows the most recent 25
|
|
|
|
Usage: list [a|num]
|
|
|
|
|
|
|
|
If `a` is given, all the in-context requests are shown. If a number is given,
|
|
|
|
that many requests will be shown.
|
|
|
|
"""
|
|
|
|
args = shlex.split(line)
|
|
|
|
if len(args) > 0:
|
|
|
|
if args[0][0].lower() == 'a':
|
|
|
|
print_count = -1
|
|
|
|
else:
|
|
|
|
try:
|
|
|
|
print_count = int(args[0])
|
|
|
|
except:
|
|
|
|
print "Please enter a valid argument for list"
|
|
|
|
return
|
|
|
|
else:
|
|
|
|
print_count = 25
|
|
|
|
|
|
|
|
rows = []
|
|
|
|
ids = yield async_main_context_ids(print_count)
|
|
|
|
for i in ids:
|
|
|
|
req = yield Request.load_request(i)
|
|
|
|
rows.append(get_req_data_row(req))
|
|
|
|
print_request_rows(rows)
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def view_request_info(line):
|
|
|
|
"""
|
|
|
|
View information about request
|
|
|
|
Usage: view_request_info <reqid(s)>
|
|
|
|
"""
|
|
|
|
args = shlex.split(line)
|
|
|
|
if not args:
|
|
|
|
raise PappyException("Request id is required")
|
|
|
|
reqids = args[0]
|
|
|
|
|
|
|
|
reqs = yield load_reqlist(reqids)
|
|
|
|
|
|
|
|
for req in reqs:
|
|
|
|
print ''
|
|
|
|
print_request_extended(req)
|
|
|
|
print ''
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def view_request_headers(line):
|
|
|
|
"""
|
|
|
|
View the headers of the request
|
|
|
|
Usage: view_request_headers <reqid(s)>
|
|
|
|
"""
|
|
|
|
args = shlex.split(line)
|
|
|
|
if not args:
|
|
|
|
raise PappyException("Request id is required")
|
|
|
|
reqid = args[0]
|
|
|
|
|
|
|
|
reqs = yield load_reqlist(reqid)
|
|
|
|
for req in reqs:
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print 'Request %s:' % req.reqid
|
|
|
|
view_full_message(req, True)
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print '-'*30
|
|
|
|
print ''
|
|
|
|
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def view_full_request(line):
|
|
|
|
"""
|
|
|
|
View the full data of the request
|
|
|
|
Usage: view_full_request <reqid(s)>
|
|
|
|
"""
|
|
|
|
args = shlex.split(line)
|
|
|
|
if not args:
|
|
|
|
raise PappyException("Request id is required")
|
|
|
|
reqid = args[0]
|
|
|
|
|
|
|
|
reqs = yield load_reqlist(reqid)
|
|
|
|
for req in reqs:
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print 'Request %s:' % req.reqid
|
|
|
|
view_full_message(req)
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print '-'*30
|
|
|
|
print ''
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def view_request_bytes(line):
|
|
|
|
"""
|
|
|
|
View the raw bytes of the request. Use this if you want to redirect output to a file.
|
|
|
|
Usage: view_request_bytes <reqid(s)>
|
|
|
|
"""
|
|
|
|
args = shlex.split(line)
|
|
|
|
if not args:
|
|
|
|
raise PappyException("Request id is required")
|
|
|
|
reqid = args[0]
|
|
|
|
|
|
|
|
reqs = yield load_reqlist(reqid)
|
|
|
|
for req in reqs:
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print 'Request %s:' % req.reqid
|
|
|
|
print req.full_message
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print '-'*30
|
|
|
|
print ''
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def pretty_print_request(line):
|
|
|
|
"""
|
|
|
|
Print the body of the request pretty printed.
|
|
|
|
Usage: pretty_print_request <format> <reqid(s)>
|
|
|
|
"""
|
|
|
|
args = shlex.split(line)
|
|
|
|
if len(args) < 2:
|
|
|
|
raise PappyException("Usage: pretty_print_request <format> <reqid(s)>")
|
|
|
|
reqids = args[1]
|
|
|
|
|
|
|
|
reqs = yield load_reqlist(reqids)
|
|
|
|
for req in reqs:
|
|
|
|
pretty_print_body(args[0], req.body)
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def view_response_headers(line):
|
|
|
|
"""
|
|
|
|
View the headers of the response
|
|
|
|
Usage: view_response_headers <reqid(s)>
|
|
|
|
"""
|
|
|
|
reqs = yield load_reqlist(line)
|
|
|
|
for req in reqs:
|
|
|
|
if req.response:
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print '-'*15 + (' %s ' % req.reqid) + '-'*15
|
|
|
|
view_full_message(req.response, True)
|
|
|
|
else:
|
|
|
|
print "Request %s does not have a response" % req.reqid
|
|
|
|
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def view_full_response(line):
|
|
|
|
"""
|
|
|
|
View the full data of the response associated with a request
|
|
|
|
Usage: view_full_response <reqid>
|
|
|
|
"""
|
|
|
|
reqs = yield load_reqlist(line)
|
|
|
|
for req in reqs:
|
|
|
|
if req.response:
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print '-'*15 + (' %s ' % req.reqid) + '-'*15
|
|
|
|
view_full_message(req.response)
|
|
|
|
else:
|
|
|
|
print "Request %s does not have a response" % req.reqid
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def view_response_bytes(line):
|
|
|
|
"""
|
|
|
|
View the full data of the response associated with a request
|
|
|
|
Usage: view_request_bytes <reqid(s)>
|
|
|
|
"""
|
|
|
|
reqs = yield load_reqlist(line)
|
|
|
|
for req in reqs:
|
|
|
|
if req.response:
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print '-'*15 + (' %s ' % req.reqid) + '-'*15
|
|
|
|
print req.response.full_message
|
|
|
|
else:
|
|
|
|
print "Request %s does not have a response" % req.reqid
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def pretty_print_response(line):
|
|
|
|
"""
|
|
|
|
Print the body of the request pretty printed.
|
|
|
|
Usage: pretty_print_request <format> <reqid(s)>
|
|
|
|
"""
|
|
|
|
args = shlex.split(line)
|
|
|
|
if len(args) < 2:
|
|
|
|
raise PappyException("Usage: pretty_print_request <format> <reqid(s)>")
|
|
|
|
reqids = args[1]
|
|
|
|
|
|
|
|
reqs = yield load_reqlist(reqids)
|
|
|
|
for req in reqs:
|
|
|
|
if req.response:
|
|
|
|
pretty_print_body(args[0], req.response.body)
|
|
|
|
else:
|
|
|
|
print 'No response associated with request %s' % req.reqid
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def print_params_cmd(line):
|
|
|
|
"""
|
|
|
|
View the headers of the request
|
|
|
|
Usage: view_request_headers <reqid(s)>
|
|
|
|
"""
|
|
|
|
args = shlex.split(line)
|
|
|
|
reqid = args[0]
|
|
|
|
if len(args) > 1:
|
|
|
|
keys = args[1:]
|
|
|
|
else:
|
|
|
|
keys = None
|
|
|
|
|
|
|
|
reqs = yield load_reqlist(reqid)
|
|
|
|
for req in reqs:
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print 'Request %s:' % req.reqid
|
|
|
|
print_params(req, keys)
|
|
|
|
if len(reqs) > 1:
|
|
|
|
print '-'*30
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def get_param_info(line):
|
|
|
|
args = shlex.split(line)
|
|
|
|
if args and args[0] == 'ct':
|
|
|
|
contains = True
|
|
|
|
args = args[1:]
|
|
|
|
else:
|
|
|
|
contains = False
|
|
|
|
|
|
|
|
if args:
|
|
|
|
params = tuple(args)
|
|
|
|
else:
|
|
|
|
params = None
|
|
|
|
|
|
|
|
def check_key(k, params, contains):
|
|
|
|
if contains:
|
|
|
|
for p in params:
|
|
|
|
if p.lower() in k.lower():
|
|
|
|
return True
|
|
|
|
else:
|
|
|
|
if params is None or k in params:
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
found_params = {}
|
|
|
|
|
|
|
|
ids = yield async_main_context_ids()
|
|
|
|
for i in ids:
|
|
|
|
req = yield Request.load_request(i)
|
|
|
|
for k, v in req.url_params.all_pairs():
|
|
|
|
if check_key(k, params, contains):
|
|
|
|
add_param(found_params, 'Url Parameter', k, v, req.reqid)
|
|
|
|
for k, v in req.post_params.all_pairs():
|
|
|
|
if check_key(k, params, contains):
|
|
|
|
add_param(found_params, 'POST Parameter', k, v, req.reqid)
|
|
|
|
for k, v in req.cookies.all_pairs():
|
|
|
|
if check_key(k, params, contains):
|
|
|
|
add_param(found_params, 'Cookie', k, v, req.reqid)
|
|
|
|
print_param_info(found_params)
|
|
|
|
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def dump_response(line):
|
|
|
|
"""
|
|
|
|
Dump the data of the response to a file.
|
|
|
|
Usage: dump_response <id> <filename>
|
|
|
|
"""
|
|
|
|
# dump the data of a response
|
|
|
|
args = shlex.split(line)
|
|
|
|
reqs = yield load_reqlist(args[0])
|
|
|
|
for req in reqs:
|
|
|
|
if req.response:
|
|
|
|
rsp = req.response
|
|
|
|
if len(args) >= 2:
|
|
|
|
fname = args[1]
|
|
|
|
else:
|
|
|
|
fname = req.path.split('/')[-1]
|
|
|
|
|
|
|
|
with open(fname, 'w') as f:
|
|
|
|
f.write(rsp.body)
|
|
|
|
print 'Response data written to %s' % fname
|
|
|
|
else:
|
|
|
|
print 'Request %s does not have a response' % req.reqid
|
|
|
|
|
|
|
|
@crochet.wait_for(timeout=None)
|
|
|
|
@defer.inlineCallbacks
|
|
|
|
def site_map(line):
|
|
|
|
"""
|
|
|
|
Print the site map. Only includes requests in the current context.
|
|
|
|
Usage: site_map
|
|
|
|
"""
|
|
|
|
args = shlex.split(line)
|
|
|
|
if len(args) > 0 and args[0] == 'p':
|
|
|
|
paths = True
|
|
|
|
else:
|
|
|
|
paths = False
|
|
|
|
ids = yield async_main_context_ids()
|
|
|
|
paths_set = set()
|
|
|
|
for reqid in ids:
|
|
|
|
req = yield Request.load_request(reqid)
|
|
|
|
if req.response and req.response.response_code != 404:
|
|
|
|
paths_set.add(req.path_tuple)
|
|
|
|
tree = sorted(list(paths_set))
|
|
|
|
if paths:
|
|
|
|
for p in tree:
|
|
|
|
print ('/'.join(list(p)))
|
|
|
|
else:
|
|
|
|
print_tree(tree)
|
|
|
|
|
|
|
|
|
|
|
|
###############
|
|
|
|
## Plugin hooks
|
|
|
|
|
|
|
|
def load_cmds(cmd):
|
|
|
|
cmd.set_cmds({
|
|
|
|
'list': (list_reqs, None),
|
|
|
|
'view_request_info': (view_request_info, None),
|
|
|
|
'view_request_headers': (view_request_headers, None),
|
|
|
|
'view_full_request': (view_full_request, None),
|
|
|
|
'view_request_bytes': (view_request_bytes, None),
|
|
|
|
'pretty_print_request': (pretty_print_request, None),
|
|
|
|
'view_response_headers': (view_response_headers, None),
|
|
|
|
'view_full_response': (view_full_response, None),
|
|
|
|
'view_response_bytes': (view_response_bytes, None),
|
|
|
|
'pretty_print_response': (pretty_print_response, None),
|
|
|
|
'print_params': (print_params_cmd, None),
|
|
|
|
'param_info': (get_param_info, None),
|
|
|
|
'site_map': (site_map, None),
|
|
|
|
'dump_response': (dump_response, None),
|
|
|
|
})
|
|
|
|
cmd.add_aliases([
|
|
|
|
('list', 'ls'),
|
|
|
|
('view_request_info', 'viq'),
|
|
|
|
('view_request_headers', 'vhq'),
|
|
|
|
('view_full_request', 'vfq'),
|
|
|
|
('view_full_request', 'kjq'),
|
|
|
|
('view_request_bytes', 'vbq'),
|
|
|
|
('pretty_print_request', 'ppq'),
|
|
|
|
('view_response_headers', 'vhs'),
|
|
|
|
('view_full_response', 'vfs'),
|
|
|
|
('view_full_response', 'kjs'),
|
|
|
|
('view_response_bytes', 'vbs'),
|
|
|
|
('pretty_print_response', 'pps'),
|
|
|
|
('print_params', 'pprm'),
|
|
|
|
('param_info', 'pri'),
|
|
|
|
('site_map', 'sm'),
|
|
|
|
#('dump_response', 'dr'),
|
|
|
|
])
|