# Copyright (C) 2014 INRIA
#
# This program is free software: you can redistribute it and/or modify
-# it under the terms of the GNU General Public License as published by
-# the Free Software Foundation, either version 3 of the License, or
-# (at your option) any later version.
+# it under the terms of the GNU General Public License version 2 as
+# published by the Free Software Foundation;
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# \ nid3.log
#
+from __future__ import print_function
+
import collections
import functools
import networkx
faces = dict()
sep = " "
- f = open(filename, "r")
-
- data = []
-
- for line in f:
- cols = line.strip().split(sep)
-
- # CCN_PEEK
- # MESSAGE interest_from
- # 1374181938.808523 ccnd[9245]: debug.4352 interest_from 6 ccnx:/test/bunny.ts (23 bytes,sim=0CDCC1D7)
- #
- # MESSAGE interest_to
- # 1374181938.812750 ccnd[9245]: debug.3502 interest_to 5 ccnx:/test/bunny.ts (39 bytes,i=2844,sim=0CDCC1D7)
- #
- # MESSAGE CONTENT FROM
- # 1374181938.868682 ccnd[9245]: debug.4643 content_from 5 ccnx:/test/bunny.ts/%FD%05%1E%85%8FVw/%00/%9E%3D%01%D9%3Cn%95%2BvZ%8
- #
- # MESSAGE CONTENT_TO
- # 1374181938.868772 ccnd[9245]: debug.1619 content_to 6 ccnx:/test/bunny.ts/%FD%05%1E%85%8FVw/%00/%9E%3D%01%D9%3Cn%95%2BvZ%8
- #
- # 1375596708.222304 ccnd[9758]: debug.3692 interest_expiry ccnx:/test/bunny.ts/%FD%05%1E%86%B1GS/%00%0A%F7 (44 bytes,c=0:1,i=2819,sim=49FA8048)
-
- # External face creation
- # 1374181452.965961 ccnd[9245]: accepted datagram client id=5 (flags=0x40012) 204.85.191.10 port 9695
-
- if line.find("accepted datagram client") > -1:
- face_id = (cols[5]).replace("id=",'')
- ip = cols[7]
- port = cols[9]
- faces[face_id] = (ip, port)
- continue
-
- # 1374181452.985296 ccnd[9245]: releasing face id 4 (slot 4)
- if line.find("releasing face id") > -1:
- face_id = cols[5]
- if face_id in faces:
- del faces[face_id]
- continue
-
- if len(cols) < 6:
- continue
-
- timestamp = cols[0]
- message_type = cols[3]
+ with open(filename, "r") as f:
+
+ data = []
+
+ for line in f:
+ cols = line.strip().split(sep)
+
+ # CCN_PEEK
+ # MESSAGE interest_from
+ # 1374181938.808523 ccnd[9245]: debug.4352 interest_from 6 ccnx:/test/bunny.ts (23 bytes,sim=0CDCC1D7)
+ #
+ # MESSAGE interest_to
+ # 1374181938.812750 ccnd[9245]: debug.3502 interest_to 5 ccnx:/test/bunny.ts (39 bytes,i=2844,sim=0CDCC1D7)
+ #
+ # MESSAGE CONTENT FROM
+ # 1374181938.868682 ccnd[9245]: debug.4643 content_from 5 ccnx:/test/bunny.ts/%FD%05%1E%85%8FVw/%00/%9E%3D%01%D9%3Cn%95%2BvZ%8
+ #
+ # MESSAGE CONTENT_TO
+ # 1374181938.868772 ccnd[9245]: debug.1619 content_to 6 ccnx:/test/bunny.ts/%FD%05%1E%85%8FVw/%00/%9E%3D%01%D9%3Cn%95%2BvZ%8
+ #
+ # 1375596708.222304 ccnd[9758]: debug.3692 interest_expiry ccnx:/test/bunny.ts/%FD%05%1E%86%B1GS/%00%0A%F7 (44 bytes,c=0:1,i=2819,sim=49FA8048)
+
+ # External face creation
+ # 1374181452.965961 ccnd[9245]: accepted datagram client id=5 (flags=0x40012) 204.85.191.10 port 9695
+
+ if line.find("accepted datagram client") > -1:
+ face_id = (cols[5]).replace("id=",'')
+ ip = cols[7]
+ port = cols[9]
+ faces[face_id] = (ip, port)
+ continue
- if message_type not in ["interest_from", "interest_to", "content_from",
- "content_to", "interest_dupnonce", "interest_expiry"]:
- continue
+ # 1374181452.985296 ccnd[9245]: releasing face id 4 (slot 4)
+ if line.find("releasing face id") > -1:
+ face_id = cols[5]
+ if face_id in faces:
+ del faces[face_id]
+ continue
- face_id = cols[4]
- content_name = cols[5]
+ if len(cols) < 6:
+ continue
- # Interest Nonce ? -> 412A74-0844-0008-50AA-F6EAD4
- nonce = ""
- if message_type in ["interest_from", "interest_to", "interest_dupnonce"]:
- last = cols[-1]
- if len(last.split("-")) == 5:
- nonce = last
+ timestamp = cols[0]
+ message_type = cols[3]
- try:
- size = int((cols[6]).replace('(',''))
- except:
- print "interest_expiry without face id!", line
- continue
+ if message_type not in ["interest_from", "interest_to", "content_from",
+ "content_to", "interest_dupnonce", "interest_expiry"]:
+ continue
- # If no external IP address was identified for this face
- # asume it is a local face
- hostip = "localhost"
+ face_id = cols[4]
+ content_name = cols[5]
- if face_id in faces:
- hostip, port = faces[face_id]
+ # Interest Nonce ? -> 412A74-0844-0008-50AA-F6EAD4
+ nonce = ""
+ if message_type in ["interest_from", "interest_to", "interest_dupnonce"]:
+ last = cols[-1]
+ if len(last.split("-")) == 5:
+ nonce = last
- data.append((content_name, timestamp, message_type, hostip, face_id,
- size, nonce, line))
+ try:
+ size = int((cols[6]).replace('(',''))
+ except:
+ print("interest_expiry without face id!", line)
+ continue
- f.close()
+ # If no external IP address was identified for this face
+ # asume it is a local face
+ peer = "localhost"
+
+ if face_id in faces:
+ peer, port = faces[face_id]
+
+ data.append((content_name, timestamp, message_type, peer, face_id,
+ size, nonce, line))
return data
return f.name
def load_content_history(fname):
- f = open(fname, "r")
- content_history = pickle.load(f)
- f.close()
+ with open(fname, "r") as f:
+ content_history = pickle.load(f)
os.remove(fname)
return content_history
def annotate_cn_node(graph, nid, ips2nid, data, content_history):
- for (content_name, timestamp, message_type, hostip, face_id,
+ for (content_name, timestamp, message_type, peer, face_id,
size, nonce, line) in data:
- # Ignore local messages for the time being.
- # They could later be used to calculate the processing times
- # of messages.
- if peer == "localhost":
- return
-
# Ignore control messages for the time being
if is_control(content_name):
- return
+ continue
if message_type == "interest_from" and \
peer == "localhost":
peer == "localhost":
graph.node[nid]["ccn_producer"] = True
+ # Ignore local messages for the time being.
+ # They could later be used to calculate the processing times
+ # of messages.
+ if peer == "localhost":
+ continue
+
# remove digest
if message_type in ["content_from", "content_to"]:
content_name = "/".join(content_name.split("/")[:-1])
content_history[content_name] = list()
peernid = ips2nid[peer]
- add_edge(graph, nid, peernid)
+ graph.add_edge(nid, peernid)
content_history[content_name].append((timestamp, message_type, nid,
peernid, nonce, size, line))
""" Adds CCN content history for each node in the topology graph.
"""
+
# Make a copy of the graph to ensure integrity
graph = graph.copy()
- ips2nids = dict()
+ ips2nid = dict()
for nid in graph.nodes():
ips = graph.node[nid]["ips"]
for ip in ips:
- ips2nids[ip] = nid
+ ips2nid[ip] = nid
+
+ found_files = False
# Now walk through the ccnd logs...
for dirpath, dnames, fnames in os.walk(logs_dir):
# Each dirpath correspond to a different node
nid = os.path.basename(dirpath)
+
+ # Cast to numeric nid if necessary
+ if int(nid) in graph.nodes():
+ nid = int(nid)
content_history = dict()
for fname in fnames:
if fname.endswith(".log"):
+ found_files = True
filename = os.path.join(dirpath, fname)
data = parse_file(filename)
- annotate_cn_node(graph, nid, ips2nids, data, content_history)
+ annotate_cn_node(graph, nid, ips2nid, data, content_history)
# Avoid storing everything in memory, instead dump to a file
# and reference the file
fname = dump_content_history(content_history)
graph.node[nid]["history"] = fname
+ if not found_files:
+ msg = "No CCND output files were found to parse at %s " % logs_dir
+ raise RuntimeError(msg)
+
if parse_ping_logs:
ping_parser.annotate_cn_graph(logs_dir, graph)
content_names[content_name]["rtt"] = rtt
content_names[content_name]["lapse"] = (interest_timestamp, content_timestamp)
- return (content_names,
+ return (graph,
+ content_names,
interest_expiry_count,
interest_dupnonce_count,
interest_count,
content_count)
-def process_content_history_logs(logs_dir, graph):
+def process_content_history_logs(logs_dir, graph, parse_ping_logs = False):
""" Parse CCN logs and aggregate content history information in graph.
Returns annotated graph and message countn and content names history.
## Process logs and analyse data
try:
graph = annotate_cn_graph(logs_dir, graph,
- parse_ping_logs = True)
+ parse_ping_logs = parse_ping_logs)
except:
- print "Skipping: Error parsing ccnd logs", logs_dir
+ print("Skipping: Error parsing ccnd logs", logs_dir)
raise
- source = consumers(graph)[0]
- target = producers(graph)[0]
+ source = ccn_consumers(graph)[0]
+ target = ccn_producers(graph)[0]
# Process the data from the ccnd logs, but do not re compute
# the link delay.
interest_count,
content_count) = process_content_history(graph)
except:
- print "Skipping: Error processing ccn data", logs_dir
+ print("Skipping: Error processing ccn data", logs_dir)
raise
return (graph,