synapse/contrib/graph/graph2.py

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

165 lines
4.8 KiB
Python
Raw Permalink Normal View History

2014-12-10 19:10:25 +03:00
#
2023-11-21 23:29:58 +03:00
# This file is licensed under the Affero General Public License (AGPL) version 3.
#
# Copyright 2014-2016 OpenMarket Ltd
2023-11-21 23:29:58 +03:00
# Copyright (C) 2023 New Vector, Ltd
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# See the GNU Affero General Public License for more details:
# <https://www.gnu.org/licenses/agpl-3.0.html>.
#
# Originally licensed under the Apache License, Version 2.0:
# <http://www.apache.org/licenses/LICENSE-2.0>.
#
# [This file includes modifications made by New Vector Limited]
2014-12-10 19:10:25 +03:00
#
#
import argparse
2014-12-10 19:10:25 +03:00
import datetime
import html
import json
import sqlite3
import pydot
2014-12-10 19:10:25 +03:00
from synapse.api.room_versions import KNOWN_ROOM_VERSIONS
from synapse.events import make_event_from_dict
from synapse.util.frozenutils import unfreeze
2014-12-10 19:10:25 +03:00
def make_graph(db_name: str, room_id: str, file_prefix: str, limit: int) -> None:
"""
Generate a dot and SVG file for a graph of events in the room based on the
topological ordering by reading from a Synapse SQLite database.
"""
2014-12-10 19:10:25 +03:00
conn = sqlite3.connect(db_name)
sql = "SELECT room_version FROM rooms WHERE room_id = ?"
c = conn.execute(sql, (room_id,))
room_version = KNOWN_ROOM_VERSIONS[c.fetchone()[0]]
2015-01-08 13:53:03 +03:00
sql = (
"SELECT json, internal_metadata FROM event_json as j "
2015-01-08 13:53:03 +03:00
"INNER JOIN events as e ON e.event_id = j.event_id "
"WHERE j.room_id = ?"
2014-12-10 19:10:25 +03:00
)
2015-01-08 13:53:03 +03:00
args = [room_id]
if limit:
2019-10-23 18:49:05 +03:00
sql += " ORDER BY topological_ordering DESC, stream_ordering DESC LIMIT ?"
2015-01-08 13:53:03 +03:00
args.append(limit)
c = conn.execute(sql, args)
events = [
make_event_from_dict(json.loads(e[0]), room_version, json.loads(e[1]))
for e in c.fetchall()
]
2014-12-10 19:10:25 +03:00
events.sort(key=lambda e: e.depth)
node_map = {}
state_groups = {}
graph = pydot.Dot(graph_name="Test")
for event in events:
c = conn.execute(
2019-10-23 18:49:05 +03:00
"SELECT state_group FROM event_to_state_groups WHERE event_id = ?",
2014-12-10 19:10:25 +03:00
(event.event_id,),
)
res = c.fetchone()
state_group = res[0] if res else None
if state_group is not None:
state_groups.setdefault(state_group, []).append(event.event_id)
t = datetime.datetime.fromtimestamp(
float(event.origin_server_ts) / 1000
).strftime("%Y-%m-%d %H:%M:%S,%f")
content = json.dumps(unfreeze(event.get_dict()["content"]))
2014-12-10 19:10:25 +03:00
label = (
"<"
"<b>%(name)s </b><br/>"
"Type: <b>%(type)s </b><br/>"
"State key: <b>%(state_key)s </b><br/>"
"Content: <b>%(content)s </b><br/>"
"Time: <b>%(time)s </b><br/>"
"Depth: <b>%(depth)s </b><br/>"
"State group: %(state_group)s<br/>"
">"
) % {
"name": event.event_id,
"type": event.type,
"state_key": event.get("state_key", None),
"content": html.escape(content, quote=True),
2014-12-10 19:10:25 +03:00
"time": t,
"depth": event.depth,
"state_group": state_group,
}
node = pydot.Node(name=event.event_id, label=label)
node_map[event.event_id] = node
graph.add_node(node)
for event in events:
for prev_id in event.prev_event_ids():
2014-12-10 19:10:25 +03:00
try:
end_node = node_map[prev_id]
except Exception:
end_node = pydot.Node(name=prev_id, label=f"<<b>{prev_id}</b>>")
2014-12-10 19:10:25 +03:00
node_map[prev_id] = end_node
graph.add_node(end_node)
edge = pydot.Edge(node_map[event.event_id], end_node)
graph.add_edge(edge)
for group, event_ids in state_groups.items():
if len(event_ids) <= 1:
continue
cluster = pydot.Cluster(str(group), label=f"<State Group: {str(group)}>")
2014-12-10 19:10:25 +03:00
for event_id in event_ids:
cluster.add_node(node_map[event_id])
graph.add_subgraph(cluster)
graph.write("%s.dot" % file_prefix, format="raw", prog="dot")
graph.write_svg("%s.svg" % file_prefix, prog="dot")
2019-06-20 12:32:02 +03:00
2014-12-10 19:10:25 +03:00
if __name__ == "__main__":
parser = argparse.ArgumentParser(
description="Generate a PDU graph for a given room by talking "
"to the given Synapse SQLite file to get the list of PDUs. \n"
2014-12-10 19:10:25 +03:00
"Requires pydot."
)
parser.add_argument(
"-p",
"--prefix",
dest="prefix",
2015-01-08 13:53:03 +03:00
help="String to prefix output files with",
default="graph_output",
2014-12-10 19:10:25 +03:00
)
2015-01-08 13:53:03 +03:00
parser.add_argument("-l", "--limit", help="Only retrieve the last N events.")
2014-12-10 19:10:25 +03:00
parser.add_argument("db")
parser.add_argument("room")
args = parser.parse_args()
2015-01-08 13:53:03 +03:00
make_graph(args.db, args.room, args.prefix, args.limit)