Compare commits
8 Commits
32dd37feea
...
main
| Author | SHA1 | Date | |
|---|---|---|---|
|
3414b6c145
|
|||
|
72c9790165
|
|||
|
6adc1e46bd
|
|||
|
7581966c88
|
|||
|
ead3d70c35
|
|||
|
b323c724c9
|
|||
|
2ef0343338
|
|||
|
4c87d4e7b0
|
@@ -26,7 +26,7 @@ def leverage(g, weight):
|
|||||||
neighbours = g.get_all_neighbours(v)
|
neighbours = g.get_all_neighbours(v)
|
||||||
ki = len(neighbours)
|
ki = len(neighbours)
|
||||||
# sum
|
# sum
|
||||||
for nv, props in neighbours:
|
for nv in neighbours:
|
||||||
other_neighbours = g.get_all_neighbours(nv)
|
other_neighbours = g.get_all_neighbours(nv)
|
||||||
kj = len(other_neighbours)
|
kj = len(other_neighbours)
|
||||||
li += (ki - kj) / (ki + kj)
|
li += (ki - kj) / (ki + kj)
|
||||||
@@ -104,7 +104,15 @@ def spatial_graph(adata):
|
|||||||
|
|
||||||
def apply(g, seed, weight, convex_hull, ax, method, method_name):
|
def apply(g, seed, weight, convex_hull, ax, method, method_name):
|
||||||
# calculate centrality values
|
# calculate centrality values
|
||||||
vp = method(g, weight=weight)
|
vp = None
|
||||||
|
if method_name == "Betweeness":
|
||||||
|
vp, ep = method(g, weight=weight)
|
||||||
|
elif method_name == "Eigenvector":
|
||||||
|
ep, vp = method(g, weight=weight)
|
||||||
|
elif method_name == "Hits":
|
||||||
|
ep, vp, hub_centrality = method(g, weight=weight)
|
||||||
|
else:
|
||||||
|
vp = method(g, weight=weight)
|
||||||
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
||||||
|
|
||||||
# normalization
|
# normalization
|
||||||
@@ -128,7 +136,7 @@ def apply(g, seed, weight, convex_hull, ax, method, method_name):
|
|||||||
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||||
)
|
)
|
||||||
# plot model containing modeled piece-wise linear function
|
# plot model containing modeled piece-wise linear function
|
||||||
plot.quantification_plot(ax, quantification, d_curve, C_curve, 'Models', aic_opt)
|
plot.quantification_plot(ax, quantification, d_curve, C_curve, method_name, aic_opt)
|
||||||
|
|
||||||
|
|
||||||
#
|
#
|
||||||
@@ -139,30 +147,33 @@ def apply(g, seed, weight, convex_hull, ax, method, method_name):
|
|||||||
# - apply centrality measure to the next axis
|
# - apply centrality measure to the next axis
|
||||||
# - Draw the corresponding resulting models into a grid
|
# - Draw the corresponding resulting models into a grid
|
||||||
#
|
#
|
||||||
points, seed = random_graph()
|
points, seed = random_graph(n=5000)
|
||||||
g, weight = spatial_graph(points)
|
g, weight = spatial_graph(points)
|
||||||
g = GraphView(g)
|
g = GraphView(g)
|
||||||
# calculate convex hull
|
# calculate convex hull
|
||||||
convex_hull = centrality.convex_hull(g)
|
convex_hull = centrality.convex_hull(g)
|
||||||
|
|
||||||
# plot graph with convex_hull
|
# plot graph with convex_hull
|
||||||
fig_graph, ax_graph = plt.subplots(figsize=(15, 5))
|
fig_graph, ax_graph = plt.subplots(figsize=(15, 12))
|
||||||
# draw without any centrality measure `vp`
|
# draw without any centrality measure `vp`
|
||||||
plot.graph_plot(fig_graph, ax_graph, g, vp, convex_hull, f"Pointcloud (seed: {seed}\n{method_name}")
|
vp = g.new_vertex_property("double")
|
||||||
|
plot.graph_plot(fig_graph, ax_graph, g, vp, convex_hull, f"Pointcloud (seed: {seed})")
|
||||||
fig_graph.savefig("Pointcloud_graph.svg", format='svg')
|
fig_graph.savefig("Pointcloud_graph.svg", format='svg')
|
||||||
|
|
||||||
fig = plt.figure(figsize=(15, 10))
|
fig = plt.figure(figsize=(15, 12))
|
||||||
axs = fig.subplots(2, 4)
|
row1, row2 = fig.subplots(2, 4)
|
||||||
|
|
||||||
for ax in axs:
|
ax1, ax2, ax3, ax4 = row1
|
||||||
# TODO select corresponding centrality measure method
|
# TODO select corresponding centrality measure method
|
||||||
apply(g, seed, weight, convex_hull, ax, closeness, "Closeness")
|
apply(g, seed, weight, convex_hull, ax1, closeness, "Closeness")
|
||||||
apply(g, seed, weight, convex_hull, ax, pagerank, "PageRank")
|
apply(g, seed, weight, convex_hull, ax2, pagerank, "PageRank")
|
||||||
apply(g, seed, weight, convex_hull, ax, betweeness, "Betweeness")
|
apply(g, seed, weight, convex_hull, ax3, betweenness, "Betweeness")
|
||||||
apply(g, seed, weight, convex_hull, ax, eigenvector, "Eigenvector")
|
apply(g, seed, weight, convex_hull, ax4, eigenvector, "Eigenvector")
|
||||||
apply(g, seed, weight, convex_hull, ax, katz, "Katz")
|
|
||||||
apply(g, seed, weight, convex_hull, ax, hits, "Hits")
|
ax1, ax2, ax3, ax4 = row2
|
||||||
apply(g, seed, weight, convex_hull, ax, leverage, "Leverage")
|
apply(g, seed, weight, convex_hull, ax1, katz, "Katz")
|
||||||
apply(g, seed, weight, convex_hull, ax, degree, "Degree")
|
apply(g, seed, weight, convex_hull, ax2, hits, "Hits")
|
||||||
|
apply(g, seed, weight, convex_hull, ax3, leverage, "Leverage")
|
||||||
|
apply(g, seed, weight, convex_hull, ax4, degree, "Degree")
|
||||||
|
|
||||||
fig.savefig(f"Comparison_Pointcloud.svg", format='svg')
|
fig.savefig(f"Comparison_Pointcloud.svg", format='svg')
|
||||||
|
|||||||
40
correction_visualization.py
Normal file
40
correction_visualization.py
Normal file
@@ -0,0 +1,40 @@
|
|||||||
|
import matplotlib.pyplot as plt
|
||||||
|
from matplotlib.collections import LineCollection
|
||||||
|
|
||||||
|
import numpy as np
|
||||||
|
|
||||||
|
b = 0.7
|
||||||
|
c0 = 0.2
|
||||||
|
m = 0.85
|
||||||
|
|
||||||
|
d_curve = np.linspace(0.3, 0.9, 500)
|
||||||
|
C_curve = np.piecewise(
|
||||||
|
d_curve,
|
||||||
|
[d_curve <= b, d_curve > b],
|
||||||
|
[lambda x: m * x + c0, lambda x: m * b + c0]
|
||||||
|
)
|
||||||
|
|
||||||
|
fig, ax = plt.subplots(figsize=(15, 12))
|
||||||
|
|
||||||
|
ax.set_xlabel('Distance to Bounding-Box')
|
||||||
|
ax.set_ylabel('Centrality Value')
|
||||||
|
|
||||||
|
ex = [0.5, 0.7, 0.5, 0.5]
|
||||||
|
ey = [m*b + c0, m*b + c0, m*0.5 + c0, m*b + c0]
|
||||||
|
ch = LineCollection([np.column_stack([ex, ey])], colors=['r'], linewidths=0.5)
|
||||||
|
ax.add_collection(ch)
|
||||||
|
ax.annotate("$𝛿$", xy=(0.48, 0.71), xytext=(0.48, 0.71), fontsize=18)
|
||||||
|
|
||||||
|
ax.scatter([0.5, 0.8, 0.85], [0.5, 0.75, 0.9], color='w', s=1)
|
||||||
|
ax.plot(d_curve, C_curve, color='k', linewidth=0.8)
|
||||||
|
ax.annotate("$x$", xy=(0.5, m*0.5 + c0 - 0.01), xytext=(0.5, m*0.5 + c0 - 0.01), fontsize=10)
|
||||||
|
ax.annotate("$b$", xy=(b, m*b + c0 + 0.005), xytext=(b, m*b + c0 + 0.005), fontsize=10)
|
||||||
|
# ax.annotate("$f(x) = m*x + c_0$", xy=(0.32, 0.58), xytext=(0.32, 0.58), fontsize=10)
|
||||||
|
# ax.annotate("$g(x) = m*b + c_0$", xy=(0.75, 0.8), xytext=(0.75, 0.8), fontsize=10)
|
||||||
|
|
||||||
|
ax.axis('off')
|
||||||
|
# ax.get_xaxis().set_visible(False)
|
||||||
|
# ax.get_yaxis().set_visible(False)
|
||||||
|
ax.set_aspect('equal')
|
||||||
|
|
||||||
|
fig.savefig("model_correction_visualization.svg", format='svg', bbox_inches='tight', pad_inches=0)
|
||||||
294
diff_comparison.py
Normal file
294
diff_comparison.py
Normal file
@@ -0,0 +1,294 @@
|
|||||||
|
import math
|
||||||
|
|
||||||
|
import matplotlib.pyplot as plt
|
||||||
|
import numpy as np
|
||||||
|
from graph_tool.all import *
|
||||||
|
|
||||||
|
from src import centrality
|
||||||
|
from src import plot
|
||||||
|
from src import fitting
|
||||||
|
|
||||||
|
|
||||||
|
def degree(g, weight):
|
||||||
|
# VertexPropertyMap
|
||||||
|
vp = g.new_vertex_property("double")
|
||||||
|
for v in g.vertices():
|
||||||
|
neighbours = g.get_all_neighbours(v)
|
||||||
|
vp[v] = len(neighbours)
|
||||||
|
return vp
|
||||||
|
|
||||||
|
|
||||||
|
def leverage(g, weight):
|
||||||
|
# VertexPropertyMap
|
||||||
|
vp = g.new_vertex_property("double")
|
||||||
|
for v in g.vertices():
|
||||||
|
li = 0.0
|
||||||
|
neighbours = g.get_all_neighbours(v)
|
||||||
|
ki = len(neighbours)
|
||||||
|
# sum
|
||||||
|
for nv in neighbours:
|
||||||
|
other_neighbours = g.get_all_neighbours(nv)
|
||||||
|
kj = len(other_neighbours)
|
||||||
|
li += (ki - kj) / (ki + kj)
|
||||||
|
li /= ki
|
||||||
|
vp[v] = li
|
||||||
|
return vp
|
||||||
|
|
||||||
|
|
||||||
|
def random_graph(n=5000, seed=None):
|
||||||
|
"""
|
||||||
|
Uniformly random point cloud generation.
|
||||||
|
`n` [int] Number of points to generate. Default 5000 seems like a good starting point in point density and corresponding runtime for the subsequent calculations.
|
||||||
|
@return [numpy.ndarray] Array of shape(n, 2) containing the coordinates for each point of the generated point cloud.
|
||||||
|
"""
|
||||||
|
if seed is None:
|
||||||
|
import secrets
|
||||||
|
seed = secrets.randbits(128)
|
||||||
|
rng = np.random.default_rng(seed=seed)
|
||||||
|
return rng.random((n, 2)), seed
|
||||||
|
|
||||||
|
|
||||||
|
def sub_spatial_graph(adata, percentage):
|
||||||
|
sub_adata = np.array([])
|
||||||
|
distance_of_center = 0.5 * percentage
|
||||||
|
for point in adata:
|
||||||
|
if point[0] > 0.5 - distance_of_center and point[0] <= 0.5 + distance_of_center:
|
||||||
|
if point[1] > 0.5 - distance_of_center and point[1] <= 0.5 + distance_of_center:
|
||||||
|
sub_adata = np.append(sub_adata, [point[0], point[1]])
|
||||||
|
|
||||||
|
sub_adata = sub_adata.reshape(sub_adata.shape[0] // 2, 2)
|
||||||
|
return spatial_graph(sub_adata)
|
||||||
|
|
||||||
|
|
||||||
|
def spatial_graph(adata):
|
||||||
|
"""
|
||||||
|
Generate the spatial graph using delaunay for the given `adata`.
|
||||||
|
`adata` will contain the calculated spatial graph contents in the keys
|
||||||
|
adata.obsm['spatial']` in case the `adata` is created from a dataset of *squidpy*.
|
||||||
|
@return [Graph] generated networkx graph from adata.obsp['spatial_distances']
|
||||||
|
"""
|
||||||
|
g, pos = graph_tool.generation.triangulation(adata, type="delaunay")
|
||||||
|
g.vp["pos"] = pos
|
||||||
|
weight = g.new_edge_property("double")
|
||||||
|
for e in g.edges():
|
||||||
|
weight[e] = math.sqrt(sum(map(abs, pos[e.source()].a - pos[e.target()].a)))**2
|
||||||
|
return g, weight
|
||||||
|
|
||||||
|
|
||||||
|
def plot_graph_diff(G, c, fig, ax, name, cmap=plt.cm.plasma):
|
||||||
|
pos = G.vp["pos"]
|
||||||
|
x = []
|
||||||
|
y = []
|
||||||
|
distance_of_center = 0.5 * percentage
|
||||||
|
for v in G.vertices():
|
||||||
|
ver = pos[v]
|
||||||
|
if ver[0] > 0.5 - distance_of_center and ver[0] <= 0.5 + distance_of_center:
|
||||||
|
if ver[1] > 0.5 - distance_of_center and ver[1] <= 0.5 + distance_of_center:
|
||||||
|
x.append(ver[0])
|
||||||
|
y.append(ver[1])
|
||||||
|
|
||||||
|
sc = ax.scatter(x, y, s=1, cmap=cmap, c=c) # map closeness values as color mapping on the verticies
|
||||||
|
ax.set_title(name)
|
||||||
|
fig.colorbar(sc, ax=ax)
|
||||||
|
|
||||||
|
|
||||||
|
def apply(g, seed, weight, convex_hull, ax, method, method_name):
|
||||||
|
# calculate centrality values
|
||||||
|
vp = None
|
||||||
|
if method_name == "Betweenness":
|
||||||
|
vp, ep = method(g, weight=weight)
|
||||||
|
elif method_name == "Eigenvector":
|
||||||
|
ep, vp = method(g, weight=weight)
|
||||||
|
elif method_name == "Hits":
|
||||||
|
ep, vp, hub_centrality = method(g, weight=weight)
|
||||||
|
else:
|
||||||
|
vp = method(g, weight=weight)
|
||||||
|
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
||||||
|
|
||||||
|
# generate model based on convex hull and associated centrality values
|
||||||
|
quantification = plot.quantification_data(g, vp, convex_hull)
|
||||||
|
|
||||||
|
# optimize model's piece-wise linear function
|
||||||
|
d = quantification[:, 0]
|
||||||
|
C = quantification[:, 1]
|
||||||
|
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||||
|
|
||||||
|
# TODO
|
||||||
|
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||||
|
d_curve = np.linspace(min(d), max(d), 500)
|
||||||
|
C_curve = np.piecewise(
|
||||||
|
d_curve,
|
||||||
|
[d_curve <= b_opt, d_curve > b_opt],
|
||||||
|
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||||
|
)
|
||||||
|
# plot model containing modeled piece-wise linear function
|
||||||
|
if ax is not None:
|
||||||
|
plot.quantification_plot(ax, quantification, d_curve, C_curve, method_name, aic_opt)
|
||||||
|
|
||||||
|
# normalization
|
||||||
|
min_val, max_val = vp.a.min(), vp.a.max()
|
||||||
|
vp.a = (vp.a - min_val) / (max_val - min_val)
|
||||||
|
|
||||||
|
return vp
|
||||||
|
|
||||||
|
|
||||||
|
def apply_corrected(g, seed, weight, convex_hull, ax, method, method_name):
|
||||||
|
# calculate centrality values
|
||||||
|
vp = None
|
||||||
|
ep = None
|
||||||
|
if method_name == "Betweenness":
|
||||||
|
vp, ep = method(g, weight=weight)
|
||||||
|
elif method_name == "Eigenvector":
|
||||||
|
ep, vp = method(g, weight=weight)
|
||||||
|
elif method_name == "Hits":
|
||||||
|
ep, vp, hub_centrality = method(g, weight=weight)
|
||||||
|
else:
|
||||||
|
vp = method(g, weight=weight)
|
||||||
|
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
||||||
|
|
||||||
|
# normalization
|
||||||
|
# min_val, max_val = vp.a.min(), vp.a.max()
|
||||||
|
# vp.a = (vp.a - min_val) / (max_val - min_val)
|
||||||
|
|
||||||
|
# generate model based on convex hull and associated centrality values
|
||||||
|
quantification = plot.quantification_data(g, vp, convex_hull)
|
||||||
|
|
||||||
|
# optimize model's piece-wise linear function
|
||||||
|
d = quantification[:, 0]
|
||||||
|
C = quantification[:, 1]
|
||||||
|
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||||
|
|
||||||
|
d_curve = np.linspace(min(d), max(d), 500)
|
||||||
|
C_curve = np.piecewise(
|
||||||
|
d_curve,
|
||||||
|
[d_curve <= b_opt, d_curve > b_opt],
|
||||||
|
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||||
|
)
|
||||||
|
# plot model containing modeled piece-wise linear function
|
||||||
|
if ax is not None:
|
||||||
|
plot.quantification_plot(ax, quantification, d_curve, C_curve, method_name, aic_opt)
|
||||||
|
|
||||||
|
vp = centrality.correct(g, vp, m_opt, c0_opt, b_opt)
|
||||||
|
# normalization
|
||||||
|
min_val, max_val = vp.a.min(), vp.a.max()
|
||||||
|
vp.a = (vp.a - min_val) / (max_val - min_val)
|
||||||
|
|
||||||
|
return vp
|
||||||
|
|
||||||
|
#
|
||||||
|
# - Create a random point cloud and calculate a triangulation on it
|
||||||
|
# - For that graph calculate the convex hull
|
||||||
|
# - Draw the graph with the convex hull
|
||||||
|
# - For each centrality measure
|
||||||
|
# - apply centrality measure to the next axis
|
||||||
|
# - Draw the corresponding resulting models into a grid
|
||||||
|
#
|
||||||
|
points, seed = random_graph(n=5000)
|
||||||
|
g, weight = spatial_graph(points)
|
||||||
|
g = GraphView(g)
|
||||||
|
|
||||||
|
# calculate convex hull
|
||||||
|
convex_hull = centrality.convex_hull(g)
|
||||||
|
|
||||||
|
# plot graph with convex_hull
|
||||||
|
fig_graph, ax_graph = plt.subplots(figsize=(15, 12))
|
||||||
|
# draw without any centrality measure `vp`
|
||||||
|
vp, ep = betweenness(g, weight=weight)
|
||||||
|
|
||||||
|
plot.graph_plot(fig_graph, ax_graph, g, vp, convex_hull, f"Pointcloud (seed: {seed})")
|
||||||
|
fig_graph.savefig("model_prediction_graph_original_betweenness_5000.svg", format='svg')
|
||||||
|
|
||||||
|
# normalization
|
||||||
|
min_val, max_val = vp.a.min(), vp.a.max()
|
||||||
|
vp.a = (vp.a - min_val) / (max_val - min_val)
|
||||||
|
vp_betweenness_original = vp
|
||||||
|
|
||||||
|
for percentage in np.arange(0.1, 1, 0.1, dtype=float):
|
||||||
|
print(f"Percentage: {percentage:.0%}")
|
||||||
|
g_sub, weight_sub = sub_spatial_graph(points, percentage)
|
||||||
|
g_sub = GraphView(g_sub)
|
||||||
|
convex_hull = centrality.convex_hull(g_sub)
|
||||||
|
# draw subgraph
|
||||||
|
fig_sub = plt.figure(figsize=(25, 12))
|
||||||
|
ax1, ax2 = fig_sub.subplots(1, 2)
|
||||||
|
vp, ep = betweenness(g_sub, weight=weight_sub)
|
||||||
|
plot.graph_plot(fig_sub, ax1, g_sub, vp, convex_hull, f"{percentage:.0%} of Pointcloud (seed: {seed})")
|
||||||
|
|
||||||
|
min_val, max_val = vp.a.min(), vp.a.max()
|
||||||
|
vp.a = (vp.a - min_val) / (max_val - min_val)
|
||||||
|
|
||||||
|
vp_betweenness_corrected = apply_corrected(g_sub, seed, weight_sub, convex_hull, None, betweenness, "Betweenness")
|
||||||
|
plot.graph_plot(fig_sub, ax2, g_sub, vp_betweenness_corrected, convex_hull, f"{percentage:.0%} of Pointcloud with applied prediction")
|
||||||
|
fig_sub.savefig(f"model_prediction_subgraph_betweenness_5000_{percentage * 100:.0f}_percent.svg", format='svg')
|
||||||
|
|
||||||
|
distance_of_center = 0.5 * percentage
|
||||||
|
|
||||||
|
sub_keys = iter(g_sub.vertices())
|
||||||
|
keys = iter(g.vertices())
|
||||||
|
|
||||||
|
scores = []
|
||||||
|
raw_sub_scores = []
|
||||||
|
sub_scores = []
|
||||||
|
raw_diff_scores = []
|
||||||
|
diff_scores = []
|
||||||
|
|
||||||
|
for sub_key in sub_keys:
|
||||||
|
key = next(keys)
|
||||||
|
position = g.vp["pos"][key]
|
||||||
|
while not (position[0] > 0.5 - distance_of_center and position[0] <= 0.5 + distance_of_center and position[1] > 0.5 - distance_of_center and position[1] <= 0.5 + distance_of_center):
|
||||||
|
key = next(keys)
|
||||||
|
position = g.vp["pos"][key]
|
||||||
|
# NOTE print corresponding position (which are identical)
|
||||||
|
# position = g.vp["pos"][key]
|
||||||
|
# sub_position = g_sub.vp["pos"][sub_key]
|
||||||
|
# print(f"position: {position} | sub_position: {sub_position}")
|
||||||
|
|
||||||
|
# calculate for betweenness
|
||||||
|
value = vp_betweenness_original[key]
|
||||||
|
pre_prediction = vp[sub_key]
|
||||||
|
sub_value = vp_betweenness_corrected[sub_key]
|
||||||
|
|
||||||
|
scores.append(value)
|
||||||
|
raw_sub_scores.append(pre_prediction)
|
||||||
|
sub_scores.append(sub_value)
|
||||||
|
raw_diff_scores.append(value - pre_prediction)
|
||||||
|
diff_scores.append(value - sub_value)
|
||||||
|
|
||||||
|
median_score = np.median(scores)
|
||||||
|
median_raw_sub_score = np.median(raw_sub_scores)
|
||||||
|
median_sub_score = np.median(sub_scores)
|
||||||
|
print(f"\tmedian score: {median_score}")
|
||||||
|
print(f"\tmedian raw_sub_score: {median_raw_sub_score}")
|
||||||
|
print(f"\tmedian sub_score: {median_sub_score}")
|
||||||
|
print(f"\tmedian delta (score - raw_sub_score): {(median_score - median_raw_sub_score)}")
|
||||||
|
print(f"\tmedian delta (score - sub_score): {(median_score - median_sub_score)}")
|
||||||
|
print("")
|
||||||
|
|
||||||
|
max_value_score = np.max(scores)
|
||||||
|
max_value_raw_sub_score = np.max(raw_sub_scores)
|
||||||
|
max_value_sub_score = np.max(sub_scores)
|
||||||
|
print(f"\tmax value score: {max_value_score}")
|
||||||
|
print(f"\tmax value raw_sub_score: {max_value_raw_sub_score}")
|
||||||
|
print(f"\tmax value sub_score: {max_value_sub_score}")
|
||||||
|
print(f"\tmax value delta (score - raw_sub_score): {(max_value_score - max_value_raw_sub_score)}")
|
||||||
|
print(f"\tmax value delta (score - sub_score): {(max_value_score - max_value_sub_score)}")
|
||||||
|
print("")
|
||||||
|
|
||||||
|
min_value_score = np.min(scores)
|
||||||
|
min_value_raw_sub_score = np.min(raw_sub_scores)
|
||||||
|
min_value_sub_score = np.min(sub_scores)
|
||||||
|
print(f"\tmin value score: {min_value_score}")
|
||||||
|
print(f"\tmin value raw_sub_score: {min_value_raw_sub_score}")
|
||||||
|
print(f"\tmin value sub_score: {min_value_sub_score}")
|
||||||
|
print(f"\tmin value delta (score - raw_sub_score): {(min_value_score - min_value_raw_sub_score)}")
|
||||||
|
print(f"\tmin value delta (score - sub_score): {(min_value_score - min_value_sub_score)}")
|
||||||
|
print("")
|
||||||
|
|
||||||
|
fig = plt.figure(figsize=(35, 10))
|
||||||
|
plot_graph_ax, plot_sub_graph_ax, plot_sub_graph_before_ax = fig.subplots(1, 3)
|
||||||
|
|
||||||
|
plot_graph_diff(g, scores, fig, plot_graph_ax, "Original Graph (region of sub graph)")
|
||||||
|
plot_graph_diff(g, diff_scores, fig, plot_sub_graph_ax, "Differences after correction of sub graph compared to original graph", plt.cm.seismic)
|
||||||
|
plot_graph_diff(g, vp.a, fig, plot_sub_graph_before_ax, "Sub Graph (extracted region of original graph) without correction")
|
||||||
|
|
||||||
|
fig.savefig(f"model_prediction_subgraph_betweenness_5000_{percentage * 100:.0f}_percentage_diff.svg", format='svg')
|
||||||
148
diff_model_comparison.py
Normal file
148
diff_model_comparison.py
Normal file
@@ -0,0 +1,148 @@
|
|||||||
|
import math
|
||||||
|
|
||||||
|
import matplotlib.pyplot as plt
|
||||||
|
import numpy as np
|
||||||
|
from graph_tool.all import *
|
||||||
|
|
||||||
|
from src import centrality
|
||||||
|
from src import plot
|
||||||
|
from src import fitting
|
||||||
|
|
||||||
|
|
||||||
|
def leverage(g, weight):
|
||||||
|
# VertexPropertyMap
|
||||||
|
vp = g.new_vertex_property("double")
|
||||||
|
for v in g.vertices():
|
||||||
|
li = 0.0
|
||||||
|
neighbours = g.get_all_neighbours(v)
|
||||||
|
ki = len(neighbours)
|
||||||
|
# sum
|
||||||
|
for nv in neighbours:
|
||||||
|
other_neighbours = g.get_all_neighbours(nv)
|
||||||
|
kj = len(other_neighbours)
|
||||||
|
li += (ki - kj) / (ki + kj)
|
||||||
|
li /= ki
|
||||||
|
vp[v] = li
|
||||||
|
return vp
|
||||||
|
|
||||||
|
|
||||||
|
def random_graph(n=5000, seed=None):
|
||||||
|
"""
|
||||||
|
Uniformly random point cloud generation.
|
||||||
|
`n` [int] Number of points to generate. Default 5000 seems like a good starting point in point density and corresponding runtime for the subsequent calculations.
|
||||||
|
@return [numpy.ndarray] Array of shape(n, 2) containing the coordinates for each point of the generated point cloud.
|
||||||
|
"""
|
||||||
|
if seed is None:
|
||||||
|
import secrets
|
||||||
|
seed = secrets.randbits(128)
|
||||||
|
rng = np.random.default_rng(seed=seed)
|
||||||
|
return rng.random((n, 2)), seed
|
||||||
|
|
||||||
|
|
||||||
|
def sub_spatial_graph(adata, percentage):
|
||||||
|
sub_adata = np.array([])
|
||||||
|
distance_of_center = 0.5 * percentage
|
||||||
|
for point in adata:
|
||||||
|
if point[0] > 0.5 - distance_of_center and point[0] <= 0.5 + distance_of_center:
|
||||||
|
if point[1] > 0.5 - distance_of_center and point[1] <= 0.5 + distance_of_center:
|
||||||
|
sub_adata = np.append(sub_adata, [point[0], point[1]])
|
||||||
|
|
||||||
|
sub_adata = sub_adata.reshape(sub_adata.shape[0] // 2, 2)
|
||||||
|
return spatial_graph(sub_adata)
|
||||||
|
|
||||||
|
|
||||||
|
def spatial_graph(adata):
|
||||||
|
"""
|
||||||
|
Generate the spatial graph using delaunay for the given `adata`.
|
||||||
|
`adata` will contain the calculated spatial graph contents in the keys
|
||||||
|
adata.obsm['spatial']` in case the `adata` is created from a dataset of *squidpy*.
|
||||||
|
@return [Graph] generated networkx graph from adata.obsp['spatial_distances']
|
||||||
|
"""
|
||||||
|
g, pos = graph_tool.generation.triangulation(adata, type="delaunay")
|
||||||
|
g.vp["pos"] = pos
|
||||||
|
weight = g.new_edge_property("double")
|
||||||
|
for e in g.edges():
|
||||||
|
weight[e] = math.sqrt(sum(map(abs, pos[e.source()].a - pos[e.target()].a)))**2
|
||||||
|
return g, weight
|
||||||
|
|
||||||
|
|
||||||
|
def apply(g, weight, convex_hull, ax, method, method_name):
|
||||||
|
# calculate centrality values
|
||||||
|
vp = None
|
||||||
|
if method_name == "Betweeness":
|
||||||
|
vp, ep = method(g, weight=weight)
|
||||||
|
elif method_name == "Eigenvector":
|
||||||
|
ep, vp = method(g, weight=weight)
|
||||||
|
elif method_name == "Hits":
|
||||||
|
ep, vp, hub_centrality = method(g, weight=weight)
|
||||||
|
else:
|
||||||
|
vp = method(g, weight=weight)
|
||||||
|
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
||||||
|
|
||||||
|
# normalization
|
||||||
|
min_val, max_val = vp.a.min(), vp.a.max()
|
||||||
|
vp.a = (vp.a - min_val) / (max_val - min_val)
|
||||||
|
|
||||||
|
# generate model based on convex hull and associated centrality values
|
||||||
|
quantification = plot.quantification_data(g, vp, convex_hull)
|
||||||
|
|
||||||
|
# optimize model's piece-wise linear function
|
||||||
|
d = quantification[:, 0]
|
||||||
|
C = quantification[:, 1]
|
||||||
|
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||||
|
|
||||||
|
# TODO
|
||||||
|
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||||
|
d_curve = np.linspace(min(d), max(d), 500)
|
||||||
|
C_curve = np.piecewise(
|
||||||
|
d_curve,
|
||||||
|
[d_curve <= b_opt, d_curve > b_opt],
|
||||||
|
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||||
|
)
|
||||||
|
# plot model containing modeled piece-wise linear function
|
||||||
|
plot.quantification_plot(ax, quantification, d_curve, C_curve, method_name, aic_opt)
|
||||||
|
|
||||||
|
|
||||||
|
#
|
||||||
|
# - Create a random point cloud and calculate a triangulation on it
|
||||||
|
# - For that graph calculate the convex hull
|
||||||
|
# - Draw the graph with the convex hull
|
||||||
|
# - For each centrality measure
|
||||||
|
# - apply centrality measure to the next axis
|
||||||
|
# - Draw the corresponding resulting models into a grid
|
||||||
|
#
|
||||||
|
points, seed = random_graph(n=5000)
|
||||||
|
g, weight = spatial_graph(points)
|
||||||
|
g = GraphView(g)
|
||||||
|
# calculate convex hull
|
||||||
|
convex_hull = centrality.convex_hull(g)
|
||||||
|
|
||||||
|
# plot graph with convex_hull
|
||||||
|
fig_graph, ax_graph = plt.subplots(figsize=(15, 12))
|
||||||
|
# draw without any centrality measure `vp`
|
||||||
|
vp = g.new_vertex_property("double")
|
||||||
|
plot.graph_plot(fig_graph, ax_graph, g, vp, convex_hull, f"Pointcloud (seed: {seed})")
|
||||||
|
fig_graph.savefig("point_cloud_diff_comparison_5000_pagerank_leverage.svg", format='svg')
|
||||||
|
|
||||||
|
fig = plt.figure(figsize=(15, 12))
|
||||||
|
row1, row2 = fig.subplots(2, 2)
|
||||||
|
|
||||||
|
ax1, ax2 = row1
|
||||||
|
apply(g, weight, convex_hull, ax1, pagerank, "PageRank")
|
||||||
|
apply(g, weight, convex_hull, ax2, leverage, "Leverage")
|
||||||
|
|
||||||
|
g_sub, weight_sub = sub_spatial_graph(points, 0.5)
|
||||||
|
g_sub = GraphView(g_sub)
|
||||||
|
convex_hull = centrality.convex_hull(g_sub)
|
||||||
|
# plot graph with convex_hull
|
||||||
|
fig_graph, ax_graph = plt.subplots(figsize=(15, 12))
|
||||||
|
# draw without any centrality measure `vp`
|
||||||
|
vp = g_sub.new_vertex_property("double")
|
||||||
|
plot.graph_plot(fig_graph, ax_graph, g_sub, vp, convex_hull, f"Pointcloud (50% of original)")
|
||||||
|
fig_graph.savefig("point_cloud_diff_comparison_5000_sub_pagerank_leverage.svg", format='svg')
|
||||||
|
|
||||||
|
ax1, ax2 = row2
|
||||||
|
apply(g_sub, weight_sub, convex_hull, ax1, pagerank, "PageRank")
|
||||||
|
apply(g_sub, weight_sub, convex_hull, ax2, leverage, "Leverage")
|
||||||
|
|
||||||
|
fig.savefig(f"model_diff_comparison_5000_pagerank_leverage.svg", format='svg')
|
||||||
73
distance_types.py
Normal file
73
distance_types.py
Normal file
@@ -0,0 +1,73 @@
|
|||||||
|
import math
|
||||||
|
|
||||||
|
import matplotlib.pyplot as plt
|
||||||
|
import numpy as np
|
||||||
|
from graph_tool.all import *
|
||||||
|
|
||||||
|
from src import centrality
|
||||||
|
from src import plot
|
||||||
|
from src import fitting
|
||||||
|
|
||||||
|
|
||||||
|
def random_graph(n=5000, seed=None):
|
||||||
|
"""
|
||||||
|
Uniformly random point cloud generation.
|
||||||
|
`n` [int] Number of points to generate. Default 5000 seems like a good starting point in point density and corresponding runtime for the subsequent calculations.
|
||||||
|
@return [numpy.ndarray] Array of shape(n, 2) containing the coordinates for each point of the generated point cloud.
|
||||||
|
"""
|
||||||
|
if seed is None:
|
||||||
|
import secrets
|
||||||
|
seed = secrets.randbits(128)
|
||||||
|
rng = np.random.default_rng(seed=seed)
|
||||||
|
return rng.random((n, 2)), seed
|
||||||
|
|
||||||
|
|
||||||
|
def spatial_graph(adata):
|
||||||
|
"""
|
||||||
|
Generate the spatial graph using delaunay for the given `adata`.
|
||||||
|
`adata` will contain the calculated spatial graph contents in the keys
|
||||||
|
adata.obsm['spatial']` in case the `adata` is created from a dataset of *squidpy*.
|
||||||
|
@return [Graph] generated networkx graph from adata.obsp['spatial_distances']
|
||||||
|
"""
|
||||||
|
g, pos = graph_tool.generation.triangulation(adata, type="delaunay")
|
||||||
|
g.vp["pos"] = pos
|
||||||
|
weight = g.new_edge_property("double")
|
||||||
|
for e in g.edges():
|
||||||
|
weight[e] = math.sqrt(sum(map(abs, pos[e.source()].a - pos[e.target()].a)))**2
|
||||||
|
return g, weight
|
||||||
|
|
||||||
|
|
||||||
|
def apply(g, seed, weight, convex_hull, ax, ax2, method):
|
||||||
|
# calculate centrality values
|
||||||
|
vp, ep = method(g, weight=weight)
|
||||||
|
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
||||||
|
|
||||||
|
# euklidian distance
|
||||||
|
quantification = plot.quantification_data(g, vp, convex_hull)
|
||||||
|
plot.quantification_plot(ax, quantification, None, None, "Euklidian Distance", None)
|
||||||
|
|
||||||
|
# generate model based on convex hull and associated centrality values
|
||||||
|
# path distance
|
||||||
|
quantification = plot.quantification_data_path_distance(g, weight, vp, convex_hull)
|
||||||
|
plot.quantification_plot(ax2, quantification, None, None, "Shortest Path Distance", None)
|
||||||
|
|
||||||
|
|
||||||
|
points, seed = random_graph(n=5000)
|
||||||
|
g, weight = spatial_graph(points)
|
||||||
|
g = GraphView(g)
|
||||||
|
# calculate convex hull
|
||||||
|
convex_hull = centrality.convex_hull(g)
|
||||||
|
|
||||||
|
fig = plt.figure(figsize=(21, 5))
|
||||||
|
ax1, ax2, ax3 = fig.subplots(1, 3)
|
||||||
|
|
||||||
|
# plot graph with convex_hull
|
||||||
|
# draw without any centrality measure `vp`
|
||||||
|
vp, ep = betweenness(g, weight=weight)
|
||||||
|
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
||||||
|
|
||||||
|
plot.graph_plot(fig, ax1, g, vp, convex_hull, f"Pointcloud (seed: {seed})")
|
||||||
|
|
||||||
|
apply(g, seed, weight, convex_hull, ax2, ax3, betweenness)
|
||||||
|
|
||||||
|
fig.savefig(f"Distance_5000_betweenness_euklidian.svg", format='svg')
|
||||||
91
effected_visualization.py
Normal file
91
effected_visualization.py
Normal file
@@ -0,0 +1,91 @@
|
|||||||
|
import math
|
||||||
|
|
||||||
|
import matplotlib.pyplot as plt
|
||||||
|
import numpy as np
|
||||||
|
from graph_tool.all import *
|
||||||
|
|
||||||
|
from src import centrality
|
||||||
|
from src import plot
|
||||||
|
from src import fitting
|
||||||
|
|
||||||
|
|
||||||
|
def random_graph(n=5000, seed=None):
|
||||||
|
"""
|
||||||
|
Uniformly random point cloud generation.
|
||||||
|
`n` [int] Number of points to generate. Default 5000 seems like a good starting point in point density and corresponding runtime for the subsequent calculations.
|
||||||
|
@return [numpy.ndarray] Array of shape(n, 2) containing the coordinates for each point of the generated point cloud.
|
||||||
|
"""
|
||||||
|
if seed is None:
|
||||||
|
import secrets
|
||||||
|
seed = secrets.randbits(128)
|
||||||
|
rng = np.random.default_rng(seed=seed)
|
||||||
|
return rng.random((n, 2)), seed
|
||||||
|
|
||||||
|
|
||||||
|
def spatial_graph(adata):
|
||||||
|
"""
|
||||||
|
Generate the spatial graph using delaunay for the given `adata`.
|
||||||
|
`adata` will contain the calculated spatial graph contents in the keys
|
||||||
|
adata.obsm['spatial']` in case the `adata` is created from a dataset of *squidpy*.
|
||||||
|
@return [Graph] generated networkx graph from adata.obsp['spatial_distances']
|
||||||
|
"""
|
||||||
|
g, pos = graph_tool.generation.triangulation(adata, type="delaunay")
|
||||||
|
g.vp["pos"] = pos
|
||||||
|
weight = g.new_edge_property("double")
|
||||||
|
for e in g.edges():
|
||||||
|
weight[e] = math.sqrt(sum(map(abs, pos[e.source()].a - pos[e.target()].a)))**2
|
||||||
|
return g, weight
|
||||||
|
|
||||||
|
|
||||||
|
points, seed = random_graph()
|
||||||
|
g, weight = spatial_graph(points)
|
||||||
|
g = GraphView(g)
|
||||||
|
|
||||||
|
# calculate centrality values
|
||||||
|
vp = closeness(g, weight=weight)
|
||||||
|
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
||||||
|
# ep.a = np.nan_to_num(ep.a) # correct floating point values
|
||||||
|
|
||||||
|
# calculate convex hull
|
||||||
|
convex_hull = centrality.convex_hull(g)
|
||||||
|
|
||||||
|
# plot graph with convex_hull
|
||||||
|
fig = plt.figure(figsize=(15, 12))
|
||||||
|
ax0 = fig.subplots(1, 1)
|
||||||
|
|
||||||
|
# generate model based on convex hull and associated centrality values
|
||||||
|
quantification = plot.quantification_data(g, vp, convex_hull)
|
||||||
|
|
||||||
|
# optimize model's piece-wise linear function
|
||||||
|
d = quantification[:, 0]
|
||||||
|
C = quantification[:, 1]
|
||||||
|
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||||
|
|
||||||
|
# TODO
|
||||||
|
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||||
|
d_curve = np.linspace(min(d), max(d), 500)
|
||||||
|
C_curve = np.piecewise(
|
||||||
|
d_curve,
|
||||||
|
[d_curve <= b_opt, d_curve > b_opt],
|
||||||
|
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||||
|
)
|
||||||
|
# plot graphs effected / uneffected nodes
|
||||||
|
plot.graph_plot_effected(fig, ax0, g, vp, convex_hull, b_opt, f"Random Graph (seed: {seed})")
|
||||||
|
|
||||||
|
# # linear regression model
|
||||||
|
# m_reg, c0_reg, b_reg, aic_reg = fitting.fit_cut(d, C)
|
||||||
|
# print(f"m_reg = {m_reg}")
|
||||||
|
|
||||||
|
# # TODO
|
||||||
|
# # should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||||
|
# d_curve = np.linspace(min(d), max(d), 500)
|
||||||
|
# C_curve = np.piecewise(
|
||||||
|
# d_curve,
|
||||||
|
# [d_curve <= b_reg, d_curve > b_reg],
|
||||||
|
# [lambda x: m_reg * x + c0_reg, lambda x: m_reg * b_reg + c0_reg]
|
||||||
|
# )
|
||||||
|
# ax1.plot(d_curve, C_curve, color='k', linewidth=1, label=f"Top Cut | AIC: {aic_reg}")
|
||||||
|
# ax1.legend()
|
||||||
|
|
||||||
|
fig.savefig(f"model_closeness_5000_effected_vs_uneffected.svg", format='svg')
|
||||||
|
|
||||||
242
example.py
242
example.py
@@ -2,7 +2,7 @@ import math
|
|||||||
|
|
||||||
import matplotlib.pyplot as plt
|
import matplotlib.pyplot as plt
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import squidpy as sq
|
# import squidpy as sq
|
||||||
from graph_tool.all import *
|
from graph_tool.all import *
|
||||||
|
|
||||||
from src import centrality
|
from src import centrality
|
||||||
@@ -29,6 +29,7 @@ def mibitof():
|
|||||||
adata = sq.datasets.mibitof()
|
adata = sq.datasets.mibitof()
|
||||||
return adata
|
return adata
|
||||||
|
|
||||||
|
|
||||||
def random_graph(n=5000, seed=None):
|
def random_graph(n=5000, seed=None):
|
||||||
"""
|
"""
|
||||||
Uniformly random point cloud generation.
|
Uniformly random point cloud generation.
|
||||||
@@ -96,207 +97,60 @@ def spatial_graph(adata):
|
|||||||
weight[e] = math.sqrt(sum(map(abs, pos[e.source()].a - pos[e.target()].a)))**2
|
weight[e] = math.sqrt(sum(map(abs, pos[e.source()].a - pos[e.target()].a)))**2
|
||||||
return g, weight
|
return g, weight
|
||||||
|
|
||||||
def merfish_example():
|
|
||||||
# generate spatial graph from a given dataset
|
|
||||||
g, weight = spatial_graph(merfish().obsm['spatial'])
|
|
||||||
g = GraphView(g)
|
|
||||||
|
|
||||||
x_spatial = []
|
points, seed = random_graph()
|
||||||
for v in g.vertices():
|
g, weight = spatial_graph(points)
|
||||||
x_spatial.append(g.vp["pos"][v][0])
|
g = GraphView(g)
|
||||||
|
|
||||||
# calculate centrality values
|
# calculate centrality values
|
||||||
vp = closeness(g, weight=weight)
|
vp = closeness(g, weight=weight)
|
||||||
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
||||||
|
# ep.a = np.nan_to_num(ep.a) # correct floating point values
|
||||||
|
|
||||||
# normalization
|
# normalization
|
||||||
min_val, max_val = vp.a.min(), vp.a.max()
|
# min_val, max_val = vp.a.min(), vp.a.max()
|
||||||
vp.a = (vp.a - min_val) / (max_val - min_val)
|
# vp.a = (vp.a - min_val) / (max_val - min_val)
|
||||||
|
|
||||||
# calculate convex hull
|
# calculate convex hull
|
||||||
convex_hull = centrality.convex_hull(g)
|
convex_hull = centrality.convex_hull(g)
|
||||||
|
|
||||||
# plot graph with convex_hull
|
# plot graph with convex_hull
|
||||||
fig = plt.figure(figsize=(15, 5))
|
fig = plt.figure(figsize=(15, 5))
|
||||||
ax0, ax1 = fig.subplots(1, 2)
|
ax0, ax1 = fig.subplots(1, 2)
|
||||||
plot.graph_plot(fig, ax0, g, vp, convex_hull, f"Merfish\nCloseness")
|
plot.graph_plot(fig, ax0, g, vp, convex_hull, f"Random Graph (seed: {seed})")
|
||||||
|
|
||||||
# generate model based on convex hull and associated centrality values
|
# generate model based on convex hull and associated centrality values
|
||||||
quantification = plot.quantification_data(g, vp, convex_hull)
|
quantification = plot.quantification_data(g, vp, convex_hull)
|
||||||
|
|
||||||
# optimize model's piece-wise linear function
|
# optimize model's piece-wise linear function
|
||||||
d = quantification[:, 0]
|
d = quantification[:, 0]
|
||||||
C = quantification[:, 1]
|
C = quantification[:, 1]
|
||||||
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||||
|
|
||||||
# TODO
|
# TODO
|
||||||
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||||
d_curve = np.linspace(min(d), max(d), 500)
|
d_curve = np.linspace(min(d), max(d), 500)
|
||||||
C_curve = np.piecewise(
|
C_curve = np.piecewise(
|
||||||
d_curve,
|
d_curve,
|
||||||
[d_curve <= b_opt, d_curve > b_opt],
|
[d_curve <= b_opt, d_curve > b_opt],
|
||||||
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||||
)
|
)
|
||||||
# plot model containing modeled piece-wise linear function
|
# plot model containing modeled piece-wise linear function
|
||||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_opt)
|
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Closeness', aic_opt)
|
||||||
|
|
||||||
# linear regression model
|
# # linear regression model
|
||||||
m_reg, c_reg, aic_reg = fitting.fit_linear_regression(d, C)
|
# m_reg, c0_reg, b_reg, aic_reg = fitting.fit_cut(d, C)
|
||||||
|
# print(f"m_reg = {m_reg}")
|
||||||
|
|
||||||
x = np.linspace(min(d), max(d), 500)
|
# # TODO
|
||||||
y = m_reg * x + c_reg
|
# # should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_reg}")
|
# d_curve = np.linspace(min(d), max(d), 500)
|
||||||
ax1.legend()
|
# C_curve = np.piecewise(
|
||||||
|
# d_curve,
|
||||||
|
# [d_curve <= b_reg, d_curve > b_reg],
|
||||||
|
# [lambda x: m_reg * x + c0_reg, lambda x: m_reg * b_reg + c0_reg]
|
||||||
|
# )
|
||||||
|
# ax1.plot(d_curve, C_curve, color='k', linewidth=1, label=f"Top Cut | AIC: {aic_reg}")
|
||||||
|
# ax1.legend()
|
||||||
|
|
||||||
fig.savefig(f"Merfish_closeness.svg", format='svg')
|
fig.savefig(f"model_closeness_5000_fitted.svg", format='svg')
|
||||||
|
|
||||||
for i in range(1, 6):
|
|
||||||
points, seed = random_graph()
|
|
||||||
g, weight = spatial_graph(points)
|
|
||||||
g = GraphView(g)
|
|
||||||
|
|
||||||
x_spatial = []
|
|
||||||
for v in g.vertices():
|
|
||||||
x_spatial.append(g.vp["pos"][v][0])
|
|
||||||
|
|
||||||
# calculate centrality values
|
|
||||||
vp = closeness(g, weight=weight)
|
|
||||||
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
|
||||||
# ep.a = np.nan_to_num(ep.a) # correct floating point values
|
|
||||||
|
|
||||||
# normalization
|
|
||||||
min_val, max_val = vp.a.min(), vp.a.max()
|
|
||||||
vp.a = (vp.a - min_val) / (max_val - min_val)
|
|
||||||
|
|
||||||
# calculate convex hull
|
|
||||||
convex_hull = centrality.convex_hull(g)
|
|
||||||
|
|
||||||
# plot graph with convex_hull
|
|
||||||
fig = plt.figure(figsize=(15, 5))
|
|
||||||
ax0, ax1 = fig.subplots(1, 2)
|
|
||||||
plot.graph_plot(fig, ax0, g, vp, convex_hull, f"Random Graph (seed: {seed})\nCloseness")
|
|
||||||
|
|
||||||
# generate model based on convex hull and associated centrality values
|
|
||||||
quantification = plot.quantification_data(g, vp, convex_hull)
|
|
||||||
|
|
||||||
# optimize model's piece-wise linear function
|
|
||||||
d = quantification[:, 0]
|
|
||||||
C = quantification[:, 1]
|
|
||||||
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
|
||||||
|
|
||||||
# TODO
|
|
||||||
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
|
||||||
d_curve = np.linspace(min(d), max(d), 500)
|
|
||||||
C_curve = np.piecewise(
|
|
||||||
d_curve,
|
|
||||||
[d_curve <= b_opt, d_curve > b_opt],
|
|
||||||
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
|
||||||
)
|
|
||||||
# plot model containing modeled piece-wise linear function
|
|
||||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_opt)
|
|
||||||
|
|
||||||
# linear regression model
|
|
||||||
m_reg, c_reg, aic_reg = fitting.fit_linear_regression(d, C)
|
|
||||||
|
|
||||||
x = np.linspace(min(d), max(d), 500)
|
|
||||||
y = m_reg * x + c_reg
|
|
||||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_reg}")
|
|
||||||
ax1.legend()
|
|
||||||
|
|
||||||
fig.savefig(f"uniform_random_point_clouds/{i}_closeness.svg", format='svg')
|
|
||||||
|
|
||||||
# ---------------------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
# calculate centrality values
|
|
||||||
vp, ep = betweenness(g, weight=weight)
|
|
||||||
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
|
||||||
# ep.a = np.nan_to_num(ep.a) # correct floating point values
|
|
||||||
|
|
||||||
# normalization
|
|
||||||
min_val, max_val = vp.a.min(), vp.a.max()
|
|
||||||
vp.a = (vp.a - min_val) / (max_val - min_val)
|
|
||||||
|
|
||||||
# calculate convex hull
|
|
||||||
convex_hull = centrality.convex_hull(g)
|
|
||||||
|
|
||||||
# plot graph with convex_hull
|
|
||||||
fig = plt.figure(figsize=(15, 5))
|
|
||||||
ax0, ax1 = fig.subplots(1, 2)
|
|
||||||
plot.graph_plot(fig, ax0, g, vp, convex_hull, f"Random Graph (seed: {seed})\nBetweenness")
|
|
||||||
|
|
||||||
# generate model based on convex hull and associated centrality values
|
|
||||||
quantification = plot.quantification_data(g, vp, convex_hull)
|
|
||||||
|
|
||||||
# optimize model's piece-wise linear function
|
|
||||||
d = quantification[:, 0]
|
|
||||||
C = quantification[:, 1]
|
|
||||||
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
|
||||||
|
|
||||||
# TODO
|
|
||||||
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
|
||||||
d_curve = np.linspace(min(d), max(d), 500)
|
|
||||||
C_curve = np.piecewise(
|
|
||||||
d_curve,
|
|
||||||
[d_curve <= b_opt, d_curve > b_opt],
|
|
||||||
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
|
||||||
)
|
|
||||||
# plot model containing modeled piece-wise linear function
|
|
||||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_opt)
|
|
||||||
|
|
||||||
# linear regression model
|
|
||||||
m_reg, c_reg, aic_reg = fitting.fit_linear_regression(d, C)
|
|
||||||
|
|
||||||
x = np.linspace(min(d), max(d), 500)
|
|
||||||
y = m_reg * x + c_reg
|
|
||||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_reg}")
|
|
||||||
ax1.legend()
|
|
||||||
|
|
||||||
fig.savefig(f"uniform_random_point_clouds/{i}_betweenness.svg", format='svg')
|
|
||||||
|
|
||||||
# ---------------------------------------------------------------------------------------------
|
|
||||||
|
|
||||||
# calculate centrality values
|
|
||||||
vp = pagerank(g, weight=weight)
|
|
||||||
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
|
||||||
# ep.a = np.nan_to_num(ep.a) # correct floating point values
|
|
||||||
|
|
||||||
# normalization
|
|
||||||
min_val, max_val = vp.a.min(), vp.a.max()
|
|
||||||
vp.a = (vp.a - min_val) / (max_val - min_val)
|
|
||||||
|
|
||||||
# calculate convex hull
|
|
||||||
convex_hull = centrality.convex_hull(g)
|
|
||||||
|
|
||||||
# plot graph with convex_hull
|
|
||||||
fig = plt.figure(figsize=(15, 5))
|
|
||||||
ax0, ax1 = fig.subplots(1, 2)
|
|
||||||
plot.graph_plot(fig, ax0, g, vp, convex_hull, f"Random Graph (seed: {seed})\nPageRank")
|
|
||||||
|
|
||||||
# generate model based on convex hull and associated centrality values
|
|
||||||
quantification = plot.quantification_data(g, vp, convex_hull)
|
|
||||||
|
|
||||||
# optimize model's piece-wise linear function
|
|
||||||
d = quantification[:, 0]
|
|
||||||
C = quantification[:, 1]
|
|
||||||
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
|
||||||
|
|
||||||
# TODO
|
|
||||||
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
|
||||||
d_curve = np.linspace(min(d), max(d), 500)
|
|
||||||
C_curve = np.piecewise(
|
|
||||||
d_curve,
|
|
||||||
[d_curve <= b_opt, d_curve > b_opt],
|
|
||||||
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
|
||||||
)
|
|
||||||
# plot model containing modeled piece-wise linear function
|
|
||||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_opt)
|
|
||||||
|
|
||||||
# linear regression model
|
|
||||||
m_reg, c_reg, aic_reg = fitting.fit_linear_regression(d, C)
|
|
||||||
|
|
||||||
x = np.linspace(min(d), max(d), 500)
|
|
||||||
y = m_reg * x + c_reg
|
|
||||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_reg}")
|
|
||||||
ax1.legend()
|
|
||||||
|
|
||||||
fig.savefig(f"uniform_random_point_clouds/{i}_pagerank.svg", format='svg')
|
|
||||||
|
|||||||
105
model_approximation.py
Normal file
105
model_approximation.py
Normal file
@@ -0,0 +1,105 @@
|
|||||||
|
import math
|
||||||
|
|
||||||
|
import matplotlib.pyplot as plt
|
||||||
|
import numpy as np
|
||||||
|
from graph_tool.all import *
|
||||||
|
|
||||||
|
from src import centrality
|
||||||
|
from src import plot
|
||||||
|
from src import fitting
|
||||||
|
|
||||||
|
|
||||||
|
def leverage(g, weight):
|
||||||
|
# VertexPropertyMap
|
||||||
|
vp = g.new_vertex_property("double")
|
||||||
|
for v in g.vertices():
|
||||||
|
li = 0.0
|
||||||
|
neighbours = g.get_all_neighbours(v)
|
||||||
|
ki = len(neighbours)
|
||||||
|
# sum
|
||||||
|
for nv in neighbours:
|
||||||
|
other_neighbours = g.get_all_neighbours(nv)
|
||||||
|
kj = len(other_neighbours)
|
||||||
|
li += (ki - kj) / (ki + kj)
|
||||||
|
li /= ki
|
||||||
|
vp[v] = li
|
||||||
|
return vp
|
||||||
|
|
||||||
|
|
||||||
|
def random_graph(n=5000, seed=None):
|
||||||
|
"""
|
||||||
|
Uniformly random point cloud generation.
|
||||||
|
`n` [int] Number of points to generate. Default 5000 seems like a good starting point in point density and corresponding runtime for the subsequent calculations.
|
||||||
|
@return [numpy.ndarray] Array of shape(n, 2) containing the coordinates for each point of the generated point cloud.
|
||||||
|
"""
|
||||||
|
if seed is None:
|
||||||
|
import secrets
|
||||||
|
seed = secrets.randbits(128)
|
||||||
|
rng = np.random.default_rng(seed=seed)
|
||||||
|
return rng.random((n, 2)), seed
|
||||||
|
|
||||||
|
|
||||||
|
def spatial_graph(adata):
|
||||||
|
g, pos = graph_tool.generation.triangulation(adata, type="delaunay")
|
||||||
|
g.vp["pos"] = pos
|
||||||
|
weight = g.new_edge_property("double")
|
||||||
|
for e in g.edges():
|
||||||
|
weight[e] = math.sqrt(sum(map(abs, pos[e.source()].a - pos[e.target()].a)))**2
|
||||||
|
return g, weight
|
||||||
|
|
||||||
|
|
||||||
|
points, seed = random_graph()
|
||||||
|
g, weight = spatial_graph(points)
|
||||||
|
g = GraphView(g)
|
||||||
|
|
||||||
|
# calculate centrality values
|
||||||
|
vp, ep = betweenness(g, weight=weight)
|
||||||
|
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
||||||
|
ep.a = np.nan_to_num(ep.a) # correct floating point values
|
||||||
|
|
||||||
|
# normalization
|
||||||
|
min_val, max_val = vp.a.min(), vp.a.max()
|
||||||
|
vp.a = (vp.a - min_val) / (max_val - min_val)
|
||||||
|
|
||||||
|
# calculate convex hull
|
||||||
|
convex_hull = centrality.convex_hull(g)
|
||||||
|
|
||||||
|
# plot graph with convex_hull
|
||||||
|
fig = plt.figure(figsize=(15, 5))
|
||||||
|
ax0, ax1 = fig.subplots(1, 2)
|
||||||
|
plot.graph_plot(fig, ax0, g, vp, convex_hull, f"Random Graph (seed: {seed})")
|
||||||
|
|
||||||
|
# generate model based on convex hull and associated centrality values
|
||||||
|
quantification = plot.quantification_data(g, vp, convex_hull)
|
||||||
|
|
||||||
|
# optimize model's piece-wise linear function
|
||||||
|
d = quantification[:, 0]
|
||||||
|
C = quantification[:, 1]
|
||||||
|
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||||
|
|
||||||
|
# TODO
|
||||||
|
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||||
|
d_curve = np.linspace(min(d), max(d), 500)
|
||||||
|
C_curve = np.piecewise(
|
||||||
|
d_curve,
|
||||||
|
[d_curve <= b_opt, d_curve > b_opt],
|
||||||
|
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||||
|
)
|
||||||
|
# plot model containing modeled piece-wise linear function
|
||||||
|
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Betweenness', aic_opt)
|
||||||
|
|
||||||
|
# linear regression model
|
||||||
|
m_reg, c0_reg, aic_reg = fitting.fit_linear_regression(d, C)
|
||||||
|
|
||||||
|
# TODO
|
||||||
|
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||||
|
d_curve = np.linspace(min(d), max(d), 500)
|
||||||
|
C_curve = np.piecewise(
|
||||||
|
d_curve,
|
||||||
|
[d_curve >= 0],
|
||||||
|
[lambda x: m_reg * x + c0_reg]
|
||||||
|
)
|
||||||
|
ax1.plot(d_curve, C_curve, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_reg}")
|
||||||
|
ax1.legend()
|
||||||
|
|
||||||
|
fig.savefig(f"model_approximation_betweenness_5000.svg", format='svg')
|
||||||
61
model_based_correction.py
Normal file
61
model_based_correction.py
Normal file
@@ -0,0 +1,61 @@
|
|||||||
|
import math
|
||||||
|
|
||||||
|
import matplotlib.pyplot as plt
|
||||||
|
import numpy as np
|
||||||
|
from graph_tool.all import *
|
||||||
|
|
||||||
|
from src import centrality
|
||||||
|
from src import plot
|
||||||
|
from src import fitting
|
||||||
|
|
||||||
|
def random_graph(n=5000, seed=None):
|
||||||
|
"""
|
||||||
|
Uniformly random point cloud generation.
|
||||||
|
`n` [int] Number of points to generate. Default 5000 seems like a good starting point in point density and corresponding runtime for the subsequent calculations.
|
||||||
|
@return [numpy.ndarray] Array of shape(n, 2) containing the coordinates for each point of the generated point cloud.
|
||||||
|
"""
|
||||||
|
if seed is None:
|
||||||
|
import secrets
|
||||||
|
seed = secrets.randbits(128)
|
||||||
|
rng = np.random.default_rng(seed=seed)
|
||||||
|
return rng.random((n, 2)), seed
|
||||||
|
|
||||||
|
|
||||||
|
def spatial_graph(adata):
|
||||||
|
g, pos = graph_tool.generation.triangulation(adata, type="delaunay")
|
||||||
|
g.vp["pos"] = pos
|
||||||
|
weight = g.new_edge_property("double")
|
||||||
|
for e in g.edges():
|
||||||
|
weight[e] = math.sqrt(sum(map(abs, pos[e.source()].a - pos[e.target()].a)))**2
|
||||||
|
return g, weight
|
||||||
|
|
||||||
|
|
||||||
|
points, seed = random_graph()
|
||||||
|
g, weight = spatial_graph(points)
|
||||||
|
g = GraphView(g)
|
||||||
|
|
||||||
|
# calculate centrality values
|
||||||
|
vp = closeness(g, weight=weight)
|
||||||
|
vp.a = np.nan_to_num(vp.a) # correct floating point values
|
||||||
|
# ep.a = np.nan_to_num(ep.a) # correct floating point values
|
||||||
|
|
||||||
|
# calculate convex hull
|
||||||
|
convex_hull = centrality.convex_hull(g)
|
||||||
|
|
||||||
|
# plot graph with convex_hull
|
||||||
|
fig = plt.figure(figsize=(15, 5))
|
||||||
|
ax0, ax1 = fig.subplots(1, 2)
|
||||||
|
plot.graph_plot(fig, ax0, g, vp, convex_hull, f"Closeness without prediction")
|
||||||
|
|
||||||
|
# generate model based on convex hull and associated centrality values
|
||||||
|
quantification = plot.quantification_data(g, vp, convex_hull)
|
||||||
|
|
||||||
|
# optimize model's piece-wise linear function
|
||||||
|
d = quantification[:, 0]
|
||||||
|
C = quantification[:, 1]
|
||||||
|
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||||
|
|
||||||
|
vp = centrality.correct(g, vp, m_opt, c0_opt, b_opt)
|
||||||
|
plot.graph_plot(fig, ax1, g, vp, convex_hull, f"Closeness with model prediction")
|
||||||
|
|
||||||
|
fig.savefig(f"model_prediction_comparison.svg", format='svg')
|
||||||
76
point_cloud_example.py
Normal file
76
point_cloud_example.py
Normal file
@@ -0,0 +1,76 @@
|
|||||||
|
import math
|
||||||
|
|
||||||
|
import matplotlib.pyplot as plt
|
||||||
|
from matplotlib.collections import LineCollection
|
||||||
|
import numpy as np
|
||||||
|
from graph_tool.all import *
|
||||||
|
|
||||||
|
|
||||||
|
def random_graph(n=5000, seed=None):
|
||||||
|
"""
|
||||||
|
Uniformly random point cloud generation.
|
||||||
|
`n` [int] Number of points to generate. Default 5000 seems like a good starting point in point density and corresponding runtime for the subsequent calculations.
|
||||||
|
@return [numpy.ndarray] Array of shape(n, 2) containing the coordinates for each point of the generated point cloud.
|
||||||
|
"""
|
||||||
|
if seed is None:
|
||||||
|
import secrets
|
||||||
|
seed = secrets.randbits(128)
|
||||||
|
rng = np.random.default_rng(seed=seed)
|
||||||
|
return rng.random((n, 2)), seed
|
||||||
|
|
||||||
|
|
||||||
|
def spatial_graph(adata):
|
||||||
|
"""
|
||||||
|
Generate the spatial graph using delaunay for the given `adata`.
|
||||||
|
`adata` will contain the calculated spatial graph contents in the keys
|
||||||
|
adata.obsm['spatial']` in case the `adata` is created from a dataset of *squidpy*.
|
||||||
|
@return [Graph] generated networkx graph from adata.obsp['spatial_distances']
|
||||||
|
"""
|
||||||
|
g, pos = graph_tool.generation.triangulation(adata, type="delaunay")
|
||||||
|
g.vp["pos"] = pos
|
||||||
|
weight = g.new_edge_property("double")
|
||||||
|
for e in g.edges():
|
||||||
|
weight[e] = math.sqrt(sum(map(abs, pos[e.source()].a - pos[e.target()].a)))**2
|
||||||
|
return g, weight
|
||||||
|
|
||||||
|
|
||||||
|
def draw_graph(G, ax, name):
|
||||||
|
pos = G.vp["pos"]
|
||||||
|
x = []
|
||||||
|
y = []
|
||||||
|
for v in G.vertices():
|
||||||
|
# print(pos[v])
|
||||||
|
ver = pos[v]
|
||||||
|
x.append(ver[0])
|
||||||
|
y.append(ver[1])
|
||||||
|
|
||||||
|
# convex hull -> Bounding-Box
|
||||||
|
# ch = LineCollection([convex_hull], colors=['g'], linewidths=1)
|
||||||
|
# ax.add_collection(ch)
|
||||||
|
|
||||||
|
# edges
|
||||||
|
for e in G.edges():
|
||||||
|
ex = [pos[e.source()][0], pos[e.target()][0]]
|
||||||
|
ey = [pos[e.source()][1], pos[e.target()][1]]
|
||||||
|
ax.add_collection(LineCollection([np.column_stack([ex, ey])], colors=['k'], linewidths=0.1))
|
||||||
|
|
||||||
|
ax.scatter(x, y, s=1) # map closeness values as color mapping on the verticies
|
||||||
|
ax.set_title(name)
|
||||||
|
|
||||||
|
|
||||||
|
#
|
||||||
|
# - Create a random point cloud and calculate a triangulation on it
|
||||||
|
# - For that graph calculate the convex hull
|
||||||
|
# - Draw the graph with the convex hull
|
||||||
|
# - For each centrality measure
|
||||||
|
# - apply centrality measure to the next axis
|
||||||
|
# - Draw the corresponding resulting models into a grid
|
||||||
|
#
|
||||||
|
points, seed = random_graph(n=3000)
|
||||||
|
g, weight = spatial_graph(points)
|
||||||
|
g = GraphView(g)
|
||||||
|
|
||||||
|
# plot graph with convex_hull
|
||||||
|
fig_graph, ax_graph = plt.subplots(figsize=(15, 12))
|
||||||
|
draw_graph(g, ax_graph, f"Pointcould (seed: {seed} | n: 500)")
|
||||||
|
fig_graph.savefig("point_cloud_example.svg", format='svg')
|
||||||
@@ -46,7 +46,7 @@ def correct(G, centrality, m_opt, c0_opt, b_opt):
|
|||||||
@param b_opt [Float] Model b value (intersection point)
|
@param b_opt [Float] Model b value (intersection point)
|
||||||
@return [VertexPropertyMap] Corrected centrality values based on @param centrality
|
@return [VertexPropertyMap] Corrected centrality values based on @param centrality
|
||||||
"""
|
"""
|
||||||
corrected_metric = {}
|
corrected_metric = centrality
|
||||||
pos = G.vp["pos"]
|
pos = G.vp["pos"]
|
||||||
x = []
|
x = []
|
||||||
y = []
|
y = []
|
||||||
@@ -56,8 +56,8 @@ def correct(G, centrality, m_opt, c0_opt, b_opt):
|
|||||||
x.append(ver[0])
|
x.append(ver[0])
|
||||||
y.append(ver[1])
|
y.append(ver[1])
|
||||||
|
|
||||||
keys = iter(centrality.a)
|
keys = iter(G.vertices())
|
||||||
hull = convex_hull(x, y)
|
hull = convex_hull(G)
|
||||||
|
|
||||||
points = np.stack((np.array(x), np.array(y)), axis=-1)
|
points = np.stack((np.array(x), np.array(y)), axis=-1)
|
||||||
for point in pos:
|
for point in pos:
|
||||||
|
|||||||
@@ -37,6 +37,7 @@ def fit_piece_wise_linear(d, C, M=1000):
|
|||||||
# Setting solver parameters for precision
|
# Setting solver parameters for precision
|
||||||
model.setParam('OptimalityTol', 1e-4)
|
model.setParam('OptimalityTol', 1e-4)
|
||||||
model.setParam('MIPGap', 0.01)
|
model.setParam('MIPGap', 0.01)
|
||||||
|
model.setParam('OutputFlag', 0)
|
||||||
|
|
||||||
for i in range(n):
|
for i in range(n):
|
||||||
# Constraints enforcing piecewise linear fit
|
# Constraints enforcing piecewise linear fit
|
||||||
|
|||||||
76
src/plot.py
76
src/plot.py
@@ -6,6 +6,7 @@ import matplotlib.colors as mcolors
|
|||||||
|
|
||||||
from matplotlib.collections import LineCollection
|
from matplotlib.collections import LineCollection
|
||||||
from src import centrality
|
from src import centrality
|
||||||
|
from graph_tool.all import *
|
||||||
|
|
||||||
class Vector:
|
class Vector:
|
||||||
"""
|
"""
|
||||||
@@ -71,6 +72,49 @@ def graph_plot(fig, ax, G, measures, convex_hull, name, show_edges=False):
|
|||||||
fig.colorbar(sc, ax=ax)
|
fig.colorbar(sc, ax=ax)
|
||||||
|
|
||||||
|
|
||||||
|
def graph_plot_effected(fig, ax, G, measures, convex_hull, b, name, show_edges=False):
|
||||||
|
"""
|
||||||
|
Plot relationship data of effected vs uneffected nodes determined through model.
|
||||||
|
"""
|
||||||
|
quantification = []
|
||||||
|
pos = G.vp["pos"]
|
||||||
|
x = []
|
||||||
|
y = []
|
||||||
|
for v in G.vertices():
|
||||||
|
# print(pos[v])
|
||||||
|
ver = pos[v]
|
||||||
|
x.append(ver[0])
|
||||||
|
y.append(ver[1])
|
||||||
|
|
||||||
|
measures = measures.a
|
||||||
|
keys = iter(measures)
|
||||||
|
|
||||||
|
points = np.stack((np.array(x), np.array(y)), axis=-1)
|
||||||
|
for point in points:
|
||||||
|
min_distance = math.inf
|
||||||
|
key = next(keys)
|
||||||
|
for edge in convex_hull:
|
||||||
|
vector = Vector.vec(point, edge)
|
||||||
|
distance = Vector.vec_len(vector)
|
||||||
|
if distance < min_distance:
|
||||||
|
min_distance = distance
|
||||||
|
quantification.append([min_distance, key])
|
||||||
|
|
||||||
|
# ax.scatter(quantification[:, 0], quantification[:, 1], c=quantification[:, 1], cmap=plt.cm.plasma, s=0.2)
|
||||||
|
c = list(map(lambda q: 'b' if q[0] > b else 'r', quantification))
|
||||||
|
# convex hull -> Bounding-Box
|
||||||
|
# ch = LineCollection([convex_hull], colors=['g'], linewidths=1)
|
||||||
|
# ax.add_collection(ch)
|
||||||
|
if show_edges:
|
||||||
|
for e in G.edges():
|
||||||
|
ex = [pos[e.source()][0], pos[e.target()][0]]
|
||||||
|
ey = [pos[e.source()][1], pos[e.target()][1]]
|
||||||
|
ax.add_collection(LineCollection([np.column_stack([ex, ey])], colors=['k'], linewidths=0.1))
|
||||||
|
|
||||||
|
sc = ax.scatter(x, y, s=1, c=c) # map closeness values as color mapping on the verticies
|
||||||
|
ax.set_title(name)
|
||||||
|
|
||||||
|
|
||||||
def normalize_dict(d):
|
def normalize_dict(d):
|
||||||
max = np.max(list(d.values()))
|
max = np.max(list(d.values()))
|
||||||
return {k: (v / max) for k, v in d.items()}
|
return {k: (v / max) for k, v in d.items()}
|
||||||
@@ -112,6 +156,38 @@ def quantification_data(G, measures, convex_hull):
|
|||||||
return np.array(quantification)
|
return np.array(quantification)
|
||||||
|
|
||||||
|
|
||||||
|
def quantification_data_path_distance(G, weights, measures, convex_hull):
|
||||||
|
quantification = []
|
||||||
|
pos = G.vp["pos"]
|
||||||
|
x = []
|
||||||
|
y = []
|
||||||
|
convex_hull_verticies = []
|
||||||
|
for v in G.vertices():
|
||||||
|
ver = pos[v]
|
||||||
|
for n in convex_hull:
|
||||||
|
if np.equal(n, np.array([ver[0], ver[1]])).all():
|
||||||
|
convex_hull_verticies.append(v)
|
||||||
|
|
||||||
|
measures = measures.a
|
||||||
|
keys = iter(measures)
|
||||||
|
|
||||||
|
points = np.stack((np.array(x), np.array(y)), axis=-1)
|
||||||
|
for v in G.vertices():
|
||||||
|
min_distance = math.inf
|
||||||
|
key = next(keys)
|
||||||
|
for h in convex_hull_verticies:
|
||||||
|
vertices, edges = graph_tool.topology.shortest_path(G, v, h, weights=weights)
|
||||||
|
# TODO calculate the total distance
|
||||||
|
path_length = sum([weights[edge] for edge in edges])
|
||||||
|
if path_length < min_distance:
|
||||||
|
min_distance = path_length
|
||||||
|
quantification.append([min_distance, key])
|
||||||
|
|
||||||
|
# sort by distance
|
||||||
|
quantification.sort(key=lambda entry: entry[0])
|
||||||
|
return np.array(quantification)
|
||||||
|
|
||||||
|
|
||||||
def quantification_plot(ax, quantification, d_curve, C_curve, metric_name, aic_score):
|
def quantification_plot(ax, quantification, d_curve, C_curve, metric_name, aic_score):
|
||||||
"""
|
"""
|
||||||
Plot relationship data.
|
Plot relationship data.
|
||||||
|
|||||||
Reference in New Issue
Block a user