add: calculate AIC when solving the model fitting problem
This commit is contained in:
133
example.py
133
example.py
@@ -42,6 +42,46 @@ def random_graph(n=5000, seed=None):
|
||||
return rng.random((n, 2)), seed
|
||||
|
||||
|
||||
def random_graph_favor_border(n=3000, seed = None):
|
||||
if seed is None:
|
||||
import secrets
|
||||
seed = secrets.randbits(128)
|
||||
rng = np.random.default_rng(seed=seed)
|
||||
vps = np.zeros((n, 2))
|
||||
for i in range(0, n):
|
||||
r_x = rng.random()
|
||||
if rng.random() > 0.5:
|
||||
while (r_x > 0.3 and r_x < 0.7):
|
||||
r_x = rng.random()
|
||||
r_y = rng.random()
|
||||
if rng.random() > 0.5:
|
||||
while (r_y > 0.3 and r_y < 0.7):
|
||||
r_y = rng.random()
|
||||
vps[i][0] = r_x
|
||||
vps[i][1] = r_y
|
||||
return vps, seed
|
||||
|
||||
|
||||
def random_graph_favor_center(n=3000, seed = None):
|
||||
if seed is None:
|
||||
import secrets
|
||||
seed = secrets.randbits(128)
|
||||
rng = np.random.default_rng(seed=seed)
|
||||
vps = np.zeros((n, 2))
|
||||
for i in range(0, n):
|
||||
r_x = rng.random()
|
||||
if rng.random() > 0.7:
|
||||
while (r_x < 0.4 or r_x > 0.6):
|
||||
r_x = rng.random()
|
||||
r_y = rng.random()
|
||||
if rng.random() > 0.7:
|
||||
while (r_y < 0.4 or r_y > 0.6):
|
||||
r_y = rng.random()
|
||||
vps[i][0] = r_x
|
||||
vps[i][1] = r_y
|
||||
return vps, seed
|
||||
|
||||
|
||||
def spatial_graph(adata):
|
||||
"""
|
||||
Generate the spatial graph using delaunay for the given `adata`.
|
||||
@@ -87,16 +127,7 @@ def merfish_example():
|
||||
# optimize model's piece-wise linear function
|
||||
d = quantification[:, 0]
|
||||
C = quantification[:, 1]
|
||||
m_opt, c0_opt, b_opt = fitting.fit_piece_wise_linear(d, C)
|
||||
|
||||
# AIC
|
||||
# AIC = 2 * k (= 2) - 2 * ln(L^~)
|
||||
# with L^~ = sum(f(x_i)) where x_i describes a data point
|
||||
# - f is *not normalized*
|
||||
sum_log = 0.0
|
||||
for x_i in x_spatial:
|
||||
sum_log += math.log(m_opt* b_opt + c0_opt if x_i >= b_opt else m_opt * x_i + c0_opt)
|
||||
aic_model = 6. - 2. * sum_log # three parameters: b_opt, m_opt, c0_opt
|
||||
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||
|
||||
# TODO
|
||||
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||
@@ -107,25 +138,18 @@ def merfish_example():
|
||||
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||
)
|
||||
# plot model containing modeled piece-wise linear function
|
||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_model)
|
||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_opt)
|
||||
|
||||
# linear regression model
|
||||
m_reg, c_reg = fitting.fit_linear_regression(d, C)
|
||||
|
||||
# AIC
|
||||
sum_log = 0.0
|
||||
for x_i in x_spatial:
|
||||
sum_log += math.log(m_reg * x_i + c_reg)
|
||||
aic_regression = 4. - 2. * sum_log # two parameter: m_reg, c_reg
|
||||
m_reg, c_reg, aic_reg = fitting.fit_linear_regression(d, C)
|
||||
|
||||
x = np.linspace(min(d), max(d), 500)
|
||||
y = m_reg * x + c_reg
|
||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_regression}")
|
||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_reg}")
|
||||
ax1.legend()
|
||||
|
||||
fig.savefig(f"Merfish_closeness.svg", format='svg')
|
||||
|
||||
|
||||
for i in range(1, 6):
|
||||
points, seed = random_graph()
|
||||
g, weight = spatial_graph(points)
|
||||
@@ -158,16 +182,7 @@ for i in range(1, 6):
|
||||
# optimize model's piece-wise linear function
|
||||
d = quantification[:, 0]
|
||||
C = quantification[:, 1]
|
||||
m_opt, c0_opt, b_opt = fitting.fit_piece_wise_linear(d, C)
|
||||
|
||||
# AIC
|
||||
# AIC = 2 * k (= 2) - 2 * ln(L^~)
|
||||
# with L^~ = sum(f(x_i)) where x_i describes a data point
|
||||
# - f is *not normalized*
|
||||
sum_log = 0.0
|
||||
for x_i in x_spatial:
|
||||
sum_log += math.log(m_opt* b_opt + c0_opt if x_i >= b_opt else m_opt * x_i + c0_opt)
|
||||
aic_model = 6. - 2. * sum_log # three parameters: b_opt, m_opt, c0_opt
|
||||
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||
|
||||
# TODO
|
||||
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||
@@ -178,20 +193,14 @@ for i in range(1, 6):
|
||||
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||
)
|
||||
# plot model containing modeled piece-wise linear function
|
||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_model)
|
||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_opt)
|
||||
|
||||
# linear regression model
|
||||
m_reg, c_reg = fitting.fit_linear_regression(d, C)
|
||||
|
||||
# AIC
|
||||
sum_log = 0.0
|
||||
for x_i in x_spatial:
|
||||
sum_log += math.log(m_reg * x_i + c_reg)
|
||||
aic_regression = 4. - 2. * sum_log # two parameter: m_reg, c_reg
|
||||
m_reg, c_reg, aic_reg = fitting.fit_linear_regression(d, C)
|
||||
|
||||
x = np.linspace(min(d), max(d), 500)
|
||||
y = m_reg * x + c_reg
|
||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_regression}")
|
||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_reg}")
|
||||
ax1.legend()
|
||||
|
||||
fig.savefig(f"uniform_random_point_clouds/{i}_closeness.svg", format='svg')
|
||||
@@ -221,16 +230,7 @@ for i in range(1, 6):
|
||||
# optimize model's piece-wise linear function
|
||||
d = quantification[:, 0]
|
||||
C = quantification[:, 1]
|
||||
m_opt, c0_opt, b_opt = fitting.fit_piece_wise_linear(d, C)
|
||||
|
||||
# AIC
|
||||
# AIC = 2 * k (= 2) - 2 * ln(L^~)
|
||||
# with L^~ = sum(f(x_i)) where x_i describes a data point
|
||||
# - f is *not normalized*
|
||||
sum_log = 0.0
|
||||
for x_i in x_spatial:
|
||||
sum_log += math.log(m_opt* b_opt + c0_opt if x_i >= b_opt else m_opt * x_i + c0_opt)
|
||||
aic_model = 6. - 2. * sum_log # three parameters: b_opt, m_opt, c0_opt
|
||||
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||
|
||||
# TODO
|
||||
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||
@@ -241,20 +241,14 @@ for i in range(1, 6):
|
||||
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||
)
|
||||
# plot model containing modeled piece-wise linear function
|
||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_model)
|
||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_opt)
|
||||
|
||||
# linear regression model
|
||||
m_reg, c_reg = fitting.fit_linear_regression(d, C)
|
||||
|
||||
# AIC
|
||||
sum_log = 0.0
|
||||
for x_i in x_spatial:
|
||||
sum_log += math.log(m_reg * x_i + c_reg)
|
||||
aic_regression = 4. - 2. * sum_log # two parameter: m_reg, c_reg
|
||||
m_reg, c_reg, aic_reg = fitting.fit_linear_regression(d, C)
|
||||
|
||||
x = np.linspace(min(d), max(d), 500)
|
||||
y = m_reg * x + c_reg
|
||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_regression}")
|
||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_reg}")
|
||||
ax1.legend()
|
||||
|
||||
fig.savefig(f"uniform_random_point_clouds/{i}_betweenness.svg", format='svg')
|
||||
@@ -284,16 +278,7 @@ for i in range(1, 6):
|
||||
# optimize model's piece-wise linear function
|
||||
d = quantification[:, 0]
|
||||
C = quantification[:, 1]
|
||||
m_opt, c0_opt, b_opt = fitting.fit_piece_wise_linear(d, C)
|
||||
|
||||
# AIC
|
||||
# AIC = 2 * k (= 2) - 2 * ln(L^~)
|
||||
# with L^~ = sum(f(x_i)) where x_i describes a data point
|
||||
# - f is *not normalized*
|
||||
sum_log = 0.0
|
||||
for x_i in x_spatial:
|
||||
sum_log += math.log(m_opt* b_opt + c0_opt if x_i >= b_opt else m_opt * x_i + c0_opt)
|
||||
aic_model = 6. - 2. * sum_log # three parameters: b_opt, m_opt, c0_opt
|
||||
m_opt, c0_opt, b_opt, aic_opt = fitting.fit_piece_wise_linear(d, C)
|
||||
|
||||
# TODO
|
||||
# should this be part of the plotting function itself, it should not be necessary for me to do this
|
||||
@@ -304,20 +289,14 @@ for i in range(1, 6):
|
||||
[lambda x: m_opt * x + c0_opt, lambda x: m_opt * b_opt + c0_opt]
|
||||
)
|
||||
# plot model containing modeled piece-wise linear function
|
||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_model)
|
||||
plot.quantification_plot(ax1, quantification, d_curve, C_curve, 'Models', aic_opt)
|
||||
|
||||
# linear regression model
|
||||
m_reg, c_reg = fitting.fit_linear_regression(d, C)
|
||||
|
||||
# AIC
|
||||
sum_log = 0.0
|
||||
for x_i in x_spatial:
|
||||
sum_log += math.log(m_reg * x_i + c_reg)
|
||||
aic_regression = 4. - 2. * sum_log # two parameter: m_reg, c_reg
|
||||
m_reg, c_reg, aic_reg = fitting.fit_linear_regression(d, C)
|
||||
|
||||
x = np.linspace(min(d), max(d), 500)
|
||||
y = m_reg * x + c_reg
|
||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_regression}")
|
||||
ax1.plot(x, y, color='k', linewidth=1, label=f"Simple Linear Regression | AIC: {aic_reg}")
|
||||
ax1.legend()
|
||||
|
||||
fig.savefig(f"uniform_random_point_clouds/{i}_pagerank.svg", format='svg')
|
||||
|
||||
Reference in New Issue
Block a user