題3.1
試析在什麼情況下f(x)=w^(T)+b中不必考慮偏置項b
首先要知道爲什麼要加偏置項?它的作用是什麼?在之前學過的一次函數中,b其實就是函數在y軸的截距,控制着函數偏離原點的距離,那麼在線性模型中應該也是類似作用。看了這個博主的文章,大概意思就是加偏置項是爲了更好的擬合數據。
其實我自己看到這個題第一反應是,如果讓 fi-fj,這樣就可以消掉b了,自然也就不用考慮它。也就是讓所有數據集都和任意一個數據集相減,最後得到一個新的數據模型f(x)’=w^(T)。
但是又感覺不是這樣的吧,莫名心虛,就有去看別人的回答主要有兩種答案,感覺都很有道理:
- ①在f(x)進行歸一化處理時消除偏置,具體方法在這裏
- ②當只需要考慮x的取值對y的影響的話,則可以不用考慮b。
題3.2
試證明,對於參數w,對率迴歸的目標函數(3.18)是非凸的,但其對數似然函數(3.27)是凸的。
題3.3
編程實現對率迴歸,並給西瓜數據集3.0a上的結果。
import numpy as np
import math
import matplotlib.pyplot as plt
data_x = [[0.697, 0.460], [0.774, 0.376], [0.634, 0.264], [0.608, 0.318], [0.556, 0.215], [0.403, 0.237],
[0.481, 0.149], [0.437, 0.211],[0.666, 0.091], [0.243, 0.267], [0.245, 0.057], [0.343, 0.099],
[0.639, 0.161], [0.657, 0.198],[0.360, 0.370], [0.593, 0.042], [0.719, 0.103]
]
data_y = [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0]
def combine(beta, x):
x = np.mat(x + [1.]).T
return beta.T * x
def predict(beta, x):
return 1 / (1 + math.exp(-combine(beta, x)))
def p1(beta, x):
return math.exp(combine(beta, x)) / (1 + math.exp(combine(beta, x)))
beta = np.mat([0.] * 3).T
steps = 50
for step in range(steps):
param_1 = np.zeros((3, 1))
for i in range(len(data_x)):
x = np.mat(data_x[i] + [1.]).T
param_1 = param_1 - x * (data_y[i] - p1(beta, data_x[i]))
param_2 = np.zeros((3, 3))
for i in range(len(data_x)):
x = np.mat(data_x[i] + [1.]).T
param_2 = param_2 + x * x.T * p1(beta, data_x[i]) * (1 - p1(beta, data_x[i]))
last_beta = beta
beta = last_beta - param_2.I * param_1
if np.linalg.norm(last_beta.T - beta.T) < 1e-6:
print(step)
break
for i in range(len(data_x)):
if data_y[i] == 1:
plt.plot(data_x[i][0], data_x[i][1], 'ob')
else:
plt.plot(data_x[i][0], data_x[i][1], '^g')
w_0 = beta[0, 0]
w_1 = beta[1, 0]
b = beta[2, 0]
print(w_0, w_1, b)
x_0 = -b / w_0 #(x_0, 0)
x_1 = -b / w_1 #(0, x_1)
plt.plot([x_0, 0], [0, x_1])
plt.show()
題3.5
編程實現線性判別分析,並給出西瓜數據集3.0a上的結果。
import numpy as np
import math
import matplotlib.pyplot as plt
data_x = [[0.697, 0.460], [0.774, 0.376], [0.634, 0.264], [0.608, 0.318], [0.556, 0.215], [0.403, 0.237],
[0.481, 0.149], [0.437, 0.211],
[0.666, 0.091], [0.243, 0.267], [0.245, 0.057], [0.343, 0.099], [0.639, 0.161], [0.657, 0.198],
[0.360, 0.370], [0.593, 0.042], [0.719, 0.103]]
data_y = [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0]
mu_0 = np.mat([0., 0.]).T
mu_1 = np.mat([0., 0.]).T
count_0 = 0
count_1 = 0
for i in range(len(data_x)):
x = np.mat(data_x[i]).T
if data_y[i] == 1:
mu_1 = mu_1 + x
count_1 = count_1 + 1
else:
mu_0 = mu_0 + x
count_0 = count_0 + 1
mu_0 = mu_0 / count_0
mu_1 = mu_1 / count_1
S_w = np.mat([[0, 0], [0, 0]])
for i in range(len(data_x)):
# 注意:西瓜書的輸入向量是列向量形式
x = np.mat(data_x[i]).T
if data_y[i] == 0:
S_w = S_w + (x - mu_0) * (x - mu_0).T
else:
S_w = S_w + (x - mu_1) * (x - mu_1).T
u, sigmav, vt = np.linalg.svd(S_w)
sigma = np.zeros([len(sigmav), len(sigmav)])
for i in range(len(sigmav)):
sigma[i][i] = sigmav[i]
sigma = np.mat(sigma)
S_w_inv = vt.T * sigma.I * u.T
w = S_w_inv * (mu_0 - mu_1)
w_0 = w[0, 0]
w_1 = w[1, 0]
tan = w_1 / w_0
sin = w_1 / math.sqrt(w_0 ** 2 + w_1 ** 2)
cos = w_0 / math.sqrt(w_0 ** 2 + w_1 ** 2)
print(w_0, w_1)
for i in range(len(data_x)):
if data_y[i] == 0:
plt.plot(data_x[i][0], data_x[i][1], "go")
else:
plt.plot(data_x[i][0], data_x[i][1], "b^")
plt.xlabel('x')
plt.ylabel('y')
plt.title('Linear Discriminant Analysis')
plt.plot(mu_0[0, 0], mu_0[1, 0], "ro")
plt.plot(mu_1[0, 0], mu_1[1, 0], "r^")
plt.plot([-0.1, 0.1], [-0.1 * tan, 0.1 * tan])
for i in range(len(data_x)):
x = np.mat(data_x[i]).T
ell = w.T * x
ell = ell[0, 0]
if data_y[i] == 0:
plt.scatter(cos * ell, sin * ell, marker='o', c='', edgecolors='g')
else:
plt.scatter(cos * ell, sin * ell, marker='^', c='', edgecolors='b')
plt.show()