梯度下降实现-绝对误差问题

3

我正在尝试在Python中实现梯度下降算法。当我绘制损失函数的历史记录时,似乎会收敛,但是我使用自己实现的方法得到的平均绝对误差要比从sklearn的linear_model获得的要糟糕得多。我没能找出我的实现有什么问题。

import matplotlib.pyplot as plt
import numpy as np
import pandas as pd
from sklearn import linear_model
from sklearn.model_selection import train_test_split
from sklearn.metrics import mean_absolute_error


def gradient_descent(x, y, theta, alpha, num_iters):
    m = len(y)
    cost_history = np.zeros(num_iters)

    for iter in range(num_iters):
        h = np.dot(x, theta)

        for i in range(len(theta)):
            theta[i] = theta[i] - (alpha/m) * np.sum((h - y) * x[:,i])

        #save the cost in every iteration
        cost_history[iter] = np.sum(np.square((h - y))) / (2 * m)

    return theta, cost_history


attributes = [...]
class_field = [...]

x_df = pd.read_csv('train.csv', usecols = attributes)
y_df = pd.read_csv('train.csv', usecols = class_field)

#normalize
x_df = (x_df - x_df.mean()) / x_df.std()

#gradient descent
alpha = 0.01
num_iters = 1000

err = 0
i = 10
for i in range(i):
    x_train, x_test, y_train, y_test = train_test_split(x_df, y_df, test_size=0.2)
    x_train = np.array(x_train)
    y_train = np.array(y_train).flatten()
    theta = np.random.sample(len(x_df.columns))
    theta, cost_history = gradient_descent(x_train, y_train, theta, alpha, num_iters)
    err = err + mean_absolute_error(y_test, np.dot(x_test, theta))
    print(np.dot(x_test, theta))
    #plt.plot(cost_history)
    #plt.show()
print(err/i)

regr = linear_model.LinearRegression()
regr.fit(x_train, y_train)
y_pred = regr.predict(x_test)
print(mean_absolute_error(y_test, y_pred))
1个回答

3

看起来你错过了偏置/截距列和系数。

线性函数的假设应该是这样的:

H = theta_0 + theta_1 * x

在您的实现中,它看起来如下:

H = theta_1 * x

网页内容由stack overflow 提供, 点击上面的
可以查看英文原文,
原文链接