逻辑回归:对象未对齐 [英] Logistic regression: objects are not aligned

查看:69
本文介绍了逻辑回归:对象未对齐的处理方法,对大家解决问题具有一定的参考价值,需要的朋友们下面随着小编来一起学习吧!

问题描述

我正在尝试对来自Coursera A Ng的machihne学习班的此数据集进行逻辑回归.

I am trying to do logistic regression on this dataset from A Ng's machihne learning class in coursera.

这个想法是我们有一个成本函数,我们需要将其最小化以找到参数theta.

The idea is that we have a cost function, which we need to minimize to find the parameters theta.

import numpy as np
from scipy.optimize import fmin_bfgs

data = np.loadtxt('ex2data1.txt',delimiter=",")
m,n = data.shape
X = np.array(np.column_stack((np.ones(m),data[:,:-1])))
y = np.array(data[:,2].reshape(m,1))
theta = np.array(np.zeros(n).reshape(n,1))

def sigmoid(z):
    return 1/(1+np.exp(-z))

def hypothesis(X,theta):
    return sigmoid( X.dot(theta) )

def cost(theta):
    print theta.shape
    h = hypothesis(X,theta)
    cost = (-y.T.dot(np.log(h))-(1-y).T.dot(np.log(1-h)))/m
    return cost

def gradient(theta):
    h = hypothesis(X,theta)
    grad = ((h-y).T.dot(X)).T/m
    return grad.flatten()

def fmin():
    initial_theta=np.zeros(n).reshape(n,1)
    theta=fmin_bfgs(cost,initial_theta,fprime=gradient)
    return theta

打印fmin()

我得到了ValueError: Objects are not aligned,但是我检查了所有实体的形状,但仍然无法弄清楚.这是回溯:

I am getting ValueError: Objects are not aligned but I have checked the shapes of all entities and still can't figure it out. Here is the traceback:

---> 32     theta=fmin_bfgs(cost,initial_theta,fprime=gradient)
     33 

/usr/lib/python2.7/dist-packages/scipy/optimize/optimize.pyc in fmin_bfgs(f, x0, fprime, args, gtol, norm, epsilon, maxiter, full_output, disp, retall, callback)
    775             'return_all': retall}
    776 
--> 777     res = _minimize_bfgs(f, x0, args, fprime, callback=callback, **opts)
    778 
    779     if full_output:

/usr/lib/python2.7/dist-packages/scipy/optimize/optimize.pyc in _minimize_bfgs(fun, x0, args, jac, callback, gtol, norm, eps, maxiter, disp, return_all, **unknown_options)
    844     gnorm = vecnorm(gfk, ord=norm)
    845     while (gnorm > gtol) and (k < maxiter):
--> 846         pk = -numpy.dot(Hk, gfk)
    847         try:
    848             alpha_k, fc, gc, old_fval, old_old_fval, gfkp1 = \

ValueError: objects are not aligned

推荐答案

我修改了代码,使用c=inf可以得到与sklearn中LogisticRegression相同的结果:

I modified your code, it can get the same result as LogisticRegression in sklearn with c=inf:

import numpy as np
from scipy.optimize import fmin_bfgs
import io
data = np.loadtxt('ex2data1.txt',delimiter=",")
m,n = data.shape
X = np.array(np.column_stack((np.ones(m),data[:,:-1])))
y = np.array(data[:,2].reshape(m,1))
theta = np.array(np.zeros(n).reshape(n,1))

def sigmoid(z):
    return 1/(1+np.exp(-z))

def hypothesis(X,theta):
    return sigmoid( X.dot(theta) )

def cost(theta):
    h = hypothesis(X,theta)
    cost = (-y.T.dot(np.log(h))-(1-y).T.dot(np.log(1-h)))/m
    r = cost[0]
    if np.isnan(r):
        return np.inf
    return r

def gradient(theta):
    theta = theta.reshape(-1, 1)
    h = hypothesis(X,theta)
    grad = ((h-y).T.dot(X)).T/m
    return grad.flatten()

def fmin():
    initial_theta=np.zeros(n)
    theta=fmin_bfgs(cost,initial_theta,fprime=gradient)
    return theta

theta = fmin()

这篇关于逻辑回归:对象未对齐的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!

查看全文
登录 关闭
扫码关注1秒登录
发送“验证码”获取 | 15天全站免登陆