我正在学习pytorch并运行玩具回归问题。令我感到困惑的是,每当我通过模型运行张量时,预测就会发生变化。显然情况不是这样,但我错过了什么?
Pytorch版本:0.4.0
我在这里没有使用GPU来消除这个潜在的问题。
代码:
from __future__ import division
import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
import torch
import torch.utils.data as utils_data
from torch.autograd import Variable
from torch import optim, nn
from torch.utils.data import Dataset
import torch.nn.functional as F
from torch.nn.init import xavier_uniform_, xavier_normal_,uniform_
from sklearn.datasets import load_boston
from sklearn.metrics import mean_squared_error
cuda=False #set to true uses GPU
#load boston data from scikit
boston = load_boston()
x=boston.data
y=boston.target
y=y.reshape(y.shape[0],1)
#change to tensors
x = torch.from_numpy(x)
y = torch.from_numpy(y)
#create dataset and use data loader
training_samples = utils_data.TensorDataset(x, y)
data_loader = utils_data.DataLoader(training_samples, batch_size=64)
#simple model
class Net(nn.Module):
def __init__(self):
super(Net, self).__init__()
#all the layers
self.fc1 = nn.Linear(x.shape[1], 20)
xavier_uniform_(self.fc1.weight.data) #this is how you can change the weight init
self.drop = nn.Dropout(p=0.5)
self.fc2 = nn.Linear(20, 1)
def forward(self, x):
x = F.relu(self.fc1(x))
x= self.drop(x)
x = self.fc2(x)
return x
net=Net()
if cuda:
net.cuda()
# create a stochastic gradient descent optimizer
optimizer = optim.Adam(net.parameters())
# create a loss function (mse)
loss = nn.MSELoss(size_average=True)
# run the main training loop
epochs =50
hold_loss=[]
for epoch in range(epochs):
cum_loss=0.
for batch_idx, (data, target) in enumerate(data_loader):
tr_x, tr_y = data.float(), target.float()
if cuda:
tr_x, tr_y = tr_x.cuda(), tr_y.cuda()
# Reset gradient
optimizer.zero_grad()
# Forward pass
fx = net(tr_x)
output = loss(fx, tr_y) #loss for this batch
cum_loss += output.item() #accumulate the loss
# Backward
output.backward()
# Update parameters based on backprop
optimizer.step()
hold_loss.append(cum_loss/len(training_samples))
#training loss
plt.plot(np.array(hold_loss))
此部分,如果重新运行,每次都会返回不同的预测,实际值不会改变,因此数据的顺序不会改变!
#score the training set
for batch_idx, (data, target) in enumerate(data_loader):
tr_x, tr_y = data.float(), target.float()
if batch_idx ==0:
hold_pred=net(tr_x).data.numpy()
hold_actual=tr_y.data.numpy().reshape(tr_y.data.numpy().shape[0],1)
else:
hold_pred =np.row_stack([hold_pred,net(tr_x).data.numpy()])
hold_actual=np.row_stack([hold_actual,tr_y.data.numpy().reshape(tr_y.data.numpy().shape[0],1)])
#view the first few predictions
print(hold_pred[0:10])
print(hold_actual[0:10])
答案 0 :(得分:3)
您的网络有一个Dropout
图层,其目的是随机抽样(此处概率为p=0.5
)在训练期间收到的数据(在推断之前设置net.train()
)。有关详细信息(用法,用途),请参阅doc。
此测试期间可以将此层短路(在推断之前设置net.eval()
)。