如何在张量流中控制LSTM单元的输出大小

时间:2017-03-15 14:49:41

标签: python tensorflow lstm recurrent-neural-network

我已经尝试构建一个简单的例子,通过Tensorflow使用LSTM RNN来预测一些目标序列的时间序列值,给定已知的输入时间序列。

Link to example problem

我正在尝试

what I try to accomplish formally

本质上我认为单元格A和下面的矩阵mult的输出应该起作用:

X = np.zeros([40,2,1])
A = np.zeros([40,1,2])
b = np.arange(0,2)

X = tf.convert_to_tensor(X)
A = tf.convert_to_tensor(A)
b = tf.convert_to_tensor(b)

Y = tf.matmul(X,A)+b

张量流代码设置为查看输出大小,而不是功能tf.graph / session:

import numpy as np
import tkinter
import matplotlib.pyplot as plt
import tensorflow as tf
n=40
x = np.linspace(0,10,n)
y1 = np.sin(x)
y2 = np.cos(x)

x1=np.random.normal(0,y1**2,n)
x2=np.random.normal(0,y2**2,n)

y1=(y1**2>0.4)*1
y2=(y2**2>0.4)*1

ys = np.vstack((y1,y2))
xs = np.vstack((x1,x2))

def plot_results_multiple(xs, ys):
    fig = plt.figure(facecolor='white')
    ax = fig.add_subplot(111)
    for i, data in enumerate(xs):
        plt.plot(data, label='x'+str(i))
        plt.legend()
    for i, data in enumerate(ys):
        plt.plot(data, label='y'+str(i))
        plt.legend()
    plt.show()

plot_results_multiple(xs,ys)

xs = xs.T
ys = ys.T

print("Shape of arrays " +str(xs.shape) + " " +str(ys.shape))


batch_size = 1
lstm_size = 1
nseries = 2
time_steps = 40
nclasses = 2

lstm = tf.contrib.rnn.BasicLSTMCell(lstm_size,state_is_tuple=True)
stacked_lstm = tf.contrib.rnn.MultiRNNCell([lstm] * 2, state_is_tuple=True)

state = lstm.zero_state(batch_size, tf.float32)
inputs = tf.unstack(xs, num=40, axis=0)

outputs = []

with tf.variable_scope("RNN"):
    for timestep in range(time_steps):
        if timestep > 0: tf.get_variable_scope().reuse_variables()
        output, state = lstm(tf.cast(tf.reshape(inputs[timestep],[1,nseries]),tf.float32), state)
        print(tf.convert_to_tensor(output).get_shape())
        outputs.append(output)

print(tf.convert_to_tensor(outputs).get_shape())
output = tf.reshape(tf.concat(outputs, 1), [-1, lstm_size])
softmax_w = tf.get_variable(
    "softmax_w", [time_steps, 1,nclasses],tf.float32)# dtype=
print(softmax_w.get_shape())
softmax_b = tf.get_variable("softmax_b", [nseries], dtype=tf.float32)
print(softmax_b.get_shape())
logits = tf.matmul(output, softmax_w) + softmax_b

print(logits.get_shape())

我认为我遇到的问题是如何修改RNN LSTM单元,因为它当前从2x1输入输出1x1张量,我预计输出2x1。非常感谢任何帮助。

1 个答案:

答案 0 :(得分:3)

它由隐藏的神经元单位控制,这是tf.contrib.rnn.BasicLSTMCell(lstm_size,state_is_tuple)的第一个参数

源代码在这里: https://github.com/tensorflow/tensorflow/blob/b0ecc7d2c1486367ec65d297e372f8935ee3ddfe/tensorflow/python/ops/rnn_cell_impl.py#254

@property
  def output_size(self):
    return self._num_units

因此,如果要更改输出大小,则需要更改tf.contrib.rnn.BasicLSTMCell的num_units。