5

我正在使用Theano 0.7,nolearn 0.6adevlasagne 0.2.dev1在 GPU 上训练神经网络(在IPython 3.2.1笔记本中)。'reduc'但是,由于第一层(

import theano
from lasagne import layers
from lasagne.updates import nesterov_momentum
from nolearn.lasagne import NeuralNet
from nolearn.lasagne import BatchIterator
from lasagne import nonlinearities
from lasagne import init
import numpy as np

testNet = NeuralNet(
    layers=[(layers.InputLayer, {"name": 'input', 'shape': (None, 12, 1000, )}),

            (layers.Conv1DLayer, {"name": 'reduc', 'filter_size': 1, 'num_filters': 4, 
                                  "nonlinearity":nonlinearities.linear,}),

            (layers.Conv1DLayer, {"name": 'conv1', 'filter_size': 25, 'num_filters': 100,
                                 'pad': 'same', }),
            (layers.MaxPool1DLayer, {'name': 'pool1', 'pool_size': 5, 'stride': 3}),



            (layers.Conv1DLayer, {"name": 'conv2', 'filter_size': 15, 'num_filters': 100,
                                 'pad': 'same', 
                                  'nonlinearity': nonlinearities.LeakyRectify(0.2)}),

            (layers.MaxPool1DLayer, {'name': 'pool2', 'pool_size': 5, 'stride': 2}),

            (layers.Conv1DLayer, {"name": 'conv3', 'filter_size': 9, 'num_filters': 100,
                                 'pad': 'same', 
                                  'nonlinearity': nonlinearities.LeakyRectify(0.2)}),

            (layers.MaxPool1DLayer, {'name': 'pool3', 'pool_size': 2}),

            (layers.Conv1DLayer, {"name": 'conv4', 'filter_size': 5, 'num_filters': 20,
                                 'pad': 'same', }),

            (layers.Conv1DLayer, {"name": 'conv5', 'filter_size': 3, 'num_filters': 20,
                                 'pad': 'same',}),


            (layers.DenseLayer, {"name": 'hidden1', 'num_units': 10, 
                                 'nonlinearity': nonlinearities.rectify}),

            (layers.DenseLayer, {"name": 'output', 'nonlinearity': nonlinearities.sigmoid, 
                                 'num_units': 5})
           ],

    # optimization method:
    update=nesterov_momentum,
    update_learning_rate=5*10**(-3),
    update_momentum=0.9,

    regression=True,  
    max_epochs=1000,
    verbose=1,
    )

testNet.fit(np.random.random([3000, 12, 1000]).astype(np.float32), 
            np.random.random([3000, 5]).astype(np.float32))

如果我注释掉第一层,训练将在几秒钟后开始。训练更复杂的网络也不是问题。知道是什么导致了这个问题吗?

编辑:奇怪的是,如果我删除conv4and conv5,培训也会在合理的时间内开始。

Edit2:更奇怪的是,如果我在 layer 中将过滤器的大小更改为 10 reduc,那么训练会在合理的时间内开始。如果在那之后我停止单元格的执行,将此值更改为 1,然后重新执行单元格,则训练进行得很好......

最后我开始使用另一个框架,但如果有人感兴趣,这里是我在 lasagne 用户组上开始的线程的链接。

4

0 回答 0