implementation:
最近实践心得: COCOB这个优化器,发如今某些网络的训练上面很是的不稳定,仍是设置Adam比较靠谱。。。。网络
lr cosine 的学习率的设计方式通常和momentum optimizer搭配~~~~~ide
记录一下本身给跪的理解力:学习
tf.nn.seperable_conv2d(inputs, depthwise_filter, pointwise_filter) depthwise_filter: [filter_height, filter_width, in_channels, channel_multiplier] pointwise_filter: [1, 1, channel_multiplier * in_channels, out_channels] output[b, i, j, k] = sum_{di, dj, q, r} input[b, strides[1] * i + di, strides[2] * j + dj, q] * depthwise_filter[di, dj, q, r] * pointwise_filter[0, 0, q * channel_multiplier + r, k] 因此其实这边对应的操做就是depthwise_filter的第i层只对input的第i层操做。