@article{retsinas_recnets_2019, title = {{RecNets}: {Channel}-wise {Recurrent} {Convolutional} {Neural} {Networks}}, shorttitle = {{RecNets}}, url = {http://arxiv.org/abs/1905.11910}, abstract = {In this paper, we introduce Channel-wise recurrent convolutional neural networks (RecNets), a family of novel, compact neural network architectures for computer vision tasks inspired by recurrent neural networks (RNNs). RecNets build upon Channel-wise recurrent convolutional (CRC) layers, a novel type of convolutional layer that splits the input channels into disjoint segments and processes them in a recurrent fashion. In this way, we simulate wide, yet compact models, since the number of parameters is vastly reduced via the parameter sharing of the RNN formulation. Experimental results on the CIFAR-10 and CIFAR-100 image classification tasks demonstrate the superior size-accuracy trade-off of RecNets compared to other compact state-of-the-art architectures.}, urldate = {2020-01-28}, journal = {arXiv:1905.11910 [cs, stat]}, author = {Retsinas, George and Elafrou, Athena and Goumas, Georgios and Maragos, Petros}, month = may, year = {2019}, note = {arXiv: 1905.11910}, keywords = {Computer Science - Machine Learning, Statistics - Machine Learning}, }