PT - JOURNAL ARTICLE AU - K. Seeliger AU - M. Fritsche AU - U. Güçlü AU - S. Schoenmakers AU - J.-M. Schoffelen AU - S. E. Bosch AU - M. A. J. van Gerven TI - CNN-based Encoding and Decoding of Visual Object Recognition in Space and Time AID - 10.1101/118091 DP - 2017 Jan 01 TA - bioRxiv PG - 118091 4099 - http://biorxiv.org/content/early/2017/03/18/118091.short 4100 - http://biorxiv.org/content/early/2017/03/18/118091.full AB - Deep convolutional neural networks (CNNs) have been put forward as neurobiologically plausible models of the visual hierarchy. Using functional magnetic resonance imaging, CNN representations of visual stimuli have previously been shown to correspond to processing stages in the ventral and dorsal streams of the visual system. Whether this correspondence between models and brain signals also holds for activity acquired at high temporal resolution has been explored less exhaustively. Here, we addressed this question by combining CNN-based encoding models with magnetoencephalography (MEG). Human participants passively viewed 1000 images of objects while MEG signals were acquired. We modelled their high temporal resolution source-reconstructed cortical activity with CNNs, and observed a feedforward sweep across the visual hierarchy between 75-200 ms after stimulus onset. This spatiotemporal cascade was captured by the network layer representations, where the increasingly abstract stimulus representation in the hierarchical network model was reflected in different parts of the visual cortex, following the visual ventral stream. We further validated the accuracy of our encoding model by decoding stimulus identity in a left-out validation set of viewed objects, achieving state-of-the-art decoding accuracy.