## Understanding CNN MouseNet Iris Shi June 2020 <!-- Put the link to this slide here so people can follow --> <!-- slide: https://hackmd.io/@zjjxsjh/Syfcim-IU --> --- ## Introduction of mousenet ![](https://i.imgur.com/f6mTj8c.png =600x400) <font size="4"> - Each cortical layer (layer 4/23/5) in each area is modelled as a conv layer - Layer 5 from all areas output to a two-layer fully connected classifier </font> --- #### MouseNet is trained on CIFAR10 dataset (32x32, 10 classes) ![](https://i.imgur.com/M5hQVdb.jpg =600x400) --- ## Visualization Stimuli ![](https://i.imgur.com/Ulm9Ttq.png =300x300) <font size="4"> - Tiny Imagenet test images - 10000 64x64 natural images (downsampled from ImageNet) </font> --- ## Visulization Methods - Gradient visualization with guided backpropagation (GBP) - Direct optimization (Optimization) - both adapted from [pytorch-cnn-visualizations](https://github.com/utkuozbulak/pytorch-cnn-visualizations) by Utku Ozbulak --- ### Conv1 (input->LGNv) all 5 channels ![](https://i.imgur.com/hWJ4x2F.png =580x400) <font size="4"> - Learned features in first layer is simpler than AlexNet - The fist channel is the most complex feature </font> --- ### Conv2 (LGNv->VISp4) all 22 channels ![](https://i.imgur.com/62S3Wtn.png =500x400) <font size="4"> - The first channel in first conv layer is highly represented by second conv layer with slightly different forms </font> --- ### Conv3 (VISp4->VISp2/3) "best" channels ![](https://i.imgur.com/0az9UBL.png =500x400) --- ### Conv3 (VISp4->VISp2/3) "worst" channels ![](https://i.imgur.com/KSDi3Wd.png =500x400) --- ### Visualization during learning ![](https://i.imgur.com/BUTPWoS.png =600x350) <font size="4"> - Lower frequency features are learned first and then refined to higher frequency features - Lower layer feature emerge earlier than higher layer </font> --- ### Summary and future direction <font size="4"> - Visualization methods only work relatively well on the first three conv layers, possibly due to - GBP: not using enough stimuli - using more stimuli, e.x. various sizes of tiling of ImageNet - Optimization: not enough regularization - adding stronger regularization prior - Intermediate layers are more distributed codes - studying groups of channels instead of single channel </font> --- ### Thank you! :sheep: :100: :muscle: :tada:
{"metaMigratedAt":"2023-06-15T05:27:17.853Z","metaMigratedFrom":"YAML","title":"Understanding CNN MouseNet","breaks":true,"description":"View the slide with \"Slide Mode\".","contributors":"[{\"id\":\"c394e885-5c37-4a17-bd13-61bf09a9d3dc\",\"add\":6503,\"del\":3954}]"}
    396 views