Table of Contents

Future Work and Open Questions

In our papers (SVCCA, Insights on Representational Similarity), we explored some questions using these technqiues, but many other open questions remain:

  1. Applying some of these methods to understand properties of generative models.
  2. Exploring low rank based compression, particularly with PLS
  3. Understanding learning dynamics and representational similarity of generalizing/memorizing networks through training.
  4. Identifying the important neurons in a layer, and determining how many there are. Relatedly, identifying which neurons are become sensitive to which classes.
  5. Pinpointing the effect of different kinds of layers on the representation.
  6. Comparing representations in biological and artificial neural networks.

We hope the code and tutorials are a helpful tool in better understanding representational properties of neural networks!


In [ ]: