A Python visualization toolkit, built with PyTorch, for neural networks in PyTorch.
Neural networks are often described as "black box". The lack of understanding on how neural networks make predictions enables unpredictable/biased models, causing real harm to society and a loss of trust in AI-assisted systems.
Feature visualization is an area of research, which aims to understand how neural networks perceive images. However, implementing such techniques is often complicated.
FlashTorch was created to solve this problem!
It is compatible with pre-trained models that come with torchvision, and seamlessly integrates with other custom models built in PyTorch.
Take a look at the quick 3min intro/demo to FlashTorch below!
Want to try?
Head over to example notebooks on Colab!
If you are installing FlashTorch for the first time:
$ pip install flashtorch
Or if you are upgrading it:
$ pip install flashtorch -U
These are currently available modules.
flashtorch.utils: some useful utility functions for data handling & transformation
ImageNetIndexclass for easy-ish retrieval of class index
Backpropclass for calculating gradients
GradientAscentclass for activation maximization
You can inspect each module with Python built-in function
help. The output of that is available on Quick API Guide for your convenience.
Below, you can find simple demos to get you started, as well as links to some handy notebooks showing additional examples of using FlashTorch.
Image handling (
- Image handling notebook
Saliency maps (
Saliency in human visual perception is a subjective quality that makes certain things within the field of view stand out from the rest and grabs our attention.
Saliency maps in computer vision provide indications of the most salient regions within images. By creating a saliency map for neural networks, we can gain some intuition on "where the network is paying the most attention to" in an input image.
The network is focusing on the sunken eyes and the round head for this owl.
Activation maximization (
Activation maximization is one form of feature visualization that allows us to visualize what CNN filters are "looking for", by applying each filter to an input image and updating the input image so as to maximize the activation of the filter of interest (i.e. treating it as a gradient ascent task with filter activation values as the loss).
Concepts such as 'eyes' (filter 45) and 'entrances (?)' (filter 271) seem to appear in the conv5_1 layer of VGG16.
Visit the notebook above to see what earlier layers do!
How to contribute
Thanks for your interest in contributing!
Please first head over to the Code of Conduct, which helps set the ground rules for participation in communities and helps build a culture of respect.
Still here? Great! There are many ways to contribute to this project. Get started here.
Talks & blog posts
Introduction and overview of feature visualization: Feature Visualization
The latest development in feature visualization: Exploring Neural Networks with Activation Atlases
Using backpropagation for gradient visualization: Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps
Guided backprobagation: Striving for Simplicity: The All Convolutional Net
Activation maximization: Visualizing Higher-Layer Features of a Deep Network
Misa Ogura (2019, September 26). MisaOgura/flashtorch: 0.1.1 (Version v0.1.1). Zenodo. http://doi.org/10.5281/zenodo.3461737