Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep...

57
Accepted manuscripts are peer-reviewed but have not been through the copyediting, formatting, or proofreading process. Copyright © 2018 the authors This Accepted Manuscript has not been copyedited and formatted. The final version may differ from this version. A link to any extended data will be provided when the final version is posted online. Research Articles: Behavioral/Cognitive Deep neural networks for modeling visual perceptual learning Li Wenliang 1 and Aaron R Seitz 2 1 Gatsby Computational Neuroscience Unit, University College London, 25 Howland Street, London W1T 4JG, United Kingdom 2 Department of Psychology, University of California–Riverside, 900 University Avenue, Riverside, CA 92521 DOI: 10.1523/JNEUROSCI.1620-17.2018 Received: 10 June 2017 Revised: 12 March 2018 Accepted: 19 March 2018 Published: 23 May 2018 Author contributions: L.K.W. and A.R.S. designed research; L.K.W. performed research; L.K.W. analyzed data; L.K.W. and A.R.S. wrote the paper. Conflict of Interest: The authors declare no competing financial interests. The authors are supported by the Gatsby Charitable Foundation (LW) and National Institute for Health (NEI 1R01EY023582) (AS). We are very grateful to Peter Dayan for extensive discussions on the use of deep neural networks and Merav Ahissar for discussions on the relationship of this work to the Reverse Hierarchy Theory; they also provided helpful comments on earlier versions of the manuscript. We thank Kirsty McNaught and Sanjeevan Ahilan for proofreading the manuscript and suggestions on clarifications. We also thank two anonymous reviewers for their valuable comments. Corresponding author: Li Wenliang, Gatsby Computational Neuroscience Unit, University College London; 25 Howland Street, London W1T 4JG, United Kingdom, +44 (0) 20 3108 8100, [email protected] Cite as: J. Neurosci ; 10.1523/JNEUROSCI.1620-17.2018 Alerts: Sign up at www.jneurosci.org/cgi/alerts to receive customized email alerts when the fully formatted version of this article is published.

Transcript of Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep...

Page 1: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

Accepted manuscripts are peer-reviewed but have not been through the copyediting, formatting, or proofreadingprocess.

Copyright © 2018 the authors

This Accepted Manuscript has not been copyedited and formatted. The final version may differ fromthis version. A link to any extended data will be provided when the final version is posted online.

Research Articles: Behavioral/Cognitive

Deep neural networks for modeling visual perceptual learning

Li Wenliang1 and Aaron R Seitz2

1Gatsby Computational Neuroscience Unit, University College London, 25 Howland Street, London W1T 4JG,United Kingdom2Department of Psychology, University of California–Riverside, 900 University Avenue, Riverside, CA 92521

DOI: 10.1523/JNEUROSCI.1620-17.2018

Received: 10 June 2017

Revised: 12 March 2018

Accepted: 19 March 2018

Published: 23 May 2018

Author contributions: L.K.W. and A.R.S. designed research; L.K.W. performed research; L.K.W. analyzeddata; L.K.W. and A.R.S. wrote the paper.

Conflict of Interest: The authors declare no competing financial interests.

The authors are supported by the Gatsby Charitable Foundation (LW) and National Institute for Health (NEI1R01EY023582) (AS). We are very grateful to Peter Dayan for extensive discussions on the use of deepneural networks and Merav Ahissar for discussions on the relationship of this work to the Reverse HierarchyTheory; they also provided helpful comments on earlier versions of the manuscript. We thank Kirsty McNaughtand Sanjeevan Ahilan for proofreading the manuscript and suggestions on clarifications. We also thank twoanonymous reviewers for their valuable comments.

Corresponding author: Li Wenliang, Gatsby Computational Neuroscience Unit, University College London; 25Howland Street, London W1T 4JG, United Kingdom, +44 (0) 20 3108 8100, [email protected]

Cite as: J. Neurosci ; 10.1523/JNEUROSCI.1620-17.2018

Alerts: Sign up at www.jneurosci.org/cgi/alerts to receive customized email alerts when the fully formattedversion of this article is published.

Page 2: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

Title: Deep neural networks for modeling visual perceptual learning 1

Abbreviated title: DNNs for VPL 2

Authors: 3

Li Wenliang (Corresponding author): 4 Gatsby Computational Neuroscience Unit, University College London; 25 Howland 5 Street, London W1T 4JG, United Kingdom 6 +44 (0) 20 3108 8100 7 [email protected] 8 9

Aaron R Seitz: 10 Department of Psychology, University of California–Riverside; 900 University Avenue, 11 Riverside, CA 92521 12 +1 (951) 827-6422 13 [email protected] 14 15

Number of pages: 47 16

Number of 17

figures: 24 18 tables: 16 19

Number of words for 20

Abstract: 249 21 Introduction: 638 22 Discussion: 1526 23

Conflict of Interest: The authors declare no competing financial interests 24

Acknowledgements: The authors are supported by the Gatsby Charitable Foundation (LW) and 25 National Institute for Health (NEI 1R01EY023582) (AS). We are very grateful to Peter Dayan 26 for extensive discussions on the use of deep neural networks and Merav Ahissar for discussions 27 on the relationship of this work to the Reverse Hierarchy Theory; they also provided helpful 28 comments on earlier versions of the manuscript. We thank Kirsty McNaught and Sanjeevan 29 Ahilan for proofreading the manuscript and suggestions on clarifications. We also thank two 30 anonymous reviewers for their valuable comments. 31

Journal section: Behavioral/Cognitive 32

33

Page 3: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

1

Abstract 34 Understanding visual perceptual learning (VPL) has become increasingly more challenging as new phenomena are 35 discovered with novel stimuli and training paradigms. While existing models aid our knowledge of critical aspects of 36 VPL, the connections shown by these models between behavioral learning and plasticity across different brain areas are 37 typically superficial. Most models explain VPL as readout from simple perceptual representations to decision areas and 38 are not easily adaptable to explain new findings. Here, we show that a well-known instance of deep neural network 39 (DNN), while not designed specifically for VPL, provides a computational model of VPL with enough complexity to be 40 studied at many levels of analyses. After learning a Gabor orientation discrimination task, the DNN model reproduced key 41 behavioral results, including increasing specificity with higher task precision, and also suggested that learning precise 42 discriminations could asymmetrically transfer to coarse discriminations when the stimulus conditions varied. In line with 43 the behavioral findings, the distribution of plasticity moved towards lower layers when task precision increased, and this 44 distribution was also modulated by tasks with different stimulus types. Furthermore, learning in the network units 45 demonstrated close resemblance to extant electrophysiological recordings in monkey visual areas. Altogether, the DNN 46 fulfilled predictions of existing theories regarding specificity and plasticity, and reproduced findings of tuning changes in 47 neurons of the primate visual areas. Although the comparisons were mostly qualitative, the DNN provides a new method 48 of studying VPL and can serve as a testbed for theories and assist in generating predictions for physiological 49 investigations. 50

Significance statement 51 Visual perceptual learning (VPL) has been found to cause changes at multiple stages of the visual hierarchy. We found 52 that training a deep neural network (DNN) on an orientation discrimination task produced similar behavioral and 53 physiological patterns found in human and monkey experiments. Unlike existing VPL models, the DNN was pre-trained 54 on natural images to reach high performance in object recognition but was not designed specifically for VPL, and yet it 55 fulfilled predictions of existing theories regarding specificity and plasticity, and reproduced findings of tuning changes in 56 neurons of the primate visual areas. When used with care, this unbiased and deep-hierarchical model can provide new 57 ways of studying VPL from behavior to physiology. 58

Introduction 59 Visual Perceptual Learning (VPL) refers to changes in sensitivity to visual stimuli through training or experience, and has 60 been demonstrated in the discrimination of simple features such as orientation, contrast and dot motion direction as well 61 as more complicated patterns (Fiorentini and Berardi, 1980; Ball and Sekuler, 1982; Karni and Sagi, 1991; Ahissar and 62 Hochstein, 1997; Mastropasqua et al., 2015). A common characteristic of VPL is its lack of transfer to untrained stimulus 63 conditions, such as when rotated by 90° (Fiorentini and Berardi, 1981; Schoups et al., 1995; Crist et al., 1997). Due to 64 their retinotopic mapping and orientation tuning (Hubel and Wiesel, 1968; Blasdel, 1992; Tootell et al., 1998), early 65 visual areas have been hypothesized to contribute to VPL and its specificity (Fahle, 2004). Despite numerous examples 66 supporting this hypothesis (Schoups et al., 2001; Bejjanki et al., 2011; Sagi, 2011; Jehee et al., 2012; Yu et al., 2016), 67 there is substantial evidence that specificity does not require low-level changes (Dosher and Lu, 1998; Ghose et al., 2002) 68 and there is great controversy regarding where learning happens in the visual hierarchy (Wang et al., 2016; Maniglia and 69 Seitz, 2018). 70

Most models of VPL are artificial neural networks with user-parametrized receptive fields and shallow network structures. 71 Trained using Hebbian-like learning rules (Sotiropoulos et al., 2011; Herzog et al., 2012; Dosher et al., 2013) or optimal 72 decoding methods (Zhaoping et al., 2003), these models can reproduce and predict behavior but rarely account for 73 physiological data. Moreover, a key limitation of these models is that they do not address how the multiple known visual 74 areas may jointly contribute to learning (Hung and Seitz, 2014). Other more conceptual models, such as the Reverse 75 Hierarchy Theory (RHT) (Ahissar and Hochstein, 1997; Ahissar and Hochstein, 2004) and the Dual Plasticity Model 76 (Watanabe and Sasaki, 2015), make predictions regarding what types of learning scenarios may lead to differential 77 plasticity across visual areas, but these descriptive models do not predict specific changes in tuning properties of neurons. 78

Page 4: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

2

Thus, there is a substantial need for a hierarchical model that can simulate learning and simultaneously produce 79 behavioral and neurological outcomes. 80

A novel approach to modeling VPL can be found in deep neural networks (DNNs) which is readily adapted to learn 81 different tasks. These DNNs have shown impressive correspondences to human behaviors and neural data from early 82 visual areas and inferior temporal cortex (IT) (Khaligh-Razavi and Kriegeskorte, 2014; Yamins et al., 2014; Guclu and 83 van Gerven, 2015; Cichy et al., 2016; Kheradpisheh et al., 2016; Eickenberg et al., 2017). This hierarchical system opens 84 up new opportunities for VPL research (Kriegeskorte, 2015). As a start, Lee & Saxe (2014) produced theoretical analyses 85 that resembled RHT predictions in a 3-layer linear network. Recently, Cohen and Weinshall (2017) used a shallow 86 network to replicate relative performances of different training conditions for a wide range of behavioral data. To date, the 87 extent to which DNNs can appropriately model physiological data of VPL remains unexplored. 88

In the present paper, we trained a DNN model modified from AlexNet (Krizhevsky et al., 2012) to perform Gabor 89 orientation and face gender discriminations. The network reflected human behavioral characteristics, such as the 90 dependence of specificity on stimulus precision (Ahissar and Hochstein, 1997; Jeter et al., 2009). Furthermore, the 91 distribution of plasticity moved towards lower layers when task precision increased, and this distribution was also 92 modulated by tasks with different types of stimulus. Most impressively, for orientation discrimination, the network units 93 showed similar changes to neurons in primate visual cortex and helped reconcile divergent physiological findings in the 94 literature (Schoups et al., 2001; Ghose et al., 2002). These results suggest that DNNs can serve as a computational model 95 for studying the relationship between behavioral learning and plasticity across the visual hierarchy during VPL, and how 96 patterns of learning vary as a function of training parameters (Maniglia and Seitz, 2018). 97

Materials and Methods 98

Model 99 An AlexNet-based DNN was used to simulate VPL. We briefly describe the network architecture here and refer readers to 100 the original paper (Krizhevsky et al., 2012) for details. The original AlexNet consists of eight layers of artificial neurons 101 (units) connected through feedforward weights. In the first five layers, each unit is connected locally to a small patch of 102 retinotopically arranged units in the previous layer or the input image. These connections are replicated spatially so that 103 the same set of features is extracted at all locations through weight sharing. The operation that combines local connection 104 with weight sharing is known as convolution. Activations in the last convolutional layer are sent to three fully connected 105 layers with the last layer corresponding to object labels in the original object classification task. Unit activations are 106 normalized in the first two convolutional layers to mimic lateral inhibition. 107

To construct the DNN model, we took only the first five convolutional layers of AlexNet and discarded the three fully 108 connected layers to reduce model complexity. An additional readout unit is added to fully connect with the units in the last 109 layer, forming a scalar representation of the stimulus in layer 6. We removed the last three layers of AlexNet because they 110 exhibited low representational similarity to early visual areas but high similarity to Inferior Temporal Cortex (IT) and 111 hence may be more relevant to object classification (Khaligh-Razavi and Kriegeskorte, 2014; Guclu and van Gerven, 112 2015), and we assume that early visual areas play a more critical role for Gabor orientation discrimination. We kept all the 113 five convolutional layers because one of our objectives was to study how learning was distributed over a visual hierarchy 114 with more levels than most VPL models which are usually limited to 2-3 levels. In addition, activations in these five 115 layers have been suggested to correspond to neural activities in V1-4 following a similar ascending order (Guclu and van 116 Gerven, 2015). 117

The resulting 6-layer network was further modified to model decision making in the two-interval two-alternative forced 118 choice (2I-2AFC) paradigm (Figure 1A). In this paradigm, a reference stimulus is first shown before a second target 119 stimulus, and the network has to judge whether the second stimulus is more clockwise or more counter-clockwise 120 compared with the reference. In our DNN model, each of the reference and target images is processed by the same 6-layer 121 network that yields a scalar readout in layer 6, and the decision is made based on the difference between the 122 representations with some noise. More precisely, two identical streams of the 6-layer network produce scalar 123

Page 5: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

3

representations for the reference and target images, denoted by and , respectively. The network outputs a clockwise 124 decision with probability (or confidence) by passing the difference through a logistic function 125

( 1 )

This construction assumes perfect memory about the two representations which are computed using the same model 126 architecture and parameters, and each choice is made with some decision noise. An advantage of using this 2I-2AFC 127 architecture is that, when tested under transfer conditions (such as a new reference orientation), the network can still 128 compare the target with the reference by taking the difference between the representations; whereas if only one stream 129 exists, the model cannot know the new reference orientation. We note that while this training paradigm is suitable for this 130 network and was thus kept consistent throughout the simulations, it is different from those used in the physiological 131 studies (Schoups et al., 2001; Ghose et al., 2002; Yang and Maunsell, 2004; Raiguel et al., 2006) with which we compare 132 our network in the Results section. Learning could also be influenced by details of those experiments beyond what is 133 accounted for by the present simulations. 134

Task and stimuli 135 All stimuli in the two experiments below were centered on 8-bit 227×227-pixel images with gray background. 136

Experiment 1 137 The network was trained to classify whether the target Gabor stimulus was tilted clockwise or counter-clockwise 138 compared to the reference. We trained the network on all combinations of the following parameters: 139

Orientation of reference Gabor: from 0° to 165° at steps of 15°; 140 Angle separation between reference and target (0.5°, 1.0°, 2.0°, 5.0°, 10.0°); 141 Spatial wavelength (10, 20, 30, 40 pixels). 142

To simulate the short period of stimulus presentation and sensory noise, we kept the contrast low and added to each image 143 isotropic Gaussian noise with the following parameters: 144

Signal contrast (20%, 30%, 40% to 50% of the dynamic range); 145 Standard deviation of the Gaussian additive noise (5 to 15 in steps of 5). 146

In addition, the standard deviation of the Gabor Gaussian window was 50 pixels. Noise was generated at run time 147 independently for each image. An example of a Gabor stimulus pair is shown in Figure 1A. 148

Experiment 2 149 The network was trained on face gender discrimination and Gabor orientation discrimination. For the face task, the 150 network was trained to classify whether the target face was more masculine or feminine (closer to the original male or 151 female in warping distance) than the reference face. Stimuli were face images with gender labels from the Photoface 152 Dataset (Zafeiriou et al., 2011). 647 male images and 74 female images with minimal facial hair were manually selected 153 from the dataset, captured in the frontal pose with the blank stare emotion. The bias in subject gender was addressed by 154 subsampling to form balanced training and testing sets. The facemorpher toolbox 155 (https://pypi.python.org/pypi/facemorpher/1.0.1) was used to create a reference that is halfway between a male and a 156 female image. 157

To manipulate task difficulty, target images were created that varied in dissimilarity to the reference image ranging from 1 158 (closest to the reference) to 5 (the original male or female image) by adjusting a warping (mixing) parameter. The 159 reference and target were morphed from the same pair of the original faces. The network was trained and tested using 12-160 fold cross-validation to simulate inter-observer variability. Each fold consists of images morphed from 49 males and 49 161 females for training (2401 pairs) and 25 males and 25 females for testing (625 pairs) randomly sampled from the full 162 dataset. Examples of face stimuli at the five dissimilarity levels are shown in Figure 1B. 163

Page 6: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

4

The Gabor stimulus had a wavelength of 20 pixels and the standard deviation of the Gaussian window was 50. The 164 reference angle ranged from 0° to 165° at steps of 15° and the target image deviated from the reference by 0.5, 1.0, 2.0, 165 5.0 or 10.0°. For both the face and Gabor tasks in this experiment, contrast was set to 50% and noise standard deviation 166 was set to 5. 167

Training procedure 168 In both experiments, network weights were initialized such that the readout weights in layer 6 were zeros and weights in 169 the other lower layers were copied from a pre-trained AlexNet (downloaded from 170 http://dl.caffe.berkeleyvision.org/bvlc_alexnet.caffemodel). The learning algorithm was stochastic gradient descent (SGD) 171 whereby the weights were changed to minimize the discrepancy between network output and stimulus label 172

( 2 )

( 3 )

where (0.0001) and (0.9) are learning rate and momentum, respectively, which are held constant through training, 173 is the network weights at iteration , and is the corresponding weight change. is the cross-entropy loss which 174 depends on the network weights, input image batch of size 20 pairs and the corresponding labels . Gradients were 175 obtained by backpropagation of the loss through layers (Rumelhart et al., 1986). 176

Under this learning rule, zero initialization in the readout weights prevents the other layers from changing in the first 177 iteration, since the weights in those layers cannot affect performance when the readout weights are zeros and hence have 178 zero gradients. This initialization can be interpreted as receiving instruction by subjects because all stimulus 179 representations in the lower layers are fixed while the network briefly learns the task on the highest decision layer. After 180 the first iteration, the readout weights will not be optimal due to small learning rate, and so weights in the lower layers 181 will start to change. Under each stimulus condition, the network was trained for 1000 iterations of 20-image batches so 182 that one iteration is analogous to a small training block for human subjects. Independent noise was generated to each 183 image at each iteration. We outline the limitations of the model in the Discussion section. 184

Behavioral performance 185 The network’s behavioral performance was estimated as the classification confidence (Equation 1) of the correct label 186 averaged over test trials. For the Gabor task, we tested the network’s performance on stimuli generated from the same 187 parameters as in training (trained condition), and also tested on stimuli generated from different parameters (transfer 188 conditions) including rotating the reference by 45 and 90°, halving or doubling the spatial frequencies and changing angle 189 separations between the reference and target (precisions). 200 pairs of Gabor stimuli were used in each test condition. For 190 the face task, performance was tested on 625 unseen validation images. Performance was measured at 20 roughly 191 logarithmically spaced iterations from 1 to 1000. 192

In Experiment 2, the contribution of DNN layers to performance was estimated using an accuracy drop measure defined 193 as follows. Under each stimulus condition, we recorded the iteration at which the fully plastic network reached 95% 194 accuracy, denoted by t95; we then trained a new network again from pretrained AlexNet weights under the same stimulus 195 condition while freezing successively more layers from layer 1, and used the accuracy drop at t95 compared to the all 196 plastic network as the contribution of the frozen layers. For example, suppose that the fully plastic network reached 95% 197 accuracy at 100th iteration (t95=100), and at this iteration, the network trained with layer 1 frozen had 90% accuracy, the 198 network trained with first two layers frozen had 85% accuracy, then the first layer contributed 5% and the first two layers 199 together contributed 10%. This accuracy drop does not indicate the contribution of each layer in isolation, but allows for 200 different interactions within the plastic higher layers when varying the number of frozen lower layers. 201

Estimating learning in layers and neurons 202 After training, weights at each layer were treated as a single vector, and learning was measured based on the difference 203 from pre-trained values. Specifically, for a particular layer with total connections to its lower layer, we denote the 204

Page 7: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

5

original -dimensional weight vector as ( and specified in the pre-trained AlexNet), the change in this vector from 205 training as , and define the layer change as 206

( 4 )

where indexes each element in the weight vector. Under this measure, scaling the weight vector by a constant gives the 207 same change regardless of dimensionality, reducing the effect of unequal dimensionalities on the magnitude of weight 208 change. For the final readout layer that was initialized as zeros, the denominator in Equation 4 was set to , effectively 209 measuring the average change per connection in this layer. When comparing weight change across layers, we focus on the 210 first five layers unless otherwise stated. In addition, the following alternative layer change measures were also used 211

( 4a )

( 4b )

( 4c )

which produced different values, but they did not change the general effects of stimulus conditions on distribution of 212 learning in weights. For the results in the main text, we report weight change in terms of unless otherwise stated. 213 Due to weight sharing, the layer change defined above is equal to the change in the subset of weights that share the same 214 receptive field. To measure learning of a single unit, we used the same equation but with being the filter of each unit 215 and being the size of the filter. 216

Tuning curves 217 For each unit in the network, we “recorded” its tuning curve before and after training by measuring its responses to Gabor 218 stimuli presented at 100 orientations evenly spaced over the 180° cycle. The stimuli had noise standard deviation 15, 219 contrast 20% and wavelength 10 pixels, and one period of the sinusoidal component of the Gabor stimulus was contained 220 within the receptive field of a layer-1 unit. The mean and standard deviation at each test orientation were obtained by 221 presenting the network with 50 realizations of noisy stimuli, followed by smoothing with a circular Gaussian kernel. The 222 gradients of tuning curves were computed by filtering the mean response with a circular Laplace filter. Both the Gaussian 223 and Laplace filters had a standard deviation of 1°. Since the receptive fields were shared across locations, we only chose 224 the units with receptive fields at the center of the image; thus, the number of units measured at each layer equals the 225 number of filter types (channels) in that layer. Also, to ensure that units were properly driven by the stimulus, we 226 excluded units that had mean activation over all orientations less than 1.0. The same procedure was repeated under 5 227 precisions and 12 reference orientations before and after training. No curve fitting was employed. On average, each 228 training condition produced the following number of units after training for analyses: 79.4 out of 96 in layer 1, 91.0 out of 229 256 in layer 2, 237.3 out of 384 in layer 3, 100.3 out of 384 in layer 4, and 16.0 out of 256 in layer 5. These numbers were 230 roughly the same for the naïve populations. Units were pooled together for the 12 reference orientations. 231

To compare with electrophysiological data in the literature, we measured the following attributes from tuning curves 232 nonparametrically. The preferred orientation was determined by the orientation at which a unit attained its peak response. 233 Tuning amplitude was taken to be the difference between the highest and lowest responses. Following (Raiguel et al., 234 2006), the selectivity index, a measure of tuning sharpness, was measured as 235

( 5 )

where is the mean activation of a unit to a Gabor stimulus presented at orientation for the ’th measured orientation 236 (from 1 to 100). The normalized variance (or Fano factor, variance ratio) of the response at a particular orientation was 237

Page 8: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

6

taken as the ratio of response variance to the mean. Following (Yang and Maunsell, 2004), we measured the best 238 discriminability of a unit by taking the minimum, over orientation, of response variance divided by tuning curve gradient 239 squared. 240

To measure how much information about orientation was contained in a layer per neuron, we computed the average linear 241 Fisher information (FI) (Seriès et al., 2004; Kanitscheider et al., 2015) at a particular orientation as 242

( 6 )

where is a vector of tuning curve gradients at orientation for units in that layer (those with receptive fields at the 243 center of the image), and is the corresponding response covariance matrix. In addition, independently for each unit, 244 we measured its FI as its tuning curve gradient squared divided by response variance at the measured orientation. For FI 245 calculation, units with activity less than 1.0 at the measured orientation were excluded to avoid very low response 246 variance. 247

Experimental design and statistical analyses 248 In Experiment 1, the network was trained on Gabor orientation discrimination under 2880 conditions (12 reference 249 orientation, 4 contrasts, 4 wavelengths, 3 noise levels and 5 angular separations). In Experiment 2, the network was 250 trained on 360 conditions in each of the Gabor and face tasks (12 reference orientations or training-testing data splits, 5 251 dissimilarity levels and 0 to 5 frozen layers). 252

We carried out our analyses on three levels. On the behavioral level, the effects of training and test angle separation on 253 performance was tested using linear regression. On the layer level, the effects of layer number and training angle 254 separation on layer change were tested also using linear regression. To see whether the distribution of learning differed 255 between tasks, we used two-way ANOVA to test whether there was an effect of task on layer change. At the unit level, we 256 tested for significant changes in various tuning curve attributes recorded in the literature, using K-S for distributional 257 changes, two-sample Mann-Whitney U for changes in tuning curve attributes from training, and two-way ANOVA when 258 neurons were grouped according to naïve/trained and their preferred orientations. Finally, to test whether there was a 259 relationship between the network’s initial sensitivity to the trained orientation, we used a regression model described with 260 the results. The significance level for all tests was 0.01. Bonferroni corrections were applied for multiple comparisons. 261

All code was written in Python with Caffe (http://caffe.berkeleyvision.org) for DNN stimulations and statsmodel (Seabold 262 and Perktold, 2010) for statistical analyses. Code and stimulated data are available on request. 263

Results 264

Behavior 265 The network was trained to discriminate whether the target Gabor patch was more clockwise or more counter-clockwise 266 to the reference, repeated in 2880 conditions (12 reference orientation, 4 contrasts, 4 wavelengths, 3 noise levels and 5 267 precisions or angle separations). The performance trajectories grouped by precision are shown in Figure 2A (top row) for 268 the trained and transfer conditions of rotated reference orientations (clockwise by 45 or 90°) and scaled spatial frequencies 269 (halved or doubled). The accuracy measured at the first iteration (analogous to naïve real subjects) indicates the initial 270 performance when only the readout layer changed. Both initial performance and learning rate under the trained condition 271 were superior for the less precise tasks, consistent with findings from the human literature (Ahissar and Hochstein, 1997; 272 Jeter et al., 2009). Percentage correct increased in a similar way to human data on Vernier discrimination (Herzog and 273 Fahlet, 1997) and motion direction discrimination (Liu and Weinshall, 2000). Convergence of performance for the 274 transfer stimuli required more training (note the logarithmic x-axis in Figure 2A) than for the trained stimuli, which may 275 imply that much more training examples are necessary to achieve mastery on the transfer stimuli, consistent with some 276 studies of tactile perceptual learning (Dempsey-Jones et al., 2016). 277

Moreover, we characterized the dynamics of transfer by calculating the transfer index as the ratio of transfer accuracy to 278 the corresponding trained accuracy. As shown in Figure 2A (bottom row), this ratio decreased initially but started to rise 279 slowly for all conditions, and the trajectory for orthogonal transfer (ori+90) under the highest precision were almost flat 280

Page 9: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

7

towards the end of training. Similar reduction in transfer index with increasing training sessions has been demonstrated in 281 human experiments (Jeter et al., 2010). 282

Figure 2B shows the final transfer performance grouped by transfer conditions and training precision. All transfer 283 accuracies were below 1.0, especially for the orientation transfers, indicating learning specificity. A linear regression on 284 transfer accuracy showed a significant positive main effect of log angle separation in all four transfer conditions 285 (p<0.0001; see Extended Data Figure 2-2 for details), implying better transfer performance for less precise training. This 286 result is consistent with experimental data and theoretical prediction that greater specificity happens in finer precision 287 tasks (Ahissar and Hochstein, 1997; Liu, 1999; Liu and Weinshall, 2000). 288

However, the test precisions were the same as used during the respective training conditions and thus varied in intrinsic 289 discriminability, which could have determined the observed transfer pattern. We then tested each trained network on all 290 angle separations to see whether we would reproduce human data (Jeter et al., 2009) where a difference in test precision 291 affects transfer more than training precision. Indeed, Figure 2C shows a strong positive main effect of test separation 292 (p<0.001 and R2>0.35 for all transfer conditions); however, we also found that training separation had a significant effect 293 (p<0.001) in all transfer conditions, and the effect size was the smallest in orthogonal orientation transfer (see Extended 294 Data Figure 2-2 for details). The more substantial transfer to 45° from trained orientations compared to 90° could be due 295 to a larger overlap between the transfer orientation and the trained orientation representation of the network units. Thus, 296 despite the observation of diminishing transfer with increasing precision when the training and test precisions are equal 297 (diagonal lines in Figure 2C), the analyses across all precision combinations predict that transfer is more pronounced from 298 precise to coarse discrimination than vice-versa, although transfer can be very small at the orthogonal orientation. 299

Overall, these behavioral findings are consistent with extant behavioral and modeling results of perceptual learning. 300 Furthermore, the DNN model can simulate a wide number of training conditions and makes predictions regarding the 301 relative performances and learning rates of the trained stimuli compared to that of transfer stimuli. However, it is expected 302 that some details of this DNN’s behavioral results will necessarily differ from that of experimental data. 303

Distribution of learning across layers 304 Learning in the weight space 305 We next examined the timecourse of learning across the layers as a function of precision, calculated using Equation 4 and 306 shown in Figure 3A. Overall, all layers changed faster at the beginning of training in coarse than in precise angle 307 separations, training precise angle separations produced greater overall changes. Of note, while the highest readout layer 308 (layer 6) changed faster than the other layers (Saxe, 2015), this is likely a consequence of zero-initialization in the readout 309 weights. This suggests that when performance was at chance level, due to naivety to the task, information about the 310 stimulus label cannot be passed on to lower layers while the performance is close to chance. Hence, we focus our 311 discussion on layers 1 to 5, whose weights were copied from a pre-trained AlexNet. 312

To characterize learning across layers, we studied when and how much each layer changed during training. To quantify 313 when significant learning happened in each layer, we estimated the iteration at which the gradient of a trajectory reached 314 its peak (peak speed iteration, PSI) which are shown in Figure 3B. In layers 1 to 5, we observed significant negative main 315 effects of log angle separation (β=-46.02, t(14396)=-52.93, p≈0.0, R2=0.40) and layer number (β=-2.07, t(14396)=-13.65, 316 p=3.6×10-42, R2=0.0031) and positive interaction of the two (β=2.93, t(14396)=11.16, p=8.0×10-29, R2=0.0050) on PSI, 317 suggesting that layer change started to asymptote later for lower layers and smaller angle separations. For individual 318 precision conditions, a linear regression analysis showed a significant negative effect of layer number on PSI only in the 319 two most precise tasks (p<0.0001, see Extended Data Figure 3-1 for details). Hence, under high precisions, the order of 320 change across layers is consistent with the Reverse Hierarchy Theory prediction that higher visual areas change before 321 earlier ones (Ahissar and Hochstein, 1997). 322

The final layer change at the end of training is shown in Figure 3C. For a better visual comparison, we calculated the 323 relative layer change under each stimulus condition by taking the ratio of layer change against the change resulted from 324 training at the coarsest angle separation, keeping other stimulus conditions fixed (Figure 3D). The results suggest that 325 changes in lower layers increased by a larger factor than higher ones except layer 5. A linear regression analysis on the 326

Page 10: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

8

changes in layers 1-5 revealed significant negative main effects of log angle separation (β=-0.0060, t=-87.25, p≈0.0, 327 R2=0.34) and layer number (8.6×10-4, p≈0.0, R2=0.092) and positive interaction of the two (0.00010, t=49.23, p≈0.0, 328 R2=0.082). The interaction of angle separation × layer number on layer change is consistent with the prediction that 329 higher-precision training induces more change lower in the hierarchy (Ahissar and Hochstein, 1997). 330

Change of information about orientation 331 While we have considered thus far the changes in the weights of the DNN, there is still a question of how the information 332 about orientation changed across layers, and how this may vary as a function of training precision. We address this by 333 showing in Figure 4 the covariance-weighted linear Fisher information (FI, Equation 6) of the trained and naïve unit 334 population at each layer and each test orientation when trained at each angle separation. The tuning curves used to 335 evaluate FI were obtained from the units with receptive fields at the center of the image (see Materials and Methods). A 336 prominent observation is that FI increased most dramatically at the highest layers under the most precise task and 337 diminished towards lower layers and coarser precisions. Lower layers saw significant FI increase only in the most precise 338 tasks, whereas higher layers increased FI in all precisions. 339

However, the quantitative trend of FI increase was contrary to the layer change where more substantial learning happened 340 in the lower layers. Notably, despite the large change in layer 1 weights (Figure 3C), there was no visible change in FI. 341 The large FI increase in higher layers may not be surprising due to a single hierarchy with readout on the top layer and the 342 accumulation of weight changes from the lower layers. 343

In addition, we observed patterns of FI over orientations. After training at the finest precision, the top layers exhibited 344 significant and substantial increase of FI around the trained orientation; FI fell off around 20° away from trained 345 orientation but remained noticeable until the orthogonal orientation. This could account for the transfer behavior predicted 346 by the network (Figure 2) where learning transferred more substantially if the Gabor stimulus was rotated by 45°, rather 347 than 90° which was a common transfer condition tested in experiments. 348

These data show that the increase of information about orientation over network layers changes as a function of training 349 precision. Later, we will discuss how this information in the pre-trained network may affect learning (Figure 10). 350

Effect of feature complexity on distribution of learning 351 Are the observed layer changes prescribed solely by the network structure and learning rule regardless of the task? To find 352 out whether these patterns were task specific, we simulated learning of a “higher-level” face gender discrimination task 353 and investigated its effect on the distribution of learning in network weights compared with Gabor orientation 354 discrimination. In the face task, difficulty was manipulated by morphing between male and female face images, and the 355 network was trained to discriminate whether the target was more masculine or feminine compared to the reference. Both 356 tasks were repeated under 360 conditions (12 reference orientations for the Gabor task or 12 training-testing data splits for 357 the face task, 5 dissimilarity levels and 0 to 5 frozen layers). By the end of training, the fully plastic network reached test 358 accuracy above 95% for all stimulus conditions. In this section, we assume that learning in stimulus representation 359 happened in the lower 5 layers and ignore the changes in layer 6. 360

Due to the hierarchical representation from earlier to later visual areas, one may hypothesize that learning in lower layers 361 of the DNN would play a more important role in performance for the Gabor task relative to the face task. To quantify the 362 contributions of layers, we measured how much accuracy dropped when more lower layers were frozen (keeping weights 363 fixed) at a particular iteration during training (see Methods and Materials). Results are shown in Figure 5A. Performance 364 in the Gabor task dropped considerably when freezing layer 2 onwards; whereas in the face task, learning was 365 significantly impaired only when freezing all the four layers or more. While this freezing technique is unnatural, and it is 366 possible that compensatory changes occurred that did not reflect properties of learning in the fully plastic network, these 367 results support the hypothesis that the higher layers are more informative for more complex stimulus judgements and the 368 earlier layers are so for more precise and simpler ones. 369

To further test whether the distribution of learning depended on task, we calculated the proportions of layer change in the 370 first five layers by normalizing these changes against their sum (Figure 5B). Since the first layer did not have a large 371 contribution to performance (Figure 5A), and response normalization happened after the first layer, we also compared the 372

Page 11: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

9

layer change proportions after training while freezing the weights in layer 1 at pre-trained values. A two-way ANOVA 373 revealed a significant interaction of layer × task on layer change proportion in both network conditions (p<0.0001, see 374 Extended Data Figure 5-2 for details), suggesting that task indeed changed the distribution of learning over layers. Post-375 hoc analysis showed a significant increase in weight change proportions in layers 4 and 5 and a significant decrease in 376 layer 2 (Mann-Whitney U, threshold p=0.01, Bonferroni-corrected for 5 or 4 layers). Therefore, more weight change 377 happened in lower layers when learning the “low-level” Gabor task and higher layers acquired more change in the “high-378 level” face task, consistent with theories of VPL (Ahissar and Hochstein, 1997; Ahissar and Hochstein, 2004; Watanabe 379 and Sasaki, 2015). 380

One should be careful when interpreting values of layer change defined by Equation 5. For instance, the layer with 381 maximum change varied between layers 1-3 depending on how these changes were calculated, although the general 382 effects of precision and task on layer change were consistent under other weight change measures (Extended Data Figure 383 5-1). In addition, it may be tempting to infer the relationship between layer contribution and change; however, freezing 384 early layers created different interactions between the higher plastic layers, which makes it difficult to compare layer 385 contribution with the layer change obtained by the fully plastic network. 386

Thus, by analyzing the weight changes in the network layers, we have shown that the distribution of learning over the 387 network hierarchy moves towards lower layers for more precise discriminations of simple features, and to higher layers 388 for less precise or more complex stimuli, such as faces. To see to what extent this DNN model can reflect changes in the 389 brain during perceptual learning, we compare activations of individual units in the network with activities of real neurons 390 in the brain recorded by electrophysiology in the following section. 391

Tuning changes of single units 392 Single units in different layers of the DNN model were examined to see whether the changes in these units were similar to 393 those in monkey neurons after VPL. A key target of this investigation was to address computationally some of the 394 significant findings in the literature that had led to diverging interpretations of plasticity within the visual hierarchy. 395 Previous research on DNNs found that the representational geometries of layers 2 and 3 (but not layer 1) in this network 396 are analogous to those in human V1-3 (Khaligh-Razavi and Kriegeskorte, 2014), so we focused our analyses on layers 2 397 and higher. 398

We compared the network units with V1-2 and V4 neurons recorded in four electrophysiological studies; animals in these 399 studies were trained to discriminate orientations of Gabor patches. Schoups et al. (2001) discovered an increase in the 400 tuning curve slope at the trained orientation for units tuned to 12-20° away from trained orientation. The same group 401 (Raiguel et al., 2006) later found in V4 a similar change along with other effects of training. These studies used a single-402 interval 2AFC training paradigm with implicit reference that was never shown. On the other hand, Ghose et al. (2012) 403 used a two-interval 2AFC training paradigm in which an irrelevant feature of the stimulus (spatial frequency) varied 404 between two values through training. Contrary to Schoups et al. (2001), they did not find significant changes in V1-2 405 regarding orientation tuning (except one case), but the final discrimination thresholds reached by the subjects were higher. 406 This group later revealed several changes in V4 using the same training paradigm (Yang and Maunsell, 2004). We 407 hypothesized that the differences between these studies may be simply explained by differences in precision during 408 training. To test this, we trained the network on a common task, the 2I-2AFC Gabor discrimination paradigm, and tested 409 whether changing training precision, holding everything else constant, was sufficient to reconcile the gross differences 410 observed between these studies. 411

Overall, it appears that V1-2 only changed when the discrimination threshold was small (0.5-1.2° by Schoups et al., 412 smaller than 3.3-7.3° by Ghose et al.), and V4 changed in both studies though the discrimination thresholds were similar 413 to each other (1.9-3.0° by Raiguel et al. and 1.9-5.4° by Yang and Maunsell). Given the pattern of change in layer Fisher 414 information demonstrated earlier (Figure 4), we hypothesized that 415

a) the contradictory results in V1-2 (corresponding to lower layers in the network) were due to the mismatch of the 416 final thresholds reached by the subjects in the two V1-2 studies, and 417

Page 12: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

10

b) the change in V4 (corresponding to higher layers in the network) should persist from fine to coarse precisions in 418 which V1-2 did not change. 419

To test the first hypothesis using the DNN model, we roughly matched the angle separations trained on the network with 420 the discrimination thresholds reached by the monkeys, 1.0° for high and 5.0° for low precisions, and compared network 421 units in layer 2 (layer 3 in Extended Data) with real V1-2 neurons. We then compared the tuning attributes of units in 422 layer 5 (layer 4 in Extended Data) with those of V4 neurons to test our second hypothesis. These choices were mainly 423 motivated by previous research on comparing this particular network with the visual brain areas (Khaligh-Razavi and 424 Kriegeskorte, 2014; Guclu and van Gerven, 2015; Eickenberg et al., 2017). It should be noted that we used the same 425 spatial frequency for the reference and target (more similar to Schoups et al. (2001) than to Ghose et al. (2002)); thus, 426 while we found that modelling the differences in thresholds is sufficient to account for many differences in physiological 427 findings between the four studies, it is likely that other task and stimulus differences also contributed to the different 428 profiles of learning. 429

Tuning curves were obtained from the units with receptive fields at the center of the image and are shown in Figure 6A 430 for layers 2 and 5. Many of the layer-2 units showed bell-shaped tuning curves with clear orientation preferences, 431 mirroring the reported similarity between these two layers with human early visual. In addition, there were also intriguing 432 tuning curves that showed more than one tuning peaks, which may not be physiologically plausible. Tuning curves in 433 layer 5 were harder to interpret with some units showing clear orientation tuning and the rest likely tuned to features other 434 than Gabor patches. 435

Lower layers trained under high precision (Schoups et al., 2001) 436 We first investigated whether this DNN model could reproduce findings of Schoups et al. (2001) on V1 neurons, who 437 found that monkeys obtained very small thresholds (around 1°) and V1 neurons showed a change in the slope of tuning 438 curve at trained orientation for neurons tuned between 12-20° from trained orientation (Figure 6B). Similar to their 439 procedure, we took the orientation of the maximum response of a unit as its preferred orientation and grouped all units by 440 preferred orientation relative to trained orientation. The slope of the tuning curve at trained orientation was evaluated after 441 normalization by maximum activation (as done for neural data in Schoups et al. (2011)). These results for layer-2 units 442 (Figure 6C) showed a similar slope increase for units tuned away from trained orientation, overlapping but broader than 443 the range found in V1 neurons, when trained under high precision but not low precision. 444

Despite a change in tuning slope, Schoups et al. (2001) found that the preferred orientations of neurons were evenly 445 distributed over all orientations before and after training (Figure 6D). This was also the case for the network units which 446 showed no significant difference between trained and naïve distributions of preferred orientation in either of the precisions 447 (Figure 6E, p>0.9, K-S, see Extended Data Figure 6-2 for details). 448

Overall, these data from lower-layer units demonstrated an impressive qualitative similarity to data from Schoups et al. 449 (2001) when the network was trained on high precision. We next look at the changes in low precision training. 450

Lower areas trained under low precision (Ghose et al., 2002) 451 Contrary to Schoups et al. (2001), Ghose et al. (2001) found very little change in V1-2 neurons after training. The tuning 452 amplitude of V1 and V2 neurons tuned around the trained orientation did not differ significantly from other neurons after 453 training (Figure 7A). However, the monkeys trained by Ghose et al. (2002) achieved relatively poorer discrimination 454 thresholds (around 5°), and when we modelled this as low precision training, we also found no significant effect of 455 preferred orientation on tuning amplitude at layers 2-3 (Figure 7B, p=0.30, Mann-Whitney U, see Extended Data Figure 456 7-2 for details). In addition, Ghose et al. (2002) found no significant change in the variance ratio for V1-2 neurons (Figure 457 7C), which was replicated by our network units at both precisions (Figure 7D, p>0.6, Mann-Whitney U, see Extended 458 Data Figure 7-4 for details). 459

Ghose et al. (2002) did observe a decrease in the number of neurons tuned to the trained orientation in V1 but not in V2 460 (not shown), contrary to Schoups et al. (2001). In our model, no such change was found in either high or low precision 461 (Figure 6D); hence, here the model did not agree with the data. 462

Page 13: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

11

Thus, under low precision, we did not find significant changes in tuning attributes in lower layers, which was also the case 463 for (Ghose et al., 2002) except for the preferred orientation distribution in V1. Together with the comparisons under high 464 precision in the previous section, our first hypothesis was supported by our simulations. The key in replicating these data 465 is the observation that the precision of training has a profound effect on the distribution of learning across layers. By 466 accounting for the different orientation thresholds found across studies and labs, the DNN model can well address the 467 different observations. In addition, the partial specificity of the network trained under low precisions (Figure 2B) did not 468 require orientation-specific changes in lower layers, in line with previous models and data (Petrov et al., 2005; 469 Sotiropoulos et al., 2011; Dosher et al., 2013). 470

Higher layers compared to data of Raiguel et al. (2006) 471 While neurons in primary visual cortex showed plasticity that was largely limited to high-precision conditions, neurons in 472 V4 generally showed more susceptibility to VPL (Yang and Maunsell, 2004; Raiguel et al., 2006). We hypothesized that 473 changes in V4 neurons should happen in both low and high precisions, and tested this hypothesis by comparing the tuning 474 attributes of units in layer 5 (layer 4 in Extended Data) with recordings in those studies. 475

We first compared the network units with the result of Raiguel et al. (2006) that, similar to V1 (Figure 6B), V4 neurons 476 tuned 22-67° away from trained orientation significantly increased their slopes at trained orientation (Figure 8A). This 477 effect was replicated in the model units (Figure 8B) which, unlike layer 2 (Figure 6C), showed significant increase in 478 tuning slope not only in high but also in low precisions (p<0.0002, two-way ANOVA, see Extended Data Figure 8-1 for 479 details), although these units were tuned much closer to trained orientation than V4 neurons. 480

Furthermore, in monkey V4, the distribution of preferred orientation became non-uniform after training (Figure 8C). In 481 our model, this distribution also became significantly different from a uniform distribution as revealed by a K-S test in 482 both conditions (p<0.003, K-S, see Extended Data Figure 8-2 for details). This is in contrast to layers 2 which only show 483 such a change for high precision (Figure 6E). There was also a substantial increase in the number of neurons tuned very 484 close to trained orientation. 485

The strength of orientation tuning, measured by the selectivity index defined in Equation 5, was found to increase after 486 training in V4 (Figure 8E), and similar results were found in the model (Figure 8F) where SI increased significantly when 487 trained under high precision (p<0.0001) but not under low precision (p=0.10, Mann-Whitney U, see Extended Data Figure 488 8-3 for details). While these results suggest that higher-layer units became sharper after training when the precision was 489 high, the shape of this distribution in layers 4 and 5 did not match real V4 neurons. 490

Raiguel et al. (2006) also discovered that training reduced response variability at the preferred orientation quantified by 491 the normalized variance (Figure 8G). We found the same in our model units (Figure 8H) at both precisions (p<0.001, 492 Mann-Whitney U, see Extended Data Figure 8-4 for details), suggesting that noisy units in the lower layers might be 493 rejected by higher ones (Dosher and Lu, 1998). 494

Higher layers compared to data of Yang & Maunsell (2004) 495 Finally, we address whether the network units also replicated the findings of Yang & Maunsell (2004) where monkeys 496 achieved a threshold comparable with (Raiguel et al., 2006). Overall, in contrast to the V1 study from the same group 497 (Ghose et al., 2002), Yang and Maunsell (2004) found many tuning changes in V4. First, tuning amplitude of V4 neurons 498 increased significantly after training (Figure 9A), and the same was observed in the model under both precisions 499 (p<0.0001, Mann-Whitney U, see Extended Data Figure 9-2 for details). Second, V4 neurons significantly lower their best 500 discriminability (Figure 9C) after training (lower implies better discriminability), suggesting that finer orientation 501 differences could be detected. Units in the model (Figure 9C) reproduced the same change in both precision levels 502 (p<0.0005, Mann-Whitney U, see Extended Data Figure 9-4 for details). 503

Yang & Maunsell (2004) went on to show that these changes were not simply a result of the scaling, but rather the 504 narrowing, of tuning curves (Figure 9E). In layer 5 of the model, the tuning widths of naïve units were already smaller 505 than trained V4 neurons; nonetheless, we found that, under high precision, the mean activation of layer-5 units (Figure 9F) 506 in the non-preferred orientation range (45° away from preferred orientation) was significantly more reduced than that in 507

Page 14: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

12

the preferred orientation range (within 45° of preferred orientation, p<0.0001, two-way ANOVA, see Extended Data 508 Figure 9-6 for details). 509

More importantly, the mismatch in the tuning width between network units and real neurons could explain several 510 quantitative discrepancies between model units and V4 neurons seen previously, including the different group of units that 511 increased the tuning slope (Figure 8A and B), the sharp peaks in preferred orientation distributions (Figure 8C and D), and 512 the higher selectivity index distributions than real neurons (Figure 8E and F). The existence of these narrow tuning curves 513 and its consequences may require more accurate physiological measurements to verify. 514

Therefore, multiple changes found in layer 5 at low precision provide strong evidence supporting our second hypothesis. 515 To conclude the comparisons with physiological studies, we find that the DNN model replicates a number of the single 516 cell results found in extant studies of primate visual cortex. In general, it appears that the network units increased their 517 responses at orientations close to the trained orientation, providing more informative response gradients essential for 518 performance. Changes are more substantial in higher layers through feedforward connections, resulting in sharper tuning 519 curves, larger tuning amplitudes and a significant accumulation of tuning preference close to the trained orientation. Also, 520 noisy neurons in lower layers may be rejected by higher ones after training, reducing response variability (Dosher and Lu, 521 1998). Nonetheless, it is important to note the quantitative differences in data between the DNN model and primates as 522 noted in the comparisons above. 523

Linking initial sensitivities to weight changes 524 A key question to understanding learning in the DNN is the extent to which learning depends on the initial conditions of 525 the network. Here, we focus on the first five layers and explore whether there is relationship between initial sensitivity to 526 the trained orientation and weight changes. 527

To test whether a larger layer sensitivity to the trained stimulus may give rise to more learning in the weights of this layer, 528 we correlated the untrained layer FI with layer change measured by Equation 4. Although the network’s initial state was 529 the same for all simulations, the reference orientations varied in 12 values to which the network were differentially 530 sensitive. We used the following regression model to test the contributions of various factors 531

( 7 )

where is the linear coefficient for FI, and are the main effects of layer and angle separation, respectively, each 532 with 5 categorical levels, and and are the interactions of layer × FI and angle separation × FI, respectively. 533 ANOVA on this model showed significant main effects of and (p<0.0001) which accounted for 21% and 46% of 534 variance, but the three effects involving FI were insignificant (p>0.1, see Extended Data Figure 10-2 for details). This 535 means that the weights in a layer did not change more when it was more sensitive to the trained reference orientation. 536

To test whether a larger unit sensitivity to the trained stimulus may give rise to more learning in the weights of the unit, 537 we used the untrained fisher information of the units (tuning gradient squared divided by variance at the trained 538 orientation) as a proxy for sensitivity, and correlated this with their weight changes. We show in Figure 10B the 539 relationship between these two quantities for each training condition after rescaling. The correlations were generally small 540 except at layer 5, and despite the positive correlation (consistent to that observed in Raiguel et al. (2006)), there was a 541 tendency for less sensitive units to also change substantially. Neurons in layer 3 showed the lowest correlation compared 542 with those in other layers even though this layer had the highest pre-train FI. The same regression analyses above revealed 543 that all effects were significant (p<0.0001, see Extended Data Figure 10-3 for details), but layer and angle separation 544 explained 2.6% and 12.4% of the variance in weight change, whereas the effects involving FI ( , and ) 545 together explained 1.1%. In addition, we also saw that the distribution of FI became more spread-out after training (Figure 546 10C), particularly for higher layers, suggesting that training did not improve FI for all units equally. Under the finest 547 precision, there was a positive correlation between the initial untrained FI and change in FI in the five layers (0.1-0.5, 548 p<0.0001). 549

Page 15: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

13

Therefore, although the network’s initial sensitivity to the trained orientation might influence the magnitude of learning, 550 its effect size was less considerable compared to layer and training precision. On the layer level, a higher initial sensitive 551 lowered the amount of learning; and on the unit level, the effect of initial sensitivity on learning was mixed. 552

Discussion 553 We find that the DNN model studied here is a highly suitable model to investigate visual perceptual learning. On the 554 behavioral level, when the network was trained on Gabor orientation discrimination, the network’s initial performance, 555 learning rate and degree of transfer to other reference orientations or spatial frequencies depended on the precision of the 556 stimuli in a similar manner as found in behavioral and theoretical accounts of perceptual learning. We found slower 557 learning with less transfer under finer training precision as predicted by Reverse Hierarchy Theory (Ahissar and 558 Hochstein, 1997; Ahissar and Hochstein, 2004); however, the model also suggests that test precision had a major 559 influence on the transfer performance and was able to account for greater transfer from precise to coarse stimuli (Jeter et 560 al., 2009). This model makes the novel prediction that high-precision training transfers more broadly to untrained and 561 coarse orientation discriminations than low-precision training. On the layer level, increasing the task precision resulted in 562 slower but more substantial changes in lower layers. In addition, learning to discriminate more complex features (e.g. face 563 gender discrimination) resulted in relatively greater changes in higher layers of the network. On the unit level, only high-564 precision tasks significantly changed tuning curve attributes at lower layers, whereas units in the higher layers showed 565 more robust changes across precisions. Various changes found in the network units mirrored many electrophysiological 566 findings of neurons in monkey visual cortex. Overall, this DNN model, while not originally designed for VPL, arrived at 567 impressively convergent solutions to behavioral, layer-level and unit-level effects of VPL found in the extant literature of 568 theories and experiments. 569

The present findings help reconcile disparate observations in the literature regarding the plasticity in early visual cortex. 570 While Schoups et al. (2001) found changes in orientation tuning curves of neurons in primary visual cortex, these results 571 were not replicated by Ghose et al. (2002). The DNN model provides a parsimonious explanation for these results, 572 accounting for the discrepancy as related to the different discrimination thresholds reached by the subjects between those 573 experiments. In Schoups et al. (2001), the subjects reached lower thresholds than those in Ghose et al. (2002), which was 574 sufficient to move plasticity down to the lower layers of the DNN model. Furthermore, through a number of observations 575 on higher-layer units that were similar to V4 neurons, the model verified our hypothesis that these neurons are more 576 susceptible to changes after VPL even under low precisions. 577

Compared to a shallower model that would have no problem learning the tasks, a deeper network has the advantage of 578 demonstrating the distribution of learning over finer levels of hierarchy. Also, it serves as an example where recurrent or 579 feedback processes may not be necessary to capture the distribution and order of learning over layers, as lower layers 580 changed before higher layers in the absence of inhomogeneous learning rate or attentional mechanisms. Moreover, despite 581 its biological implausibility, weight sharing in the convolutional layers did not result in substantial unwanted transfer over 582 reference orientation or spatial frequency in this study; location specificity was also demonstrated in a shallow 583 convolutional network (Cohen and Weinshall, 2017), although breaking this weight sharing may still be necessary to 584 better interpret learning. 585

Despite the striking resemblance between the DNN model output and primate data, it is worth noting that a number of 586 choices we made regarding training paradigm, noise injection, learning rule and learning rate may have significant 587 consequences to the results reported here. First, the network was trained on fixed differences which differed from using 588 staircases as done in many VPL studies; the 2I-2AFC procedure, which avoided explicit definition of label in transfer, 589 may produce different learning outcomes compared to the 2AFC without a reference used by Schoups et al. (2001) or the 590 varied spatial frequencies used by Ghose et al. (2002). Second, there was no source of internal noise in the middle layers 591 to generate behavioral variability (Acerbi et al., 2014), and the readout weights were zero-initialized to minimize learning 592 variability in contrast to previous network models that used random initialization to simulate multiple subjects 593 (Sotiropoulos et al., 2011; Talluri et al., 2015). Third, our learning rule SGD does not compare favorably with Hebbian-594 like learning methods (Sotiropoulos et al., 2011; Dosher et al., 2013; Talluri et al., 2015) that are more biologically 595 plausible, although more biologically plausible versions have been proposed (Lillicrap et al., 2016; Scellier and Bengio, 596

Page 16: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

14

2017). Other studies suggested that small differences in training paradigms, including precision as shown in the present 597 data, can have a significant impact on learning and plasticity (Hung and Seitz, 2014), and it will be important target for 598 future studies to research the contributions of the numerous other differences between these studies. 599

Another issue regards the distinction between representation learning in the lower layers and task or decision learning in 600 the readout layer. This DNN can perform very well even if only the final layer is plastic, in which case both forms of 601 learning are mixed into the same layer. In our simulations, the use of a small learning rate was necessary to ensure 602 learning stability on precise tasks, but this also had the effect of distributing more change to the lower layers. Other 603 schemes can be used to control learning between layers, such as pre-training on an easier task or readout weight 604 regularization. Direct connections from lower layers to the readout layer are also possible given a reasonable weight 605 initialization. Future research will be necessary to examine the consequence of such alternative schemes on the predictions 606 regarding distribution of learning. 607

A long-standing topic in research of neural coding regards the efficiency of sparse representations in visual cortex 608 (Barlow, 1961; Olshausen and Field, 1997). This raises a question of whether perceptual learning “raises all boats” or 609 “makes the rich get richer” and mostly the best tuned neurons change the most. Some support for the latter possibility is 610 found in physiological studies where neural responses changed primarily in the most informative neurons. The present 611 DNN model appears insufficient to well-address sparsity in learning. While we found (Figure 10) small positive 612 correlations between initial Fisher information and weights changes from training, consistent with the notion of “the rich 613 get richer”, there were also substantial changes in many of the insensitive units across layers. This may be due to the fact 614 that the network was only trained on one task and was not consistently performing the many visual tasks involving natural 615 stimuli that humans and animals must perform on a daily basis. Hence, to better address learning sparsity, a network may 616 need to be trained simultaneously on a number of tasks. 617

Furthermore, one must be cautious in inferring homologies between layers of DNNs and areas in the brain. The similarity 618 between the network and visual areas depends on the layer parameters (such as number of filters, receptive field size, etc.) 619 in a subtle manner (Pinto et al., 2009; Yamins and DiCarlo, 2016). It is also unknown how much our analyses on the 620 changes in the weights (instead of unit activity) can inform us about the synaptic changes caused by VPL. Comparing 621 results across different DNNs may help us understand which results are robust against details of model architecture. 622

The simulations shown in the present manuscript just touched the surface of the vast VPL literature. While beyond the 623 scope of the present study, future modeling targets can be considered in pursuit of many of perceptual learning 624 phenomena. For example, DNNs may be used to replicate other psychophysical phenomena, including disruption (Seitz et 625 al., 2005), roving (Zhang et al., 2008; Tartaglia et al., 2009; Hussain et al., 2012), double training (Xiao et al., 2008; 626 Zhang et al., 2010) and the effects of attention (Ahissar and Hochstein, 1993; Byers and Serences, 2012; Bays et al., 2015; 627 Donovan et al., 2015) and adaptation (Harris et al., 2012). Moreover, small variations in training procedures can lead to 628 dramatic changes in learning outcome (Hung and Seitz, 2014); therefore, it is important for future simulations to 629 understand how such details may impact learning in DNNs. 630

In addition, DNNs may provide a straightforward way to model the ‘when’ and ‘where’ aspects of VPL that would be 631 otherwise difficult to test experimentally on subjects. As discussed in previous reviews (Seitz, 2017; Watanabe and 632 Sasaki, 2015), it is likely that VPL involves more areas than the two or three-layer models of pattern-matching 633 representation and nonlinear readout that typify the field (Dosher and Lu, 2017). The distribution and timecourse of 634 plasticity could be further examined in other tasks, using other DNNs or layer change measures. 635

In conclusion, we find that deep neural networks provide an appropriate framework to study VPL. An advantage of DNNs 636 is that they can be flexibly adapted to different tasks, stimulus types and training paradigms. Also, layer- and unit-specific 637 changes resulting from learning can be examined and related to fMRI and electrophysiological data. While some caution 638 is needed in interpreting the relationship between these models and biological systems, the striking similarities with many 639 studies suggests that DNNs may provide solutions to learning and representation problems faced by biological systems 640 and as such may be useful in generating testable predictions to constrain and guide perceptual learning research within 641 living systems. 642

Page 17: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

15

Reference 643 Acerbi L, Vijayakumar S, Wolpert DM (2014) On the Origins of Suboptimality in Human Probabilistic Inference. PLoS 644

Computational Biology 10:e1003661. 645

Ahissar M, Hochstein S (1993) Attentional control of early perceptual learning. Proceedings of the National Academy of 646 Sciences 90:5718–5722. 647

Ahissar M, Hochstein S (1997) Task difficulty and the specificity of perceptual learning. Nature 387:401–406. 648

Ahissar M, Hochstein S (2004) The reverse hierarchy theory of visual perceptual learning. Trends in Cognitive 649 Sciences 8:457–464. 650

Ball K, Sekuler R (1982) A specific and enduring improvement in visual motion discrimination. Science 218:697–8. 651

Barlow HB (1961) Possible Principles Underlying the Transformations of Sensory Messages In Sensory Communication, 652 pp. 216–234. The MIT Press. 653

Bays BC, Visscher KM, Dantec CCL, Seitz AR (2015) Alpha-band EEG activity in perceptual learning. Journal of 654 Vision 15:1–12. 655

Bejjanki VR, Beck JM, Lu ZL, Pouget A (2011) Perceptual learning as improved probabilistic inference in early sensory 656 areas. Nature neuroscience 14:642–648. 657

Blasdel GG (1992) Orientation selectivity, preference, and continuity in monkey striate cortex. The Journal of 658 Neuroscience 12:3139–3161. 659

Byers A, Serences JT (2012) Exploring the relationship between perceptual learning and top-down attentional control. 660 Vision Research 74:30–39. 661

Cichy RM, Khosla A, Pantazis D, Torralba A, Oliva A (2016) Comparison of deep neural networks to spatio-temporal 662 cortical dynamics of human visual object recognition reveals hierarchical correspondence. Scientific Reports 6:27755. 663

Cohen G, Weinshall D (2017) Hidden Layers in Perceptual Learning In The IEEE Conference on Computer Vision and 664 Pattern Recognition (CVPR), pp. 4554–62, Honolulu, HI. 665

Crist RE, Kapadia MK, Westheimer G, Gilbert CD (1997) Perceptual learning of spatial localization: specificity for 666 orientation, position, and context. Journal of Neurophysiology 78:2889–2894. 667

Dempsey-Jones H, Harrar V, Oliver J, Johansen-Berg H, Spence C, Makin TR (2016) Transfer of tactile perceptual 668 learning to untrained neighboring fingers reflects natural use relationships. Journal of Neurophysiology 115:1088–669 1097. 670

Donovan I, Szpiro S, Carrasco M (2015) Exogenous attention facilitates location transfer of perceptual learning. Journal 671 of Vision 15:11. 672

Dosher BA, Lu ZL (1998) Perceptual learning reflects external noise filtering and internal noise reduction through 673 channel reweighting. Proceedings of the National Academy of Sciences 95:13988–13993. 674

Dosher BA, Jeter P, Liu J, Lu ZL (2013) An integrated reweighting theory of perceptual learning. Proceedings of the 675 National Academy of Sciences 110:13678–13683. 676

Eickenberg M, Gramfort A, Varoquaux G, Thirion B (2017) Seeing it all: Convolutional network layers map the function 677 of the human visual system. NeuroImage 152:184–194. 678

Fahle M (2004) Perceptual learning: A case for early selection. Journal of Vision 4:4. 679

Fiorentini A, Berardi N (1980) Perceptual learning specific for orientation and spatial frequency. Nature 287:43–44. 680

Page 18: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

16

Fiorentini A, Berardi N (1981) Learning in grating waveform discrimination: Specificity for orientation and spatial 681 frequency. Vision Research 21:1149–1158. 682

Ghose GM, Yang T, Maunsell JHR (2002) Physiological Correlates of Perceptual Learning in Monkey V1 and V2. J 683 Neurophysiol 87:1867–1888. 684

Guclu U, van Gerven MAJ (2015) Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations 685 across the Ventral Stream. Journal of Neuroscience 35:10005–10014. 686

Harris H, Gliksberg M, Sagi D (2012) Generalized Perceptual Learning in the Absence of Sensory Adaptation. Current 687 Biology 22:1813–1817. 688

Herzog MH, Aberg KC, Frémaux N, Gerstner W, Sprekeler H (2012) Perceptual learning, roving and the unsupervised 689 bias. Vision Research 61:95–99. 690

Herzog MH, Fahlet M (1997) The Role of Feedback in Learning a Vernier Discrimination Task. Vision 691 Research 37:2133–2141. 692

Hubel DH, Wiesel TN (1968) Receptive fields and functional architecture of monkey striate cortex. The Journal of 693 Physiology 195:215–243. 694

Hung SC, Seitz AR (2014) Prolonged Training at Threshold Promotes Robust Retinotopic Specificity in Perceptual 695 Learning. Journal of Neuroscience 34:8423–8431. 696

Hussain Z, Bennett PJ, Sekuler AB (2012) Versatile perceptual learning of textures after variable exposures. Vision 697 Research 61:89–94. 698

Janssen D, Schuett H, Wichmann F (2016) Some observations on the psychophysics of Deep Neural Networks. Journal of 699 Vision 16:963–963. 700

Jehee JFM, Ling S, Swisher JD, van Bergen RS, Tong F (2012) Perceptual Learning Selectively Refines Orientation 701 Representations in Early Visual Cortex. Journal of Neuroscience 32:16747–16753. 702

Jeter PE, Dosher BA, Liu SH, Lu ZL (2010) Specificity of perceptual learning increases with increased training. Vision 703 Research 50:1928–1940. 704

Jeter PE, Dosher BA, Petrov A, Lu ZL (2009) Task precision at transfer determines specificity of perceptual learning. 705 Journal of Vision 9:1. 706

Kanitscheider I, Coen-Cagli R, Kohn A, Pouget A (2015) Measuring Fisher Information Accurately in Correlated Neural 707 Populations. PLoS Computational Biology 11:e1004218. 708

Karni A, Sagi D (1991) Where practice makes perfect in texture discrimination: evidence for primary visual cortex 709 plasticity. Proceedings of the National Academy of Sciences 88:4966–70. 710

Khaligh-Razavi SM, Kriegeskorte N (2014) Deep Supervised, but Not Unsupervised, Models May Explain IT Cortical 711 Representation. PLoS Computational Biology 10:e1003915. 712

Kheradpisheh SR, Ghodrati M, Ganjtabesh M, Masquelier T (2016) Deep Networks Can Resemble Human Feed-forward 713 Vision in Invariant Object Recognition. Scientific Reports 6:32672. 714

Kriegeskorte N (2015) Deep Neural Networks: A New Framework for Modeling Biological Vision and Brain Information 715 Processing. Annual Review of Vision Science 1:417–446. 716

Krizhevsky A, Sutskever I, Hinton GE (2012) ImageNet classification with deep convolutional neural networks In 717 Advances in Neural Information Processing Systems 25, pp. 1097–1105. 718

Page 19: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

17

Lillicrap TP, Cownden D, Tweed DB, Akerman CJ (2016) Random synaptic feedback weights support error 719 backpropagation for deep learning. Nature Communications 7:13276. 720

Liu Z (1999) Perceptual learning in motion discrimination that generalizes across motion directions. Proceedings of the 721 National Academy of Sciences 96:14085–14087. 722

Liu Z, Weinshall D (2000) Mechanisms of generalization in perceptual learning. Vision Research 40:97–109. 723

Maniglia M, Seitz AR (2018) Towards a whole brain model of Perceptual Learning. Current Opinion in Behavioral 724 Sciences 20:47–55. 725

Mastropasqua T, Galliussi J, Pascucci D, Turatto M (2015) Location transfer of perceptual learning: Passive stimulation 726 and double training. Vision Research 108:93–102. 727

Olshausen BA, Field DJ (1997) Sparse coding with an overcomplete basis set: A strategy employed by V1? Vision 728 Research 37:3311–3325. 729

Petrov AA, Dosher BA, Lu ZL (2005) The Dynamics of Perceptual Learning: An Incremental Reweighting Model. 730 Psychological Review 112:715–743. 731

Pinto N, Doukhan D, DiCarlo JJ, Cox DD (2009) A high-throughput screening approach to discovering good forms of 732 biologically inspired visual representation. PLoS Computational Biology 5. 733

Raiguel S, Vogels R, Mysore SG, Orban Ga (2006) Learning to see the difference specifically alters the most informative 734 V4 neurons. The Journal of neuroscience 26:6589–6602. 735

Robert G, Janssen D, Schütt H, Bethge M, Wichmann F (2017) Of Human Observers and Deep Neural Networks: A 736 Detailed Psychophysical Comparison In Vision Sciences Society. 737

Rumelhart DE, Hinton GE, Williams RJ (1986) Learning representations by back-propagating errors. Nature 323:533–738 536. 739

Sagi D (2011) Perceptual learning in Vision Research. Vision Research 51:1552–1566. 740

Saxe A (2015) The Effect of Pooling in a Deep Learning Model of Perceptual Learning In Cosyne Abstracts, Salt Lake 741 City USA. 742

Scellier B, Bengio Y (2017) Equilibrium Propagation: Bridging the Gap between Energy-Based Models and 743 Backpropagation. Frontiers in Computational Neuroscience 11. 744

Schoups AA, Vogels R, Orban GA (1995) Human perceptual learning in identifying the oblique orientation: retinotopy, 745 orientation specificity and monocularity. The Journal of Physiology 483:797–810. 746

Schoups A, Vogels R, Qian N, Orban G (2001) Practising orientation identification improves orientation coding in V1 747 neurons. Nature 412:549–553. 748

Seabold S, Perktold J (2010) Statsmodels: econometric and statistical modeling with Python. Proceedings of the 9th 749 Python in Science Conference pp. 57–61. 750

Seitz AR, Yamagishi N, Werner B, Goda N, Kawato M, Watanabe T (2005) Task-specific disruption of perceptual 751 learning. Proceedings of the National Academy of Sciences 102:14895–14900. 752

Seriès P, Latham PE, Pouget A (2004) Tuning curve sharpening for orientation selectivity: coding efficiency and the 753 impact of correlations. Nature Neuroscience 7:1129–1135. 754

Sotiropoulos G, Seitz AR, Seriès P (2011) Perceptual learning in visual hyperacuity: A reweighting model. Vision 755 Research 51:585–599. 756

Page 20: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

18

Talluri BC, Hung SC, Seitz AR, Seriès P (2015) Confidence-based integrated reweighting model of task-difficulty 757 explains location-based specificity in perceptual learning. Journal of vision 15:17. 758

Tartaglia EM, Aberg KC, Herzog MH (2009) Perceptual learning and roving: Stimulus types and overlapping neural 759 populations. Vision Research 49:1420–1427. 760

Tootell RB, Hadjikhani NK, Vanduffel W, Liu AK, Mendola JD, Sereno MI, Dale AM (1998) Functional analysis of 761 primary visual cortex (V1) in humans. Proceedings of the National Academy of Sciences 95:811–7. 762

Wang R, Wang J, Zhang JY, Xie XY, Yang YX, Luo SH, Yu C, Li W (2016) Perceptual Learning at a Conceptual Level. 763 Journal of Neuroscience 36:2238–2246. 764

Watanabe T, Sasaki Y (2015) Perceptual Learning: Toward a Comprehensive Theory. Annual Review of 765 Psychology 66:197–221. 766

Xiao LQ, Zhang JY, Wang R, Klein SA, Levi DM, Yu C (2008) Complete Transfer of Perceptual Learning across Retinal 767 Locations Enabled by Double Training. Current Biology 18:1922–1926. 768

Yamins DLK, DiCarlo JJ (2016) Using goal-driven deep learning models to understand sensory cortex. Nature 769 Neuroscience 19:356–365. 770

Yamins DLK, Hong H, Cadieu CF, Solomon EA, Seibert D, DiCarlo JJ (2014) Performance-optimized hierarchical 771 models predict neural responses in higher visual cortex. Proceedings of the National Academy of Sciences 111:8619–772 8624. 773

Yang T, Maunsell JH (2004) The effect of perceptual learning on neuronal responses in monkey visual area V4. 774 J.Neurosci. 24:1617–1626. 775

Yu Q, Zhang P, Qiu J, Fang F (2016) Perceptual Learning of Contrast Detection in the Human Lateral Geniculate 776 Nucleus. Current Biology 26:3176–3182. 777

Zafeiriou S, Hansen M, Atkinson G, Argyriou V, Petrou M, Smith M, Smith L (2011) The Photoface database In CVPR 778 2011 Workshops, pp. 132–139. IEEE. 779

Zhang JY, Kuai SG, Xiao LQ, Klein SA, Levi DM, Yu C (2008) Stimulus Coding Rules for Perceptual Learning. PLoS 780 Biology 6:e197. 781

Zhang T, Xiao LQ, Klein SA, Levi DM, Yu C (2010) Decoupling location specificity from perceptual learning of 782 orientation discrimination. Vision Research 50:368–374. 783

Zhaoping L, Herzog M, Dayan P (2003) Nonlinear ideal observation and recurrent preprocessing in perceptual learning. 784 Network: Computation in Neural Systems 14:233–247. 785

786

Page 21: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

19

Figures 787

Figure 1. Model structure and stimulus examples. A, Model architecture and a pair of Gabor stimuli. The network consists of two identical processing streams producing scalar representations, one for the reference hr and the other for the target ht, and the difference of the two is used to obtain a probability of the target being more clockwise p(CW) through the sigmoid function. Darker colors indicate higher layers. Layers 1 to 5 consists of multiple units arranged in retinotopic order (rectangles) and convolutional weights (triple triangles, not indicative of the actual filter sizes or counts) to their previous layers or image, and layer 6 has a single unit (dark orange circles) fully connected to layer-5 units (single triangles). Weights at each layer are shared between the two streams so that the representations of the two images are generated by the same parameters. Feedback is provided at the last sigmoidal unit. The Gabor examples have the following parameters: reference orientation: 30°, target orientation: 20°, contrast: 50%, wavelength: 20 pixels, noise standard deviation: 5. B, Face examples morphed from three males and three females. The reference (ref) is paired with either a more masculine (M) or more feminine(F) target image, both morphed from the same two originals (M5 and F5) with the reference being the halfway morph. The number following the label indicates dissimilarity with the reference.

788

Page 22: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

20

789

Figure 2. Performance of the model when trained under various angle separations and tested at the trained and transfer conditions: reference orientation rotated clockwise by 45 (ori + 45) or 90° (ori + 90) and spatial frequency halved (SF × 0.5) or doubled (SF × 2.0). See Extended Data Figure 2-2 for statistical details. A, Accuracy (top row) and transfer index (bottom row) trajectories against training iterations. Darker blue indicates finer precision. Error envelope represents 1 sem and is hardly visible. See Extended Data Figure 2-1 for accuracies plotted as mean±std during the first 50 iterations. B, Final performance under the four transfer conditions. There was a significant positive main effect of log angle separation in each of the four transfer conditions (p<0.0001, R2>0.2 in all conditions), indicating greater transfer for coarser precisions. Error bar indicates 1 std. C, Final mean accuracies when the network was trained and tested on all combinations of training and test precisions. The diagonal lines in the four transfer conditions indicate equal training and test precision for which the accuracies are also shown in panel B. For each transfer condition, there was a strong positive main effect of log test separation (p<0.0001, R2>0.35 in all conditions) shown as increasing color gradient from bottom to top, and the log training separation also had a weaker but significant negative effect (p<0.0001 in all conditions, R2>0.05 in all conditions except for angle+90 where R2=0.018) shown as decreasing color gradient from left to right. Higher training precisions enhanced performance at transfer to low precisions, shown as higher accuracy on top-left quadrants compared to lower-right quadrants.

Page 23: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

21

790

Figure 3. Layer change under different training precisions. A, Layer change (Equation 4) trajectories during learning. Lighter colors indicate larger angle separations. One-sem envelops are hardly visible. B, Iteration at which the rate of change peaked (peak speed iteration, PSI). Excluding layer 6, there were significant negative main effects of log angle separation (β=-37.24, t(14397)=-100.02, p≈0.0, R2=0.41) and layer number (β=-1.07, t(14397)=-8.73, p=2.9×10-18, R2=0.0031) on PSI, suggesting that layer change started to asymptote earlier in higher layers and finer precisions. For individual precisions, layer number had a significant effect only for the two smallest angle separations (p<0.0001, see Figure 3-1 for details). C, Final layer change. Ignoring layer 6, for which the change was measured differently, a linear regression analysis on the logarithm of layer change yielded significant negative main effects of log angle separation (β=-1.0, t(14397)=-208.4, p≈0.0, R2=0.66) and layer number (β=-0.15, t(14397)=-91.2, p≈0.0, R2=0.13), implying greater layer change in lower layers and finer precisions. D, Ratio of final layer change relative to the change under the easiest condition (10.0°). Changes in lower layers increased by a larger factor than higher layers when precision was high. B-D, Error bar represents 1 sem.

Page 24: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

22

791

Figure 4. Linear Fisher information (FI) defined in Equation 6 of the trained (black) and naïve (gray) populations at each layer and each test orientation and when trained at each precision. Only units with receptive fields at the center of the image and with minimum activation of 1.0 are included. (*) indicates significant increase in mean FI within 10° of the trained orientation (threshold p=0.01, Mann-Whitney U, Bonferroni-corrected for 5 layers × 5 angle separations). The oscillations were caused by pooling over 12 reference orientations and the inhomogeneity in orientation sensitivity of AlexNet. The FI values at layer 5 do not reflect real discrimination thresholds because the readout was noisy.

Page 25: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

23

792

Figure 5. Effect of different tasks (Gabor orientation and face gender discriminations) on performance and layer change. A, Accuracy drop as successive low layers were frozen at the iteration where the fully plastic network reached 95% accuracy for the two tasks. (*) indicates significant drop from zero (threshold p=0.01, 1-sample t-test against zero, Bonferroni-corrected for 5 frozen layers × 5 dissimilarities). Performance was impaired when freezing layer 2 onwards in the Gabor task and when freezing layer 4 onwards in the face task. The largest incremental performance drop happened in layer 3 for the Gabor task and layer 5 for the face task. B, Distribution of learning over layers when the network was trained on the two tasks if the network was fully plastic (left) or if layer 1 was frozen (right). There was a significant interaction of layer × task on layer change proportion (p<0.0001, 2-way ANOVA, see Extended Data Figure 5-2 for details). See Extended Data Figure 5-1 for demonstration of robustness to other measures of layer change. (*) indicates significant difference in the layer change proportion between the two tasks (threshold p=0.01, Mann-Whitney U, Bonferroni-corrected for 5 or 4 layers). Error bar represents 1 sem.

Page 26: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

24

793

Figure 6. A, Tuning curve examples of network units before (gray dashed) and after (black solid) training in layers 2 and 5. B-E, Comparison between V1 neurons trained under high precision in Schoups et al. (2001) with model units in layer 2 trained under high (1.0°, matching with experiment) and low (5.0°) precisions. B, Slope at trained orientation for trained and naïve V1 neurons grouped according to preferred orientation. C, Same as B but from model units. Units tuned around 20° increased their slope magnitude at trained orientation only under high precision. See Extended Data Figure 6-2 for statistical test details; see Extended Data Figure 6-1 for layer 3 and other precisions. D, Distribution of preferred orientation in V1 was roughly uniform before and after training. E, Same as D but from model units. There was no significant difference in the distribution of preferred orientation between the trained and naïve populations under either of the two precisions (p>0.9, K-S, see Extended Data Figure 6-4 for details; see Extended Data Figure 6-3 for layer 3 and other precisions). B and D, Adapted from Schoups et al. (2001) with permission.

Page 27: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

25

794

Figure 7. Comparison between V1-2 neurons trained under low precision in Ghose et al. (2002) with model units in layers 2 and 3 trained under high (1.0°) and low (5.0°, matching with experiment) precisions. Black bar indicates the orientation bin that contains the trained orientation. A, No significant effect of preferred orientation was found on tuning amplitude. B, Same as A but from model units. Preferred orientation had a significant effect only in layer 3 when trained under high precision (p<0.0001, one-way ANOVA, see Extended Data Figure 7-2 for details; see Extended Data Figure 7-1 for layer 3 and other precisions). C, No significant effect of preferred orientation was found on variance ratio. D, Same as C but from model units. No significant effect of preferred orientation was found on variance ratio (p>0.6, one-way ANOVA, see Extended Data Figure 7-4 for details; see Extended Data Figure 7-3 for layer 3 and other precisions) for either layer and precision. A and C, Adapted from Ghose et al. (2002) with permission.

Page 28: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

26

Figure 8. Comparison between V4 neurons in Raiguel et al. (2006) with model units in layer 5 trained under high (1.0°) and low (5.0°) precisions. Black and white bars indicate trained and naïve populations, respectively. A, Neurons tuned 22-67° away from trained orientation increased their slopes at trained orientation. (*) indicates significant increase in slope before and after training. B, Same as A but from model units. There was a significant interaction of training × preferred orientation under either of the precisions (p<0.0002, two-way ANOVA, see Extended Data Figure 8-2 for details; see Extended Data Figure 8-1 for layer 4 and other precisions). (*) indicates significant increase in slope before and after training (threshold p=0.01, Mann-Whitney U, Bonferroni-corrected for 6 orientation bins). Only neurons tuned within 48° of trained orientation are shown; other neurons did not change significantly after training. C, Distribution of preferred orientation shifted away from uniform after training. D, Same as C but from model units. The units under both precisions altered their preferred orientation distribution which became significantly different from a uniform distribution (p<0.003, K-S, see Extended Data Figure 8-4 for details; see Extended Data Figure 8-3 for layer 4

Page 29: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

27

795

and other precisions). E-H, Black and gray dashed lines indicate trained and naïve distribution means, respectively. E, Selectivity index of V4 neurons increased significantly after training. F, Same as E but from model units. Training produced a significant increase in selectivity index under high precision (p<0.0001, Mann-Whitney U, see Extended Data Figure 8-6 for details; see Extended Data Figure 8-5 for layer 4 and other precisions). G, Training significantly reduced normalized variance of V4 neurons. H, Same as G but from model units. Normalized variance was significantly reduced after training under both precisions (p<0.001, Mann-Whitney U, see Extended Data Figure 8-8 for details; see Extended Data Figure 8-7 for layer 4 and other precisions). A, C, E and G, Adapted from Raiguel et al. (2006)

Page 30: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

28

796

Figure 9. Comparison between V4 neurons in Yang & Maunsell (2004) with model units in layer 5 trained under high (1.0°) and low (5.0°) precisions. A-D, Black and white bars indicate trained and naïve populations, respectively; black and gray dashed lines indicate trained and naïve distribution medians, respectively. A, Training significantly increased the tuning amplitude of V4 neurons. B, Same as A but from model units. Training significantly increased response amplitude for both precisions (p<0.0001, Mann-Whitney U, see Extended Data Figure 9-2 for details, see Extended Data Figure 9-1 for layer 4 and other precisions). C, Training produced a significant reduction in best discriminability (lower indicates better discriminability) for V4 neurons. D, Same as B but from model units. Training significantly reduced the best discriminability for both precisions (p<0.0005, Mann-Whitney U, see Extended Data Figure 9-4 for details; see Extended Data Figure 9-3 for layer 4 and other precisions). E, Training resulted in narrower normalize tuning curves (by maximum response). F, Same as E but from model units. Activation was significantly lower after training for the non-preferred orientation range (>45° away of trained orientation) than preferred orientation range in high precision (p<0.0001, two-way ANOVA, see Extended Data Figure 9-6 for details; see Extended Data Figure 9-5 for layer 4 and other precisions). Red lines indicate orientations with significant reduction in activation (threshold p=0.01, Mann-Whitney U, Bonferroni-corrected for 100 test orientations). Curves are mean with one-sem envelope. A, C, E and G, Adapted from Yang & Maunsell (2004).

Page 31: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

29

797

798

Figure 10. A-B, Effect of the network’s initial sensitivity to trained orientation (TO) on the magnitude of learning. A, Relationship between layer weight change and layer-wise pre-train linear fisher information at TO (FI). Color indicates different layers, and darker color indicates higher training precision. Using a regression analysis on the layer change under the main effects of layer, precision, FI and the interactions of layer × FI and precision × FI, the effects of layer and precision were significant (p<0.0001) while the main effect of FI and its two-way interactions with layer and precision were not (p>0.1, see Extended Data Figure 10-2 for details; see Extended Data Figure 10-1 for results under different measures of weight change). B, Relationship between the change in the weights and pre-train FI at TO for network units, both of which are rescaled to between 0 and 1 after dividing by the respective maximum change for each angle separation and each layer. Significant Pearson’s correlation is shown for each layer and angle separation (Bonferroni-corrected for 5 layer × 5 angle separations). Despite the general positive correlation, there was a tendency for units with lower FI to change more when training precision increased. See Extended Data Figure 10-3 for results of regression analyses comparing effects related to FI and layer. C, Distribution of network unit fisher information at TO (FI) before (white) and after (black) training. In layers 1 to 2, most units had very low FI before and after training, whereas the distributions for units in layers 3 and 4 are more spread-out and training increased FI for many neurons.

Page 32: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

30

Extended Data 799

800

Figure 2-1. Learning performance trajectory for first 50 iterations, same data as Figure 2 but plotted as mean±std of 801 accuracy (left) and d-prime (right). Performance increased approximately linearly at the beginning and more slowly for 802 harder tasks. 803

Page 33: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

31

Statistical test on training and test precision on transfer performance 804

transfer conditions ori + 45 ori + 90 SF × 0.5 SF × 2.0 train and test at the same angle

separations

effect of log angle separation

β=0.15 t(2878)=27.58 p=8.9×10-149 R2=0.21

β=0.184 t(2878)=34.57 p=2.2×10-219 R2=0.29

β=0.21 t(2878)=59.29 p≈0.0 R2=0.55

β=0.17 t(2878)=42.78 p=5.7×10-310 R2=0.39

train and test at all combinations of angle separations

effect of log training angle

separation

β=-0.086 t(14397)=-36.09 p≈3.9×10-273 R2=0.052

β=-0.050 t(14397)=-20.94 p≈6.8×10-96 R2=0.018

β=-0.12 t(14397)=-78.93 p≈0.0 R2=0.11

β=-0.12 t(14397)=-71.41 p≈0.0 R2=0.11

effect of log testing angle separation

β=0.23 t(14397)=96.32 p≈0.0 R2=0.37

β=0.23 t(14397)=77.50 p≈0.0 R2=0.40

β=0.29 t(14397)=191.92 p≈0.0 R2=0.64

β=0.27 t(14397)=160.26 p≈0.0 R=0.57

Figure 2-2. Linear regression test details of training and test (log) angle separations on transfer performance. When 805 transfer was tested on the same angle separation as training, the transfer performance increased as angle separation 806 increased. When the network was trained and tested on all combinations of angle separations, significant effects of both 807 predictors were seen in all transfer conditions, but the effect size (R2) of training angle separation in orthogonal 808 orientation transfer (ori+90) was smaller than in other conditions. Some p-vales are less than machine precision and are 809 indicated as p≈0.0. 810

Statistical test details for effect of layer number on PSI 811

separation 0.5° 1.0° 2.0° 5.0° 10.0°

effect of layer number (1-5)

β=-4.73 t(2878)=-8.44 p=4.95×10-17 R2=0.024

β=-0.52 t(2878)=-4.27 p=2.0×10-5 R2=0.0063

β=-0.094 t(2878)=-1.42 p=0.16 R2=0.00070

β=-0.014 t(2878)=-0.32 p=0.75 R2=3.6×10-5

β=-0.013 t(2878)=-0.46 p=0.65 R2=7.3×10-5

Figure 3-1. Linear regression test details for effect of layer number (excluding layer 6) on the iteration at which a layer 812 reached its maximum speed of change (PSI). Significant effects of layer number were found under the two most precise 813 conditions (bold), implying that higher layers started to asymptote before early layers. 814

815

Page 34: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

32

816

Figure 5-1. Weights change proportions at each layer when all layers were plastic (top row) and when all but the first 817 layers were plastic (second row) using four different layer change measures (columns). The first normalization operation 818 in the network was after the first layer, so freezing layer 1 avoided learning directly on the pixel values. The four 819 measures are defined in Equations 4, 4a, 4b and 4c, respectively. These changes were then normalized to sum to 1 to 820 obtain the proportions. The interaction of layer × task on layer change proportion was significant in both conditions using 821 any of the four measures (p<0.0001, see Figure 5-1 for details). Learning the face task produced more change in the 822 higher layers when the first layer was frozen. (*) indicates significant difference in layer change proportion (threshold 823 p=0.01, Mann-Whitney U, Bonferroni-corrected for 4 or 5 plastic layers). 824

Statistical test details for the interaction of layer × task on layer change proportion 825

Freeze layer 1?

No F(4,590)=263.65 p=8.7×10-130

R2=0.12

F(4,590)=198.09 p=1.4×10-107 R2=0.058

F(4,590)=48.89 p=1.6×10-35 R2=0.0055

F(4,590)=135.89 p=3.1×10-82 R2=0.052

Yes F(3,472)=908.80 p=1.2×10-195 R2=0.13

F(3,472)=240.37 p=1.2×10-94 R2=0.11

F(3,472)=990.63 p=3.3×10-203 R2=0.1

F(3,472)=252.61 p=9.6×10-98 R2=0.074

Figure 5-2. ANOVA test details for the interaction of layer × task on layer change proportion. The individual main effects 826 were included in the analyses. All interactions were significant. 827

828

829

Page 35: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

33

830

Figure 6-1. Slope of (max-normalized) tuning curve at the trained orientation (TO) grouped by preferred orientation (PO) 831 relative for the trained (black solid) and naïve (gray dashed) model unit populations in the two lower layers. Two rows 832 correspond to layers 3 and 2; columns are different training angle separations. There was a significant interaction of 833 orientation × training on the slope at TO for units in layer 3 when trained at angle separations of at 0.5 and 1.0° (p<0.01) 834 but not for larger angle separations (p>0.07, see Figure 6-2 for details). Layer 2 failed to reach significance for all angle 835 separations (p>0.1). (*) indicate significant increase in slope at TO (threshold p=0.01, Mann-Whitney U, Bonferroni-836 corrected for 7 bins). 837

Statistical test details for the interaction of orientation × training on slope at TO 838

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer 3 F(6,3425)=5.17

p=2.7×105 F(6,3417)=2.99 p=0.0065

F(6,3356)=1.90 p=0.077

F(6,3317)=0.47 p=0.83

F(6,3330)=0.39 p=0.89

2 F(6,1311)=1.71 p=0.11

F(6,1303)=0.82 p=0.55

F(6,1284)=0.15 p=0.99

F(6,1283)=0.24 p=0.96

F(6,1272)=0.51 p=0.8

Figure 6-2. ANOVA test details for the interaction of orientation × training on slope at TO. The individual effects were 839 included in the analyses. There was a significant interaction of orientation × training on the slope at TO for units in layer 3 840 when trained at angle separations of at 0.5 and 1.0° but not for larger angle separations. 841

842

Page 36: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

34

843

Figure 6-3. Distributions of preferred orientation (PO) relative to trained orientation (TO) for the trained (black solid) and 844 naïve (gray dashed) model unit population in the two lower layers. Two rows correspond to layers 3 and 2; columns are 845 different training angle separations. No significant changes in distribution were found (p>0.05, see Figure 6-4 for details). 846

Statistical test details for change in distribution of preferred orientation 847

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer

3 (nnaive=2848)

d=0.0353 p=0.059 n=2787

d=0.0278 p=0.22 n=2839

d=0.0144 p=0.93 n=2845

d=0.00805 p=1.0 n=2845

d=0.00888 p=1.0 n=2843

2 (nnaive=1092)

d=0.0383 p=0.4 n=1078

d=0.0231 p=0.93 n=1089

d=0.0126 p=1.0 n=1103

d=0.0113 p=1.0 n=1105

d=0.0210 p=0.97 n=1100

Figure 6-4. K-S test details for the interaction of orientation × training on slope at TO. No significant changes in 848 distribution were found (p>0.05 for all layers and angle separations). 849

850

851

Page 37: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

35

852

Figure 7-1. Tuning amplitude grouped by preferred orientation (PO) relative to trained orientation (TO) for the trained 853 model unit population in the two lower layers. Two rows correspond to layers 3 and 2; columns are different training 854 angle separations. Black bar indicates the orientation bin that contains the trained orientation. Preferred orientation had 855 significant effects on tuning amplitude only in layer 3 when trained up to 2.0° (p<0.01, see Figure 7-2 for details). 856

Statistical test details for effect of orientation on tuning amplitude 857

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer 3 F(9,2777)=6.94

p=4.4×10-9 F(9,2829)=4.00 p=0.00010

F(9,2835)=2.71 p=0.0057

F(9,2835)=1.19 p=0.30

F(9,2833)=1.00 p=0.44

2 F(9,1068)=1.70 p=0.095

F(9,1079)=0.72 p=0.67

F(9,1093)=0.26 p=0.98

F(9,1095)=0.44 p=0.90

F(9,1090)=0.19 p=0.99

Figure 7-2. ANOVA test details for effect of orientation on tuning amplitude. Preferred orientation had significant effects 858 on tuning amplitude only in layer 3 when trained up to 2.0°. 859

860

Page 38: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

36

861

Figure 7-3. Variance ratio grouped by preferred orientation (PO) relative to trained orientation (TO) for the trained model 862 unit population in the lower two layers. Two rows correspond to layers 3 and 2; columns are different training angle 863 separations. Black bar indicates the bin that contains the trained orientation. PO had a significant effect on variance ratio 864 only in layer 3 when trained at 0.5° (p=0.0028, see Figure 7-4 for details). 865

Statistical test details for the effect of PO on variance ratio 866

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer 3 F(9,2777)=3.62

p=0.00034 F(9,2829)=0.67 p=0.72

F(9,2835)=0.68 p=0.71

F(9,2835)=0.61 p=0.77

F(9,2833)=0.14 p=1.0

2 F(9,1068)=1.85 p=0.064

F(9,1079)=1.00 p=0.44

F(9,1093)=0.20 p=0.99

F(9,1095)=0.77 p=0.69

F(9,1090)=0.16 p=1.0

Figure 7-4. ANOVA test details for the effect of preferred orientation (PO) on variance ratio. PO had a significant effect 867 on variance ratio only in layer 3 when trained at 0.5° 868

869

Page 39: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

37

870

Figure 8-1. Slope at trained orientation (TO) grouped by preferred orientation (PO) for the trained (black) and naïve 871 (white) model unit populations in the two higher layers. Two rows correspond to layers 5 and 4; columns are different 872 training angle separations. Significant interactions of PO × training on the slope at TO were found for angle separations 873 up to 5.0° in layer 5 and up to 2.0° in layer 4 (p<0.001, see Figure 8-2 for details). (*) indicates significant increase in 874 slope at TO (threshold p=0.01, Mann-Whitney U, Bonferroni-corrected for 6 bins). 875

Statistical test details for interaction of PO × training on the slope at TO 876

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer

5 F(5,196)=3.68 p=0.0033

F(5,202)=5.33 p=0.00013

F(5,203)=8.26 p=4×10-7

F(5,198)=5.61 p=7.4×10-5

F(5,199)=2.79 p=0.018

4 F(5,1244)=12.69 p=4.8×10-12

F(5,1234)=9.48 p=7×10-9

F(5,1220)=5.78 p=2.8×10-5

F(5,1215)=1.51 p=0.18

F(5,1224)=0.21 p=0.96

Figure 8-2. ANOVA test details for interaction of preferred orientation (PO) × training on the slope at trained orientation 877 (TO). The individual main effects were included in the analyses. Significant interactions of PO × training on the slope at 878 TO were found for angle separations up to 5.0° in layer 5 and up to 2.0° in layer 4. 879

880

Page 40: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

38

881

Figure 8-3. Distribution of preferred orientation (PO) relative to trained orientation (TO) for the trained (black) and naïve 882 (white) model unit populations in the two higher layers. Two rows correspond to layers 5 and 4; columns are different 883 training angle separations. Significant deviations in distribution from uniform were found after training for angle 884 separations up to 5.0° in layer 5 and up to 2.0° in layer 4 (p<0.003, see Figure 8-4). 885

Statistical test details for distribution of preferred orientation compared against a uniform distribution 886

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer

5 d=0.21 p=1.3×10-11 n=280

d=0.21 p=5.2×10-12 n=295

d=0.15 p=1.7×10-5 n=264

d=0.12 p=0.0029 n=222

d=0.083 p=0.10 n=213

4 d=0.095 p=6.4×10-11 n=1334

d=0.085 p=8×10-9 n=1318

d=0.051 p=0.0023 n=1301

d=0.029 p=0.24 n=1250

d=0.017 p=0.88 n=1230

Figure 8-4. K-S test details for distribution of preferred orientation (PO) compared against a uniform distribution. 887 Significant deviations in distribution from uniform were found after training for angle separations up to 5.0° in layer 5 and 888 up to 2.0° in layer 4. 889

890

Page 41: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

39

891

Figure 8-5. Distribution of selective index for the trained (black) and naïve (white) model unit populations in the two 892 higher layers. Two rows correspond to layers 5 and 4; columns are different training angle separations. Black and gray 893 dashed lines indicate the trained and naïve population means, respectively. Significant increases in selective index were 894 found for angle separations up to 2.0° in layer 5 and up to 1.0° in layer 4 (p<0.002, See Figure 8-3 for details). 895

Statistical test details for increase in selectivity index 896

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer

5 (MDnaive=0.55,

nnaive=192)

MD=0.72 U=2.0×104 p=1.7×10-6 n=280

MD=0.75 U=2.1×104 p=1.1×10-6 n=295

MD=0.66 U=2.1×104 p=0.00038 n=264

MD=0.58 U=2×104 p=0.1.0 n=222

MD=0.56 U=1.9×104 p=0.21 n=213

4 (MDnaive=0.54,

nnaive=1203)

MD=0.60 U=7.2×105 p=2.0×10-6 n=1334

MD=0.57 U=7.4×105 p=0.0018 n=1318

MD=0.56 U=7.6×105 p=0.067 n=1301

MD=0.55 U=7.4×105

p=0.34 n=1250

MD=0.54 U=7.4×105 p=0.44 n=1230

Figure 8-6. Mann-Whitney U test details for increase in selectivity index. Significant increases in selective index were 897 found for angle separations up to 2.0° in layer 5 and up to 1.0° in layer 4. 898

899

Page 42: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

40

900

Figure 8-7. Distribution of normalized variance (or Fano factor) for the trained (black) and naïve (white) model unit 901 populations in the two higher layers. Two rows correspond to layers 5 and 4; columns are different training angle 902 separations. Black and gray dashed lines indicate the trained and naïve population means, respectively. Significant 903 reductions in normalized variance were found for angle separations up to 5.0° in layer 5 and up to 2.0° in layer 4 904 (p<0.001, see Figure 8-8 for details). 905

Statistical test details for decrease in normalized variance 906

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer

5 (MDnaive=0.45,

nnaive=192)

MD=0.16 U=1.02×104 p=7.3×10-31 n=280

MD=0.22 U=1.58×104 p=6.5×10-17 n=295

MD=0.29 U=1.84×104 p=2.7×10-7 n=264

MD=0.34 U=1.75×104 p=0.00083 n=222

MD=0.38 U=1.84×104 p=0.041 n=213

4 (MDnaive=0.66,

nnaive=1203)

MD=0.39 U=4.87×105 p=5.7×10-66 n=1334

MD=0.46 U=5.81×105 p=2.3×10-31 n=1318

MD=0.58 U=6.93×105 p=3.5×10-7 n=1301

MD=0.65 U=7.37×105 p=0.20 n=1250

MD=0.66 U=7.35×105 p=0.39 n=1230

Figure 8-8. Mann-Whitney U test details for decrease in normalized variance. Significant reductions in normalized 907 variance were found for angle separations up to 5.0° in layer 5 and up to 2.0° in layer 4. 908

909

Page 43: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

41

910

Figure 9-1. Distribution of tuning amplitude for the trained (black) and naïve (white) model units in the two higher layers. 911 Two rows correspond to layers 5 and 4; columns are different training angle separations. Black and gray dashed lines 912 indicate the trained and naïve population medians, respectively. Significant increases in tuning amplitude were found for 913 angle separations under all precisions in layer 5 and up to 2.0° in layer 4 (p<0.003, see Figure 9-2 for details). 914

Statistical test details for increase in response amplitude 915

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer

5 (MDnaive=14.64,

nnaive=192)

MD=25.28 U=1.54×104 p=1.7×10-15 n=280

MD=22.09 U=1.90×104 p=4.2×10-10 n=295

MD=20.77 U=1.88×104 p=1.2×10-6 n=264

MD=18.88 U=1.62×104 p=1.5×10-5 n=222

MD=16.72 U=1.71×104 p=0.0020 n=213

4 (MDnaive=19.79,

nnaive=1203)

MD=26.92 U=6.41×105 p=1.1×10-18 n=1334

MD=25.85 U=6.64×105 p=9.5×10-13 n=1318

MD=23.10 U=7.15×105 p=0.0001 n=1301

MD=21.73 U=7.25×105 p=0.062 n=1250

MD=21.22 U=7.24×105 p=0.17 n=1230

Figure 9-2. Mann-Whitney U test details for increase in response amplitude. Significant increases in tuning amplitude 916 were found for angle separations under all precisions in layer 5 and up to 2.0° in layer 4. 917

Page 44: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

42

918

Figure 9-3. Distribution of best discriminability for the trained (black) and naïve (white) model units in the two higher 919 layers. Two rows correspond to layers 5 and 4; columns are different training angle separations. Black and gray dashed 920 lines indicate the trained and naïve population medians, respectively. Significant reductions in best discriminability were 921 found for angle separations up to 5.0° in both layers (p<0.005, see Figure 9-4 for details). 922

Statistical test details for decrease in best discriminability 923

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer

5 (MDnaive=0.8,

nnaive=192)

MD=0.08 U=8.08×103 p=1.9×10-38 n=280

MD=0.11 U=9.48×103 p=1.1×10-35 n=295

MD=0.33 U=1.46×104 p=4.8×10-15 n=264

MD=0.55 U=1.76×104 p=0.0011 n=222

MD=0.74 U=1.90×104 p=0.10 n=213

4 (MDnaive=1.53,

nnaive=1203)

MD=0.27 U=4.19×105 p=2.3×10-96 n=1334

MD=0.46 U=5.13×105 p=2.2×10-53 n=1318

MD=0.80 U=6.34×105 p=1.1×10-16 n=1301

MD=1.20 U=7.04×105 p=0.0033 n=1250

MD=1.31 U=7.18×105 p=0.10 n=1230

Figure 9-4. Mann-Whitney U test details for increase in best discriminability. Significant reductions in best 924 discriminability were found for angle separations up to 5.0° in both layers. 925

926

Page 45: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

43

927

Figure 9-5. Average normalized tuning curves in for the trained (black) and naïve (gray) model units in the two higher 928 layers. Two rows correspond to layers 5 and 4; columns are different training angle separations. Training reduced the 929 activation away preferred orientations (PO) in layer 5 for small angle separations. By grouping the activations into two 930 groups according to whether the test stimulus was oriented within 45° of the PO, two-way ANOVA revealed a significant 931 interaction of orientation group × training on activation in layer 5 for up to 2.0° and in layer 4 for 0.5° angle separation 932 (p<0.01, two-way ANOVA, see Figure 9-6 for details), suggesting that the activation of non-POs was significantly more 933 reduced than that of POs. Red lines indicate orientations where the trained activation was significantly less than naïve 934 (threshold p=0.01, Mann-Whitney U, Bonferroni-corrected for 100 test orientations). 935

Statistical test details for interaction orientation group × training on activation 936

separation 0.5° 1.0° 2.0° 5.0° 10.0°

layer 5 F(1,940)=13.02

p=0.00032 F(1,970)=15.86 p=7.3×10-5

F(1,908)=6.73 p=0.0096

F(1,824)=0.36 p=0.55

F(1,806)=0.04 p=0.84

4 F(1,5070)=7.19 p=0.0074

F(1,5038)=3.54 p=0.06

F(1,5004)=1.74 p=0.19

F(1,4902)=0.03 p=0.87

F(1,4862)=0.00 p=0.98

Figure 9-6. ANOVA test details for interaction orientation group × training on activation. The individual main effects 937 were included in the analysis. By grouping the activations into two groups according to whether the test stimulus was 938 oriented within 45° of the preferred orientation, the analysis revealed a significant interaction of orientation group × 939 training on activation in layer 5 for up to 2.0° and in layer 4 for 0.5° angle separation. 940

Page 46: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

44

941

Figure 10-1. Weight change at each layer under the four different measures (columns) against pre-train linear fisher 942 information (FI) at the trained orientation (TO). Darker color indicates smaller angle separation. The data were obtained 943 from the simulations where unit tuning curves were taken (noise standard deviation 15, contrast 20% and wavelength 10 944 pixels). Units with responses less than 0.1 were excluded. The four measures are defined in Equations 4, 4a, 4b and 4c, 945 respectively. Under all layer change measures, we conducted regression analyses using Equation 7, which revealed 946 significant main effects of layer and angle separation (p<0.0001), but the effects involving FI were insignificant except for 947

and . See Figure 10-2 for statistical test details. 948

Statistical test details for various effects on layer weight change 949

Factors

pre-train Fisher information (FI)

F(1,282)=0.0069 p=0.93 R2=7.7×10-6

F(1,282)=0.031 p=0.86 R2=1.9×10-5

F(1,282)=0.00032 p=0.99 R2=5.6×10-7

F(1,282)=0.00077 p=0.98 R2=1.3×10-6

Layer (L) F(4,282)=48.27 p=6.7×10-31 R2=0.21

F(4,282)=50.51 p=5×10-32 R2=0.13

F(4,282)=40.58 p=7.6×10-27 R2=0.28

F(4,282)=33.83 p=4.7×10-23 R2=0.22

angle separation (S) F(4,282)=103.3 p=4.7×10-54 R2=0.46

F(4,282)=275.6 p=4.3×10-96 R2=0.69

F(4,282)=28.38 p=8×10-20 R2=0.2

F(4,282)=42.84 p=4.6×10-28 R2=0.28

L × FI F(4,282)=0.02025 p=1.00 R2=9×10-5

F(4,282)=0.1487 p=0.96 R2=0.00037

F(4,282)=0.001148 p=1.00 R2=8×10-6

F(4,282)=0.005635 p=1.00 R2=3.7×10-5

S × FI F(4,282)=2.887 p=0.023 R2=0.013

F(4,282)=0.5435 p=0.70 R2=0.0014

F(4,282)=3.83 p=0.0048 R2=0.027

F(4,282)=3.704 p=0.0059 R2=0.025

Figure 10-2. Statistical test details for various effects on layer weight change. The data were obtained from the 950 simulations where unit tuning curves were taken (noise standard deviation 15, contrast 20% and wavelength 10 pixels). 951 The four measures in the title row are defined in Equations 4, 4a, 4b and 4c, respectively. Bold indicates significant 952 effects. 953

954

955

956

Page 47: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs

45

Statistical test details for various effects on unit weight change 957

Factors pre-train Fisher

information (FI)

F(1,18984)=127.76 p=1.6×10-29 R2=0.0053

F(1,18984)=177.70 p=2.3×10-40 R2=0.0074

F(1,18984)=130.61 p=3.8×10-30 R2=0.0038

F(1,18984)=198.96 p=5.9×10-45 R2=0.006

Layer (L) F(4,18984)=158.16 p=2.2×10-133 R2=0.026

F(4,18984)=32.60 p=4×10-27 R2=0.0054

F(4,18984)=819.25 p=0 R2=0.095

F(4,18984)=478.17 p=0 R2=0.058

angle separation

(S)

F(4,18984)=750.58 p≈0.0 R2=0.12

F(4,18984)=973.87 p≈0.0 R2=0.16

F(4,18984)=1287.37 p≈0.0 R2=0.15

F(4,18984)=1625.12 p≈0.0 R2=0.2

L × FI F(4,18984)=22.47 p=1.5×10-18 R2=0.0037

F(4,18984)=24.03 p=7.4×10-20 R2=0.004

F(4,18984)=191.88 p=1.5×10-161 R2=0.022

F(4,18984)=168.17 p=9.4×10-142 R2=0.02

S × FI F(4,18984)=32.31 p=7×10-27 R2=0.0054

F(4,18984)=38.87 p=1.8×10-32 R2=0.0065

F(4,18984)=35.11 p=2.9×10-29 R2=0.0041

F(4,18984)=48.62 p=9.4×10-41 R2=0.0059

Figure 10-3. Statistical test details for various effects on unit weight change. The data were obtained from the simulations 958 where unit tuning curves were taken (noise standard deviation 15, contrast 20% and wavelength 10 pixels). The four 959 measures in the title row are defined in Equations 4, 4a, 4b and 4c, respectively. All effects are significant under all 960 change measures, but angle separation accounted for the majority of variance. 961

Page 48: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs
Page 49: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs
Page 50: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs
Page 51: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs
Page 52: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs
Page 53: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs
Page 54: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs
Page 55: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs
Page 56: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs
Page 57: Deep neural networks for modeling visual perceptual learning · 2018. 5. 23. · í Title: Deep neural networks for mode ling visual perceptual learning î Abbreviated title: DNNs