-
Notifications
You must be signed in to change notification settings - Fork 0
/
MyLibrary.bib
9897 lines (9237 loc) · 958 KB
/
MyLibrary.bib
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
@article{field_scale-invariance_1993,
title = {Scale-Invariance and Self-Similar ‘Wavelet’ Transforms: {{An}} Analysis of Natural Scenes and Mammalian Visual Systems},
url = {https://www.researchgate.net/publication/266310150_Scale-invariance_and_self-similar_'wavelet'_transforms_An_analysis_of_natural_scenes_and_mammalian_visual_systems},
shorttitle = {Scale-Invariance and Self-Similar ‘Wavelet’ Transforms},
abstract = {The processing of spatial patterns by the mammalian visual system shows a number of similarities to the ‘wavelet transforms’ which have recently attracted considerable interest outside of the...},
journaltitle = {ResearchGate},
urldate = {2016-07-28},
date = {1993},
author = {Field, D. J.},
file = {/Users/fergalcotter/Dropbox/Papers/Field_Scale-invariance and self-similar ‘wavelet’ transforms.pdf;/Users/fergalcotter/Zotero/storage/SAPHZHU3/266310150_Scale-invariance_and_self-similar_'wavelet'_transforms_An_analysis_of_natural_scenes_.html},
note = {00216}
}
@article{gu_recent_2015,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1512.07108},
primaryClass = {cs},
title = {Recent {{Advances}} in {{Convolutional Neural Networks}}},
url = {http://arxiv.org/abs/1512.07108},
abstract = {In the last few years, deep learning has led to very good performance on a variety of problems, such as visual recognition, speech recognition and natural language processing. Among different types of deep neural networks, convolutional neural networks have been most extensively studied. Due to the lack of training data and computing power in early days, it is hard to train a large high-capacity convolutional neural network without overfitting. After the rapid growth in the amount of the annotated data and the recent improvements in the strengths of graphics processor units (GPUs), the research on convolutional neural networks has been emerged swiftly and achieved state-of-the-art results on various tasks. In this paper, we provide a broad survey of the recent advances in convolutional neural networks. Besides, we also introduce some applications of convolutional neural networks in computer vision.},
urldate = {2016-08-09},
date = {2015-12-22},
keywords = {Computer Science - Computer Vision and Pattern Recognition,Computer Science - Learning,Computer Science - Neural and Evolutionary Computing},
author = {Gu, Jiuxiang and Wang, Zhenhua and Kuen, Jason and Ma, Lianyang and Shahroudy, Amir and Shuai, Bing and Liu, Ting and Wang, Xingxing and Wang, Gang},
file = {/Users/fergalcotter/Dropbox/Papers/Gu et al_2015_Recent Advances in Convolutional Neural Networks.pdf;/Users/fergalcotter/Zotero/storage/72QQ5TWA/1512.html},
note = {00002}
}
@incollection{plate_avoiding_2012,
langid = {english},
title = {Avoiding {{Roundoff Error}} in {{Backpropagating Derivatives}}},
isbn = {978-3-642-35288-1 978-3-642-35289-8},
url = {http://link.springer.com/chapter/10.1007/978-3-642-35289-8_15},
abstract = {One significant source of roundoff error in backpropagation networks is the calculation of derivatives of unit outputs with respect to their total inputs. The roundoff error can lead result in high relative error in derivatives, and in particular, derivatives being calculated to be zero when in fact they are small but non-zero. This roundoff error is easily avoided with a simple programming trick which has a small memory overhead (one or two extra floating point numbers per unit) and an insignificant computational overhead.},
number = {7700},
booktitle = {Neural {{Networks}}: {{Tricks}} of the {{Trade}}},
series = {Lecture {{Notes}} in {{Computer Science}}},
publisher = {{Springer Berlin Heidelberg}},
urldate = {2016-08-09},
date = {2012},
pages = {225-230},
keywords = {Algorithm Analysis and Problem Complexity,Artificial Intelligence (incl. Robotics),Complexity,Computation by Abstract Devices,Information Systems Applications (incl. Internet),pattern recognition},
author = {Plate, Tony},
editor = {Montavon, Grégoire and Orr, Geneviève B. and Müller, Klaus-Robert},
file = {/Users/fergalcotter/Dropbox/Papers/Plate_2012_Avoiding Roundoff Error in Backpropagating Derivatives.pdf;/Users/fergalcotter/Zotero/storage/5VEV4NP7/978-3-642-35289-8_15.html},
doi = {10.1007/978-3-642-35289-8_15},
note = {00001}
}
@unpublished{kingsbury_visualisation_2015,
venue = {{Adelaide University}},
title = {Visualisation of {{Convolutional Networks}} and {{Multiscale Scatter}}-{{Nets}}},
date = {2015-11},
keywords = {Unread},
author = {Kingsbury, Nick},
file = {/Users/fergalcotter/Dropbox/Papers/DeconvNets&ScatterNetsTalk1.pdf},
note = {00000}
}
@article{kawaguchi_deep_2016,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1605.07110},
primaryClass = {cs, math, stat},
title = {Deep {{Learning}} without {{Poor Local Minima}}},
url = {http://arxiv.org/abs/1605.07110},
abstract = {In this paper, we prove a conjecture published in 1989 and also partially address an open problem announced at the Conference on Learning Theory (COLT) 2015. For an expected loss function of a deep nonlinear neural network, we prove the following statements under the independence assumption adopted from recent work: 1) the function is non-convex and non-concave, 2) every local minimum is a global minimum, 3) every critical point that is not a global minimum is a saddle point, and 4) the property of saddle points differs for shallow networks (with three layers) and deeper networks (with more than three layers). Moreover, we prove that the same four statements hold for deep linear neural networks with any depth, any widths and no unrealistic assumptions. As a result, we present an instance, for which we can answer to the following question: how difficult to directly train a deep model in theory? It is more difficult than the classical machine learning models (because of the non-convexity), but not too difficult (because of the nonexistence of poor local minima and the property of the saddle points). We note that even though we have advanced the theoretical foundations of deep learning, there is still a gap between theory and practice.},
urldate = {2016-05-25},
date = {2016-05-23},
keywords = {Computer Science - Learning,Mathematics - Optimization and Control,Statistics - Machine Learning},
author = {Kawaguchi, Kenji},
file = {C:\\Users\\fbc23\\Google Drive\\Papers\\May 16\\Kawaguchi_2016_Deep Learning without Poor Local Minima.pdf;/Users/fergalcotter/Zotero/storage/8RAQN7N8/1605.html},
note = {00000}
}
@inproceedings{nguyen_deep_2015,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1412.1897},
location = {{Boston, MA, USA}},
title = {Deep {{Neural Networks}} Are {{Easily Fooled}}: {{High Confidence Predictions}} for {{Unrecognizable Images}}},
url = {http://arxiv.org/abs/1412.1897},
shorttitle = {Deep {{Neural Networks}} Are {{Easily Fooled}}},
abstract = {Deep neural networks (DNNs) have recently been achieving state-of-the-art performance on a variety of pattern-recognition tasks, most notably visual classification problems. Given that DNNs are now able to classify objects in images with near-human-level performance, questions naturally arise as to what differences remain between computer and human vision. A recent study revealed that changing an image (e.g. of a lion) in a way imperceptible to humans can cause a DNN to label the image as something else entirely (e.g. mislabeling a lion a library). Here we show a related result: it is easy to produce images that are completely unrecognizable to humans, but that state-of-the-art DNNs believe to be recognizable objects with 99.99\% confidence (e.g. labeling with certainty that white noise static is a lion). Specifically, we take convolutional neural networks trained to perform well on either the ImageNet or MNIST datasets and then find images with evolutionary algorithms or gradient ascent that DNNs label with high confidence as belonging to each dataset class. It is possible to produce images totally unrecognizable to human eyes that DNNs believe with near certainty are familiar objects, which we call "fooling images" (more generally, fooling examples). Our results shed light on interesting differences between human vision and current DNNs, and raise questions about the generality of DNN computer vision.},
eventtitle = {2015 {{IEEE Conference}} on {{Computer Vision}} and {{Pattern Recognition}} ({{CVPR}})},
booktitle = {Proceedings of 2015 {{IEEE Conference}} on {{Computer Vision}} and {{Pattern Recognition}} ({{CVPR}})},
publisher = {{IEEE}},
urldate = {2016-08-24},
date = {2015-06},
keywords = {Computer Science - Artificial Intelligence,Computer Science - Computer Vision and Pattern Recognition,Computer Science - Neural and Evolutionary Computing},
author = {Nguyen, Anh and Yosinski, Jason and Clune, Jeff},
file = {/Users/fergalcotter/Dropbox/Papers/Nguyen et al_2014_Deep Neural Networks are Easily Fooled.pdf;/Users/fergalcotter/Zotero/storage/2FUQVBDB/1412.html},
note = {00164}
}
@article{magarey_motion_1998,
title = {Motion Estimation Using a Complex-Valued Wavelet Transform},
volume = {46},
doi = {10.1109/78.668557},
abstract = {This paper describes a new motion estimation algorithm that is potentially useful for both computer vision and video compression applications, It is hierarchical in structure, using a separable two-dimensional (2-D) discrete wavelet transform (DWT) on each frame to efficiently construct a multiresolution pyramid of subimages, The DWT is based on a complex-valued pair of four-tap FIR filters with Gabor-like characteristics. The resulting complex DWT (CDWT) effectively implements an analysis by an ensemble of Gabor-like filters with a variety of orientations and scales, The phase difference between the subband coefficients of each frame at a given subpel bears a predictable relation to a local translation in the region of the reference frame subtended by that subpel, That relation is used to estimate the displacement field at the coarsest scale of the multiresolution pyramid, Each estimate is accompanied by a directional confidence measure in the form of the parameters of a quadratic matching surface, The initial estimate field is progressively refined by a coarse-to-fine strategy in which finer scale information is appropriately incorporated at each stage, The accuracy, efficiency, and robustness of the new algorithm are demonstrated in comparison testing against hierarchical implementations of intensity gradient-based and fractional-precision block matching motion estimators.},
number = {4},
journaltitle = {IEEE Transactions on Signal Processing},
date = {1998-04},
pages = {1069-1084},
author = {Magarey, J. and Kingsbury, N.},
file = {/Users/fergalcotter/Dropbox/Papers/Magarey_Kingsbury_1998_Motion estimation using a complex-valued wavelet transform.pdf},
note = {00228}
}
@incollection{yosinski_how_2014,
title = {How Transferable Are Features in Deep Neural Networks?},
url = {http://papers.nips.cc/paper/5347-how-transferable-are-features-in-deep-neural-networks.pdf},
booktitle = {Advances in {{Neural Information Processing Systems}} 27},
publisher = {{Curran Associates, Inc.}},
urldate = {2016-07-15},
date = {2014},
pages = {3320--3328},
keywords = {_tablet},
author = {Yosinski, Jason and Clune, Jeff and Bengio, Yoshua and Lipson, Hod},
editor = {Ghahramani, Z. and Welling, M. and Cortes, C. and Lawrence, N. D. and Weinberger, K. Q.},
file = {/Users/fergalcotter/Dropbox/Papers/Yosinski et al_2014_How transferable are features in deep neural networks.pdf;/Users/fergalcotter/Zotero/storage/K87W6KT8/5347-how-transferable-are-features-in-deep-neural-networks.html},
note = {00217}
}
@article{bruna_signal_2013,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1311.4025},
primaryClass = {stat},
title = {Signal {{Recovery}} from {{Pooling Representations}}},
url = {http://arxiv.org/abs/1311.4025},
abstract = {In this work we compute lower Lipschitz bounds of \$\textbackslash{}ell\_p\$ pooling operators for \$p=1, 2, \textbackslash{}infty\$ as well as \$\textbackslash{}ell\_p\$ pooling operators preceded by half-rectification layers. These give sufficient conditions for the design of invertible neural network layers. Numerical experiments on MNIST and image patches confirm that pooling layers can be inverted with phase recovery algorithms. Moreover, the regularity of the inverse pooling, controlled by the lower Lipschitz constant, is empirically verified with a nearest neighbor regression.},
urldate = {2016-02-01},
date = {2013-11-16},
keywords = {Statistics - Machine Learning,Unread},
author = {Bruna, Joan and Szlam, Arthur and LeCun, Yann},
file = {/Users/fergalcotter/Dropbox/Papers/Bruna et al_2013_Signal Recovery from Pooling Representations.pdf;/Users/fergalcotter/Zotero/storage/AHQCZXAD/1311.html},
note = {00008}
}
@inproceedings{oyallon_deep_2015,
location = {{Boston, MA, USA}},
title = {Deep {{Roto}}-{{Translation Scattering}} for {{Object Classification}}},
url = {http://www.cv-foundation.org/openaccess/content_cvpr_2015/html/Oyallon_Deep_Roto-Translation_Scattering_2015_CVPR_paper.html},
eventtitle = {2015 {{IEEE Conference}} on {{Computer Vision}} and {{Pattern Recognition}} ({{CVPR}})},
booktitle = {Proceedings of 2015 {{IEEE Conference}} on {{Computer Vision}} and {{Pattern Recognition}} ({{CVPR}})},
publisher = {{IEEE}},
urldate = {2016-03-01},
date = {2015-06},
pages = {2865-2873},
author = {Oyallon, Edouard and Mallat, Stephane},
file = {/Users/fergalcotter/Dropbox/Papers/Oyallon_Mallat_2015_Deep Roto-Translation Scattering for Object Classification.pdf;/Users/fergalcotter/Zotero/storage/IB99F9KA/Oyallon_Deep_Roto-Translation_Scattering_2015_CVPR_paper.html},
note = {00011}
}
@misc{hinton_recognize_2006,
title = {To {{Recognize Shapes}}, {{First Learn}} to {{Generate Images}}},
date = {2006-10},
keywords = {Unread},
author = {Hinton, Geoff},
file = {/Users/fergalcotter/Dropbox/Papers/Geoff Hinton_2006_To Recognize Shapes, First Learn to Generate Images.pdf},
note = {00000}
}
@incollection{rumelhart_parallel_1986,
location = {{Cambridge, MA, USA}},
title = {Parallel {{Distributed Processing}}: {{Explorations}} in the {{Microstructure}} of {{Cognition}}, {{Vol}}. 1},
isbn = {978-0-262-68053-0},
url = {http://dl.acm.org/citation.cfm?id=104279.104293},
shorttitle = {Parallel {{Distributed Processing}}},
publisher = {{MIT Press}},
urldate = {2016-08-24},
date = {1986},
pages = {318--362},
author = {Rumelhart, D. E. and Hinton, G. E. and Williams, R. J.},
editor = {Rumelhart, David E. and McClelland, James L. and PDP Research Group, CORPORATE},
file = {/Users/fergalcotter/Dropbox/Papers/Rumelhart et al_1986_Parallel Distributed Processing.pdf},
note = {00087}
}
@incollection{lawrence_neural_2012,
langid = {english},
title = {Neural {{Network Classification}} and {{Prior Class Probabilities}}},
isbn = {978-3-642-35288-1 978-3-642-35289-8},
url = {http://link.springer.com/chapter/10.1007/978-3-642-35289-8_19},
abstract = {A commonly encountered problem in MLP (multi-layer perceptron) classification problems is related to the prior probabilities of the individual classes – if the number of training examples that correspond to each class varies significantly between the classes, then it may be harder for the network to learn the rarer classes in some cases. Such practical experience does not match theoretical results which show that MLPs approximate Bayesian a posteriori probabilities (independent of the prior class probabilities). Our investigation of the problem shows that the difference between the theoretical and practical results lies with the assumptions made in the theory (accurate estimation of Bayesian a posteriori probabilities requires the network to be large enough, training to converge to a global minimum, infinite training data, and the a priori class probabilities of the test set to be correctly represented in the training set). Specifically, the problem can often be traced to the fact that efficient MLP training mechanisms lead to sub-optimal solutions for most practical problems. In this chapter, we demonstrate the problem, discuss possible methods for alleviating it, and introduce new heuristics which are shown to perform well on a sample ECG classification problem. The heuristics may also be used as a simple means of adjusting for unequal misclassification costs.},
number = {7700},
booktitle = {Neural {{Networks}}: {{Tricks}} of the {{Trade}}},
series = {Lecture {{Notes}} in {{Computer Science}}},
publisher = {{Springer Berlin Heidelberg}},
urldate = {2016-08-09},
date = {2012},
pages = {295-309},
keywords = {Algorithm Analysis and Problem Complexity,Artificial Intelligence (incl. Robotics),Complexity,Computation by Abstract Devices,Information Systems Applications (incl. Internet),pattern recognition},
author = {Lawrence, Steve and Burns, Ian and Back, Andrew and Tsoi, Ah Chung and Giles, C. Lee},
editor = {Montavon, Grégoire and Orr, Geneviève B. and Müller, Klaus-Robert},
file = {/Users/fergalcotter/Dropbox/Papers/Lawrence et al_2012_Neural Network Classification and Prior Class Probabilities.pdf;/Users/fergalcotter/Zotero/storage/S29BKNUN/978-3-642-35289-8_19.html},
doi = {10.1007/978-3-642-35289-8_19},
note = {00117}
}
@article{kingma_adam:_2014,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1412.6980},
primaryClass = {cs},
title = {Adam: {{A Method}} for {{Stochastic Optimization}}},
url = {http://arxiv.org/abs/1412.6980},
shorttitle = {Adam},
abstract = {We introduce Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments. The method is straightforward to implement, is computationally efficient, has little memory requirements, is invariant to diagonal rescaling of the gradients, and is well suited for problems that are large in terms of data and/or parameters. The method is also appropriate for non-stationary objectives and problems with very noisy and/or sparse gradients. The hyper-parameters have intuitive interpretations and typically require little tuning. Some connections to related algorithms, on which Adam was inspired, are discussed. We also analyze the theoretical convergence properties of the algorithm and provide a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework. Empirical results demonstrate that Adam works well in practice and compares favorably to other stochastic optimization methods. Finally, we discuss AdaMax, a variant of Adam based on the infinity norm.},
urldate = {2016-08-07},
date = {2014-12-22},
keywords = {Computer Science - Learning},
author = {Kingma, Diederik and Ba, Jimmy},
file = {/Users/fergalcotter/Dropbox/Papers/Kingma_Ba_2014_Adam.pdf;/Users/fergalcotter/Zotero/storage/33XDZ7EF/1412.html}
}
@article{soatto_visual_2014,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1411.7676},
primaryClass = {cs},
title = {Visual {{Representations}}: {{Defining Properties}} and {{Deep Approximations}}},
url = {http://arxiv.org/abs/1411.7676},
shorttitle = {Visual {{Representations}}},
abstract = {Visual representations are defined in terms of minimal sufficient statistics of visual data, for a class of tasks, that are also invariant to nuisance variability. Minimal sufficiency guarantees that we can store a representation in lieu of raw data with smallest complexity and no performance loss on the task at hand. Invariance guarantees that the statistic is constant with respect to uninformative transformations of the data. We derive analytical expressions for such representations and show they are related to feature descriptors commonly used in computer vision, as well as to convolutional neural networks. This link highlights the assumptions and approximations tacitly assumed by these methods and explains empirical practices such as clamping, pooling and joint normalization.},
urldate = {2016-08-06},
date = {2014-11-27},
keywords = {Computer Science - Computer Vision and Pattern Recognition},
author = {Soatto, Stefano and Chiuso, Alessandro},
file = {/Users/fergalcotter/Dropbox/Papers/Soatto_Chiuso_2014_Visual Representations.pdf;/Users/fergalcotter/Zotero/storage/A9CR6RD2/1411.html},
note = {00000}
}
@incollection{lecun_efficient_2012,
langid = {english},
title = {Efficient {{BackProp}}},
isbn = {978-3-642-35288-1},
url = {http://link.springer.com/chapter/10.1007/978-3-642-35289-8_3},
abstract = {The convergence of back-propagation learning is analyzed so as to explain common phenomenon observed by practitioners. Many undesirable behaviors of backprop can be avoided with tricks that are rarely exposed in serious technical publications. This paper gives some of those tricks, and offers explanations of why they work. Many authors have suggested that second-order optimization methods are advantageous for neural net training. It is shown that most “classical” second-order methods are impractical for large neural networks. A few methods are proposed that do not have these limitations.},
number = {7700},
booktitle = {Neural {{Networks}}: {{Tricks}} of the {{Trade}}},
series = {Lecture {{Notes}} in {{Computer Science}}},
publisher = {{Springer Berlin Heidelberg}},
urldate = {2016-08-09},
date = {2012},
pages = {9-48},
keywords = {Algorithm Analysis and Problem Complexity,Artificial Intelligence (incl. Robotics),Complexity,Computation by Abstract Devices,Information Systems Applications (incl. Internet),pattern recognition},
author = {LeCun, Yann A. and Bottou, Léon and Orr, Genevieve B. and Müller, Klaus-Robert},
editor = {Montavon, Grégoire and Orr, Geneviève B. and Müller, Klaus-Robert},
file = {/Users/fergalcotter/Dropbox/Papers/LeCun et al_2012_Efficient BackProp.pdf;/Users/fergalcotter/Zotero/storage/TPNIZZTW/978-3-642-35289-8_3.html},
doi = {10.1007/978-3-642-35289-8_3},
note = {01085}
}
@article{saxe_exact_2013,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1312.6120},
primaryClass = {cond-mat, q-bio, stat},
title = {Exact Solutions to the Nonlinear Dynamics of Learning in Deep Linear Neural Networks},
url = {http://arxiv.org/abs/1312.6120},
abstract = {Despite the widespread practical success of deep learning methods, our theoretical understanding of the dynamics of learning in deep neural networks remains quite sparse. We attempt to bridge the gap between the theory and practice of deep learning by systematically analyzing learning dynamics for the restricted case of deep linear neural networks. Despite the linearity of their input-output map, such networks have nonlinear gradient descent dynamics on weights that change with the addition of each new hidden layer. We show that deep linear networks exhibit nonlinear learning phenomena similar to those seen in simulations of nonlinear networks, including long plateaus followed by rapid transitions to lower error solutions, and faster convergence from greedy unsupervised pretraining initial conditions than from random initial conditions. We provide an analytical description of these phenomena by finding new exact solutions to the nonlinear dynamics of deep learning. Our theoretical analysis also reveals the surprising finding that as the depth of a network approaches infinity, learning speed can nevertheless remain finite: for a special class of initial conditions on the weights, very deep networks incur only a finite, depth independent, delay in learning speed relative to shallow networks. We show that, under certain conditions on the training data, unsupervised pretraining can find this special class of initial conditions, while scaled random Gaussian initializations cannot. We further exhibit a new class of random orthogonal initial conditions on weights that, like unsupervised pre-training, enjoys depth independent learning times. We further show that these initial conditions also lead to faithful propagation of gradients even in deep nonlinear networks, as long as they operate in a special regime known as the edge of chaos.},
urldate = {2016-08-09},
date = {2013-12-20},
keywords = {Computer Science - Computer Vision and Pattern Recognition,Computer Science - Learning,Computer Science - Neural and Evolutionary Computing,Condensed Matter - Disordered Systems and Neural Networks,Quantitative Biology - Neurons and Cognition,Statistics - Machine Learning},
author = {Saxe, Andrew M. and McClelland, James L. and Ganguli, Surya},
file = {/Users/fergalcotter/Dropbox/Papers/Saxe et al_2013_Exact solutions to the nonlinear dynamics of learning in deep linear neural.pdf;/Users/fergalcotter/Zotero/storage/BKC39HTK/1312.html},
note = {00087}
}
@inproceedings{zhang_image_2009,
location = {{Taipei, Taiwan}},
title = {Image Deconvolution Using a {{Gaussian Scale Mixtures}} Model to Approximate the Wavelet Sparseness Constraint},
url = {http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=4959675},
eventtitle = {2009 {{IEEE International Conference}} on {{Acoustics}}, {{Speech}}, and {{Signal Processing}} ({{ICASSP}})},
booktitle = {Proceedings of 2009 {{IEEE International Conference}} on {{Acoustics}}, {{Speech}}, and {{Signal Processing}} ({{ICASSP}})},
publisher = {{IEEE}},
urldate = {2015-11-03},
date = {2009-04},
pages = {681--684},
author = {Zhang, Yingsong and Kingsbury, Nick},
file = {/Users/fergalcotter/Dropbox/Papers/Zhang_Kingsbury_2009_Image deconvolution using a Gaussian Scale Mixtures model to approximate the.pdf},
note = {00009}
}
@report{radoslaw_mantiuk_-it-yourself_????,
title = {Do-{{It}}-{{Yourself Eye Tracker}}: {{Low}}-{{Cost Pupil}}-{{Based Eye Tracker}} for {{Computer Graphics Applications}}},
abstract = {Eye tracking technologies offer sophisticated methods for
capturing humans’ gaze direction but their popularity in multimedia and
computer graphics systems is still low. One of the main reasons for this
are the high cost of commercial eye trackers that comes to 25,000 euros.
Interestingly, this price seems to stem from the costs incurred in research
rather than the value of used hardware components. In this work we show
that an eye tracker of a satisfactory precision can be built in the budget
of 30 euros. In the paper detailed instruction on how to construct a low
cost pupil-based eye tracker and utilise open source software to control
its behaviour is presented. We test the accuracy of our eye tracker and reveal
that its precision is comparable to commercial video-based devices.
We give an example of application in which our eye tracker is used to
control the depth-of-field rendering in real time virtual environment.},
institution = {{West Pomeranian University of Technology in Szczecin, Faculty of Computer Science}},
author = {{Radoslaw Mantiuk} and {Michal Kowalik} and {Adam Nowosielski} and {Bartosz Bazyluk}},
file = {/Users/fergalcotter/Dropbox/Papers/Radoslaw Mantiuk et al_Do-It-Yourself Eye Tracker.pdf},
note = {00032}
}
@article{ghahramani_probabilistic_2015,
langid = {english},
title = {Probabilistic Machine Learning and Artificial Intelligence},
volume = {521},
issn = {0028-0836},
url = {http://www.nature.com/nature/journal/v521/n7553/full/nature14541.html},
doi = {10.1038/nature14541},
abstract = {How can a machine learn from experience? Probabilistic modelling provides a framework for understanding what learning is, and has therefore emerged as one of the principal theoretical and practical approaches for designing machines that learn from data acquired through experience. The probabilistic framework, which describes how to represent and manipulate uncertainty about models and predictions, has a central role in scientific data analysis, machine learning, robotics, cognitive science and artificial intelligence. This Review provides an introduction to this framework, and discusses some of the state-of-the-art advances in the field, namely, probabilistic programming, Bayesian optimization, data compression and automatic model discovery.},
number = {7553},
journaltitle = {Nature},
shortjournal = {Nature},
urldate = {2016-10-25},
date = {2015-05-28},
pages = {452-459},
keywords = {Computer science,Mathematics and computing,Neuroscience},
author = {Ghahramani, Zoubin},
file = {/Users/fergalcotter/Dropbox/Papers/Ghahramani_2015_Probabilistic machine learning and artificial intelligence.pdf;/Users/fergalcotter/Zotero/storage/2BDTM2GI/nature14541.html}
}
@inproceedings{shi_real-time_2016,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1609.05158},
location = {{Las Vegas, NV}},
title = {Real-{{Time Single Image}} and {{Video Super}}-{{Resolution Using}} an {{Efficient Sub}}-{{Pixel Convolutional Neural Network}}},
url = {http://arxiv.org/abs/1609.05158},
abstract = {Recently, several models based on deep neural networks have achieved great success in terms of both reconstruction accuracy and computational performance for single image super-resolution. In these methods, the low resolution (LR) input image is upscaled to the high resolution (HR) space using a single filter, commonly bicubic interpolation, before reconstruction. This means that the super-resolution (SR) operation is performed in HR space. We demonstrate that this is sub-optimal and adds computational complexity. In this paper, we present the first convolutional neural network (CNN) capable of real-time SR of 1080p videos on a single K2 GPU. To achieve this, we propose a novel CNN architecture where the feature maps are extracted in the LR space. In addition, we introduce an efficient sub-pixel convolution layer which learns an array of upscaling filters to upscale the final LR feature maps into the HR output. By doing so, we effectively replace the handcrafted bicubic filter in the SR pipeline with more complex upscaling filters specifically trained for each feature map, whilst also reducing the computational complexity of the overall SR operation. We evaluate the proposed approach using images and videos from publicly available datasets and show that it performs significantly better (+0.15dB on Images and +0.39dB on Videos) and is an order of magnitude faster than previous CNN-based methods.},
eventtitle = {2016 {{IEEE Conference}} on {{Computer Vision}} and {{Pattern Recognition}} ({{CVPR}})},
booktitle = {Proceedings of 2016 {{IEEE Conference}} on {{Computer Vision}} and {{Pattern Recognition}} ({{CVPR}})},
publisher = {{IEEE}},
urldate = {2016-10-18},
date = {2016-09-16},
keywords = {Computer Science - Computer Vision and Pattern Recognition,Statistics - Machine Learning},
author = {Shi, Wenzhe and Caballero, Jose and Huszár, Ferenc and Totz, Johannes and Aitken, Andrew P. and Bishop, Rob and Rueckert, Daniel and Wang, Zehan},
file = {/Users/fergalcotter/Dropbox/Papers/1609.07009.pdf;/Users/fergalcotter/Dropbox/Papers/Shi et al_2016_Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel.pdf;/Users/fergalcotter/Zotero/storage/E2HRVWZA/1609.html}
}
@article{sifre_rigid-motion_2014,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1403.1687},
primaryClass = {cs},
title = {Rigid-{{Motion Scattering}} for {{Texture Classification}}},
url = {http://arxiv.org/abs/1403.1687},
abstract = {A rigid-motion scattering computes adaptive invariants along translations and rotations, with a deep convolutional network. Convolutions are calculated on the rigid-motion group, with wavelets defined on the translation and rotation variables. It preserves joint rotation and translation information, while providing global invariants at any desired scale. Texture classification is studied, through the characterization of stationary processes from a single realization. State-of-the-art results are obtained on multiple texture data bases, with important rotation and scaling variabilities.},
urldate = {2015-11-30},
date = {2014-03-07},
keywords = {Computer Science - Computer Vision and Pattern Recognition,Unread},
author = {Sifre, Laurent and Mallat, Stéphane},
file = {/Users/fergalcotter/Dropbox/Papers/Sifre_Mallat_2014_Rigid-Motion Scattering for Texture Classification.pdf;/Users/fergalcotter/Zotero/storage/BGFJPMFX/1403.html},
note = {00003}
}
@article{srivastava_training_2015,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1507.06228},
primaryClass = {cs},
title = {Training {{Very Deep Networks}}},
url = {http://arxiv.org/abs/1507.06228},
abstract = {Theoretical and empirical evidence indicates that the depth of neural networks is crucial for their success. However, training becomes more difficult as depth increases, and training of very deep networks remains an open problem. Here we introduce a new architecture designed to overcome this. Our so-called highway networks allow unimpeded information flow across many layers on information highways. They are inspired by Long Short-Term Memory recurrent networks and use adaptive gating units to regulate the information flow. Even with hundreds of layers, highway networks can be trained directly through simple gradient descent. This enables the study of extremely deep and efficient architectures.},
urldate = {2016-08-07},
date = {2015-07-22},
keywords = {68T01,Computer Science - Learning,Computer Science - Neural and Evolutionary Computing,G.1.6,I.2.6},
author = {Srivastava, Rupesh Kumar and Greff, Klaus and Schmidhuber, Jürgen},
file = {/Users/fergalcotter/Dropbox/Papers/Srivastava et al_2015_Training Very Deep Networks.pdf;/Users/fergalcotter/Zotero/storage/3CX7XUPJ/1507.html},
note = {00064}
}
@article{hull_database_1994,
title = {A Database for Handwritten Text Recognition Research},
volume = {16},
issn = {0162-8828},
doi = {10.1109/34.291440},
abstract = {An image database for handwritten text recognition research is described. Digital images of approximately 5000 city names, 5000 state names, 10000 ZIP Codes, and 50000 alphanumeric characters are included. Each image was scanned from mail in a working post office at 300 pixels/in in 8-bit gray scale on a high-quality flat bed digitizer. The data were unconstrained for the writer, style, and method of preparation. These characteristics help overcome the limitations of earlier databases that contained only isolated characters or were prepared in a laboratory setting under prescribed circumstances. Also, the database is divided into explicit training and testing sets to facilitate the sharing of results among researchers as well as performance comparisons},
number = {5},
journaltitle = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
date = {1994-05},
pages = {550-554},
keywords = {8 bit,Character recognition,Cities and towns,Gray-scale,Handwriting recognition,Image databases,Performance analysis,Postal services,Testing,Text recognition,Writing,alphanumeric characters,digital images,flat bed digitizer,gray scale,handwritten text recognition,image database,performance comparisons,performance evaluation,style,visual databases,writer},
author = {Hull, J. J.},
file = {/Users/fergalcotter/Dropbox/Papers/Hull_1994_A database for handwritten text recognition research.pdf;/Users/fergalcotter/Zotero/storage/PG7BZX6P/abs_all.html},
note = {00905}
}
@inproceedings{kingsbury_design_2003,
location = {{Barcelona, Spain}},
title = {Design of {{Q}}-Shift Complex Wavelets for Image Processing Using Frequency Domain Energy Minimization},
volume = {1},
doi = {10.1109/ICIP.2003.1247137},
abstract = {This paper proposes a new method of designing finite-support wavelet filters, based on minimization of energy in key parts of the frequency domain. In particular this technique is shown to be very effective for designing families of filters that are suitable for use in the shift-invariant dual-tree complex wavelet structure that has been developed by the author recently, and has been shown to be important for a range of image processing applications. The dual-tree structure requires most of the wavelet filters to have a well-controlled group delay, equivalent to one quarter of a sample period, in order to achieve optimal shift invariance. The proposed new design technique allows this requirement to be included along with the usual smoothness and perfect reconstruction properties to yield wavelet filters with a unique combination of features: linear phase, tight frame, compact spatial support, good frequency domain selectivity with low sidelobe levels, approximate shift invariance, and good directional selectivity in two or more dimensions.},
eventtitle = {2003 {{IEEE International Conference}} on {{Image Processing}} ({{ICIP}})},
booktitle = {2003 {{IEEE International Conference}} on {{Image Processing}} ({{ICIP}})},
date = {2003-09},
keywords = {Continuous wavelet transforms,Design engineering,Discrete wavelet transforms,Filters,Frequency domain analysis,frequency domain energy minimization,frequency domain selectivity,group delay,image processing,low sidelobe level,minimisation,Minimization methods,Multidimensional signal processing,optimal shift invariance,Propagation delay,Q-shift complex wavelet transform,shift-invariant dual-tree complex wavelet structure,Wavelet domain,wavelet filter,wavelet transforms},
author = {Kingsbury, N.},
file = {/Users/fergalcotter/Dropbox/Papers/Kingsbury_2003_Design of Q-shift complex wavelets for image processing using frequency domain.pdf;/Users/fergalcotter/Zotero/storage/JHCS4WSW/1247137.html},
note = {00115}
}
@thesis{waldspurger_wavelet_2012,
location = {{Paris, France}},
title = {Wavelet Transform Modulus: Phase Retrieval and Scattering},
abstract = {Automatically understanding the content of a natural signal, like a sound or an image, is in
general a difficult task. In their naive representation, signals are indeed complicated objects,
belonging to high-dimensional spaces. With a different representation, they can however be
easier to interpret.
This thesis considers a representation commonly used in these cases, in particular for the
analysis of audio signals: the modulus of the wavelet transform. To better understand the
behaviour of this operator, we study, from a theoretical as well as algorithmic point of view, the
corresponding inverse problem: the reconstruction of a signal from the modulus of its wavelet
transform.
This problem belongs to a wider class of inverse problems: phase retrieval problems. In a
first chapter, we describe a new algorithm, PhaseCut, which numerically solves a generic phase
retrieval problem. Like the similar algorithm PhaseLift, PhaseCut relies on a convex relaxation
of the phase retrieval problem, which happens to be of the same form as relaxations of the widely
studied problem MaxCut. We compare the performances of PhaseCut and PhaseLift, in terms
of precision and complexity.
In the next two chapters, we study the specific case of phase retrieval for the wavelet transform.
We show that any function with no negative frequencies is uniquely determined (up to
a global phase) by the modulus of its wavelet transform, but that the reconstruction from the
modulus is not stable to noise, for a strong notion of stability. However, we prove a local stability
property. We also present a new non-convex phase retrieval algorithm, which is specific to the
case of the wavelet transform, and we numerically study its performances.
Finally, in the last two chapters, we study a more sophisticated representation, built from
the modulus of the wavelet transform: the scattering transform. Our goal is to understand
which properties of a signal are characterized by its scattering transform. We first prove that
the energy of scattering coefficients of a signal, at a given order, is upper bounded by the energy
of the signal itself, convolved with a high-pass filter that depends on the order. We then study
a generalization of the scattering transform, for stationary processes. We show that, in finite
dimension, this generalized transform preserves the norm. In dimension one, we also show that
the generalized scattering coefficients of a process characterize the tail of its distribution.},
pagetotal = {210},
institution = {{École Normale Supérieure}},
type = {PhD Thesis},
date = {2012-11},
keywords = {Unread},
author = {Waldspurger, Irene},
file = {/Users/fergalcotter/Dropbox/Papers/Waldspurger_2012_Wavelet transform modulus.pdf},
note = {00000}
}
@article{selesnick_hilbert_2001,
title = {Hilbert Transform Pairs of Wavelet Bases},
volume = {8},
issn = {1070-9908},
doi = {10.1109/97.923042},
abstract = {This paper considers the design of pairs of wavelet bases where the wavelets form a Hilbert transform pair. The derivation is based on the limit functions defined by the infinite product formula. It is found that the scaling filters should be offset from one another by a half sample. This gives an alternative derivation and explanation for the result by Kingsbury (1999), that the dual-tree DWT is (nearly) shift-invariant when the scaling filters satisfy the same offset.},
number = {6},
journaltitle = {IEEE Signal Processing Letters},
date = {2001-06},
pages = {170-173},
keywords = {Delay,Discrete transforms,Discrete wavelet transforms,Fourier transforms,Hilbert transform pairs,Hilbert transforms,Transient analysis,Wavelet analysis,dual-tree DWT,filter bank,filtering theory,half sample,infinite product formula,limit functions,scaling filters,shift-invariant,signal processing,wavelet bases,wavelet transforms,Encoding},
author = {Selesnick, I.W.},
file = {/Users/fergalcotter/Dropbox/Papers/Wavelets and DTCWT/Selesnick_2001_Hilbert transform pairs of wavelet bases.pdf;/Users/fergalcotter/Zotero/storage/GW5DTFXB/Selesnick_2001_Hilbert transform pairs of wavelet bases.pdf;/Users/fergalcotter/Zotero/storage/XE22KAT7/abs_all.html},
note = {00351}
}
@inproceedings{kingsbury_shift_1999,
location = {{Phoenix, AZ, USA}},
title = {Shift Invariant Properties of the Dual-Tree Complex Wavelet Transform},
abstract = {We discuss the shift invariant properties of a new implementation of the Discrete Wavelet Transform, which employs a dual tree of wavelet filters to obtain the real and imaginary parts of complex: wavelet coefficients. This introduces limited redundancy (2(m):1 for m-dimensional signals) and allows the transform to provide approximate shift invariance and directionally selective filters (properties lacking in the traditional wavelet transform) while preserving the usual properties of perfect reconstruction and computational efficiency with good well-balanced frequency responses.},
eventtitle = {1999 {{IEEE International Conference}} on {{Acoustics}}, {{Speech}}, and {{Signal Processing}} ({{ICASSP}})},
booktitle = {Proceedings of 1999 {{IEEE International Conference}} on {{Acoustics}}, {{Speech}}, and {{Signal Processing}} ({{ICASSP}})},
publisher = {{IEEE}},
date = {1999},
pages = {1221-1224},
author = {Kingsbury, N.},
file = {/Users/fergalcotter/Dropbox/Papers/Kingsbury_1999_Shift invariant properties of the dual-tree complex wavelet transform.pdf},
note = {00234}
}
@article{gatys_neural_2015,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1508.06576},
primaryClass = {cs, q-bio},
title = {A {{Neural Algorithm}} of {{Artistic Style}}},
url = {http://arxiv.org/abs/1508.06576},
abstract = {In fine art, especially painting, humans have mastered the skill to create unique visual experiences through composing a complex interplay between the content and style of an image. Thus far the algorithmic basis of this process is unknown and there exists no artificial system with similar capabilities. However, in other key areas of visual perception such as object and face recognition near-human performance was recently demonstrated by a class of biologically inspired vision models called Deep Neural Networks. Here we introduce an artificial system based on a Deep Neural Network that creates artistic images of high perceptual quality. The system uses neural representations to separate and recombine content and style of arbitrary images, providing a neural algorithm for the creation of artistic images. Moreover, in light of the striking similarities between performance-optimised artificial neural networks and biological vision, our work offers a path forward to an algorithmic understanding of how humans create and perceive artistic imagery.},
urldate = {2016-02-02},
date = {2015-08-26},
keywords = {Computer Science - Computer Vision and Pattern Recognition,Computer Science - Neural and Evolutionary Computing,Quantitative Biology - Neurons and Cognition},
author = {Gatys, Leon A. and Ecker, Alexander S. and Bethge, Matthias},
file = {/Users/fergalcotter/Dropbox/Papers/Gatys et al_2015_A Neural Algorithm of Artistic Style.pdf;/Users/fergalcotter/Zotero/storage/I3FS73XK/1508.html},
note = {00007}
}
@inproceedings{szegedy_going_2015,
location = {{Boston, MA, USA}},
title = {Going {{Deeper With Convolutions}}},
url = {http://www.cv-foundation.org/openaccess/content_cvpr_2015/html/Szegedy_Going_Deeper_With_2015_CVPR_paper.html},
eventtitle = {2015 {{IEEE Conference}} on {{Computer Vision}} and {{Pattern Recognition}} ({{CVPR}})},
booktitle = {Proceedings of 2015 {{IEEE Conference}} on {{Computer Vision}} and {{Pattern Recognition}} ({{CVPR}})},
publisher = {{IEEE}},
urldate = {2015-11-29},
date = {2015-06},
pages = {1-9},
keywords = {Key Paper,Unread},
author = {Szegedy, Christian and Liu, Wei and Jia, Yangqing and Sermanet, Pierre and Reed, Scott and Anguelov, Dragomir and Erhan, Dumitru and Vanhoucke, Vincent and Rabinovich, Andrew},
file = {/Users/fergalcotter/Dropbox/Papers/Szegedy et al_2015_Going Deeper With Convolutions.pdf;/Users/fergalcotter/Zotero/storage/E7T97XTB/Szegedy_Going_Deeper_With_2015_CVPR_paper.html},
note = {00569}
}
@book{cormen_introduction_2009,
langid = {english},
location = {{Cambridge, Mass}},
title = {Introduction to {{Algorithms}}, 3rd {{Edition}}},
edition = {3rd edition},
isbn = {978-0-262-03384-8},
abstract = {Some books on algorithms are rigorous but incomplete; others cover masses of material but lack rigor. Introduction to Algorithms uniquely combines rigor and comprehensiveness. The book covers a broad range of algorithms in depth, yet makes their design and analysis accessible to all levels of readers. Each chapter is relatively self-contained and can be used as a unit of study. The algorithms are described in English and in a pseudocode designed to be readable by anyone who has done a little programming. The explanations have been kept elementary without sacrificing depth of coverage or mathematical rigor.The first edition became a widely used text in universities worldwide as well as the standard reference for professionals. The second edition featured new chapters on the role of algorithms, probabilistic analysis and randomized algorithms, and linear programming. The third edition has been revised and updated throughout. It includes two completely new chapters, on van Emde Boas trees and multithreaded algorithms, substantial additions to the chapter on recurrence (now called "Divide-and-Conquer"), and an appendix on matrices. It features improved treatment of dynamic programming and greedy algorithms and a new notion of edge-based flow in the material on flow networks. Many new exercises and problems have been added for this edition. As of the third edition, this textbook is published exclusively by the MIT Press.},
pagetotal = {1312},
publisher = {{The MIT Press}},
date = {2009-07-31},
author = {Cormen, Thomas and Leiserson, Charles and Rivest, Ronald and Stein, Clifford},
file = {C:\\Users\\fbc23\\Google Drive\\Papers\\Books and Notes\\introduction-to-algorithms-3rd-edition.pdf},
note = {00000}
}
@inproceedings{kingsbury_dual-tree_1998,
location = {{Utah}},
title = {The {{Dual}}-{{Tree Complex Wavelet Transform}}: {{A New Technique For Shift Invariance And Directional Filters}}},
shorttitle = {The {{Dual}}-{{Tree Complex Wavelet Transform}}},
abstract = {A new implementation of the Discrete Wavelet Transform is presented, suitable for a range of signal and image processing applications. It employs a dual tree of wavelet filters to obtain the real and imaginary parts of complex wavelet coefficients. This introduces limited redundancy (4:1 for 2-dimensional signals) and allows the transform to provide approximate shift invariance and directionally selective filters (properties lacking in the traditional wavelet transform) while preserving the usual properties of perfect reconstruction and computational efficiency. An application to texture synthesis is presented. 1. INTRODUCTION Although the Discrete Wavelet Transform (DWT) in its maximally decimated form (Mallat's dyadic filter tree [1]) has established an impressive reputation as a tool for image compression, its use for other signal analysis and reconstruction tasks has been hampered by two main disadvantages: ffl Lack of shift invariance, which means that small shifts in the input...},
eventtitle = {1998 8th {{International Conference}} on {{Digital Signal Processing}} ({{DSP}})},
booktitle = {1998 8th {{International Conference}} on {{Digital Signal Processing}} ({{DSP}})},
publisher = {{IEEE}},
date = {1998-08},
pages = {319-322},
author = {Kingsbury, Nick},
file = {/Users/fergalcotter/Dropbox/Papers/Kingsbury_1998_The Dual-Tree Complex Wavelet Transform.pdf;/Users/fergalcotter/Zotero/storage/ZXPH2C6I/summary.html},
note = {00586}
}
@incollection{yaeger_combining_2012,
langid = {english},
title = {Combining {{Neural Networks}} and {{Context}}-{{Driven Search}} for {{On}}-Line, {{Printed Handwriting Recognition}} in the {{Newton}}},
isbn = {978-3-642-35288-1 978-3-642-35289-8},
url = {http://link.springer.com/chapter/10.1007/978-3-642-35289-8_18},
abstract = {While on-line handwriting recognition is an area of long-standing and ongoing research, the recent emergence of portable, pen-based computers has focused urgent attention on usable, practical solutions. We discuss a combination and improvement of classical methods to produce robust recognition of hand-printed English text, for a recognizer shipping in new models of Apple Computer’s Newton MessagePad® and eMate®. Combining an artificial neural network (ANN), as a character classifier, with a context-driven search over segmentation and word recognition hypotheses provides an effective recognition system. Long-standing issues relative to training, generalization, segmentation, models of context, probabilistic formalisms, etc., need to be resolved, however, to get excellent performance. We present a number of recent innovations in the application of ANNs as character classifiers for word recognition, including integrated multiple representations, normalized output error, negative training, stroke warping, frequency balancing, error emphasis, and quantized weights. User-adaptation and extension to cursive recognition pose continuing challenges.},
number = {7700},
booktitle = {Neural {{Networks}}: {{Tricks}} of the {{Trade}}},
series = {Lecture {{Notes}} in {{Computer Science}}},
publisher = {{Springer Berlin Heidelberg}},
urldate = {2016-08-09},
date = {2012},
pages = {271-293},
keywords = {Algorithm Analysis and Problem Complexity,Artificial Intelligence (incl. Robotics),Complexity,Computation by Abstract Devices,Information Systems Applications (incl. Internet),pattern recognition},
author = {Yaeger, Larry S. and Webb, Brandyn J. and Lyon, Richard F.},
editor = {Montavon, Grégoire and Orr, Geneviève B. and Müller, Klaus-Robert},
file = {/Users/fergalcotter/Dropbox/Papers/Yaeger et al_2012_Combining Neural Networks and Context-Driven Search for On-line, Printed.pdf;/Users/fergalcotter/Zotero/storage/IZ5CECMA/978-3-642-35289-8_18.html},
doi = {10.1007/978-3-642-35289-8_18},
note = {00000}
}
@article{shuman_emerging_2013,
title = {The Emerging Field of Signal Processing on Graphs: {{Extending}} High-Dimensional Data Analysis to Networks and Other Irregular Domains},
volume = {30},
issn = {1053-5888},
doi = {10.1109/MSP.2012.2235192},
shorttitle = {The Emerging Field of Signal Processing on Graphs},
abstract = {In applications such as social, energy, transportation, sensor, and neuronal networks, high-dimensional data naturally reside on the vertices of weighted graphs. The emerging field of signal processing on graphs merges algebraic and spectral graph theoretic concepts with computational harmonic analysis to process such signals on graphs. In this tutorial overview, we outline the main challenges of the area, discuss different ways to define graph spectral domains, which are the analogs to the classical frequency domain, and highlight the importance of incorporating the irregular structures of graph data domains when processing signals on graphs. We then review methods to generalize fundamental operations such as filtering, translation, modulation, dilation, and downsampling to the graph setting and survey the localized, multiscale transforms that have been proposed to efficiently extract information from high-dimensional data on graphs. We conclude with a brief discussion of open issues and possible extensions.},
number = {3},
journaltitle = {IEEE Signal Processing Magazine},
date = {2013-05},
pages = {83-98},
keywords = {Biological neural networks,Frequency domain analysis,Harmonic analysis,Spectral analysis,Tutorials,classical frequency domain,computational harmonic analysis,data analysis,data structures,graph spectral domains,graph theory,high-dimensional data analysis,high-dimensional graph data,information extraction,irregular graph data structures,open issues,signal processing,spectral graph theoretic concepts,weighted graphs,Feature extraction},
author = {Shuman, D. I. and Narang, S. K. and Frossard, P. and Ortega, A. and Vandergheynst, P.},
file = {/Users/fergalcotter/Dropbox/Papers/Shuman et al_2013_The emerging field of signal processing on graphs.pdf;/Users/fergalcotter/Zotero/storage/3FAS2I8W/6494675.html},
note = {00414}
}
@article{bruna_invariant_2013,
title = {Invariant {{Scattering Convolution Networks}}},
volume = {35},
issn = {0162-8828},
doi = {10.1109/TPAMI.2012.230},
abstract = {A wavelet scattering network computes a translation invariant image representation which is stable to deformations and preserves high-frequency information for classification. It cascades wavelet transform convolutions with nonlinear modulus and averaging operators. The first network layer outputs SIFT-type descriptors, whereas the next layers provide complementary invariant information that improves classification. The mathematical analysis of wavelet scattering networks explains important properties of deep convolution networks for classification. A scattering representation of stationary processes incorporates higher order moments and can thus discriminate textures having the same Fourier power spectrum. State-of-the-art classification results are obtained for handwritten digits and texture discrimination, with a Gaussian kernel SVM and a generative PCA classifier.},
number = {8},
journaltitle = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
date = {2013-08},
pages = {1872-1886},
keywords = {Classification,Computer architecture,Convolution,Fourier power spectrum,Fourier transforms,Gaussian kernel SVM,Gaussian processes,Key Paper,SIFT-type descriptors,Scattering,Similar Work,Wavelet coefficients,averaging operators,complementary invariant information,convolution networks,deep convolution networks,deformations,generative PCA classifier,handwritten character recognition,handwritten digits,high-frequency information,image classification,image texture,invariant scattering convolution networks,invariants,mathematical analysis,network layer,nonlinear modulus,principal component analysis,scattering representation,state-of-the-art classification,stationary process,support vector machines,texture discrimination,translation invariant image representation,wavelet scattering network,wavelet transform convolutions,wavelet transforms,wavelets,Image representation},
author = {Bruna, J. and Mallat, S.},
file = {/Users/fergalcotter/Dropbox/Papers/Bruna_Mallat_2013_Invariant Scattering Convolution Networks.pdf;/Users/fergalcotter/Zotero/storage/RCVIVNU4/articleDetails.html},
note = {00150}
}
@inproceedings{lecun_convolutional_2010,
title = {Convolutional Networks and Applications in Vision},
doi = {10.1109/ISCAS.2010.5537907},
abstract = {Intelligent tasks, such as visual perception, auditory perception, and language understanding require the construction of good internal representations of the world (or "features")? which must be invariant to irrelevant variations of the input while, preserving relevant information. A major question for Machine Learning is how to learn such good features automatically. Convolutional Networks (ConvNets) are a biologically-inspired trainable architecture that can learn invariant features. Each stage in a ConvNets is composed of a filter bank, some nonlinearities, and feature pooling layers. With multiple stages, a ConvNet can learn multi-level hierarchies of features. While ConvNets have been successfully deployed in many commercial applications from OCR to video surveillance, they require large amounts of labeled training samples. We describe new unsupervised learning algorithms, and new non-linear stages that allow ConvNets to be trained with very few labeled samples. Applications to visual object recognition and vision navigation for off-road mobile robots are described.},
eventtitle = {Proceedings of 2010 {{IEEE International Symposium}} on {{Circuits}} and {{Systems}} ({{ISCAS}})},
booktitle = {Proceedings of 2010 {{IEEE International Symposium}} on {{Circuits}} and {{Systems}} ({{ISCAS}})},
date = {2010-05},
pages = {253-256},
keywords = {ConvNets,Key Paper,Learning systems,Navigation,Optical character recognition software,Unread,Video surveillance,Visual perception,_tablet,biologically-inspired architecture,convolutional networks,feature pooling layers,filter bank,intelligent tasks,internal representations,labeled training samples,machine learning,mobile robots,multilevel hierarchies,object recognition,off-road mobile robots,robot vision,unsupervised learning,vision navigation,visual object recognition},
author = {LeCun, Y. and Kavukcuoglu, K. and Farabet, C.},
file = {/Users/fergalcotter/Dropbox/Papers/LeCun et al_2010_Convolutional networks and applications in vision.pdf;/Users/fergalcotter/Zotero/storage/JM4MPRJC/abs_all.html},
note = {00261}
}
@article{lake_human-level_2015,
langid = {english},
title = {Human-Level Concept Learning through Probabilistic Program Induction},
volume = {350},
issn = {0036-8075, 1095-9203},
url = {http://science.sciencemag.org/content/350/6266/1332},
doi = {10.1126/science.aab3050},
abstract = {Handwritten characters drawn by a model
Not only do children learn effortlessly, they do so quickly and with a remarkable ability to use what they have learned as the raw material for creating new stuff. Lake et al. describe a computational model that learns in a similar fashion and does so better than current deep learning algorithms. The model classifies, parses, and recreates handwritten characters, and can generate new letters of the alphabet that look “right” as judged by Turing-like tests of the model's output in comparison to what real humans produce.
Science, this issue p. 1332
People learning new concepts can often generalize successfully from just a single example, yet machine learning algorithms typically require tens or hundreds of examples to perform with similar accuracy. People can also use learned concepts in richer ways than conventional algorithms—for action, imagination, and explanation. We present a computational model that captures these human learning abilities for a large class of simple visual concepts: handwritten characters from the world’s alphabets. The model represents concepts as simple programs that best explain observed examples under a Bayesian criterion. On a challenging one-shot classification task, the model achieves human-level performance while outperforming recent deep learning approaches. We also present several “visual Turing tests” probing the model’s creative generalization abilities, which in many cases are indistinguishable from human behavior.
Combining the capacity to handle noise with probabilistic learning yields humanlike performance in a computational model.
Combining the capacity to handle noise with probabilistic learning yields humanlike performance in a computational model.},
number = {6266},
journaltitle = {Science},
urldate = {2016-02-25},
date = {2015-12-11},
pages = {1332-1338},
keywords = {Read Now},
author = {Lake, Brenden M. and Salakhutdinov, Ruslan and Tenenbaum, Joshua B.},
file = {/Users/fergalcotter/Dropbox/Papers/Lake et al_2015_Human-level concept learning through probabilistic program induction.pdf;/Users/fergalcotter/Zotero/storage/J46EU8T2/1332.html},
eprinttype = {pmid},
eprint = {26659050},
note = {00006}
}
@article{porat_localized_1989,
title = {Localized Texture Processing in Vision: Analysis and Synthesis in the {{Gaborian}} Space},
volume = {36},
issn = {0018-9294},
doi = {10.1109/10.16457},
shorttitle = {Localized Texture Processing in Vision},
abstract = {Recent studies of cortical simple cell function suggest that the primitives of image representation in vision have a wavelet form similar to Gabor elementary functions (EFs). It is shown that textures and fully textured images can be practically decomposed into, and synthesized from, a finite set of EFs. Textured-images can be synthesized from a set of EFs using an image coefficient library. Alternatively, texturing of contoured (cartoonlike) images is analogous to adding chromaticity information to contoured images. A method for texture discrimination and image segmentation using local features based on the Gabor approach is introduced. Features related to the EF's parameters provide efficient means for texture discrimination and classification. This method is invariant under rotation and translation. The performance of the classification appears to be robust with respect to noisy conditions. The results show the insensitivity of the discrimination to relatively high noise levels, comparable to the performances of the human observer.{$<>$}},
number = {1},
journaltitle = {IEEE Transactions on Biomedical Engineering},
date = {1989-01},
pages = {115-129},
keywords = {Artificial Intelligence,Bandwidth,Computer Simulation,Depth Perception,Frequency,Gaborian space,Humans,Image Processing; Computer-Assisted,Image edge detection,Image texture analysis,Libraries,Models; Neurological,Noise level,Visual perception,cartoonlike images,chromaticity information,contoured images,cortical simple cell function,image coefficient library,image representation primitives,image segmentation,localized texture processing,noise robustness,noisy conditions,texture discrimination,vision,visual analysis,visual processing,wavelet form,Image representation},
author = {Porat, M. and Zeevi, Y. Y.},
file = {/Users/fergalcotter/Dropbox/Papers/Porat_Zeevi_1989_Localized texture processing in vision.pdf;/Users/fergalcotter/Zotero/storage/2EMBC5J6/articleDetails.html},
note = {00223}
}
@article{serre_robust_2007,
title = {Robust {{Object Recognition}} with {{Cortex}}-{{Like Mechanisms}}},
volume = {29},
issn = {0162-8828},
doi = {10.1109/TPAMI.2007.56},
abstract = {We introduce a new general framework for the recognition of complex visual scenes, which is motivated by biology: We describe a hierarchical system that closely follows the organization of visual cortex and builds an increasingly complex and invariant feature representation by alternating between a template matching and a maximum pooling operation. We demonstrate the strength of the approach on a range of recognition tasks: From invariant single object recognition in clutter to multiclass categorization problems and complex scene understanding tasks that rely on the recognition of both shape-based as well as texture-based objects. Given the biological constraints that the system had to satisfy, the approach performs surprisingly well: It has the capability of learning from only a few training examples and competes with state-of-the-art systems. We also discuss the existence of a universal, redundant dictionary of features that could handle the recognition of most object categories. In addition to its relevance for computer vision, the success of this approach suggests a plausibility proof for a class of feedforward models of object recognition in cortex},
number = {3},
journaltitle = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
date = {2007-03},
pages = {411-426},
keywords = {Algorithms,Artificial Intelligence,Biomimetics,Brain modeling,Computer Simulation,Face detection,Gabor filters,Humans,Image Enhancement,Image Interpretation; Computer-Assisted,Layout,Models; Biological,Neuroscience,Pattern Recognition; Automated,Pattern Recognition; Visual,Reproducibility of Results,Robustness,Sensitivity and Specificity,Streaming media,Unread,complex visual scenes,computer vision,cortex-like mechanisms,image matching,model,multiclass categorization,neural network.,object recognition,robust object recognition,scene understanding,template matching,visual cortex},
author = {Serre, T. and Wolf, L. and Bileschi, S. and Riesenhuber, M. and Poggio, T.},
file = {/Users/fergalcotter/Dropbox/Papers/Serre et al_2007_Robust Object Recognition with Cortex-Like Mechanisms.pdf;/Users/fergalcotter/Zotero/storage/UB4KQ2DT/abs_all.html},
note = {01215}
}
@article{lecun_deep_2015,
langid = {english},
title = {Deep Learning},
volume = {521},
issn = {0028-0836},
url = {http://www.nature.com/nature/journal/v521/n7553/full/nature14539.html},
doi = {10.1038/nature14539},
abstract = {Deep learning allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. These methods have dramatically improved the state-of-the-art in speech recognition, visual object recognition, object detection and many other domains such as drug discovery and genomics. Deep learning discovers intricate structure in large data sets by using the backpropagation algorithm to indicate how a machine should change its internal parameters that are used to compute the representation in each layer from the representation in the previous layer. Deep convolutional nets have brought about breakthroughs in processing images, video, speech and audio, whereas recurrent nets have shone light on sequential data such as text and speech.},
number = {7553},
journaltitle = {Nature},
shortjournal = {Nature},
urldate = {2015-11-19},
date = {2015-05-28},
pages = {436-444},
keywords = {Computer science,Key Paper,Mathematics and computing},
author = {LeCun, Yann and Bengio, Yoshua and Hinton, Geoffrey},
file = {/Users/fergalcotter/Dropbox/Papers/Other Networks/LeCun et al_2015_Deep learning.pdf;/Users/fergalcotter/Zotero/storage/EDFAJTE5/nature14539.html},
note = {00049}
}
@article{hinton_fast_2006,
title = {A {{Fast Learning Algorithm}} for {{Deep Belief Nets}}},
volume = {18},
issn = {0899-7667},
url = {http://dx.doi.org/10.1162/neco.2006.18.7.1527},
doi = {10.1162/neco.2006.18.7.1527},
abstract = {We show how to use "complementary priors" to eliminate the explaining-away effects that make inference difficult in densely connected belief nets that have many hidden layers. Using complementary priors, we derive a fast, greedy algorithm that can learn deep, directed belief networks one layer at a time, provided the top two layers form an undirected associative memory. The fast, greedy algorithm is used to initialize a slower learning procedure that fine-tunes the weights using a contrastive version of the wake-sleep algorithm. After fine-tuning, a network with three hidden layers forms a very good generative model of the joint distribution of handwritten digit images and their labels. This generative model gives better digit classification than the best discriminative learning algorithms. The low-dimensional manifolds on which the digits lie are modeled by long ravines in the free-energy landscape of the top-level associative memory, and it is easy to explore these ravines by using the directed connections to display what the associative memory has in mind.},
number = {7},
journaltitle = {Neural Comput.},
urldate = {2016-07-28},
date = {2006-07},
pages = {1527--1554},
author = {Hinton, Geoffrey E. and Osindero, Simon and Teh, Yee-Whye},
file = {/Users/fergalcotter/Dropbox/Papers/Hinton et al_2006_A Fast Learning Algorithm for Deep Belief Nets.pdf},
note = {04328}
}
@article{porter_robust_1997,
title = {Robust Rotation-Invariant Texture Classification: Wavelet, {{Gabor}} Filter and {{GMRF}} Based Schemes},
volume = {144},
issn = {1350-245X},
doi = {10.1049/ip-vis:19971182},
shorttitle = {Robust Rotation-Invariant Texture Classification},
abstract = {Three novel feature extraction schemes for texture classification are proposed. The schemes employ the wavelet transform, a circularly symmetric Gabor filter or a Gaussian Markov random field with a circular neighbour set to achieve rotation-invariant texture classification. The schemes are shown to give a high level of classification accuracy compared to most existing schemes, using both fewer features (four) and a smaller area of analysis (16×16). Furthermore, unlike most existing schemes, the proposed schemes are shown to be rotation invariant demonstrate a high level of robustness noise. The performances of the three schemes are compared, indicating that the wavelet-based approach is the most accurate, exhibits the best noise performance and has the lowest computational complexity},
number = {3},
journaltitle = {IEE Proceedings - Vision, Image and Signal Processing},
date = {1997-06},
pages = {180-188},
keywords = {GMRF,Gaussian Markov random field,Gaussian processes,Markov processes,circular neighbour set,circularly symmetric Gabor filter,classification accuracy,computational complexity,filtering theory,image classification,image texture,noise,noise performance,noise robustness,random processes,robust rotation invariant texture classification,wavelet transform,wavelet transforms,Feature extraction},
author = {Porter, R. and Canagarajah, N.},
file = {/Users/fergalcotter/Dropbox/Papers/Porter_Canagarajah_1997_Robust rotation-invariant texture classification.pdf;/Users/fergalcotter/Zotero/storage/ASMUBX96/articleDetails.html},
note = {00210}
}
@inproceedings{pickering_object_2011,
location = {{London, UK}},
title = {Object Search Using Wavelet-Based Polar Matching for Aerial Imagery},
url = {http://digital-library.theiet.org/content/conferences/10.1049/ic.2011.0167},
eventtitle = {Sensor {{Signal Processing}} for {{Defence}} ({{SSPD}})},
booktitle = {Sensor {{Signal Processing}} for {{Defence}} ({{SSPD}})},
publisher = {{IEEE}},
urldate = {2015-11-03},
date = {2011-09},
author = {Pickering, Andy and Kingsbury, Nick},
file = {/Users/fergalcotter/Dropbox/Papers/Pickering_Kingsbury_2011_Object search using wavelet-based polar matching for aerial imagery.pdf},
note = {00000}
}
@article{simonyan_deep_2014,
title = {Deep {{Inside Convolutional Networks}}: {{Visualising Image Classification Models}} and {{Saliency Maps}}},
url = {http://arxiv.org/abs/1312.6034},
shorttitle = {Deep {{Inside Convolutional Networks}}},
abstract = {This paper addresses the visualisation of image classification models, learnt using deep Convolutional Networks (ConvNets). We consider two visualisation techniques, based on computing the gradient of the class score with respect to the input image. The first one generates an image, which maximises the class score [Erhan et al., 2009], thus visualising the notion of the class, captured by a ConvNet. The second technique computes a class saliency map, specific to a given image and class. We show that such maps can be employed for weakly supervised object segmentation using classification ConvNets. Finally, we establish the connection between the gradient-based ConvNet visualisation methods and deconvolutional networks [Zeiler et al., 2013].},
journaltitle = {Proceedings of the International Conference on Learning Representations (ICLR)},
urldate = {2015-12-01},
date = {2014},
keywords = {Computer Science - Computer Vision and Pattern Recognition,Unread,_tablet},
author = {Simonyan, Karen and Vedaldi, Andrea and Zisserman, Andrew},
file = {/Users/fergalcotter/Dropbox/Papers/Simonyan et al_2014_Deep Inside Convolutional Networks.pdf;/Users/fergalcotter/Zotero/storage/7GECRCNB/1312.html},
note = {00092}
}
@incollection{larsen_adaptive_2012,
langid = {english},
title = {Adaptive {{Regularization}} in {{Neural Network Modeling}}},
isbn = {978-3-642-35288-1 978-3-642-35289-8},
url = {http://link.springer.com/chapter/10.1007/978-3-642-35289-8_8},
abstract = {In this paper we address the important problem of optimizing regularization parameters in neural network modeling. The suggested optimization scheme is an extended version of the recently presented algorithm [25]. The idea is to minimize an empirical estimate - like the cross-validation estimate - of the generalization error with respect to regularization parameters. This is done by employing a simple iterative gradient descent scheme using virtually no additional programming overhead compared to standard training. Experiments with feed-forward neural network models for time series prediction and classification tasks showed the viability and robustness of the algorithm. Moreover, we provided some simple theoretical examples in order to illustrate the potential and limitations of the proposed regularization framework.},
number = {7700},
booktitle = {Neural {{Networks}}: {{Tricks}} of the {{Trade}}},
series = {Lecture {{Notes}} in {{Computer Science}}},
publisher = {{Springer Berlin Heidelberg}},
urldate = {2016-08-09},
date = {2012},
pages = {111-130},
keywords = {Algorithm Analysis and Problem Complexity,Artificial Intelligence (incl. Robotics),Complexity,Computation by Abstract Devices,Information Systems Applications (incl. Internet),pattern recognition},
author = {Larsen, Jan and Svarer, Claus and Andersen, Lars Nonboe and Hansen, Lars Kai},
editor = {Montavon, Grégoire and Orr, Geneviève B. and Müller, Klaus-Robert},
file = {/Users/fergalcotter/Dropbox/Papers/Larsen et al_2012_Adaptive Regularization in Neural Network Modeling.pdf;/Users/fergalcotter/Zotero/storage/EBC28MWN/978-3-642-35289-8_8.html},
doi = {10.1007/978-3-642-35289-8_8},
note = {00056}
}
@article{autor_why_2015,
title = {Why {{Are There Still So Many Jobs}}? {{The History}} and {{Future}} of {{Workplace Automation}}},
volume = {29},
issn = {0895-3309},
url = {https://www.aeaweb.org/articles?id=10.1257/jep.29.3.3},
doi = {10.1257/jep.29.3.3},
shorttitle = {Why {{Are There Still So Many Jobs}}?},
abstract = {In this essay, I begin by identifying the reasons that automation has not wiped out a majority of jobs over the decades and centuries.
Automation does indeed substitute for labor—as it is typically intended to do.
However, automation also complements labor, raises output in ways that leads to higher demand for labor, and interacts with adjustments in labor supply.
Journalists and even expert commentators tend to overstate the extent of machine substitution for human labor and ignore the strong complementarities between automation and labor that increase productivity, raise earnings, and augment demand for labor.
Changes in technology do alter the types of jobs available and what those jobs pay.
In the last few decades, one noticeable change has been a "polarization" of the labor market, in which wage gains went disproportionately to those at the top and at the bottom of the income and skill distribution, not to those in the middle; however, I also argue, this polarization is unlikely to continue very far into future.
The final section of this paper reflects on how recent and future advances in artificial intelligence and robotics should shape our thinking about the likely trajectory of occupational change and employment growth.
I argue that the interplay between machine and human comparative advantage allows computers to substitute for workers in performing routine, codifiable tasks while amplifying the comparative advantage of workers in supplying problem-solving skills, adaptability, and creativity.},
number = {3},
journaltitle = {Journal of Economic Perspectives},
urldate = {2016-07-20},
date = {2015-08},
pages = {3-30},
author = {Autor, David H.},
file = {/Users/fergalcotter/Dropbox/Papers/Autor_2015_Why Are There Still So Many Jobs.pdf;/Users/fergalcotter/Zotero/storage/VAATQBII/articles.html},
note = {00000}
}
@article{bovik_multichannel_1990,
title = {Multichannel Texture Analysis Using Localized Spatial Filters},
volume = {12},
issn = {0162-8828},
doi = {10.1109/34.41384},
abstract = {A computational approach for analyzing visible textures is described. Textures are modeled as irradiance patterns containing a limited range of spatial frequencies, where mutually distinct textures differ significantly in their dominant characterizing frequencies. By encoding images into multiple narrow spatial frequency and orientation channels, the slowly varying channel envelopes (amplitude and phase) are used to segregate textural regions of different spatial frequency, orientation, or phase characteristics. Thus, an interpretation of image texture as a region code, or carrier of region information, is emphasized. The channel filters used, known as the two-dimensional Gabor functions, are useful for these purposes in several senses: they have tunable orientation and radial frequency bandwidths and tunable center frequencies, and they optimally achieve joint resolution in space and in spatial frequency. By comparing the channel amplitude responses, one can detect boundaries between textures. Locating large variations in the channel phase responses allows discontinuities in the texture phase to be detected. Examples are given of both types of texture processing using a variety of real and synthetic textures},
number = {1},
journaltitle = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
date = {1990-01},
pages = {55-73},
keywords = {Demodulation,Frequency,Image analysis,Image texture analysis,Layout,Shape,Spatial filters,Surface texture,channel amplitude responses,discontinuities,filtering and prediction theory,image segmentation,image texture,irradiance patterns,joint resolution,localized spatial filters,multichannel texture analysis,pattern recognition,radial frequency bandwidths,region code,region information,tunable center frequencies,tunable orientation,two-dimensional Gabor functions,visible textures,Encoding},
author = {Bovik, A. C. and Clark, M. and Geisler, W. S.},
file = {/Users/fergalcotter/Dropbox/Papers/Bovik et al_1990_Multichannel texture analysis using localized spatial filters.pdf;/Users/fergalcotter/Zotero/storage/U83GQ6MT/abs_all.html},
note = {01720}
}
@inproceedings{glorot_understanding_2010,
location = {{Sardinia, Italy}},
title = {Understanding the Difficulty of Training Deep Feedforward Neural Networks},
abstract = {Whereas before 2006 it appears that deep multilayer neural networks were not successfully trained, since then several algorithms have been shown to successfully train them, with experimental results showing the superiority of deeper vs less deep architectures. All these experimental results were obtained with new initialization or training mechanisms. Our objective here is to understand better why standard gradient descent from random initialization is doing so poorly with deep neural networks, to better understand these recent relative successes and help design better algorithms in the future. We first observe the influence of the non-linear activations functions. We find that the logistic sigmoid activation is unsuited for deep networks with random initialization because of its mean value, which can drive especially the top hidden layer into saturation. Surprisingly, we find that saturated units can move out of saturation by themselves, albeit slowly, and explaining the plateaus sometimes seen when training neural networks. We find that a new non-linearity that saturates less can often be beneficial. Finally, we study how activations and gradients vary across layers and during training, with the idea that training may be more difficult when the singular values of the Jacobian associated with each layer are far from 1. Based on these considerations, we propose a new initialization scheme that brings substantially faster convergence. 1 Deep Neural Networks Deep learning methods aim at learning feature hierarchies with features from higher levels of the hierarchy formed by the composition of lower level features. They include},
eventtitle = {International {{Conference}} on {{Artificial Intelligence}} and {{Statistics}} ({{AISTATS}})},
booktitle = {Proceedings of the {{International Conference}} on {{Artificial Intelligence}} and {{Statistics}} ({{AISTATS}})},
date = {2010-05},
author = {Glorot, Xavier and Bengio, Yoshua},
file = {/Users/fergalcotter/Dropbox/Papers/Glorot_Bengio_2010_Understanding the difficulty of training deep feedforward neural networks.pdf;/Users/fergalcotter/Zotero/storage/TQFGSP6G/summary.html},
note = {00649}
}
@incollection{mairal_convolutional_2014-1,
title = {Convolutional {{Kernel Networks}}},
url = {http://papers.nips.cc/paper/5348-convolutional-kernel-networks.pdf},
booktitle = {Advances in {{Neural Information Processing Systems}} 27},
publisher = {{Curran Associates, Inc.}},
urldate = {2016-02-01},
date = {2014},
pages = {2627--2635},
keywords = {Unread},
author = {Mairal, Julien and Koniusz, Piotr and Harchaoui, Zaid and Schmid, Cordelia},
editor = {Ghahramani, Z. and Welling, M. and Cortes, C. and Lawrence, N. D. and Weinberger, K. Q.},
file = {/Users/fergalcotter/Dropbox/Papers/Mairal et al_2014_Convolutional Kernel Networks.pdf;/Users/fergalcotter/Zotero/storage/6R3EB42S/5348-convolutional-kernel-networks.html},
note = {00030}
}
@article{hyeonwoo_noh_learning_????,
title = {Learning {{Deconvolution Network}} for {{Semantic Segmentation}}},
author = {{Hyeonwoo Noh} and {Seunghoon Hong} and {Bohyung Han}},
file = {/Users/fergalcotter/Dropbox/Papers/CNNs/1505.04366v1.pdf},
note = {00073}
}
@inproceedings{kingsbury_rotation-invariant_2006,
location = {{Florence, Italy}},
title = {Rotation-Invariant Local Feature Matching with Complex Wavelets},
url = {http://link.eng.cam.ac.uk/foswiki/pub/Main/NGK/Kingsbury_Eusipco06.pdf},
eventtitle = {Proc. {{European Conference}} on {{Signal Processing}} ({{EUSIPCO}})},
booktitle = {14th {{European Signal Processing Conference14th European Signal Processing Conference}}},
publisher = {{IEEE}},
urldate = {2015-11-03},
date = {2006-09},
pages = {901--904},
keywords = {Useful},
author = {Kingsbury, Nick},
file = {/Users/fergalcotter/Dropbox/Papers/Kingsbury_2006_Rotation-invariant local feature matching with complex wavelets.pdf},
note = {00052}
}
@software{vedaldi_matconvnet_2016,
title = {{{MatConvNet Software}}},
language = {Matlab},
url = {https://github.com/vlfeat/matconvnet/releases/tag/v1.0-beta20},
abstract = {MatConvNet is an implementation of Convolutional Neural Networks (CNNs)
for MATLAB. The toolbox is designed with an emphasis on simplicity and flexibility.
It exposes the building blocks of CNNs as easy-to-use MATLAB functions, providing
routines for computing linear convolutions with filter banks, feature pooling, and many
more. In this manner, MatConvNet allows fast prototyping of new CNN architectures;
at the same time, it supports efficient computation on CPU and GPU allowing
to train complex models on large datasets such as ImageNet ILSVRC. This document
provides an overview of CNNs and how they are implemented in MatConvNet and
gives the technical details of each computational block in the toolbox.},
version = {1.0 beta 20},
date = {2016-05},
author = {Vedaldi, Andrea and Lenc, Karel and Gupta, Ankush},
file = {/Users/fergalcotter/Dropbox/Papers/Andrea Vedaldi et al_2016_MatConvNet.pdf},
note = {00253}
}
@incollection{orr_speeding_2012,
langid = {english},
title = {Speeding {{Learning}}},
isbn = {978-3-642-35288-1 978-3-642-35289-8},
url = {http://link.springer.com/chapter/10.1007/978-3-642-35289-8_2},
abstract = {There are those who argue that developing fast algorithms is no longer necessary because computers have become so fast. However, we believe that the complexity of our algorithms and the size of our problems will always expand to consume all cycles available, regardless of the speed of ourmachines.Thus, there will never come a time when computational efficiency can or should be ignored. Besides, in the quest to find solutions faster, we also often find better and more stable solutions as well. This section is devoted to techniques for making the learning process in backpropagation (BP) faster and more efficient. It contains a single chapter based on a workshop by Leon Bottou and Yann LeCun. While many alternative learning systems have emerged since the time BP was first introduced, BP is still the most widely used learning algorithm.The reason for this is its simplicity, efficiency, and its general effectiveness on a wide range of problems. Even so, there are many pitfalls in applying it, which is where all these tricks enter.},
number = {7700},
booktitle = {Neural {{Networks}}: {{Tricks}} of the {{Trade}}},
series = {Lecture {{Notes}} in {{Computer Science}}},
publisher = {{Springer Berlin Heidelberg}},
urldate = {2016-08-09},
date = {2012},
pages = {7-8},
keywords = {Algorithm Analysis and Problem Complexity,Artificial Intelligence (incl. Robotics),Complexity,Computation by Abstract Devices,Information Systems Applications (incl. Internet),pattern recognition},
author = {Orr, Geneviève B. and Müller, Klaus-Robert},
editor = {Montavon, Grégoire and Orr, Geneviève B. and Müller, Klaus-Robert},
file = {/Users/fergalcotter/Dropbox/Papers/Orr_Müller_2012_Speeding Learning.pdf;/Users/fergalcotter/Zotero/storage/7ARKTK9R/978-3-642-35289-8_2.html},
doi = {10.1007/978-3-642-35289-8_2},
note = {00000}
}
@incollection{prechelt_early_2012,
langid = {english},
title = {Early {{Stopping}} — {{But When}}?},
isbn = {978-3-642-35288-1 978-3-642-35289-8},
url = {http://link.springer.com/chapter/10.1007/978-3-642-35289-8_5},
abstract = {Validation can be used to detect when overfitting starts during supervised training of a neural network; training is then stopped before convergence to avoid the overfitting (“early stopping”). The exact criterion used for validation-based early stopping, however, is usually chosen in an ad-hoc fashion or training is stopped interactively. This trick describes how to select a stopping criterion in a systematic fashion; it is a trick for either speeding learning procedures or improving generalization, whichever is more important in the particular situation. An empirical investigation on multi-layer perceptrons shows that there exists a tradeoff between training time and generalization: From the given mix of 1296 training runs using different 12 problems and 24 different network architectures I conclude slower stopping criteria allow for small improvements in generalization (here: about 4\% on average), but cost much more training time (here: about factor 4 longer on average).},
number = {7700},
booktitle = {Neural {{Networks}}: {{Tricks}} of the {{Trade}}},
series = {Lecture {{Notes}} in {{Computer Science}}},
publisher = {{Springer Berlin Heidelberg}},
urldate = {2016-08-09},
date = {2012},
pages = {53-67},
keywords = {Algorithm Analysis and Problem Complexity,Artificial Intelligence (incl. Robotics),Complexity,Computation by Abstract Devices,Information Systems Applications (incl. Internet),pattern recognition},
author = {Prechelt, Lutz},
editor = {Montavon, Grégoire and Orr, Geneviève B. and Müller, Klaus-Robert},
file = {/Users/fergalcotter/Dropbox/Papers/Prechelt_2012_Early Stopping — But When.pdf;/Users/fergalcotter/Zotero/storage/WKZI47V7/978-3-642-35289-8_5.html},
doi = {10.1007/978-3-642-35289-8_5},
note = {00206}
}
@article{russakovsky_imagenet_2015,
langid = {english},
title = {{{ImageNet Large Scale Visual Recognition Challenge}}},
volume = {115},
issn = {0920-5691, 1573-1405},
url = {http://link.springer.com/article/10.1007/s11263-015-0816-y},
doi = {10.1007/s11263-015-0816-y},
abstract = {The ImageNet Large Scale Visual Recognition Challenge is a benchmark in object category classification and detection on hundreds of object categories and millions of images. The challenge has been run annually from 2010 to present, attracting participation from more than fifty institutions. This paper describes the creation of this benchmark dataset and the advances in object recognition that have been possible as a result. We discuss the challenges of collecting large-scale ground truth annotation, highlight key breakthroughs in categorical object recognition, provide a detailed analysis of the current state of the field of large-scale image classification and object detection, and compare the state-of-the-art computer vision accuracy with human accuracy. We conclude with lessons learned in the 5 years of the challenge, and propose future directions and improvements.},
number = {3},
journaltitle = {International Journal of Computer Vision},
shortjournal = {Int J Comput Vis},
urldate = {2016-08-13},
date = {2015-04-11},
pages = {211-252},
author = {Russakovsky, Olga and Deng, Jia and Su, Hao and Krause, Jonathan and Satheesh, Sanjeev and Ma, Sean and Huang, Zhiheng and Karpathy, Andrej and Khosla, Aditya and Bernstein, Michael and Berg, Alexander C. and Fei-Fei, Li},
file = {/Users/fergalcotter/Dropbox/Papers/Russakovsky et al_2015_ImageNet Large Scale Visual Recognition Challenge.pdf;/Users/fergalcotter/Zotero/storage/IXN6Q53D/s11263-015-0816-y.html},
note = {01188}
}
@article{carandini_linearity_1997,
langid = {english},
title = {Linearity and {{Normalization}} in {{Simple Cells}} of the {{Macaque Primary Visual Cortex}}},
volume = {17},
issn = {0270-6474, 1529-2401},
url = {http://www.jneurosci.org/content/17/21/8621},
abstract = {Simple cells in the primary visual cortex often appear to compute a weighted sum of the light intensity distribution of the visual stimuli that fall on their receptive fields. A linear model of these cells has the advantage of simplicity and captures a number of basic aspects of cell function. It, however, fails to account for important response nonlinearities, such as the decrease in response gain and latency observed at high contrasts and the effects of masking by stimuli that fail to elicit responses when presented alone. To account for these nonlinearities we have proposed a normalization model, which extends the linear model to include mutual shunting inhibition among a large number of cortical cells. Shunting inhibition is divisive, and its effect in the model is to normalize the linear responses by a measure of stimulus energy. To test this model we performed extracellular recordings of simple cells in the primary visual cortex of anesthetized macaques. We presented large stimulus sets consisting of (1) drifting gratings of various orientations and spatiotemporal frequencies; (2) plaids composed of two drifting gratings; and (3) gratings masked by full-screen spatiotemporal white noise. We derived expressions for the model predictions and fitted them to the physiological data. Our results support the normalization model, which accounts for both the linear and the nonlinear properties of the cells. An alternative model, in which the linear responses are subject to a compressive nonlinearity, did not perform nearly as well.},
number = {21},
journaltitle = {The Journal of Neuroscience},
shortjournal = {J. Neurosci.},
urldate = {2016-07-27},
date = {1997-01-11},
pages = {8621-8644},
keywords = {contrast,gain control,masking,noise,nonlinearity,normalization,visual cortex},
author = {Carandini, Matteo and Heeger, David J. and Movshon, J. Anthony},
file = {/Users/fergalcotter/Dropbox/Papers/Carandini et al_1997_Linearity and Normalization in Simple Cells of the Macaque Primary Visual Cortex.pdf;/Users/fergalcotter/Zotero/storage/GI5ZS2GS/8621.html},
eprinttype = {pmid},
eprint = {9334433},
note = {00741}
}
@article{grun_taxonomy_2016,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1606.07757},
primaryClass = {cs},
title = {A {{Taxonomy}} and {{Library}} for {{Visualizing Learned Features}} in {{Convolutional Neural Networks}}},
url = {http://arxiv.org/abs/1606.07757},
abstract = {Over the last decade, Convolutional Neural Networks (CNN) saw a tremendous surge in performance. However, understanding what a network has learned still proves to be a challenging task. To remedy this unsatisfactory situation, a number of groups have recently proposed different methods to visualize the learned models. In this work we suggest a general taxonomy to classify and compare these methods, subdividing the literature into three main categories and providing researchers with a terminology to base their works on. Furthermore, we introduce the FeatureVis library for MatConvNet: an extendable, easy to use open source library for visualizing CNNs. It contains implementations from each of the three main classes of visualization methods and serves as a useful tool for an enhanced understanding of the features learned by intermediate layers, as well as for the analysis of why a network might fail for certain examples.},
urldate = {2016-07-27},
date = {2016-06-24},
keywords = {Computer Science - Computer Vision and Pattern Recognition},
author = {Grun, Felix and Rupprecht, Christian and Navab, Nassir and Tombari, Federico},
file = {/Users/fergalcotter/Dropbox/Papers/Grün et al_2016_A Taxonomy and Library for Visualizing Learned Features in Convolutional Neural.pdf;/Users/fergalcotter/Zotero/storage/ABMFD94A/1606.html},
note = {00000}
}
@article{lecun_tutorial_2006,
title = {A {{Tutorial}} on {{Energy Based Learning}}},
url = {http://yann.lecun.com/exdb/publis/pdf/lecun-06.pdf},
journaltitle = {Predicting Structured Data},
urldate = {2016-05-06},
date = {2006-08-19},
author = {LeCun, Yann and Chopra, Sumit and Hadsell, Raia and Ranzato, Marc' Aurelio and Huang, Fu Jie},
file = {/Users/fergalcotter/Dropbox/Papers/LeCun et al_2006_A Tutorial on Energy Based Learning.pdf;/Users/fergalcotter/Zotero/storage/N6SCVTS8/lecun-06.html},
note = {00206}
}
@article{tang_deep_2013,
archivePrefix = {arXiv},
eprinttype = {arxiv},
eprint = {1306.0239},
primaryClass = {cs, stat},
title = {Deep {{Learning}} Using {{Linear Support Vector Machines}}},
url = {http://arxiv.org/abs/1306.0239},