Skip to content

Commit dcff4d2

Browse files
committed
177
1 parent dc06749 commit dcff4d2

File tree

6 files changed

+86
-31
lines changed

6 files changed

+86
-31
lines changed

Artificial Intelligence.md

Lines changed: 35 additions & 17 deletions
Large diffs are not rendered by default.

Deep Learning.md

Lines changed: 24 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -404,6 +404,8 @@
404404

405405
["Deep Generative Models"](http://www.deeplearningbook.org/contents/generative_models.html) chapter of "Deep Learning" book by Goodfellow, Bengio, Courville
406406

407+
["Deep Generative Modelling: A Comparative Review of VAEs, GANs, Normalizing Flows, Energy-Based and Autoregressive Models"](https://arxiv.org/abs/2103.04922) by Bond-Taylor et al. `paper`
408+
407409
["Do we still need models or just more data and compute?"](https://staff.fnwi.uva.nl/m.welling/wp-content/uploads/Model-versus-Data-AI-1.pdf) by Max Welling
408410

409411
----
@@ -420,10 +422,6 @@
420422
[course](https://youtube.com/playlist?list=PLk4h7dmY2eYHjH6wjO1Wytb2ipXJX-aWj) from MIPT `video` `in russian` `2020`
421423
[course](https://youtube.com/playlist?list=PLk4h7dmY2eYH9RtoKGzxHKji0GLiBzSlZ) from MIPT `video` `in russian` `2019`
422424

423-
----
424-
425-
[**interesting recent papers**](https://github.com/brylevkirill/notes/blob/master/interesting%20recent%20papers.md#generative-models)
426-
427425
----
428426

429427
"A generative model should be able to draw samples from p(x); however estimating p(x) may be computationally intractable. Instead, we often learn a function that maps a vector to an image sample x. The vector may be either be a noise vector, z, drawn from a prior distribution, a label vector, y, or a combination of the two. Probabilistically, these may be interpreted as conditional probabilities: p(x|z), p(x|y) or p(x|z,y). By sampling these conditional probabilities appropriately, novel samples of x may be generated."
@@ -919,7 +917,9 @@
919917

920918
[overview](https://youtube.com/watch?v=x5Vxk9twXlE) by Geoffrey Hinton `video`
921919

922-
["How to Represent Part-whole Hierarchies in a Neural Network"](https://arxiv.org/abs/2102.12627) by Hinton `paper` ([overview](https://youtube.com/watch?v=cllFzkvrYmE) `video`)
920+
[**"How to Represent Part-whole Hierarchies in a Neural Network"**](#how-to-represent-part-whole-hierarchies-in-a-neural-network) by Hinton `paper` `summary` *(GLOM)*
921+
["Canonical Capsules: Unsupervised Capsules in Canonical Pose"](https://arxiv.org/abs/2012.04718) by Sun et al. `paper`
922+
["Unsupervised Part Representation by Flow Capsules"](https://arxiv.org/abs/2011.13920) by Sabour et al. `paper`
923923
[**"Stacked Capsule Autoencoders"**](#stacked-capsule-autoencoders-kosiorek-sabour-teh-hinton) by Kosiorek et al. `paper` `summary`
924924
[**"Matrix Capsules with EM Routing"**](#matrix-capsules-with-em-routing-hinton-sabour-frosst) by Hinton et al. `paper` `summary`
925925
[**"Dynamic Routing Between Capsules"**](#dynamic-routing-between-capsules-sabour-frosst-hinton) by Hinton et al. `paper` `summary`
@@ -1946,6 +1946,7 @@ Yoshua Bengio:
19461946
> "Unlike recent language representation models (Peters et al., 2018; Radford et al., 2018), BERT is designed to pre-train deep bidirectional representations by jointly conditioning on both left and right context in all layers. As a result, the pre-trained BERT representations can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks, such as question answering and language inference, without substantial task-specific architecture modifications."
19471947
19481948
- `post` <https://ai.googleblog.com/2018/11/open-sourcing-bert-state-of-art-pre.html>
1949+
- `video` <https://youtube.com/watch?v=knTc-NQSjKA> (Devlin)
19491950
- `video` <https://youtube.com/watch?v=BWP3DyJ6k3g> (Chang)
19501951
- `video` <https://youtube.com/watch?v=OR0wfP2FD3c> (Shorten)
19511952
- `video` <https://youtube.com/watch?v=-9evrZnBorM> (Kilcher)
@@ -2657,7 +2658,6 @@ Yoshua Bengio:
26572658
----
26582659
> "Hopfield construction would allow iterating the attention, which Transformer doesn't do as it does the attention in a single step, but the authors find that iterating the Hopfield attention doesn't seem to be useful as it converges in a single step. The Hopfield layer also has beta in place of Transformer's 1/sqrt(d_k) scaling factor, and the authors note that this factor has a significant effect on the Hopfield network attraction point types, so it might lead to future understanding about what is the optimal value for this. Regardless they use the scaling factor from Transformer, 1/sqrt(d_k) in all the experiments."
26592660
2660-
26612661
- `post` <https://research.googleblog.com/2017/08/transformer-novel-neural-network.html>
26622662
- `video` <https://youtube.com/watch?v=bYmeuc5voUQ> (Vaswani)
26632663
- `video` <https://facebook.com/nipsfoundation/videos/1554654864625747?t=1107> (Vaswani, Shazeer)
@@ -2686,6 +2686,7 @@ Yoshua Bengio:
26862686
- `video` <https://youtube.com/watch?v=7e4LxIVENZA> (Sapunov) `in russian`
26872687
- `video` <https://youtube.com/watch?v=znyvj3ijl7Q> (Sapunov) `in russian`
26882688
- `audio` <https://soundcloud.com/nlp-highlights/36-attention-is-all-you-need-with-ashish-vaswani-and-jakob-uszkoreit> (Vaswani, Uszkoreit)
2689+
- `post` <https://people.idsia.ch/~juergen/fast-weight-programmer-1991-transformer.html>
26892690
- `post` <https://lilianweng.github.io/lil-log/2020/04/07/the-transformer-family.html>
26902691
- `post` <https://jalammar.github.io/illustrated-transformer/>
26912692
- `post` <https://danieltakeshi.github.io/2019/03/30/transformers/>
@@ -2701,15 +2702,26 @@ Yoshua Bengio:
27012702
- `notes` <https://medium.com/@sharaf/a-paper-a-day-24-attention-is-all-you-need-26eb2da90a91>
27022703
- `code` <https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/models/transformer.py>
27032704
- `code` <https://colab.research.google.com/notebook#fileId=/v2/external/notebooks/t2t/hello_t2t.ipynb>
2705+
- `paper` ["Linear Transformers Are Secretly Fast Weight Memory Systems"](https://arxiv.org/abs/2102.11174) by Schlag, Irie, Schmidhuber ([post](https://people.idsia.ch/~juergen/fast-weight-programmer-1991-transformer.html))
27042706
- `paper` ["Hopfield Networks is All You Need"](https://arxiv.org/abs/2008.02217) by Ramsauer et al. ([post](https://jku.at/index.php?id=18677), [overview](https://youtube.com/watch?v=nv6oFDp6rNQ) (Kilcher) `video`)
27052707
- `paper` [**"Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"**](https://github.com/brylevkirill/notes/blob/master/Natural%20Language%20Processing.md#exploring-the-limits-of-transfer-learning-with-a-unified-text-to-text-transformer-raffel-et-al) by Raffel et al. `summary`
27062708
- `paper` [**"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding"**](#bert-pre-training-of-deep-bidirectional-transformers-for-language-understanding-devlin-chang-lee-toutanova) by Devlin et al. `summary`
27072709
- `paper` [**"Language Models are Unsupervised Multitask Learners"**](https://github.com/brylevkirill/notes/blob/master/Natural%20Language%20Processing.md#language-models-are-unsupervised-multitask-learners-radford-wu-child-luan-amodei-sutskever) by Radford et al. `summary`
27082710
- `paper` ["An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale"](https://arxiv.org/abs/2010.11929) `paper` ([overview](https://youtube.com/watch?v=TrdevFK_am4) by Kilcher `video`)
27092711

27102712

2713+
#### ["How to Represent Part-whole Hierarchies in a Neural Network"](https://arxiv.org/abs/2102.12627) Hinton
2714+
`GLOM` `Capsule Networks`
2715+
> "This paper does not describe a working system. Instead, it presents a single idea about representation which allows advances made by several different groups to be combined into an imaginary system called GLOM. The advances include transformers, neural fields, contrastive representation learning, distillation and capsules. GLOM answers the question: How can a neural network with a fixed architecture parse an image into a part-whole hierarchy which has a different structure for each image? The idea is simply to use islands of identical vectors to represent the nodes in the parse tree. If GLOM can be made to work, it should significantly improve the interpretability of the representations produced by transformer-like systems when applied to vision or language."
2716+
2717+
- `video` <https://youtube.com/watch?v=ygInBb1fS9Y> (Hinton)
2718+
- `video` <https://youtube.com/watch?v=eEXnJOHQ_Xw> (Hinton)
2719+
- `video` <https://youtube.com/watch?v=cllFzkvrYmE> (Kilcher)
2720+
- `video` <https://youtube.com/watch?v=fkqiQuD6k8k> (Lewis)
2721+
2722+
27112723
#### ["Stacked Capsule Autoencoders"](https://arxiv.org/abs/1906.06818) Kosiorek, Sabour, Teh, Hinton
2712-
`Capsule Network` `CapsNet` `NeurIPS 2019`
2724+
`Capsule Networks` `NeurIPS 2019`
27132725
> "An object can be seen as a geometrically organized set of interrelated parts. A system that makes explicit use of these geometric relationships to recognize objects should be naturally robust to changes in viewpoint, because the intrinsic geometric relationships are viewpoint-invariant. We describe an unsupervised version of capsule networks, in which a neural encoder, which looks at all of the parts, is used to infer the presence and poses of object capsules. The encoder is trained by backpropagating through a decoder, which predicts the pose of each already discovered part using a mixture of pose predictions. The parts are discovered directly from an image, in a similar manner, by using a neural encoder, which infers parts and their affine transformations. The corresponding decoder models each image pixel as a mixture of predictions made by affine-transformed parts. We learn object- and their part-capsules on unlabeled data, and then cluster the vectors of presences of object capsules. When told the names of these clusters, we achieve state-of-the-art results for unsupervised classification on SVHN (55%) and near state-of-the-art on MNIST (98.5%)."
27142726

27152727
----
@@ -2780,14 +2792,15 @@ Yoshua Bengio:
27802792
> Once the vectors have been suitably transformed, additional processing is used to predict the objects."
27812793
27822794
- `post` <http://akosiorek.github.io/ml/2019/06/23/stacked_capsule_autoencoders.html>
2795+
- `video` <https://youtu.be/N0ER1MC9cqM?t=2m22s> (Hinton)
27832796
- `video` <https://youtu.be/UX8OubxsY8w?t=3m8s> (Hinton)
27842797
- `video` <https://youtu.be/zRg3IuxaJ6I?t=27m49s> (Sabour)
27852798
- `notes` <https://medium.com/syncedreview/geoffrey-hintons-unsupervised-capsule-networks-achieve-sota-results-on-svhn-ffe05e871249>
27862799
- `paper` ["Set Transformer: A Framework for Attention-basedPermutation-Invariant Neural Networks"](https://arxiv.org/abs/1810.00825) by Lee et al. ([overview](https://facebook.com/icml.imls/videos/3253466301345987?t=1814) by Jungtaek Kim `video`)
27872800

27882801

27892802
#### ["Matrix Capsules with EM Routing"](https://openreview.net/forum?id=HJWLfGWRb) Hinton, Sabour, Frosst
2790-
`Capsule Network` `CapsNet` `ICLR 2018`
2803+
`Capsule Networks` `ICLR 2018`
27912804
> "A capsule is a group of neurons whose outputs represent different properties of the same entity. Each layer in a capsule network contains many capsules [a group of capsules forms a capsule layer and can be used in place of a traditional layer in a neural net]. We describe a version of capsules in which each capsule has a logistic unit to represent the presence of an entity and a 4x4 matrix which could learn to represent the relationship between that entity and the viewer (the pose). A capsule in one layer votes for the pose matrix of many different capsules in the layer above by multiplying its own pose matrix by trainable viewpoint-invariant transformation matrices that could learn to represent part-whole relationships. Each of these votes is weighted by an assignment coefficient. These coefficients are iteratively updated for each image using the Expectation-Maximization algorithm such that the output of each capsule is routed to a capsule in the layer above that receives a cluster of similar votes. The transformation matrices are trained discriminatively by backpropagating through the unrolled iterations of EM between each pair of adjacent capsule layers. On the smallNORB benchmark, capsules reduce the number of test errors by 45\% compared to the state-of-the-art. Capsules also show far more resistance to white box adversarial attack than our baseline convolutional neural network."
27922805

27932806
- `video` <https://youtu.be/x5Vxk9twXlE?t=30m46s> (Hinton)
@@ -2803,7 +2816,7 @@ Yoshua Bengio:
28032816

28042817

28052818
#### ["Dynamic Routing Between Capsules"](https://arxiv.org/abs/1710.09829) Sabour, Frosst, Hinton
2806-
`Capsule Network` `CapsNet` `NIPS 2017`
2819+
`Capsule Networks` `NIPS 2017`
28072820
> "A capsule is a group of neurons whose activity vector represents the instantiation parameters of a specific type of entity such as an object or object part. We use the length of the activity vector to represent the probability that the entity exists and its orientation to represent the instantiation paramters. Active capsules at one level make predictions, via transformation matrices, for the instantiation parameters of higher-level capsules. When multiple predictions agree, a higher level capsule becomes active. We show that a discrimininatively trained, multi-layer capsule system achieves state-of-the-art performance on MNIST and is considerably better than a convolutional net at recognizing highly overlapping digits. To achieve these results we use an iterative routing-by-agreement mechanism: A lower-level capsule prefers to send its output to higher level capsules whose activity vectors have a big scalar product with the prediction coming from the lower-level capsule."
28082821
28092822
> "Dynamic routing can be viewed as a parallel attention mechanism that allows each capsule at one level to attend to some active capsules at the level below and to ignore others. This should allow the model to recognize multiple objects in the image even if objects overlap. The routing-by-agreement should make it possible to use a prior about shape of objects to help segmentation and it should obviate the need to make higher-level segmentation decisions in the domain of pixels."
@@ -2875,6 +2888,7 @@ Yoshua Bengio:
28752888
----
28762889
> "Highway Networks have a gated connection in the depth dimension analogous to the gated connection LSTMs have in the time dimension. There are also Grid LSTMs, which have these gated connections in both dimensions."
28772890
2891+
- `post` <https://people.idsia.ch/~juergen/highway-networks.html>
28782892
- `post` <http://people.idsia.ch/~rupesh/very_deep_learning/>
28792893
- `video` <http://research.microsoft.com/apps/video/default.aspx?id=259633> (9:00) (Srivastava)
28802894
- `post` <https://medium.com/jim-fleming/highway-networks-with-tensorflow-1e6dfa667daa>
@@ -3247,6 +3261,7 @@ Yoshua Bengio:
32473261

32483262

32493263
#### ["Backpropagation and the Brain"](https://www.nature.com/articles/s41583-020-0277-3) Lillicrap, Santoro, Marris, Akerman, Hinton
3264+
`NGRAD`
32503265
> "During learning, the brain modifies synapses to improve behaviour. In the cortex, synapses are embedded within multilayered networks, making it difficult to determine the effect of an individual synaptic modification on the behaviour of the system. The backpropagation algorithm solves this problem in deep artificial neural networks, but historically it has been viewed as biologically problematic. Nonetheless, recent developments in neuroscience and the successes of artificial neural networks have reinvigorated interest in whether backpropagation offers insights for understanding learning in the cortex. The backpropagation algorithm learns quickly by computing synaptic updates using feedback connections to deliver error signals. Although feedback connections are ubiquitous in the cortex, it is difficult to see how they could deliver the error signals required by strict formulations of backpropagation. Here we build on past and recent developments to argue that feedback connections may instead induce neural activities whose differences can be used to locally approximate these signals and hence drive effective learning in deep networks in the brain."
32513266

32523267
- `video` <https://youtube.com/watch?v=a0f07M2uj_A> (Kilcher)

Machine Learning.md

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -344,8 +344,9 @@
344344

345345
----
346346

347-
[overview](https://youtube.com/watch?v=nlgA2gnwscQ) by Alex Polozov `video`
347+
[overview](https://youtube.com/watch?v=2RvxsJYn2UI) by Rishabh Singh `video`
348348
[overview](https://youtube.com/watch?v=Fs7FquuLprM) by Rishabh Singh `video`
349+
[overview](https://youtube.com/watch?v=nlgA2gnwscQ) by Alex Polozov `video`
349350
[overview](https://facebook.com/nipsfoundation/videos/1552060484885185?t=5412) by Scott Reed `video`
350351
[overview](https://youtu.be/vzDuVhFMB9Q?t=2m40s) by Alex Gaunt `video`
351352

@@ -388,12 +389,14 @@
388389

389390
----
390391

392+
[overview](http://people.idsia.ch/~juergen/metalearning.html) by Juergen Schmidhuber
393+
391394
[overview](https://youtu.be/3FIo6evmweo?t=4m6s) by Juergen Schmidhuber `video` *(meta-learning vs transfer learning)*
392395
[overview](https://vimeo.com/250399374) by Juergen Schmidhuber `video`
393396
[overview](https://youtube.com/watch?v=nqiUFc52g78) by Juergen Schmidhuber `video`
394397

395-
[overview](http://scholarpedia.org/article/Metalearning) by Tom Schaul and Juergen Schmidhuber
396398
[overview](http://people.idsia.ch/~juergen/metalearner.html) by Juergen Schmidhuber
399+
[overview](http://scholarpedia.org/article/Metalearning) by Tom Schaul and Juergen Schmidhuber
397400

398401
[**Goedel Machine**](https://github.com/brylevkirill/notes/blob/master/Artificial%20Intelligence.md#meta-learning---goedel-machine)
399402

0 commit comments

Comments
 (0)