1. Adversarial images for the primate brain
Li Yuan, Will Xiao, Gabriel Kreiman, Francis E.H. Tay, Jiashi Feng, Margaret S. Livingstone
- retweets: 240, favorites: 92 (11/13/2020 08:40:46)
- links: abs | pdf
- q-bio.NC | cs.CV | cs.NE | eess.IV
Deep artificial neural networks have been proposed as a model of primate vision. However, these networks are vulnerable to adversarial attacks, whereby introducing minimal noise can fool networks into misclassifying images. Primate vision is thought to be robust to such adversarial images. We evaluated this assumption by designing adversarial images to fool primate vision. To do so, we first trained a model to predict responses of face-selective neurons in macaque inferior temporal cortex. Next, we modified images, such as human faces, to match their model-predicted neuronal responses to a target category, such as monkey faces. These adversarial images elicited neuronal responses similar to the target category. Remarkably, the same images fooled monkeys and humans at the behavioral level. These results challenge fundamental assumptions about the similarity between computer and primate vision and show that a model of neuronal activity can selectively direct primate visual behavior.
"Adversarial images for the primate brain" (!)
— Chaz Firestone (@chazfirestone) November 12, 2020
1. Train a model to predict responses of face-selective neurons
2. Modify images to fool that model
3. Test actual behavioral & neural responses to those images
Voila: Macaques (and humans) are fooled!https://t.co/FfoN8zbPvN pic.twitter.com/E38aJfLRjQ
2. DeepI2I: Enabling Deep Hierarchical Image-to-Image Translation by Transferring from GANs
Yaxing Wang, Lu Yu, Joost van de Weijer
Image-to-image translation has recently achieved remarkable results. But despite current success, it suffers from inferior performance when translations between classes require large shape changes. We attribute this to the high-resolution bottlenecks which are used by current state-of-the-art image-to-image methods. Therefore, in this work, we propose a novel deep hierarchical Image-to-Image Translation method, called DeepI2I. We learn a model by leveraging hierarchical features: (a) structural information contained in the shallow layers and (b) semantic information extracted from the deep layers. To enable the training of deep I2I models on small datasets, we propose a novel transfer learning method, that transfers knowledge from pre-trained GANs. Specifically, we leverage the discriminator of a pre-trained GANs (i.e. BigGAN or StyleGAN) to initialize both the encoder and the discriminator and the pre-trained generator to initialize the generator of our model. Applying knowledge transfer leads to an alignment problem between the encoder and generator. We introduce an adaptor network to address this. On many-class image-to-image translation on three datasets (Animal faces, Birds, and Foods) we decrease mFID by at least 35% when compared to the state-of-the-art. Furthermore, we qualitatively and quantitatively demonstrate that transfer learning significantly improves the performance of I2I systems, especially for small datasets. Finally, we are the first to perform I2I translations for domains with over 100 classes.
DeepI2I: Enabling Deep Hierarchical Image-to-Image Translation by Transferring from GANs
— AK (@ak92501) November 12, 2020
pdf: https://t.co/qSG98rl2LZ
abs: https://t.co/d2VoOoTpk9 pic.twitter.com/univDmJcFP
3. Don’t Read Too Much into It: Adaptive Computation for Open-Domain Question Answering
Yuxiang Wu, Sebastian Riedel, Pasquale Minervini, Pontus Stenetorp
Most approaches to Open-Domain Question Answering consist of a light-weight retriever that selects a set of candidate passages, and a computationally expensive reader that examines the passages to identify the correct answer. Previous works have shown that as the number of retrieved passages increases, so does the performance of the reader. However, they assume all retrieved passages are of equal importance and allocate the same amount of computation to them, leading to a substantial increase in computational cost. To reduce this cost, we propose the use of adaptive computation to control the computational budget allocated for the passages to be read. We first introduce a technique operating on individual passages in isolation which relies on anytime prediction and a per-layer estimation of an early exit probability. We then introduce SkylineBuilder, an approach for dynamically deciding on which passage to allocate computation at each step, based on a resource allocation policy trained via reinforcement learning. Our results on SQuAD-Open show that adaptive computation with global prioritisation improves over several strong static and adaptive methods, leading to a 4.3x reduction in computation while retaining 95% performance of the full model.
Don’t Read Too Much Into It: Adaptive Computation for Open-Domain Question Answering
— Aran Komatsuzaki (@arankomatsuzaki) November 12, 2020
Achieves 4.3x reduction in computation for fine-tuning a LM on SQuAD-Open while retaining 95% performance of the full model. https://t.co/7yKf3PncAp pic.twitter.com/fZtKBv8SxO
4. End-to-End Chinese Landscape Painting Creation Using Generative Adversarial Networks
Alice Xue
Current GAN-based art generation methods produce unoriginal artwork due to their dependence on conditional input. Here, we propose Sketch-And-Paint GAN (SAPGAN), the first model which generates Chinese landscape paintings from end to end, without conditional input. SAPGAN is composed of two GANs: SketchGAN for generation of edge maps, and PaintGAN for subsequent edge-to-painting translation. Our model is trained on a new dataset of traditional Chinese landscape paintings never before used for generative research. A 242-person Visual Turing Test study reveals that SAPGAN paintings are mistaken as human artwork with 55% frequency, significantly outperforming paintings from baseline GANs. Our work lays a groundwork for truly machine-original art generation.
End-to-End Chinese Landscape Painting Creation Using
— AK (@ak92501) November 12, 2020
Generative Adversarial Networks
pdf: https://t.co/w8RaBfc6A7
abs: https://t.co/qQ3L0GNJU2 pic.twitter.com/89vJLMs2V1
5. The Impact of Text Presentation on Translator Performance
Samuel Läubli, Patrick Simianer, Joern Wuebker, Geza Kovacs, Rico Sennrich, Spence Green
Widely used computer-aided translation (CAT) tools divide documents into segments such as sentences and arrange them in a side-by-side, spreadsheet-like view. We present the first controlled evaluation of these design choices on translator performance, measuring speed and accuracy in three experimental text processing tasks. We find significant evidence that sentence-by-sentence presentation enables faster text reproduction and within-sentence error identification compared to unsegmented text, and that a top-and-bottom arrangement of source and target sentences enables faster text reproduction compared to a side-by-side arrangement. For revision, on the other hand, our results suggest that presenting unsegmented text results in the highest accuracy and time efficiency. Our findings have direct implications for best practices in designing CAT tools.
The way in which CAT tools divide documents into segments and arrange these segments on your screen probably impact your productivity. Check out the results of our full-day experiment with 20 professional translators: https://t.co/89cxQ53YF4
— Samuel Läubli (@samlaeubli) November 12, 2020