1. Modern Methods for Text Generation
Dimas Munoz Montesinos
Synthetic text generation is challenging and has limited success. Recently, a new architecture, called Transformers, allow machine learning models to understand better sequential data, such as translation or summarization. BERT and GPT-2, using Transformers in their cores, have shown a great performance in tasks such as text classification, translation and NLI tasks. In this article, we analyse both algorithms and compare their output quality in text generation tasks.
Here is a nice summary report of modern methods for text generation.
— elvis (@omarsar0) September 11, 2020
Very easy to read.https://t.co/Qf2h33Yzf7 pic.twitter.com/cG1QJoEYy2
2. Understanding the Role of Individual Units in a Deep Neural Network
David Bau, Jun-Yan Zhu, Hendrik Strobelt, Agata Lapedriza, Bolei Zhou, Antonio Torralba
Deep neural networks excel at finding hierarchical representations that solve complex tasks over large data sets. How can we humans understand these learned representations? In this work, we present network dissection, an analytic framework to systematically identify the semantics of individual hidden units within image classification and image generation networks. First, we analyze a convolutional neural network (CNN) trained on scene classification and discover units that match a diverse set of object concepts. We find evidence that the network has learned many object classes that play crucial roles in classifying scene classes. Second, we use a similar analytic method to analyze a generative adversarial network (GAN) model trained to generate scenes. By analyzing changes made when small sets of units are activated or deactivated, we find that objects can be added and removed from the output scenes while adapting to the context. Finally, we apply our analytic framework to understanding adversarial attacks and to semantic image editing.
Understanding the Role of Individual Units in a Deep Neural Network
— AK (@ak92501) September 11, 2020
pdf: https://t.co/BZl1kDRfUd
abs: https://t.co/mntNScqLfl
project page: https://t.co/G2VP12WfvT
github: https://t.co/GIfyBPVzTz pic.twitter.com/wMVcBlcUHl
3. Brain2Word: Decoding Brain Activity for Language Generation
Nicolas Affolter, Beni Egressy, Damian Pascual, Roger Wattenhofer
Brain decoding, understood as the process of mapping brain activities to the stimuli that generated them, has been an active research area in the last years. In the case of language stimuli, recent studies have shown that it is possible to decode fMRI scans into an embedding of the word a subject is reading. However, such word embeddings are designed for natural language processing tasks rather than for brain decoding. Therefore, they limit our ability to recover the precise stimulus. In this work, we propose to directly classify an fMRI scan, mapping it to the corresponding word within a fixed vocabulary. Unlike existing work, we evaluate on scans from previously unseen subjects. We argue that this is a more realistic setup and we present a model that can decode fMRI data from unseen subjects. Our model achieves 5.22% Top-1 and 13.59% Top-5 accuracy in this challenging task, significantly outperforming all the considered competitive baselines. Furthermore, we use the decoded words to guide language generation with the GPT-2 model. This way, we advance the quest for a system that translates brain activities into coherent text.
Brain2Word: Decoding Brain Activity for Language Generation
— AK (@ak92501) September 11, 2020
pdf: https://t.co/Fh7a0QpeZG
abs: https://t.co/Xbc8XhZSTm pic.twitter.com/NhJKvkjNy2
4. The Grievance Dictionary: Understanding Threatening Language Use
Isabelle van der Vegt, Maximilian Mozes, Bennett Kleinberg, Paul Gill
This paper introduces the Grievance Dictionary, a psycholinguistic dictionary which can be used to automatically understand language use in the context of grievance-fuelled violence threat assessment. We describe the development the dictionary, which was informed by suggestions from experienced threat assessment practitioners. These suggestions and subsequent human and computational word list generation resulted in a dictionary of 20,502 words annotated by 2,318 participants. The dictionary was validated by applying it to texts written by violent and non-violent individuals, showing strong evidence for a difference between populations in several dictionary categories. Further classification tasks showed promising performance, but future improvements are still needed. Finally, we provide instructions and suggestions for the use of the Grievance Dictionary by security professionals and (violence) researchers.
Our pre-print "The Grievance Dictionary: Understanding Threatening Language Use" is now available: https://t.co/AkSaEoj59P We developed this open source tool for use in automated linguistic analysis, specifically in the violence and extremism (research) domain. cc: @Grievance_ERC
— Isabelle (@Isabellevdv) September 11, 2020
Our latest, led by @Isabellevdv , is now out on pre-print - The Grievance Dictionary: Understanding Threatening Language Use - https://t.co/c4F0EcfO0R - Develops a psycholinguistic dictionary to automatically understand language use in the context of grievance-fuelled violence
— Paul Gill (@paulgill_ucl) September 11, 2020
5. Pay Attention when Required
Swetha Mandava, Szymon Migacz, Alex Fit Florea
Transformer-based models consist of interleaved feed-forward blocks - that capture content meaning, and relatively more expensive self-attention blocks - that capture context meaning. In this paper, we explored trade-offs and ordering of the blocks to improve upon the current Transformer architecture and proposed PAR Transformer. It needs 35% lower compute time than Transformer-XL achieved by replacing ~63% of the self-attention blocks with feed-forward blocks, and retains the perplexity on WikiText-103 language modelling benchmark. We further validated our results on text8 and enwiki8 datasets, as well as on the BERT model.
Pushed my first paper to arXiv 🤓
— Swetha Mandava (@swethmandava) September 11, 2020
Introducing PAR Transformer that needs 35% lower compute time than Transformer-XL by replacing ~63% of the self-attention blocks with feed-forward blockshttps://t.co/ZmbzsMrikS pic.twitter.com/4gSfuD6xHA
6. Fully Convolutional Graph Neural Networks for Parametric Virtual Try-On
Raquel Vidaurre, Igor Santesteban, Elena Garces, Dan Casas
We present a learning-based approach for virtual try-on applications based on a fully convolutional graph neural network. In contrast to existing data-driven models, which are trained for a specific garment or mesh topology, our fully convolutional model can cope with a large family of garments, represented as parametric predefined 2D panels with arbitrary mesh topology, including long dresses, shirts, and tight tops. Under the hood, our novel geometric deep learning approach learns to drape 3D garments by decoupling the three different sources of deformations that condition the fit of clothing: garment type, target body shape, and material. Specifically, we first learn a regressor that predicts the 3D drape of the input parametric garment when worn by a mean body shape. Then, after a mesh topology optimization step where we generate a sufficient level of detail for the input garment type, we further deform the mesh to reproduce deformations caused by the target body shape. Finally, we predict fine-scale details such as wrinkles that depend mostly on the garment material. We qualitatively and quantitatively demonstrate that our fully convolutional approach outperforms existing methods in terms of generalization capabilities and memory requirements, and therefore it opens the door to more general learning-based models for virtual try-on applications.
Fully Convolutional Graph Neural Networks for Parametric Virtual Try-On
— AK (@ak92501) September 11, 2020
pdf: https://t.co/mzDg8j5Z87
abs: https://t.co/eQKu0NG32C
project page: https://t.co/TzeqmjYWEo pic.twitter.com/FQU95vsWZl