1. Parametric UMAP: learning embeddings with deep neural networks for representation and semi-supervised learning
Tim Sainburg, Leland McInnes, Timothy Q Gentner
- retweets: 1484, favorites: 137 (09/30/2020 09:24:37)
- links: abs | pdf
- cs.LG | cs.CG | q-bio.QM | stat.ML
We propose Parametric UMAP, a parametric variation of the UMAP (Uniform Manifold Approximation and Projection) algorithm. UMAP is a non-parametric graph-based dimensionality reduction algorithm using applied Riemannian geometry and algebraic topology to find low-dimensional embeddings of structured data. The UMAP algorithm consists of two steps: (1) Compute a graphical representation of a dataset (fuzzy simplicial complex), and (2) Through stochastic gradient descent, optimize a low-dimensional embedding of the graph. Here, we replace the second step of UMAP with a deep neural network that learns a parametric relationship between data and embedding. We demonstrate that our method performs similarly to its non-parametric counterpart while conferring the benefit of a learned parametric mapping (e.g. fast online embeddings for new data). We then show that UMAP loss can be extended to arbitrary deep learning applications, for example constraining the latent distribution of autoencoders, and improving classifier accuracy for semi-supervised learning by capturing structure in unlabeled data. Our code is available at https://github.com/timsainb/ParametricUMAP_paper.
New paper "Parametric UMAP: learning embeddings with deep neural networks for representation and semi-supervised learning" with @leland_mcinnes and @TqGentner! 1/https://t.co/2x5YTZPSXU pic.twitter.com/WJ7ECE0DI4
— Tim Sainburg (@tim_sainburg) September 29, 2020
2. EvolGAN: Evolutionary Generative Adversarial Networks
Baptiste Roziere, Fabien Teytaud, Vlad Hosu, Hanhe Lin, Jeremy Rapin, Mariia Zameshina, Olivier Teytaud
We propose to use a quality estimator and evolutionary methods to search the latent space of generative adversarial networks trained on small, difficult datasets, or both. The new method leads to the generation of significantly higher quality images while preserving the original generator’s diversity. Human raters preferred an image from the new version with frequency 83.7pc for Cats, 74pc for FashionGen, 70.4pc for Horses, and 69.2pc for Artworks, and minor improvements for the already excellent GANs for faces. This approach applies to any quality scorer and GAN generator.
EvolGAN: Evolutionary Generative Adversarial Networks
— AK (@ak92501) September 29, 2020
pdf: https://t.co/pEewN8dy9v
abs: https://t.co/pY6czuphHE pic.twitter.com/77A5uZ08h3
3. Answering Complex Open-Domain Questions with Multi-Hop Dense Retrieval
Wenhan Xiong, Xiang Lorraine Li, Srini Iyer, Jingfei Du, Patrick Lewis, William Yang Wang, Yashar Mehdad, Wen-tau Yih, Sebastian Riedel, Douwe Kiela, Barlas Oğuz
We propose a simple and efficient multi-hop dense retrieval approach for answering complex open-domain questions, which achieves state-of-the-art performance on two multi-hop datasets, HotpotQA and multi-evidence FEVER. Contrary to previous work, our method does not require access to any corpus-specific information, such as inter-document hyperlinks or human-annotated entity markers, and can be applied to any unstructured text corpus. Our system also yields a much better efficiency-accuracy trade-off, matching the best published accuracy on HotpotQA while being 10 times faster at inference time.
Answering Complex Open-Domain Questions with Multi-Hop Dense Retrieval
— Aran Komatsuzaki (@arankomatsuzaki) September 29, 2020
Achieves the SOTA performance-computes trade-off in multi-hop open-domain QA (better than FiD). Best published accuracy on HotpotQA w/ 10x faster inference.https://t.co/o5aOrMLH4A pic.twitter.com/dck9Cc1Ym1
😉Happy to share our work on multi-hop (open-domain) QA (https://t.co/7BhmCqX64m). TL;DR: you don't need the Wikipedia hyperlinks to achieve SoTA performance on HotpotQA(@qi2peng2 )! A shared RoBERTa encoder (for both Q and docs) is all you need to retrieve SP passages! 1/3 pic.twitter.com/B9JO85Vd1l
— Wenhan Xiong (@xwhan_) September 29, 2020
4. A Diagnostic Study of Explainability Techniques for Text Classification
Pepa Atanasova, Jakob Grue Simonsen, Christina Lioma, Isabelle Augenstein
Recent developments in machine learning have introduced models that approach human performance at the cost of increased architectural complexity. Efforts to make the rationales behind the models’ predictions transparent have inspired an abundance of new explainability techniques. Provided with an already trained model, they compute saliency scores for the words of an input instance. However, there exists no definitive guide on (i) how to choose such a technique given a particular application task and model architecture, and (ii) the benefits and drawbacks of using each such technique. In this paper, we develop a comprehensive list of diagnostic properties for evaluating existing explainability techniques. We then employ the proposed list to compare a set of diverse explainability techniques on downstream text classification tasks and neural network architectures. We also compare the saliency scores assigned by the explainability techniques with human annotations of salient input regions to find relations between a model’s performance and the agreement of its rationales with human ones. Overall, we find that the gradient-based explanations perform best across tasks and model architectures, and we present further insights into the properties of the reviewed explainability techniques.
When an NLP model makes a prediction, to what degree did each input token contribute towards making that prediction?@atanasovapepa @IAugenstein compare various explainability techniques (for different models and on different datasets) in https://t.co/fIT9uF8tSP #emnlp2020 pic.twitter.com/mMsNdJX6P6
— Jay Alammar (@JayAlammar) September 29, 2020
A preprint of this #emnlp2020 paper by @atanasovapepa et al. is now available: https://t.co/l5fuq8qJQw#XAI #ML #NLProc https://t.co/vhgWXsmQ6F
— Isabelle Augenstein (@IAugenstein) September 29, 2020
5. Normalization Techniques in Training DNNs: Methodology, Analysis and Application
Lei Huang, Jie Qin, Yi Zhou, Fan Zhu, Li Liu, Ling Shao
Normalization techniques are essential for accelerating the training and improving the generalization of deep neural networks (DNNs), and have successfully been used in various applications. This paper reviews and comments on the past, present and future of normalization methods in the context of DNN training. We provide a unified picture of the main motivation behind different approaches from the perspective of optimization, and present a taxonomy for understanding the similarities and differences between them. Specifically, we decompose the pipeline of the most representative normalizing activation methods into three components: the normalization area partitioning, normalization operation and normalization representation recovery. In doing so, we provide insight for designing new normalization technique. Finally, we discuss the current progress in understanding normalization methods, and provide a comprehensive review of the applications of normalization for particular tasks, in which it can effectively solve the key issues.
Normalization Techniques in Training DNNs:
— phalanx (@ZFPhalanx) September 29, 2020
Methodology, Analysis and Applicationhttps://t.co/wNzCQT1dUB pic.twitter.com/GIiLYeKuBs
6. Sim-to-Real Transfer in Deep Reinforcement Learning for Robotics: a Survey
Wenshuai Zhao, Jorge Peña Queralta, Tomi Westerlund
Deep reinforcement learning has recently seen huge success across multiple areas in the robotics domain. Owing to the limitations of gathering real-world data, i.e., sample inefficiency and the cost of collecting it, simulation environments are utilized for training the different agents. This not only aids in providing a potentially infinite data source, but also alleviates safety concerns with real robots. Nonetheless, the gap between the simulated and real worlds degrades the performance of the policies once the models are transferred into real robots. Multiple research efforts are therefore now being directed towards closing this sim-to-real gap and accomplish more efficient policy transfer. Recent years have seen the emergence of multiple methods applicable to different domains, but there is a lack, to the best of our knowledge, of a comprehensive review summarizing and putting into context the different methods. In this survey paper, we cover the fundamental background behind sim-to-real transfer in deep reinforcement learning and overview the main methods being utilized at the moment: domain randomization, domain adaptation, imitation learning, meta-learning and knowledge distillation. We categorize some of the most relevant recent works, and outline the main application scenarios. Finally, we discuss the main opportunities and challenges of the different approaches and point to the most promising directions.
Sim-to-Real Transfer in Deep Reinforcement Learning for Robotics: a Surveyhttps://t.co/DRUQ8Y1LgZ pic.twitter.com/vkkPu27vYH
— sim2real (@sim2realAIorg) September 29, 2020
7. Texture Memory-Augmented Deep Patch-Based Image Inpainting
Rui Xu, Minghao Guo, Jiaqi Wang, Xiaoxiao Li, Bolei Zhou, Chen Change Loy
Patch-based methods and deep networks have been employed to tackle image inpainting problem, with their own strengths and weaknesses. Patch-based methods are capable of restoring a missing region with high-quality texture through searching nearest neighbor patches from the unmasked regions. However, these methods bring problematic contents when recovering large missing regions. Deep networks, on the other hand, show promising results in completing large regions. Nonetheless, the results often lack faithful and sharp details that resemble the surrounding area. By bringing together the best of both paradigms, we propose a new deep inpainting framework where texture generation is guided by a texture memory of patch samples extracted from unmasked regions. The framework has a novel design that allows texture memory retrieval to be trained end-to-end with the deep inpainting network. In addition, we introduce a patch distribution loss to encourage high-quality patch synthesis. The proposed method shows superior performance both qualitatively and quantitatively on three challenging image benchmarks, i.e., Places, CelebA-HQ, and Paris Street-View datasets.
Texture Memory-Augmented Deep Patch-Based Image Inpainting
— AK (@ak92501) September 29, 2020
pdf: https://t.co/hT0SJScW1s
abs: https://t.co/be9R1hY9Xm pic.twitter.com/JfkHIfz0Ow
8. Is Reinforcement Learning More Difficult Than Bandits? A Near-optimal Algorithm Escaping the Curse of Horizon
Zihan Zhang, Xiangyang Ji, Simon S. Du
Episodic reinforcement learning and contextual bandits are two widely studied sequential decision-making problems. Episodic reinforcement learning generalizes contextual bandits and is often perceived to be more difficult due to long planning horizon and unknown state-dependent transitions. The current paper shows that the long planning horizon and the unknown state-dependent transitions (at most) pose little additional difficulty on sample complexity. We consider the episodic reinforcement learning with states, actions, planning horizon , total reward bounded by , and the agent plays for episodes. We propose a new algorithm, \textbf{M}onotonic \textbf{V}alue \textbf{P}ropagation (MVP), which relies on a new Bernstein-type bonus. The new bonus only requires tweaking the \emph{constants} to ensure optimism and thus is significantly simpler than existing bonus constructions. We show MVP enjoys an regret, approaching the lower bound of \emph{contextual bandits}. Notably, this result 1) \emph{exponentially} improves the state-of-the-art polynomial-time algorithms by Dann et al. [2019], Zanette et al. [2019] and Zhang et al. [2020] in terms of the dependency on , and 2) \emph{exponentially} improves the running time in [Wang et al. 2020] and significantly improves the dependency on , and in sample complexity.
Reinforcement Learning is a strict generalization of bandits due to long planning horizon and unknown transition, but is RL fundamentally harder? Our new algorithm for tabular RL, Monotonic Value Propogation (MVP), approaches the bandits' lower bound!https://t.co/inwDpzSOQ1 pic.twitter.com/qk544zqfeV
— Simon Shaolei Du (@SimonShaoleiDu) September 29, 2020
9. SceneGen: Generative Contextual Scene Augmentation using Scene Graph Priors
Mohammad Keshavarzi, Aakash Parikh, Xiyu Zhai, Melody Mao, Luisa Caldas, Allen Yang
Spatial computing experiences are constrained by the real-world surroundings of the user. In such experiences, augmenting virtual objects to existing scenes require a contextual approach, where geometrical conflicts are avoided, and functional and plausible relationships to other objects are maintained in the target environment. Yet, due to the complexity and diversity of user environments, automatically calculating ideal positions of virtual content that is adaptive to the context of the scene is considered a challenging task. Motivated by this problem, in this paper we introduce SceneGen, a generative contextual augmentation framework that predicts virtual object positions and orientations within existing scenes. SceneGen takes a semantically segmented scene as input, and outputs positional and orientational probability maps for placing virtual content. We formulate a novel spatial Scene Graph representation, which encapsulates explicit topological properties between objects, object groups, and the room. We believe providing explicit and intuitive features plays an important role in informative content creation and user interaction of spatial computing settings, a quality that is not captured in implicit models. We use kernel density estimation (KDE) to build a multivariate conditional knowledge model trained using prior spatial Scene Graphs extracted from real-world 3D scanned data. To further capture orientational properties, we develop a fast pose annotation tool to extend current real-world datasets with orientational labels. Finally, to demonstrate our system in action, we develop an Augmented Reality application, in which objects can be contextually augmented in real-time.
SceneGen is a framework to place virtual objects in a target scene. The placement is guided by the generative model that captures functional and topological relationships learned from real-world datasets.https://t.co/druyXewJta pic.twitter.com/4VATxNXBo1
— Ankur Handa (@ankurhandos) September 29, 2020
SceneGen: Generative Contextual Scene Augmentation using Scene Graph Priors
— AK (@ak92501) September 29, 2020
pdf: https://t.co/fLaNd6rdWK
abs: https://t.co/B3zVCsk0se pic.twitter.com/4kb2lxIot1
10. COVID-19’s (mis)information ecosystem on Twitter: How partisanship boosts the spread of conspiracy narratives on German speaking Twitter
Morteza Shahrezaye, Miriam Meckel, Léa Steinacker, Viktor Suter
In late 2019, the gravest pandemic in a century began spreading across the world. A state of uncertainty related to what has become known as SARS-CoV-2 has since fueled conspiracy narratives on social media about the origin, transmission and medical treatment of and vaccination against the resulting disease, COVID-19. Using social media intelligence to monitor and understand the proliferation of conspiracy narratives is one way to analyze the distribution of misinformation on the pandemic. We analyzed more than 9.5M German language tweets about COVID-19. The results show that only about 0.6% of all those tweets deal with conspiracy theory narratives. We also found that the political orientation of users correlates with the volume of content users contribute to the dissemination of conspiracy narratives, implying that partisan communicators have a higher motivation to take part in conspiratorial discussions on Twitter. Finally, we showed that contrary to other studies, automated accounts do not significantly influence the spread of misinformation in the German speaking Twitter sphere. They only represent about 1.31% of all conspiracy-related activities in our database.
In our new paper on the spread of COVID-19 misinformation on 🇩🇪-Twitter we show that
— Léa Steinacker (@leasteinacker) September 29, 2020
🔸partisan accounts contribute relatively more to conspiratorial narratives
🔸bots don't significantly influence the spread of misinformation in 🇩🇪-Twitter (1.31%)https://t.co/Ih8B2tJFwb pic.twitter.com/hBLBZcoZCd
11. Operads for Designing Systems of Systems
John C. Baez, John Foley
System of systems engineering seeks to analyze, design and deploy collections of systems that together can flexibly address an array of complex tasks. In the Complex Adaptive System Composition and Design Environment program, we developed “network operads” as a tool for designing and tasking systems of systems, and applied them to domains including maritime search and rescue. The network operad formalism offers new ways to handle changing levels of abstraction in system-of-system design and tasking.
https://t.co/PFW5M1MYJs (2 page extended abstract) pic.twitter.com/yoEBINJYTe
— julesh (@_julesh_) September 29, 2020
12. Deep Transformers with Latent Depth
Xian Li, Asa Cooper Stickland, Yuqing Tang, Xiang Kong
The Transformer model has achieved state-of-the-art performance in many sequence modeling tasks. However, how to leverage model capacity with large or variable depths is still an open challenge. We present a probabilistic framework to automatically learn which layer(s) to use by learning the posterior distributions of layer selection. As an extension of this framework, we propose a novel method to train one shared Transformer network for multilingual machine translation with different layer selection posteriors for each language pair. The proposed method alleviates the vanishing gradient issue and enables stable training of deep Transformers (e.g. 100 layers). We evaluate on WMT English-German machine translation and masked language modeling tasks, where our method outperforms existing approaches for training deeper Transformers. Experiments on multilingual machine translation demonstrate that this approach can effectively leverage increased model capacity and bring universal improvement for both many-to-one and one-to-many translation with diverse language pairs.
Deep Transformers with Latent Depth
— Aran Komatsuzaki (@arankomatsuzaki) September 29, 2020
Depth-adaptive conditional computation of Transformers. Improves performance-computes trade-off on multilingual translation and MLM. https://t.co/9dewhvQGJF pic.twitter.com/zBmfj9kNJv
13. Dissecting Lottery Ticket Transformers: Structural and Behavioral Study of Sparse Neural Machine Translation
Rajiv Movva, Jason Y. Zhao
Recent work on the lottery ticket hypothesis has produced highly sparse Transformers for NMT while maintaining BLEU. However, it is unclear how such pruning techniques affect a model’s learned representations. By probing sparse Transformers, we find that complex semantic information is first to be degraded. Analysis of internal activations reveals that higher layers diverge most over the course of pruning, gradually becoming less complex than their dense counterparts. Meanwhile, early layers of sparse models begin to perform more encoding. Attention mechanisms remain remarkably consistent as sparsity increases.
Our paper “Dissecting Lottery Ticket Transformers: Structural and Behavioral Study of Sparse Neural Machine Translation” is accepted to BlackboxNLP 2020 (at EMNLP this year)! arXiv link: https://t.co/XOX6ZHKWUD. Brief summary to follow: 1/15
— Raj Movva (@rajivmovva) September 29, 2020
14. Semi-Supervised Learning for In-Game Expert-Level Music-to-Dance Translation
Yinglin Duan, Tianyang Shi, Zhengxia Zou, Jia Qin, Yifei Zhao, Yi Yuan, Jie Hou, Xiang Wen, Changjie Fan
Music-to-dance translation is a brand-new and powerful feature in recent role-playing games. Players can now let their characters dance along with specified music clips and even generate fan-made dance videos. Previous works of this topic consider music-to-dance as a supervised motion generation problem based on time-series data. However, these methods suffer from limited training data pairs and the degradation of movements. This paper provides a new perspective for this task where we re-formulate the translation problem as a piece-wise dance phrase retrieval problem based on the choreography theory. With such a design, players are allowed to further edit the dance movements on top of our generation while other regression based methods ignore such user interactivity. Considering that the dance motion capture is an expensive and time-consuming procedure which requires the assistance of professional dancers, we train our method under a semi-supervised learning framework with a large unlabeled dataset (20x than labeled data) collected. A co-ascent mechanism is introduced to improve the robustness of our network. Using this unlabeled dataset, we also introduce self-supervised pre-training so that the translator can understand the melody, rhythm, and other components of music phrases. We show that the pre-training significantly improves the translation accuracy than that of training from scratch. Experimental results suggest that our method not only generalizes well over various styles of music but also succeeds in expert-level choreography for game players.
Semi-Supervised Learning for In-Game Expert-Level Music-to-Dance Translation
— AK (@ak92501) September 29, 2020
pdf: https://t.co/orathZ9c9i
abs: https://t.co/ID3dLih2Ps pic.twitter.com/sK08JhHnoo