1. Generative Language Modeling for Automated Theorem Proving
Stanislas Polu, Ilya Sutskever
- retweets: 242, favorites: 1126 (09/10/2020 11:36:02)
- links: abs | pdf
- cs.LG | cs.AI | cs.CL | stat.ML
We explore the application of transformer-based language models to automated theorem proving. This work is motivated by the possibility that a major limitation of automated theorem provers compared to humans — the generation of original mathematical terms — might be addressable via generation from language models. We present an automated prover and proof assistant, GPT-f, for the Metamath formalization language, and analyze its performance. GPT-f found new short proofs that were accepted into the main Metamath library, which is to our knowledge, the first time a deep-learning based system has contributed proofs that were adopted by a formal mathematics community.
Posted my first paper on arXiv💥🙌
— Stanislas Polu (@spolu) September 9, 2020
GPT-f is a Transformer-based automated theorem prover. We show that Transformer + Search is suitable to formal reasoning and continuous self-improvement 🦾https://t.co/VllDcCV3Kc pic.twitter.com/5ttVX0MNBC
Generative Language Modeling for Automated
— Aran Komatsuzaki (@arankomatsuzaki) September 9, 2020
Theorem Proving
GPT-f, a Transformer LM pretrained on arXiv and trained through continuous self improvement for theorem proving, finds novel short proofs. https://t.co/NKIY7WgVRm pic.twitter.com/iYzqnHAddJ
GPT-fはTransformerを使った言語モデルで自動定理証明を行う。ウェブ収集された数学関連の論文などで事前学習し、証明目標で条件付した証明経路の生成確率を使って探索する。発見された中で23個の証明はこれまでの証明より短くわかりやすくMetamath証明ライブラリに加えられたhttps://t.co/RiMM5QGQQB
— Daisuke Okanohara (@hillbig) September 9, 2020
OpenAIのAI研究(GPT-f)
— 小猫遊りょう(たかにゃし・りょう) (@jaguring1) September 9, 2020
言語モデルを用いて自動定理証明。GPT-fが、Metamathライブラリに受け入れられた新しい短い証明を発見。これは、ディープラーニングがおそらく初めて正式な数学コミュニティに採用された証明を提供した。トランスフォーマーは形式推論にも適してるhttps://t.co/z2PMMgyoLJ
Generative Language Modeling for Automated Theorem Proving
— AK (@ak92501) September 9, 2020
pdf: https://t.co/gmD9HK1R80
abs: https://t.co/qstcD7UYyZ pic.twitter.com/nKKECgFhpT
OpenAIのGPT研究 (pdf)
— 小猫遊りょう(たかにゃし・りょう) (@jaguring1) September 9, 2020
①GPT(2018/6/11)https://t.co/zJoazUBUez………
②GPT-2(2019/2/14)https://t.co/gVN6kkxAHX………
③GPT-3(2020/5/28)https://t.co/4qlMV7VDoO……
④image GPT(2020/6/17)https://t.co/2LOWDz78LV
⑤GPT-f(2020/9/7)https://t.co/sNWyPweIoJ
Compute is eating AI. @OpenAI trains a state of the art theorem prover using a GPT architecture. I was waiting for this to happen, neat!
— Connor Leahy (@NPCollapse) September 9, 2020
Though for real...wtf is this proof? Maybe the gains were just that you need an AI to understand this language haha!https://t.co/VXV4gWZcNK pic.twitter.com/cNpqNiZmcc
2. Randomness Concerns When Deploying Differential Privacy
Simson L. Garfinkel, Philip Leclerc
The U.S. Census Bureau is using differential privacy (DP) to protect confidential respondent data collected for the 2020 Decennial Census of Population & Housing. The Census Bureau’s DP system is implemented in the Disclosure Avoidance System (DAS) and requires a source of random numbers. We estimate that the 2020 Census will require roughly 90TB of random bytes to protect the person and household tables. Although there are critical differences between cryptography and DP, they have similar requirements for randomness. We review the history of random number generation on deterministic computers, including von Neumann’s “middle-square” method, Mersenne Twister (MT19937) (previously the default NumPy random number generator, which we conclude is unacceptable for use in production privacy-preserving systems), and the Linux /dev/urandom device. We also review hardware random number generator schemes, including the use of so-called “Lava Lamps” and the Intel Secure Key RDRAND instruction. We finally present our plan for generating random bits in the Amazon Web Services (AWS) environment using AES-CTR-DRBG seeded by mixing bits from /dev/urandom and the Intel Secure Key RDSEED instruction, a compromise of our desire to rely on a trusted hardware implementation, the unease of our external reviewers in trusting a hardware-only implementation, and the need to generate so many random bits.
This paper contains the "philosophy of differential privacy" stuff I've been wanting to get out, as well as a fun history of PRNGs and it even tells you what we're doing for the 2020 Census! https://t.co/VGUlIkAWBh
— Simson Garfinkel (@xchatty) September 9, 2020
(Thread)
— Gautam Kamath (@thegautamkamath) September 9, 2020
Some highlights from this paper by Garfinkel (@xchatty) and Leclerc: https://t.co/vYgfyt8nkL.
First, you WON'T BELIEVE the amount of randomness required! While the raw data for the 2020 US Census takes only "a few tens of [GB]," *90 TB* of random data will be used! 1/n pic.twitter.com/nRp85gGQFb
3. Robust Conversational AI with Grounded Text Generation
Jianfeng Gao, Baolin Peng, Chunyuan Li, Jinchao Li, Shahin Shayandeh, Lars Liden, Heung-Yeung Shum
This article presents a hybrid approach based on a Grounded Text Generation (GTG) model to building robust task bots at scale. GTG is a hybrid model which uses a large-scale Transformer neural network as its backbone, combined with symbol-manipulation modules for knowledge base inference and prior knowledge encoding, to generate responses grounded in dialog belief state and real-world knowledge for task completion. GTG is pre-trained on large amounts of raw text and human conversational data, and can be fine-tuned to complete a wide range of tasks. The hybrid approach and its variants are being developed simultaneously by multiple research teams. The primary results reported on task-oriented dialog benchmarks are very promising, demonstrating the big potential of this approach. This article provides an overview of this progress and discusses related methods and technologies that can be incorporated for building robust conversational AI systems.
A method to improve conversational AI systems through a hybrid approach for grounded text generation.
— elvis (@omarsar0) September 9, 2020
It combines Transformers and modules that encode prior knowledge & knowledge base inference which helps to generate fluent & on-topic responses.https://t.co/Sk7QsoJNgq pic.twitter.com/5CLWISoQpo
Robust Conversational AI with Grounded Text Generation
— AK (@ak92501) September 9, 2020
pdf: https://t.co/Okofj9kWcC
abs: https://t.co/AFYHal00FG pic.twitter.com/06U29v10n3
4. Deep Cyclic Generative Adversarial Residual Convolutional Networks for Real Image Super-Resolution
Rao Muhammad Umer, Christian Micheloni
Recent deep learning based single image super-resolution (SISR) methods mostly train their models in a clean data domain where the low-resolution (LR) and the high-resolution (HR) images come from noise-free settings (same domain) due to the bicubic down-sampling assumption. However, such degradation process is not available in real-world settings. We consider a deep cyclic network structure to maintain the domain consistency between the LR and HR data distributions, which is inspired by the recent success of CycleGAN in the image-to-image translation applications. We propose the Super-Resolution Residual Cyclic Generative Adversarial Network (SRResCycGAN) by training with a generative adversarial network (GAN) framework for the LR to HR domain translation in an end-to-end manner. We demonstrate our proposed approach in the quantitative and qualitative experiments that generalize well to the real image super-resolution and it is easy to deploy for the mobile/embedded devices. In addition, our SR results on the AIM 2020 Real Image SR Challenge datasets demonstrate that the proposed SR approach achieves comparable results as the other state-of-art methods.
Deep Cyclic Generative Adversarial Residual Convolutional Networks for Real Image Super-Resolution
— AK (@ak92501) September 9, 2020
pdf: https://t.co/GhYBfgMctN
abs: https://t.co/BNXeKciPAn
github: https://t.co/rRYf5BbJ3m pic.twitter.com/CZkE70OPqr