In this work, we introduce Boolformer, the first Transformer architectur...
Datasets that pair Knowledge Graphs (KG) and text together (KG-T) can be...
A fairly reliable trend in deep reinforcement learning is that the
perfo...
Diffusion models have demonstrated excellent potential for generating di...
Autoregressive models for text sometimes generate repetitive and low-qua...
Diffusion models have recently become the de-facto approach for generati...
Training stability is of great importance to Transformers. In this work,...
Recent Self-Supervised Learning (SSL) methods are able to learn feature
...
Novel view synthesis from a single image requires inferring occluded reg...
Diffusion models (DMs) have recently emerged as SoTA tools for generativ...
We introduce GAUDI, a generative model capable of capturing the distribu...
Self-attention mechanisms model long-range context by using pairwise
att...
In this paper, we study the representation of neural networks from the v...
Modeling the world can benefit robot learning by providing a rich traini...
Including memory banks in a natural language processing architecture
inc...
We introduce Attention Free Transformer (AFT), an efficient variant of
T...
We study the problem of directly optimizing arbitrary non-differentiable...
We propose a framework for learning neural scene representations directl...
In most machine learning training paradigms a fixed, often handcrafted, ...
With recent progress in graphics, it has become more tractable to train
...