matlok
's Collections
Papers - Google
updated
Lumiere: A Space-Time Diffusion Model for Video Generation
Paper
•
2401.12945
•
Published
•
86
Long-form factuality in large language models
Paper
•
2403.18802
•
Published
•
24
ObjectDrop: Bootstrapping Counterfactuals for Photorealistic Object
Removal and Insertion
Paper
•
2403.18818
•
Published
•
25
TC4D: Trajectory-Conditioned Text-to-4D Generation
Paper
•
2403.17920
•
Published
•
16
Transforming and Combining Rewards for Aligning Large Language Models
Paper
•
2402.00742
•
Published
•
11
Leverage the Average: an Analysis of KL Regularization in RL
Paper
•
2003.14089
•
Published
•
2
Snap-it, Tap-it, Splat-it: Tactile-Informed 3D Gaussian Splatting for
Reconstructing Challenging Surfaces
Paper
•
2403.20275
•
Published
•
8
Gecko: Versatile Text Embeddings Distilled from Large Language Models
Paper
•
2403.20327
•
Published
•
47
Localizing Paragraph Memorization in Language Models
Paper
•
2403.19851
•
Published
•
13
Streaming Dense Video Captioning
Paper
•
2404.01297
•
Published
•
11
FormNetV2: Multimodal Graph Contrastive Learning for Form Document
Information Extraction
Paper
•
2305.02549
•
Published
•
6
FormNet: Structural Encoding beyond Sequential Modeling in Form Document
Information Extraction
Paper
•
2203.08411
•
Published
•
1
ETC: Encoding Long and Structured Inputs in Transformers
Paper
•
2004.08483
•
Published
•
1
GQA: Training Generalized Multi-Query Transformer Models from Multi-Head
Checkpoints
Paper
•
2305.13245
•
Published
•
5
Sparse Upcycling: Training Mixture-of-Experts from Dense Checkpoints
Paper
•
2212.05055
•
Published
•
5
Exploring the Limits of Transfer Learning with a Unified Text-to-Text
Transformer
Paper
•
1910.10683
•
Published
•
9
Bigger is not Always Better: Scaling Properties of Latent Diffusion
Models
Paper
•
2404.01367
•
Published
•
20
Mixture-of-Depths: Dynamically allocating compute in transformer-based
language models
Paper
•
2404.02258
•
Published
•
104
Training LLMs over Neurally Compressed Text
Paper
•
2404.03626
•
Published
•
21
BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions
Paper
•
1905.10044
•
Published
•
1
Program Induction by Rationale Generation : Learning to Solve and
Explain Algebraic Word Problems
Paper
•
1705.04146
•
Published
•
1
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
Paper
•
2201.11903
•
Published
•
9
Premise Order Matters in Reasoning with Large Language Models
Paper
•
2402.08939
•
Published
•
25
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language
Understanding
Paper
•
1804.07461
•
Published
•
4
Program of Thoughts Prompting: Disentangling Computation from Reasoning
for Numerical Reasoning Tasks
Paper
•
2211.12588
•
Published
•
3
BERT Rediscovers the Classical NLP Pipeline
Paper
•
1905.05950
•
Published
•
2
Prompt-to-Prompt Image Editing with Cross Attention Control
Paper
•
2208.01626
•
Published
•
2
WIT: Wikipedia-based Image Text Dataset for Multimodal Multilingual
Machine Learning
Paper
•
2103.01913
•
Published
•
2
No "Zero-Shot" Without Exponential Data: Pretraining Concept Frequency
Determines Multimodal Model Performance
Paper
•
2404.04125
•
Published
•
27
Leave No Context Behind: Efficient Infinite Context Transformers with
Infini-attention
Paper
•
2404.07143
•
Published
•
103
RecurrentGemma: Moving Past Transformers for Efficient Open Language
Models
Paper
•
2404.07839
•
Published
•
42
PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive
Summarization
Paper
•
1912.08777
•
Published
•
2
Leveraging Pre-trained Checkpoints for Sequence Generation Tasks
Paper
•
1907.12461
•
Published
•
1
Phenaki: Variable Length Video Generation From Open Domain Textual
Description
Paper
•
2210.02399
•
Published
•
3
Scaling Up Models and Data with t5x and seqio
Paper
•
2203.17189
•
Published
•
1
Reuse Your Rewards: Reward Model Transfer for Zero-Shot Cross-Lingual
Alignment
Paper
•
2404.12318
•
Published
•
14
Does Gaussian Splatting need SFM Initialization?
Paper
•
2404.12547
•
Published
•
8
Batch Normalization: Accelerating Deep Network Training by Reducing
Internal Covariate Shift
Paper
•
1502.03167
•
Published
•
1
Hyper-X: A Unified Hypernetwork for Multi-Task Multilingual Transfer
Paper
•
2205.12148
•
Published
•
2
Capabilities of Gemini Models in Medicine
Paper
•
2404.18416
•
Published
•
23
Stylus: Automatic Adapter Selection for Diffusion Models
Paper
•
2404.18928
•
Published
•
14
DOCCI: Descriptions of Connected and Contrasting Images
Paper
•
2404.19753
•
Published
•
11
What needs to go right for an induction head? A mechanistic study of
in-context learning circuits and their formation
Paper
•
2404.07129
•
Published
•
3
Greedy Growing Enables High-Resolution Pixel-Based Diffusion Models
Paper
•
2405.16759
•
Published
•
7
TACT: Advancing Complex Aggregative Reasoning with Information
Extraction Tools
Paper
•
2406.03618
•
Published
•
2
Neural Autoregressive Distribution Estimation
Paper
•
1605.02226
•
Published
•
1
Tx-LLM: A Large Language Model for Therapeutics
Paper
•
2406.06316
•
Published
•
15
Classifier-Free Diffusion Guidance
Paper
•
2207.12598
•
Published
•
2
Improve Mathematical Reasoning in Language Models by Automated Process
Supervision
Paper
•
2406.06592
•
Published
•
25
Revisiting Unreasonable Effectiveness of Data in Deep Learning Era
Paper
•
1707.02968
•
Published
•
1
Inception-v4, Inception-ResNet and the Impact of Residual Connections on
Learning
Paper
•
1602.07261
•
Published
•
1
Data curation via joint example selection further accelerates multimodal
learning
Paper
•
2406.17711
•
Published
•
3
TokenFormer: Rethinking Transformer Scaling with Tokenized Model
Parameters
Paper
•
2410.23168
•
Published
•
22