Unnamed: 0
int64 0
217
| id
int64 1,526,373,200B
1,546,707,910B
| tweet_text
stringlengths 76
140
| paper_reference
stringlengths 20
113
| like_count
int64 8
2.72k
|
---|---|---|---|---|
0 | 1,546,707,909,748,342,800 | High-resource Language-specific Training for Multilingual Neural Machine Translation
abs: https://t.co/fYrwIPVpV2 https://t.co/b23EVZ6J5O | High-resource Language-specific Training for Multilingual Neural Machine Translation | 11 |
1 | 1,546,669,556,789,387,300 | Exploring Length Generalization in Large Language Models
abs: https://t.co/7Gphb7Q8jJ https://t.co/cCpLTSrXfR | Exploring Length Generalization in Large Language Models | 17 |
2 | 1,546,667,351,885,729,800 | LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action
abs:… https://t.co/lCk3P8KIwM | LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action | 32 |
3 | 1,546,665,636,734,140,400 | Scaling the Number of Tasks in Continual Learning
abs: https://t.co/F4HxAxGUpI https://t.co/cyvXSBKthk | Scaling the Number of Tasks in Continual Learning | 47 |
4 | 1,546,707,909,748,342,800 | High-resource Language-specific Training for Multilingual Neural Machine Translation
abs: https://t.co/fYrwIPVpV2 https://t.co/b23EVZ6J5O | High-resource Language-specific Training for Multilingual Neural Machine Translation | 11 |
5 | 1,546,669,556,789,387,300 | Exploring Length Generalization in Large Language Models
abs: https://t.co/7Gphb7Q8jJ https://t.co/cCpLTSrXfR | Exploring Length Generalization in Large Language Models | 17 |
6 | 1,546,667,351,885,729,800 | LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action
abs:… https://t.co/lCk3P8KIwM | LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action | 32 |
7 | 1,546,665,636,734,140,400 | Scaling the Number of Tasks in Continual Learning
abs: https://t.co/F4HxAxGUpI https://t.co/cyvXSBKthk | Scaling the Number of Tasks in Continual Learning | 47 |
8 | 1,546,379,163,803,721,700 | CausalAgents: A Robustness Benchmark for Motion Forecasting using Causal Relationships
abs: https://t.co/ozIrQ7gx68 https://t.co/gSGfnsZbji | CausalAgents: A Robustness Benchmark for Motion Forecasting using Causal Relationships | 53 |
9 | 1,546,376,106,122,567,700 | The Harvard USPTO Patent Dataset: A Large-Scale, Well-Structured, and Multi-Purpose Corpus of Patent Applications
a… https://t.co/TOPpVPQbM8 | The Harvard USPTO Patent Dataset: A Large-Scale, Well-Structured, and Multi-Purpose Corpus of Patent Applications | 11 |
10 | 1,546,375,104,262,725,600 | Code Translation with Compiler Representations
abs: https://t.co/nTT3dmXH4c
method improves upon the state of the… https://t.co/wD4SozbilN | Code Translation with Compiler Representations | 127 |
11 | 1,546,363,822,121,820,200 | End-to-End Binaural Speech Synthesis
abs: https://t.co/tR86cSAjQO
project page: https://t.co/nB1iSV68U2
end-to-end… https://t.co/OTzfVZTFqb | End-to-End Binaural Speech Synthesis | 58 |
12 | 1,545,243,820,496,937,000 | Cross-Scale Vector Quantization for Scalable Neural Speech Coding
abs: https://t.co/AbE9rP0ApQ https://t.co/pZXUTNipgs | Cross-Scale Vector Quantization for Scalable Neural Speech Coding | 25 |
13 | 1,545,240,373,328,593,000 | Finding Fallen Objects Via Asynchronous Audio-Visual Integration
abs: https://t.co/mv9Rvl0hFA
project page:… https://t.co/N8l4zaP9bH | Finding Fallen Objects Via Asynchronous Audio-Visual Integration | 33 |
14 | 1,545,228,848,391,938,000 | Back to the Source: Diffusion-Driven Test-Time Adaptation
abs: https://t.co/5jmESOLQxG https://t.co/cI5UFyQI0B | Back to the Source: Diffusion-Driven Test-Time Adaptation | 82 |
15 | 1,544,897,525,664,170,000 | When does Bias Transfer in Transfer Learning?
abs: https://t.co/tf8FWyf8Ge https://t.co/0l6vy8RHXI | When does Bias Transfer in Transfer Learning? | 135 |
16 | 1,544,865,587,343,630,300 | Transformers are Adaptable Task Planners
abs: https://t.co/6lgFJD2Olt
TTP can be pre-trained on multiple preferenc… https://t.co/XrolcxlV22 | Transformers are Adaptable Task Planners | 82 |
17 | 1,544,853,650,316,599,300 | Ultra-Low-Bitrate Speech Coding with Pretrained Transformers
abs: https://t.co/rYRe5N7Bqu https://t.co/zOsCY53r2s | Ultra-Low-Bitrate Speech Coding with Pretrained Transformers | 34 |
18 | 1,544,721,641,049,145,300 | CLEAR: Improving Vision-Language Navigation with Cross-Lingual, Environment-Agnostic Representations
abs:… https://t.co/6ng3UArKdE | CLEAR: Improving Vision-Language Navigation with Cross-Lingual, Environment-Agnostic Representations | 52 |
19 | 1,544,521,037,274,046,500 | An Empirical Study of Implicit Regularization in Deep Offline RL
abs: https://t.co/rCjHkQ2jwL https://t.co/8hJOsVA6D0 | An Empirical Study of Implicit Regularization in Deep Offline RL | 45 |
20 | 1,544,519,268,234,154,000 | Offline RL Policies Should be Trained to be Adaptive
abs: https://t.co/kC7TPSOTt2 https://t.co/Ox2D028P33 | Offline RL Policies Should be Trained to be Adaptive | 34 |
21 | 1,544,491,557,293,854,700 | Efficient Representation Learning via Adaptive Context Pooling
abs: https://t.co/zZzezhvbN7 https://t.co/xJoStGBSqp | Efficient Representation Learning via Adaptive Context Pooling | 163 |
22 | 1,544,488,616,734,429,200 | CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning
abs:… https://t.co/HqXmDpaUEh | CodeRL: Mastering Code Generation through Pretrained Models and Deep Reinforcement Learning | 102 |
23 | 1,544,485,593,991,811,000 | How Much More Data Do I Need? Estimating Requirements for Downstream Tasks
abs: https://t.co/RNXT4IRIaL https://t.co/uJGrEfgaAv | How Much More Data Do I Need? Estimating Requirements for Downstream Tasks | 230 |
24 | 1,544,483,235,542,990,800 | Neural Networks and the Chomsky Hierarchy
abs: https://t.co/u6Jl2WvKMr
sota architectures, such as LSTMs and Trans… https://t.co/DyHnH8Q8z7 | Neural Networks and the Chomsky Hierarchy | 209 |
25 | 1,544,207,617,102,332,000 | GlowVC: Mel-spectrogram space disentangling model for language-independent text-free voice conversion
abs:… https://t.co/kFYdKhrhSA | GlowVC: Mel-spectrogram space disentangling model for language-independent text-free voice conversion | 19 |
26 | 1,544,201,186,739,458,000 | Object Representations as Fixed Points: Training Iterative Refinement Algorithms with Implicit Differentiation
abs:… https://t.co/yL9kWlUYfs | Object Representations as Fixed Points: Training Iterative Refinement Algorithms with Implicit Differentiation | 112 |
27 | 1,544,193,877,053,161,500 | WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents
abs: https://t.co/8hZyMt90Rv
pro… https://t.co/eHzGN2GHqj | WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents | 52 |
28 | 1,544,127,293,660,037,000 | UserLibri: A Dataset for ASR Personalization Using Only Text
abs: https://t.co/0bug7OWU42 https://t.co/OMqJSGlqDx | UserLibri: A Dataset for ASR Personalization Using Only Text | 9 |
29 | 1,543,981,460,964,708,400 | LaserMix for Semi-Supervised LiDAR Semantic Segmentation
abs: https://t.co/SvqHy1y7LI
project page:… https://t.co/jbQtQiDbDy | LaserMix for Semi-Supervised LiDAR Semantic Segmentation | 74 |
30 | 1,543,766,808,309,670,000 | Rethinking Optimization with Differentiable Simulation from a Global Perspective
abs: https://t.co/trEcw4VZb2
proje… https://t.co/1UsI0q03IL | Rethinking Optimization with Differentiable Simulation from a Global Perspective | 94 |
31 | 1,543,763,117,515,182,000 | Visual Pre-training for Navigation: What Can We Learn from Noise?
abs: https://t.co/Rn5UGvvMMz
github:… https://t.co/eKeMSlBxVx | Visual Pre-training for Navigation: What Can We Learn from Noise? | 134 |
32 | 1,543,759,817,449,390,000 | DeepSpeed Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale
abs:… https://t.co/IbF6IdUDj7 | DeepSpeed Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale | 120 |
33 | 1,543,757,524,356,272,000 | When Does Differentially Private Learning Not Suffer in High Dimensions?
abs: https://t.co/yws7BhoBaP https://t.co/bD2Gz6B3GU | When Does Differentially Private Learning Not Suffer in High Dimensions? | 28 |
34 | 1,542,740,430,084,792,300 | Implicit Neural Spatial Filtering for Multichannel Source Separation in the Waveform Domain
abs:… https://t.co/3cNoOlr5SD | Implicit Neural Spatial Filtering for Multichannel Source Separation in the Waveform Domain | 31 |
35 | 1,542,713,456,268,304,400 | Denoised MDPs: Learning World Models Better Than the World Itself
abs: https://t.co/CPwlF0soWZ
project page:… https://t.co/5BBwGXYZ2l | Denoised MDPs: Learning World Models Better Than the World Itself | 98 |
36 | 1,542,712,192,746,782,700 | Forecasting Future World Events with Neural Networks
abs: https://t.co/tD8F0ZC1rC
github: https://t.co/v8HZgye0ZH… https://t.co/eJaakYSUSw | Forecasting Future World Events with Neural Networks | 77 |
37 | 1,542,709,853,516,431,400 | Learning Iterative Reasoning through Energy Minimization
abs: https://t.co/WDLx1hKPqG
project page:… https://t.co/oDEClr0ho1 | Learning Iterative Reasoning through Energy Minimization | 125 |
38 | 1,542,709,029,964,849,200 | Improving the Generalization of Supervised Models
abs: https://t.co/3CzEuuxvHt
project page: https://t.co/uSjiKvSMN8 https://t.co/ffUkpTL7Ng | Improving the Generalization of Supervised Models | 189 |
39 | 1,542,325,850,036,752,400 | RegMixup: Mixup as a Regularizer Can Surprisingly Improve Accuracy and Out Distribution Robustness
abs:… https://t.co/iFAou98U0X | RegMixup: Mixup as a Regularizer Can Surprisingly Improve Accuracy and Out Distribution Robustness | 172 |
40 | 1,542,316,111,743,664,000 | Masked World Models for Visual Control
abs: https://t.co/eZx53zuqnm
project page: https://t.co/hgZwrV3zO5
Can MAE… https://t.co/UfybFx81uj | Masked World Models for Visual Control | 83 |
41 | 1,542,313,347,835,732,000 | Beyond neural scaling laws: beating power law scaling via data pruning
abs: https://t.co/OFYkTt5b2d https://t.co/7SKXMClaR8 | Beyond neural scaling laws: beating power law scaling via data pruning | 164 |
42 | 1,542,312,585,768,435,700 | 3D-Aware Video Generation
abs: https://t.co/N64ARXFKMJ
project page: https://t.co/5MoGVKqItn https://t.co/uZdLIXWc1P | 3D-Aware Video Generation | 122 |
43 | 1,541,957,148,070,011,000 | DayDreamer: World Models for Physical Robot Learning
abs: https://t.co/quyTQGcjEA
project page:… https://t.co/DD67NUzgJy | DayDreamer: World Models for Physical Robot Learning | 182 |
44 | 1,541,948,699,559,006,200 | Long Range Language Modeling via Gated State Spaces
abs: https://t.co/HEd2lwlGan https://t.co/tPOHv7dP0T | Long Range Language Modeling via Gated State Spaces | 124 |
45 | 1,541,945,827,035,332,600 | ProGen2: Exploring the Boundaries of Protein Language Models
abs: https://t.co/kelWMlhH8r
github:… https://t.co/nzvei5pMJR | ProGen2: Exploring the Boundaries of Protein Language Models | 64 |
46 | 1,541,626,617,490,837,500 | Multitask vocal burst modeling with ResNets and pre-trained paralinguistic Conformers
abs: https://t.co/QZLcoFOeSz https://t.co/315WfiVVRr | Multitask vocal burst modeling with ResNets and pre-trained paralinguistic Conformers | 11 |
47 | 1,541,599,748,624,351,200 | Programmatic Concept Learning for Human Motion Description and Synthesis
abs: https://t.co/uIoxGozwhD
project page:… https://t.co/MmCMQouLF7 | Programmatic Concept Learning for Human Motion Description and Synthesis | 83 |
48 | 1,541,592,312,094,101,500 | Prompting Decision Transformer for Few-Shot Policy Generalization
abs: https://t.co/bD2f4SjRP6
project page:… https://t.co/ZfAxxx6zCu | Prompting Decision Transformer for Few-Shot Policy Generalization | 48 |
49 | 1,541,590,513,241,006,000 | Repository-Level Prompt Generation for Large Language Models of Code
abs: https://t.co/GG1YHoCQdf
github:… https://t.co/Z9fUO4r8sU | Repository-Level Prompt Generation for Large Language Models of Code | 56 |
50 | 1,541,588,372,631,818,200 | Your Autoregressive Generative Model Can be Better If You Treat It as an Energy-Based One
abs:… https://t.co/uJuKxO7XJC | Your Autoregressive Generative Model Can be Better If You Treat It as an Energy-Based One | 121 |
51 | 1,541,226,747,533,922,300 | PSP: Million-level Protein Sequence Dataset for Protein Structure Prediction
abs: https://t.co/yXdFTqRWF3
dataset… https://t.co/ZDNMPI2NVR | PSP: Million-level Protein Sequence Dataset for Protein Structure Prediction | 94 |
52 | 1,541,219,433,259,176,000 | Megapixel Image Generation with Step-Unrolled Denoising Autoencoders
abs: https://t.co/6fX9PseXBT
obtain FID score… https://t.co/HPodJ8xzPx | Megapixel Image Generation with Step-Unrolled Denoising Autoencoders | 147 |
53 | 1,540,184,734,390,706,200 | Walk the Random Walk: Learning to Discover and Reach Goals Without Supervision
abs: https://t.co/NO2vzfdYdS https://t.co/WoN73BzgeQ | Walk the Random Walk: Learning to Discover and Reach Goals Without Supervision | 66 |
54 | 1,540,176,838,017,917,000 | Offline RL for Natural Language Generation with Implicit Language Q Learning
abs: https://t.co/wYTtUgdryZ
project p… https://t.co/xS8JCODxwP | Offline RL for Natural Language Generation with Implicit Language Q Learning | 43 |
55 | 1,540,161,095,930,880,000 | MaskViT: Masked Visual Pre-Training for Video Prediction
abs: https://t.co/uhMEB6ashb
project page:… https://t.co/gbnxrCxUrc | MaskViT: Masked Visual Pre-Training for Video Prediction | 147 |
56 | 1,540,156,319,923,060,700 | The ArtBench Dataset: Benchmarking Generative Models with Artworks
abs: https://t.co/Zzq0A2i5ob
github:… https://t.co/SfQlvTLrk3 | The ArtBench Dataset: Benchmarking Generative Models with Artworks | 200 |
57 | 1,539,811,680,359,796,700 | TiCo: Transformation Invariance and Covariance Contrast for Self-Supervised Visual Representation Learning
abs:… https://t.co/UArbr7zhRE | TiCo: Transformation Invariance and Covariance Contrast for Self-Supervised Visual Representation Learning | 85 |
58 | 1,539,794,210,190,155,800 | Jointist: Joint Learning for Multi-instrument Transcription and Its Applications
abs: https://t.co/xeuPUBcr01
proje… https://t.co/QmyCioKviJ | Jointist: Joint Learning for Multi-instrument Transcription and Its Applications | 18 |
59 | 1,539,780,412,297,330,700 | GEMv2: Multilingual NLG Benchmarking in a Single Line of Code
abs: https://t.co/pKS5mgoDkG
GEMv2 supports 40 docum… https://t.co/qMitHzTlO0 | GEMv2: Multilingual NLG Benchmarking in a Single Line of Code | 18 |
60 | 1,539,777,865,688,010,800 | reStructured Pre-training
abs: https://t.co/mYm7qbt59N https://t.co/O5T3tSY4PL | reStructured Pre-training | 32 |
61 | 1,539,672,920,456,298,500 | Scaling Autoregressive Models for Content-Rich Text-to-Image Generation
paper: https://t.co/NKkTeHttLd
project page… https://t.co/CcKxsWPmjR | Scaling Autoregressive Models for Content-Rich Text-to-Image Generation | 137 |
62 | 1,539,480,179,151,712,300 | Intra-Instance VICReg: Bag of Self-Supervised Image Patch Embedding
abs: https://t.co/Bq3GUQywPV https://t.co/iLTaoXm0yC | Intra-Instance VICReg: Bag of Self-Supervised Image Patch Embedding | 66 |
63 | 1,539,460,213,211,910,100 | EnvPool: A Highly Parallel Reinforcement Learning Environment Execution Engine
abs: https://t.co/F4XkHLRxPi
github:… https://t.co/JiwSuMdkZH | EnvPool: A Highly Parallel Reinforcement Learning Environment Execution Engine | 34 |
64 | 1,539,459,120,667,021,300 | EpiGRAF: Rethinking training of 3D GANs
abs: https://t.co/RcY2vQr0NH
project page: https://t.co/kuXPKA00bZ https://t.co/CVCsseAS21 | EpiGRAF: Rethinking training of 3D GANs | 145 |
65 | 1,539,453,554,578,055,200 | Unbiased Teacher v2: Semi-supervised Object Detection for Anchor-free and Anchor-based Detectors
abs:… https://t.co/noluSxtqzu | Unbiased Teacher v2: Semi-supervised Object Detection for Anchor-free and Anchor-based Detectors | 72 |
66 | 1,539,435,374,103,220,200 | Global Context Vision Transformers
abs: https://t.co/d6go0yv7fu
github: https://t.co/rUYFs09ReC
On ImageNet-1K dat… https://t.co/HJnw5wclQV | Global Context Vision Transformers | 89 |
67 | 1,539,421,251,076,247,600 | (Certified!!) Adversarial Robustness for Free!
abs: https://t.co/NTU6lioyII
show how to achieve sota certified adv… https://t.co/2VW1CDARya | (Certified!!) Adversarial Robustness for Free! | 42 |
68 | 1,539,076,449,788,997,600 | A Closer Look at Smoothness in Domain Adversarial Training
abs: https://t.co/GgKE9695vj
github:… https://t.co/33MX6TZhjt | A Closer Look at Smoothness in Domain Adversarial Training | 97 |
69 | 1,538,710,356,444,471,300 | Fast Finite Width Neural Tangent Kernel
abs: https://t.co/iY1lFoYMjA https://t.co/hWzzcCd5OZ | Fast Finite Width Neural Tangent Kernel | 23 |
70 | 1,538,706,936,211,951,600 | What do navigation agents learn about their environment?
abs: https://t.co/eXelV0REgZ
github:… https://t.co/TGSzEQ1v1c | What do navigation agents learn about their environment? | 37 |
71 | 1,538,698,653,493,338,000 | Bootstrapped Transformer for Offline Reinforcement Learning
abs: https://t.co/YiEY3uiTgL https://t.co/yle4hPgMmf | Bootstrapped Transformer for Offline Reinforcement Learning | 137 |
72 | 1,538,695,457,550,921,700 | Bridge-Tower: Building Bridges Between Encoders in Vision-Language Representation Learning
abs:… https://t.co/uLQLmf4l3M | Bridge-Tower: Building Bridges Between Encoders in Vision-Language Representation Learning | 42 |
73 | 1,538,692,524,830,769,200 | MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge
abs: https://t.co/etfGL1xnum
project pa… https://t.co/Fv1aLuEJSV | MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge | 265 |
74 | 1,538,687,423,722,541,000 | Lossy Compression with Gaussian Diffusion
abs: https://t.co/tw5YiZAN3B
implement a proof of concept and find that… https://t.co/4nvLjhIX4e | Lossy Compression with Gaussian Diffusion | 102 |
75 | 1,538,686,489,491,648,500 | NU-Wave 2: A General Neural Audio Upsampling Model for Various Sampling Rates
abs: https://t.co/4S8sBXq6Ko
a diffu… https://t.co/xd3eQ0ApQJ | NU-Wave 2: A General Neural Audio Upsampling Model for Various Sampling Rates | 87 |
76 | 1,538,006,265,363,738,600 | iBoot: Image-bootstrapped Self-Supervised Video Representation Learning
abs: https://t.co/dkZUd4QC81 https://t.co/pJFpxd7ckU | iBoot: Image-bootstrapped Self-Supervised Video Representation Learning | 73 |
77 | 1,538,000,649,933,115,400 | Neural Scene Representation for Locomotion on Structured Terrain
abs: https://t.co/68xY622f4w https://t.co/W3wTYp31f6 | Neural Scene Representation for Locomotion on Structured Terrain | 83 |
78 | 1,537,924,151,389,737,000 | Programmatic Concept Learning for Human Motion Description and Synthesis
paper: https://t.co/Qemk23gUHX
project pag… https://t.co/ImHeYQC5vj | Programmatic Concept Learning for Human Motion Description and Synthesis | 60 |
79 | 1,537,640,654,968,324,000 | Spatially-Adaptive Multilayer Selection for GAN Inversion and Editing
abs: https://t.co/9tpvhXuaRw
project page:… https://t.co/XxpZg5PGke | Spatially-Adaptive Multilayer Selection for GAN Inversion and Editing | 73 |
80 | 1,537,637,590,274,277,400 | MoDi: Unconditional Motion Synthesis from Diverse Data
abs: https://t.co/YBV9jSUemo https://t.co/o1uvG18RSk | MoDi: Unconditional Motion Synthesis from Diverse Data | 70 |
81 | 1,537,630,146,244,518,000 | OmniMAE: Single Model Masked Pretraining on Images and Videos
abs: https://t.co/j9a3imUEJ6
single pretrained model… https://t.co/OiR2pY5emm | OmniMAE: Single Model Masked Pretraining on Images and Videos | 146 |
82 | 1,537,622,879,386,456,000 | SAVi++: Towards End-to-End Object-Centric Learning from Real-World Videos
abs: https://t.co/0MkpFJiUzM
using spars… https://t.co/x1Hvgf13qE | SAVi++: Towards End-to-End Object-Centric Learning from Real-World Videos | 54 |
83 | 1,537,621,348,339,572,700 | BYOL-Explore: Exploration by Bootstrapped Prediction
abs: https://t.co/xXQtolzjlP
BYOL-Explore achieves superhuman… https://t.co/uZvAbVd1Bb | BYOL-Explore: Exploration by Bootstrapped Prediction | 79 |
84 | 1,537,618,457,365,303,300 | Know your audience: specializing grounded language models with the game of Dixit
abs: https://t.co/T8d5ir8LDQ https://t.co/zSk5oR2F9D | Know your audience: specializing grounded language models with the game of Dixit | 39 |
85 | 1,537,323,042,380,124,200 | VCT: A Video Compression Transformer
abs: https://t.co/llH1L1ooKa
presented an elegantly simple transformer-based… https://t.co/ErovCWVDg3 | VCT: A Video Compression Transformer | 68 |
86 | 1,537,314,480,056,672,300 | Contrastive Learning as Goal-Conditioned Reinforcement Learning
abs: https://t.co/6dv7PNn0qq
project page:… https://t.co/vRSdekL9If | Contrastive Learning as Goal-Conditioned Reinforcement Learning | 77 |
87 | 1,537,288,570,880,368,600 | Masked Siamese ConvNets
abs: https://t.co/YMG1O1ZZ5N https://t.co/LCVqVvFNfR | Masked Siamese ConvNets | 83 |
88 | 1,537,265,816,609,116,200 | Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone
abs: https://t.co/UgdYW9Cf1g
project page:… https://t.co/v2sTfFBq5r | Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone | 89 |
89 | 1,537,257,011,657,814,000 | Variable Bitrate Neural Fields
abs: https://t.co/Rp1t2LaQaW
project page: https://t.co/e2t8OrznxI https://t.co/6hw7OwbjZN | Variable Bitrate Neural Fields | 162 |
90 | 1,537,254,679,188,488,200 | A Unified Sequence Interface for Vision Tasks
abs: https://t.co/hXbVXdqHh1
explore a unified sequence interface fo… https://t.co/QG5UxvIgS4 | A Unified Sequence Interface for Vision Tasks | 50 |
91 | 1,537,252,952,666,087,400 | Prefix Language Models are Unified Modal Learners
abs: https://t.co/BD4b3rQnKg https://t.co/2ofScnMIKN | Prefix Language Models are Unified Modal Learners | 66 |
92 | 1,537,248,480,074,293,200 | Diffusion Models for Video Prediction and Infilling
abs: https://t.co/MwfxwKXG4z
project page:… https://t.co/rnwB8eGFAs | Diffusion Models for Video Prediction and Infilling | 103 |
93 | 1,536,879,515,883,946,000 | ReCo: Retrieve and Co-segment for Zero-shot Transfer
abs: https://t.co/YwxkCGGyG1
project page:… https://t.co/WzVhmfhWCz | ReCo: Retrieve and Co-segment for Zero-shot Transfer | 58 |
94 | 1,536,872,875,885,580,300 | Object Scene Representation Transformer
abs: https://t.co/SUfNIBGAxt
project page: https://t.co/j8ebSAeM8v
scales… https://t.co/wa4vo3RJAK | Object Scene Representation Transformer | 97 |
95 | 1,536,871,347,372,052,500 | Adversarial Audio Synthesis with Complex-valued Polynomial Networks
abs: https://t.co/ekeC0nKIhR
APOLLO results in… https://t.co/sDcl2nydkt | Adversarial Audio Synthesis with Complex-valued Polynomial Networks | 23 |
96 | 1,536,526,888,289,575,000 | Large-Scale Retrieval for Reinforcement Learning
abs: https://t.co/fjzGvI3ZXB https://t.co/eFRHt8yXoq | Large-Scale Retrieval for Reinforcement Learning | 86 |
97 | 1,536,522,198,785,183,700 | GLIPv2: Unifying Localization and Vision-Language Understanding
abs: https://t.co/3GomrHG8xq
github:… https://t.co/bD68NZk4Lp | GLIPv2: Unifying Localization and Vision-Language Understanding | 73 |
98 | 1,536,521,362,898,145,300 | Self-critiquing models for assisting human evaluators
abs: https://t.co/8Zy2xfA5Qz https://t.co/qndZMS9zXa | Self-critiquing models for assisting human evaluators | 19 |
99 | 1,536,515,535,202,136,000 | Multi-instrument Music Synthesis with Spectrogram Diffusion
abs: https://t.co/UNDV4e7A6R
use a simple two-stage pr… https://t.co/AebIraqLF2 | Multi-instrument Music Synthesis with Spectrogram Diffusion | 87 |