|
1 | 1 | [
|
| 2 | + { |
| 3 | + "title": "Training Language Model Agents without Modifying Language Models", |
| 4 | + "authors": [ |
| 5 | + "Shaokun Zhang*", |
| 6 | + "Jieyu Zhang*", |
| 7 | + "Jiale Liu", |
| 8 | + "Linxin Song", |
| 9 | + "Chi Wang", |
| 10 | + "Ranjay Krishna", |
| 11 | + "Qingyun Wu" |
| 12 | + ], |
| 13 | + "year": "2024", |
| 14 | + "venue": "ICML 2024", |
| 15 | + "links": { |
| 16 | + "pdf": "https://arxiv.org/abs/2402.11359" |
| 17 | + }, |
| 18 | + }, |
| 19 | + { |
| 20 | + "title": "Iterated Learning Improves Compositionality in Large Vision-Language Models", |
| 21 | + "authors": [ |
| 22 | + "Chenhao Zheng", |
| 23 | + "Jieyu Zhang", |
| 24 | + "Aniruddha Kembhavi", |
| 25 | + "Ranjay Krishna" |
| 26 | + ], |
| 27 | + "year": "2024", |
| 28 | + "venue": "CVPR 2024", |
| 29 | + "links": { |
| 30 | + "pdf": "https://arxiv.org/abs/2404.02145" |
| 31 | + }, |
| 32 | + }, |
| 33 | + { |
| 34 | + "title": "SHARCS: Efficient Transformers through Routing with Dynamic Width Sub-networks", |
| 35 | + "authors": [ |
| 36 | + "Mohammadreza Salehi", |
| 37 | + "Sachin Mehta", |
| 38 | + "Aditya Kusupati", |
| 39 | + "Ali Farhadi", |
| 40 | + "Hannaneh Hajishirzi" |
| 41 | + ], |
| 42 | + "year": "2023", |
| 43 | + "venue": "EMNLP Findings 2023", |
| 44 | + "links": { |
| 45 | + "pdf": "https://arxiv.org/abs/2310.12126" |
| 46 | + }, |
| 47 | + }, |
| 48 | + { |
| 49 | + "title": "MatFormer: Nested Transformer for Elastic Inference", |
| 50 | + "authors": [ |
| 51 | + "Ethan Shen", |
| 52 | + "Alan Fan", |
| 53 | + "Sarah Pratt", |
| 54 | + "Jae Sung Park", |
| 55 | + "Matthew Wallingford", |
| 56 | + "Sham M. Kakade", |
| 57 | + "Ari Holtzman", |
| 58 | + "Ranjay Krishna", |
| 59 | + "Ali Farhadi", |
| 60 | + "Aditya Kusupati" |
| 61 | + ], |
| 62 | + "year": "2024", |
| 63 | + "venue": "NeurIPS 2024", |
| 64 | + "links": { |
| 65 | + "pdf": "https://arxiv.org/abs/2405.18400" |
| 66 | + }, |
| 67 | + }, |
| 68 | + { |
| 69 | + "title": "Objaverse-XL: A Universe of 10M+ 3D Objects", |
| 70 | + "authors": [ |
| 71 | + "Matt Deitke", |
| 72 | + "Ruoshi Liu", |
| 73 | + "Matthew Wallingford", |
| 74 | + "Huong Ngo", |
| 75 | + "Oscar Michel", |
| 76 | + "Aditya Kusupati", |
| 77 | + "Alan Fan", |
| 78 | + "Christian Laforte", |
| 79 | + "Vikram Voleti", |
| 80 | + "Samir Yitzhak Gadre", |
| 81 | + "Eli VanderBilt", |
| 82 | + "Aniruddha Kembhavi", |
| 83 | + "Carl Vondrick", |
| 84 | + "Georgia Gkioxari", |
| 85 | + "Kiana Ehsani", |
| 86 | + "Ludwig Schmidt", |
| 87 | + "Ali Farhadi" |
| 88 | + ], |
| 89 | + "year": "2023", |
| 90 | + "venue": "NeurIPS 2023", |
| 91 | + "links": { |
| 92 | + "pdf": "https://arxiv.org/abs/2307.05663", |
| 93 | + "code": "https://github.com/allenai/objaverse-xl", |
| 94 | + "project page": "https://objaverse.allenai.org/" |
| 95 | + }, |
| 96 | + "thumbnail": "/objaverse-xl.jpg" |
| 97 | + }, |
| 98 | + { |
| 99 | + "title": "SugarCrepe: Fixing Hackable Benchmarks for Vision-Language Compositionality", |
| 100 | + "authors": [ |
| 101 | + "Cheng-Yu Hsieh*", |
| 102 | + "Jieyu Zhang*", |
| 103 | + "Zixian Ma", |
| 104 | + "Aniruddha Kembhavi", |
| 105 | + "Ranjay Krishna" |
| 106 | + ], |
| 107 | + "year": "2023", |
| 108 | + "venue": "NeurIPS 2023", |
| 109 | + "links": { |
| 110 | + "pdf": "https://arxiv.org/abs/2306.14610", |
| 111 | + "code": "https://github.com/RAIVNLab/sugar-crepe" |
| 112 | + }, |
| 113 | + "thumbnail": "/sugar-crepe.jpg" |
| 114 | + }, |
| 115 | + { |
| 116 | + "title": "DataComp: In search of the next generation of multimodal datasets", |
| 117 | + "authors": [ |
| 118 | + "Samir Yitzhak Gadre", |
| 119 | + "Gabriel Ilharco", |
| 120 | + "Alex Fang", |
| 121 | + "Jonathan Hayase", |
| 122 | + "Georgios Smyrnis", |
| 123 | + "Thao Nguyen", |
| 124 | + "Ryan Marten", |
| 125 | + "Mitchell Wortsman", |
| 126 | + "Dhruba Ghosh", |
| 127 | + "Jieyu Zhang", |
| 128 | + "Eyal Orgad", |
| 129 | + "Rahim Entezari", |
| 130 | + "Giannis Daras", |
| 131 | + "Sarah Pratt", |
| 132 | + "Vivek Ramanujan", |
| 133 | + "Yonatan Bitton", |
| 134 | + "Kalyani Marathe", |
| 135 | + "Stephen Mussmann", |
| 136 | + "Richard Vencu", |
| 137 | + "Mehdi Cherti", |
| 138 | + "Ranjay Krishna", |
| 139 | + "Pang Wei Koh", |
| 140 | + "Olga Saukh", |
| 141 | + "Alexander Ratner", |
| 142 | + "Shuran Song", |
| 143 | + "Hannaneh Hajishirzi", |
| 144 | + "Ali Farhadi", |
| 145 | + "Romain Beaumont", |
| 146 | + "Sewoong Oh", |
| 147 | + "Alex Dimakis", |
| 148 | + "Jenia Jitsev", |
| 149 | + "Yair Carmon", |
| 150 | + "Vaishaal Shankar", |
| 151 | + "Ludwig Schmidt." |
| 152 | + ], |
| 153 | + "year": "2023", |
| 154 | + "venue": "NeurIPS 2023", |
| 155 | + "links": { |
| 156 | + "pdf": "https://arxiv.org/abs/2304.14108", |
| 157 | + "project page": "https://datacomp.ai/" |
| 158 | + }, |
| 159 | + "thumbnail": "/datacomp.jpg" |
| 160 | + }, |
| 161 | + { |
| 162 | + "title": "On the Connection between Pre-training Data Diversity and Fine-tuning Robustness", |
| 163 | + "authors": [ |
| 164 | + "Vivek Ramanujan", |
| 165 | + "Thao Nguyen", |
| 166 | + "Sewoong Oh", |
| 167 | + "Ludwig Schmidt", |
| 168 | + "Ali Farhadi" |
| 169 | + ], |
| 170 | + "year": "2023", |
| 171 | + "venue": "NeurIPS 2023", |
| 172 | + "links": { |
| 173 | + "pdf": "https://arxiv.org/abs/2307.12532" |
| 174 | + }, |
| 175 | + }, |
| 176 | + { |
| 177 | + "title": "OBJECT 3DIT: Language-guided 3D-aware Image Editing", |
| 178 | + "thumbnail": "/rotate.gif", |
| 179 | + "authors": [ |
| 180 | + "Oscar Michel", |
| 181 | + "Anand Bhattad", |
| 182 | + "Eli VanderBilt", |
| 183 | + "Ranjay Krishna", |
| 184 | + "Aniruddha Kembhavi", |
| 185 | + "Tanmay Gupta" |
| 186 | + ], |
| 187 | + "year": "2023", |
| 188 | + "venue": "NeurIPS 2023", |
| 189 | + "links": { |
| 190 | + "pdf": "https://arxiv.org/abs/2307.11073", |
| 191 | + "project page": "https://prior.allenai.org/projects/object-edit" |
| 192 | + }, |
| 193 | + "imageWidth": "calc(7/12*100%)" |
| 194 | + }, |
| 195 | + { |
| 196 | + "title": "MADLAD-400: Monolingual And Document-level Large Audited Dataset", |
| 197 | + "authors": [ |
| 198 | + "Sneha Kudugunta", |
| 199 | + "Isaac Rayburn Caswell", |
| 200 | + "Biao Zhang", |
| 201 | + "Xavier Garcia", |
| 202 | + "Derrick Xin", |
| 203 | + "Aditya Kusupati", |
| 204 | + "Romi Stella", |
| 205 | + "Ankur Bapna", |
| 206 | + "Orhan Firat" |
| 207 | + ], |
| 208 | + "year": "2023", |
| 209 | + "venue": "NeurIPS 2023", |
| 210 | + "links": { |
| 211 | + "pdf": "https://arxiv.org/abs/2309.04662" |
| 212 | + } |
| 213 | + }, |
| 214 | + { |
| 215 | + "title": "Neural Priming for Sample-Efficient Adaptation", |
| 216 | + "authors": [ |
| 217 | + "Matthew Wallingford", |
| 218 | + "Vivek Ramanujan", |
| 219 | + "Alex Fang", |
| 220 | + "Aditya Kusupati", |
| 221 | + "Roozbeh Mottaghi", |
| 222 | + "Aniruddha Kembhavi", |
| 223 | + "Ludwig Schmidt", |
| 224 | + "Ali Farhadi" |
| 225 | + ], |
| 226 | + "year": "2023", |
| 227 | + "venue": "NeurIPS 2023", |
| 228 | + "links": { |
| 229 | + "pdf": "https://arxiv.org/abs/2306.10191", |
| 230 | + "code": "https://github.com/RAIVNLab/neural-priming" |
| 231 | + } |
| 232 | + }, |
| 233 | + { |
| 234 | + "title": "Quilt-1M: One Million Image-Text Pairs for Histopathology", |
| 235 | + "thumbnail": "/quilt-1m.jpg", |
| 236 | + "authors": [ |
| 237 | + "Wisdom Oluchi Ikezogwo", |
| 238 | + "Mehmet Saygin Seyfioglu", |
| 239 | + "Fatemeh Ghezloo", |
| 240 | + "Dylan Stefan Chan Geva", |
| 241 | + "Fatwir Sheikh Mohammed", |
| 242 | + "Pavan Kumar Anand", |
| 243 | + "Ranjay Krishna", |
| 244 | + "Linda Shapiro" |
| 245 | + ], |
| 246 | + "year": "2023", |
| 247 | + "venue": "NeurIPS 2023", |
| 248 | + "links": { |
| 249 | + "pdf": "http://arxiv.org/abs/2306.11207", |
| 250 | + "code": "https://github.com/wisdomikezogwo/quilt1m" |
| 251 | + } |
| 252 | + }, |
2 | 253 | {
|
3 | 254 | "title": "Cola: How to adapt vision-language models to Compose Objects Localized with Attributes?",
|
4 | 255 | "authors": [
|
|
2723 | 2974 | }
|
2724 | 2975 | }
|
2725 | 2976 | ]
|
2726 |
| - |
2727 |
| -] |
|
0 commit comments