-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathindex.html
445 lines (412 loc) · 19.1 KB
/
index.html
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
<!DOCTYPE html>
<!-- adapted from http://bayesiandeeplearning.org/ -->
<html class="mel_workshop"><head><meta http-equiv="Content-Type" content="text/html; charset=UTF-8">
<title>Embodied Multimodal Learning Workshop | ICLR 2021</title>
<meta http-equiv="Cache-Control" content="no-cache, no-store, must-revalidate">
<meta http-equiv="Pragma" content="no-cache">
<meta http-equiv="Expires" content="0">
<meta name="description" content="Embodied Multimodal Learning Workshop | ICLR 2021">
<meta name="keywords" content="Embodied,Multimodal,Learning,Workshop,ICLR,2021">
<meta name="viewport" content="width=device-width, initial-scale=1">
<link rel="stylesheet" href="./EML_files/main.css">
<meta property="og:title" content="Embodied Multimodal Learning (EML)| ICLR 2021">
<meta property="og:type" content="website">
<meta property="og:url" content="http://eml-workshop.org">
<meta property="og:description" content="Embodied Multimodal Learning Workshop at ICLR 2021 (Virtual) — Friday, May 7th, 2021">
</head>
<body data-gr-c-s-loaded="true" class="">
<!-- Wrapper -->
<div id="wrapper">
<!-- Header -->
<header id="header" class="alt">
<h1>Embodied Multimodal Learning (EML)</h1>
<h2><b>ICLR 2021 Workshop (Virtual)</b></h2>
<h2>Friday, May 7th, 2021</h2>
<h2><a href="https://iclr.cc/virtual/2021/workshop/2134" target="_blank"><font color="red">Link to ICLR Workshop Virtual Site (Join Zoom)</font></a></h2>
</header>
<!-- Nav -->
<nav id="nav" class="">
<ul>
<li><a href="https://eml-workshop.github.io/#abstract" class="active">Abstract</a></li>
<li><a href="https://eml-workshop.github.io/#speakers" class="">Invited Speakers</a></li>
<li><a href="https://eml-workshop.github.io/#cfp" class="">Call for Papers</a></li>
<li><a href="https://eml-workshop.github.io/#schedule" class="">Schedule</a></li>
<li><a href="https://eml-workshop.github.io/#organizers" class="">Organizers</a></li>
</ul>
</nav>
<!-- Main -->
<div id="main">
<!-- Introduction -->
<section id="abstract" class="main">
<div>
<div class="content">
<header class="major">
<center>
<h2>Abstract</h2>
</center>
</header>
<h3> Despite encouraging progress in embodied learning over the past two decades, there is still a large gap between embodied agents' perception and human perception. Humans have remarkable capabilities combining all our multisensory inputs. To close the gap, embodied agents should also be enabled to see, hear, touch, and interact with their surroundings in order to select the appropriate actions. However, today's learning algorithms primarily operate on a single modality. In order for Artificial Intelligence to make progress in understanding the world around us, it needs to be able to interpret such multimodal signals jointly. The goal of this workshop is to share recent progress and discuss current challenges on embodied learning with multiple modalities.<br><br>
The EML workshop will bring together researchers in different subareas of embodied multimodal learning including computer vision, robotics, machine learning, natural language processing, and cognitive science to examine the challenges and opportunities emerging from the design of embodied agents that unify their multisensory inputs. We will review the current state and identify the research infrastructure needed to enable a stronger collaboration between researchers working on different modalities.
</h3>
</div>
<!-- <span class="image"></span> -->
</div>
</section>
<section id="speakers" class="main">
<div>
<div class="content">
<header class="major">
<center>
<h2>Invited Speakers</h2>
</center>
</header>
<div class="row uniform" align="center">
<div class="3u 12u$(small)">
<a href="https://ai.stanford.edu/~cdarpino/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/claudia.jpg" alt="">
</span>
<h2>Claudia Pérez D'Arpino<br> (Stanford) </h2>
</a>
</div>
<div class="3u 12u$(small)">
<a href="http://www.cs.cmu.edu/~abhinavg/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/abhinav.jpg" alt="">
</span>
<h2>Abhinav Gupta<br> (CMU & FAIR) </h2>
</a>
</div>
<div class="3u 12u$(small)">
<a href="https://fh295.github.io/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/felix.jpg" alt="">
</span>
<h2>Felix Hill<br> (DeepMind) </h2>
</a>
</div>
<div class="3u 12u$(small)">
<a href="http://www.csc.kth.se/~danik/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/danica.jpg" alt="">
</span>
<h2>Danica Kragic<br> (KTH) </h2>
</a>
</div>
</div>
<div class="row uniform" align="center">
<div class="3u 12u$(small)">
<a href="https://www.is.mpg.de/~kjk" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/katherine.jpg" alt="">
</span>
<h2>Katherine Kuchenbecker<br> (MPI-IS) </h2>
</a>
</div>
<div class="3u 12u$(small)">
<a href="https://people.eecs.berkeley.edu/~svlevine/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/levine.jpg" alt="">
</span>
<h2>Sergey Levine<br> (UC Berkeley & Google) </h2>
</a>
</div>
<div class="3u 12u$(small)">
<a href="https://people.eecs.berkeley.edu/~malik/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/malik.jpg" alt="">
</span>
<h2>Jitendra Malik<br> (UC Berkeley & FAIR) </h2>
</a>
</div>
<div class="3u 12u$(small)">
<a href="https://psych.indiana.edu/directory/faculty/smith-linda.html" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/smith.jpg" alt="">
</span>
<h2>Linda Smith<br> (Indiana University) </h2>
</a>
</div>
</div>
</div>
</section>
<section id="cfp" class="main">
<div>
<div class="content">
<header class="major">
<center>
<h2>Call for Papers</h2>
</center>
</header>
<h3>We invite submissions of 2-4 pages extended abstracts in topics related to (but not limited to):
<ul>
<li>audio-visual embodied learning</li>
<li>touch sensing and embodied learning</li>
<li>language and embodied learning</li>
<li>speech and embodied learning</li>
<li>self-supervised/semi-supervised learning with multiple modalities</li>
<li>multimodal reinforcement learning</li>
<li>meta-learning with multiple modalities</li>
<li>novel multimodal datasets/simulators/tasks for embodied agents</li>
<li>combining multisensory inputs for robot perception</li>
<li>bio-inspired approaches for multimodal perception</li>
</ul>
A submission should take the form of an extended abstract (2-4 pages long excluding references) in PDF format using the <a href="https://eml-workshop.github.io/ICLR2021_EML_Workshop.zip" target="_blank">ICLR style</a>. We will accept submissions of (1) papers that have not been previously published or accepted for publication in substantially similar form; (2) papers that have been published or accepted for publication in recent venues including journal, conference, workshop, and arXiv; and (3) research proposals for future work with a focus on well-defined concepts and ideas. All submissions will be reviewed with single blind policy. Accepted extended abstracts will not appear in ICLR proceedings, and hence will not affect future publication of the work. We will publish all accepted extended abstracts on the workshop webpage.
</h3>
<br>
<h2>CMT submissions website: <a href="https://cmt3.research.microsoft.com/EML2021">https://cmt3.research.microsoft.com/EML2021</a><h2>
<br>
<h2>Key Dates:
<h3>
<ul>
<li>Extended abstract submission deadline: <del>March 5th, 2021 (11:59 PM PST)</del></li>
<li>Late submission deadline: <del>March 22nd, 2021 (11:59 PM PST)</del></li>
<li>Notification to authors: <del>March 26th, 2021</del></li>
<li>Workshop date: May 7th, 2021</li>
</ul>
</h3>
</h2>
<h2>Program Committee:
<h3>
Unnat Jain (UIUC), Michelle Lee (Stanford), Paul Pu Liang (CMU), Senthil Purushwalkam (CMU), Santhosh Kumar Ramakrishnan (UT Austin), Mohit Shridhar (UW), Tianmin Shu (MIT), Shaoxiong Wang (MIT)
</h3>
</h2>
</div>
</div>
</section>
<section id="schedule" class="main">
<div>
<div class="content">
<header class="major">
<center>
<h2>Schedule</h2>
</center>
</header>
<div class="table-wrapper">
<table class="alt">
<tbody>
<col width="20%">
<col width="20%">
<col width="25%">
<tr>
<td>07:55 am - 08:00 am (PDT)</td>
<td>Introduction and Opening Remarks</td>
<td></td>
<td></td>
</tr>
<tr>
<td>08:00 am - 08:30 am (PDT)</td>
<td>Invited Talk</td>
<td>Katherine Kuchenbecker<br><font size="2">(MPI-IS)</font></td>
<td></td>
</tr>
<tr>
<td>08:30 am - 09:00 am (PDT) </td>
<td>Invited Talk</td>
<td>Danica Kragic<br><font size="2">(KTH)</font></td>
<td></td>
</tr>
<tr>
<td>09:00 am - 09:30 am (PDT) </td>
<td><b>Paper Session A</b></td>
<td>A1 - A5</td>
<td></td>
</tr>
<tr>
<td>09:30 am - 09:40 am (PDT) </td>
<td><b>Paper Session A Q&A</b></td>
<td></td>
<td></td>
</tr>
<tr>
<td>09:40 am - 10:00 am (PDT)</td>
<td>Break</td>
<td></td>
<td></td>
</tr>
<tr>
<td>10:00 am - 10:30 am (PDT)</td>
<td>Invited Talk</td>
<td>Linda Smith<br><font size="2">(Indiana University)</font></td>
<td></td>
</tr>
<tr>
<td>10:30 am - 11:00 am (PDT)</td>
<td>Invited Talk</td>
<td>Felix Hill<br><font size="2">(DeepMind)</font></td>
<td></td>
</tr>
<tr>
<td>11:00 am - 12:00 pm (PDT)</td>
<td><b>Panel Discussion</b></td>
<td>Kristen Grauman, Felix Hill, Katherine Kuchenbecker, Sergey Levine, Jitendra Malik, Linda Smith</td>
<td>Having a question for the panelists? Ask <a href="https://app.sli.do/event/0bmttghf/live/questions">here</a>!</td>
</tr>
<tr>
<td>12:00 pm - 12:30 pm (PDT)</td>
<td>Break</td>
<td></td>
<td></td>
</tr>
<tr>
<td>12:30 pm - 01:00 pm (PDT)</td>
<td>Invited Talk</td>
<td>Abhinav Gupta<br><font size="2">(CMU & FAIR)</font></td>
<td></td>
</tr>
<tr>
<td>01:00 pm - 01:30 pm (PDT)</td>
<td>Invited Talk</td>
<td>Sergey Levine<br><font size="2">(UC Berkeley & Google)</font></td>
<td></td>
</tr>
<tr>
<td>01:30 pm - 02:00 pm (PDT)</td>
<td><b>Paper Session B</b></td>
<td>B1 - B4</td>
<td></td>
</tr>
<tr>
<td>02:00 pm - 02:10 pm (PDT) </td>
<td><b>Paper Session B Q&A</b></td>
<td></td>
<td></td>
</tr>
<tr>
<td>02:10 pm - 02:30 pm (PDT) </td>
<td>Break</td>
<td></td>
<td></td>
</tr>
<tr>
<td>02:30 pm - 03:00 pm (PDT)</td>
<td>Invited Talk</td>
<td>Jitendra Malik<br><font size="2">(UC Berkeley & FAIR)</font></td>
<td></td>
</tr>
<tr>
<td>03:00 pm - 03:30 pm (PDT)</td>
<td>Invited Talk</td>
<td>Claudia Pérez D'Arpino<br><font size="2">(Stanford University)</font></td>
<td></td>
</tr>
<tr>
<td>03:30 pm - 03:35 pm (PDT)</td>
<td>Closing Remarks</td>
<td></td>
<td></td>
</tr>
</tbody>
</table>
</div>
<center>
<h2>Accepted Papers</h2>
</center>
<table>
<tbody>
<col width="40%">
<col width="60%">
<tr><td><b>
Title
</b></td><td><b>
Authors
</b></td><td><b>
Paper Session
</b></td></tr>
<tr><td><b><a href="./Papers/A1.pdf">ABC Problem: An Investigation of Offline RL for Vision-Based Dynamic Manipulation</a></b></td><td>Kamyar Ghassemipour, Igor Mordatch, Shixiang Shane Gu</td><td><b>A1</b></td></tr>
<tr><td><b><a href="./Papers/A2.pdf">Language Acquisition is Embodied, Interactive, Emotive: a Research Proposal</a></b></td><td>Casey Kennington</td><td><b>A2</b></td></tr>
<tr><td><b><a href="./Papers/A3.pdf">Ask & Explore: Grounded Question Answering for Curiosity-Driven Exploration</a></b></td><td>Jivat Neet Kaur, Yiding Jiang, Paul Pu Liang</td><td><b>A3</b></td></tr>
<tr><td><b><a href="./Papers/A4.pdf">Towards Teaching Machines with Language: Interactive Learning From Only Language Descriptions of Activities</a></b></td><td>Khanh Nguyen, Dipendra Misra, Robert Schapire, Miroslav Dudik, Patrick Shafto</td><td><b>A4</b></td></tr>
<tr><td><b><a href="./Papers/A5.pdf">YouRefIt: Embodied Reference Understanding with Language and Gesture</b></td><td>Yixin Chen, Qing Li, Deqian Kong, Yik Lun Kei, Tao Gao, Yixin Zhu, Song-Chun Zhu, Siyuan Huang</td><td></a><b>A5</b></td></tr>
<tr><td><b><a href="./Papers/B1.pdf">Learning to Set Waypoints for Audio-Visual Navigation</b></td><td>Changan Chen, Sagnik Majumder, Ziad Al-Halah, Ruohan Gao, Santhosh K. Ramakrishnan, Kristen Grauman</td><td></a><b>B1</b></td></tr>
<tr><td><b><a href="./Papers/B2.pdf">Semantic Audio-Visual Navigation</b></td><td>Changan Chen, Ziad Al-Halah, Kristen Grauman</a></td><td><b>B2</b></td></tr>
<tr><td><b>Attentive Feature Reuse for Multi Task Meta learning</b></td><td>Kiran Lekkala, Laurent Itti</a></td><td><b>B3</b></td></tr>
<tr><td><b><a href="./Papers/B4.pdf">SeLaVi: self-labelling videos without any annotations from scratch</b></td><td>Yuki Asano, Mandela Patric, Christian Rupprecht, Andrea Vedaldi</td><td><b>B4</b></td></tr>
</tbody>
</table>
</div>
</div>
</section>
<section id="organizers" class="main special">
<div>
<div class="content">
<header class="major">
<h2>Organizers</h2>
</header>
<div class="row uniform">
<div class="2u 12u$(small)">
<a href="https://ai.stanford.edu/~rhgao/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/ruohan.jpg" alt="">
</span>
<h3>Ruohan Gao <br> (Stanford)</h2>
</a>
</div>
<div class="2u 12u$(small)">
<a href="http://andrewowens.com/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/andrew.jpg" alt="">
</span>
<h3>Andrew Owens <br> (UMich)</h2>
</a>
</div>
<div class="2u 12u$(small)">
<a href="https://www.seas.upenn.edu/~dineshj/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/dinesh.jpg" alt="">
</span>
<h3>Dinesh Jayaraman<br> (UPenn)</h2>
</a>
</div>
<div class="2u 12u$(small)">
<a href="https://www.cs.utexas.edu/~yukez/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/yuke.jpg" alt="">
</span>
<h3>Yuke Zhu<br> (UT Austin & Nvidia)</h2>
</a>
</div>
<div class="2u 12u$(small)">
<a href="https://jiajunwu.com/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/jiajun.jpg" alt="">
</span>
<h3>Jiajun Wu <br> (Stanford)</h2>
</a>
</div>
<div class="2u 12u$(small)">
<a href="http://www.cs.utexas.edu/users/grauman/" target="_blank" class="image">
<span class="image fit">
<img src="./EML_files/grauman.jpg" alt="">
</span>
<h3>Kristen Grauman <br> (UT Austin & FAIR)</h2>
</a>
</div>
</div>
<br>
</div>
</div>
</section>
</div>
<!-- Footer -->
<footer id="footer">
<p class="copyright">Website design adapted from <a href="http://bayesiandeeplearning.org/">Yarin Gal </a> and based on <a href="https://html5up.net/">HTML5 UP</a>.</p>
</footer>
</div>
<!-- Scripts -->
<script async="" src="./MEL_files/analytics.js"></script><script src="./EML_files/jquery.min.js"></script>
<script src="./EML_files/jquery.scrollex.min.js"></script>
<script src="./EML_files/jquery.scrolly.min.js"></script>
<script src="./EML_files/skel.min.js"></script>
<script src="./EML_files/util.js"></script>
<!--[if lte IE 8]><script src="assets/js/ie/respond.min.js"></script><![endif]-->
<script src="./EML_files/main.js"></script>
<!-- Global site tag (gtag.js) - Google Analytics -->
<script async src="https://www.googletagmanager.com/gtag/js?id=G-JC1QVS8GW2"></script>
<script>
window.dataLayer = window.dataLayer || [];
function gtag(){dataLayer.push(arguments);}
gtag('js', new Date());
gtag('config', 'G-JC1QVS8GW2');
</script>
</body></html>