-
-
Notifications
You must be signed in to change notification settings - Fork 0
/
uncbiag-bibliography.bib
4853 lines (4646 loc) · 406 KB
/
uncbiag-bibliography.bib
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
@article{DBLP:journals/mia/GerberNEPDSAE23,
author = {Samuel Gerber and
Marc Niethammer and
Ebrahim Ebrahim and
Joseph Piven and
Stephen R. Dager and
Martin Styner and
Stephen R. Aylward and
Andinet Enquobahrie},
title = {Optimal transport features for morphometric population analysis},
journal = {Medical Image Anal.},
volume = {84},
pages = {102696},
year = {2023},
url = {https://doi.org/10.1016/j.media.2022.102696},
doi = {10.1016/j.media.2022.102696},
timestamp = {Sun, 12 Feb 2023 18:49:22 +0100},
biburl = {https://dblp.org/rec/journals/mia/GerberNEPDSAE23.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Brain pathologies often manifest as partial or complete loss of tissue. The goal of many neuroimaging studies is to capture the location and amount of tissue changes with respect to a clinical variable of interest, such as disease progression. Morphometric analysis approaches capture local differences in the distribution of tissue or other quantities of interest in relation to a clinical variable. We propose to augment morphometric analysis with an additional feature extraction step based on unbalanced optimal transport. The optimal transport feature extraction step increases statistical power for pathologies that cause spatially dispersed tissue loss, minimizes sensitivity to shifts due to spatial misalignment or differences in brain topology, and separates changes due to volume differences from changes due to tissue location. We demonstrate the proposed optimal transport feature extraction step in the context of a volumetric morphometric analysis of the OASIS-1 study for Alzheimer’s disease. The results demonstrate that the proposed approach can identify tissue changes and differences that are not otherwise measurable.},
keywords = {brain,MEDIA}
}
@article{DBLP:journals/corr/abs-2303-06493,
author = {Qin Liu and
Meng Zheng and
Benjamin Planche and
Zhongpai Gao and
Terrence Chen and
Marc Niethammer and
Ziyan Wu},
title = {Exploring Cycle Consistency Learning in Interactive Volume Segmentation},
journal = {CoRR},
volume = {abs/2303.06493},
year = {2023},
url = {https://doi.org/10.48550/arXiv.2303.06493},
doi = {10.48550/arXiv.2303.06493},
eprinttype = {arXiv},
eprint = {2303.06493},
timestamp = {Thu, 16 Mar 2023 16:04:57 +0100},
biburl = {https://dblp.org/rec/journals/corr/abs-2303-06493.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Interactive volume segmentation can be approached via two decoupled modules: interaction-to-segmentation and segmentation propagation. Given a medical volume, a user first segments a slice (or several slices) via the interaction module and then propagates the segmentation(s) to the remaining slices. The user may repeat this process multiple times until a sufficiently high volume segmentation quality is achieved. However, due to the lack of human correction during propagation, segmentation errors are prone to accumulate in the intermediate slices and may lead to sub-optimal performance. To alleviate this issue, we propose a simple yet effective cycle consistency loss that regularizes an intermediate segmentation by referencing the accurate segmentation in the starting slice. To this end, we introduce a backward segmentation path that propagates the intermediate segmentation back to the starting slice using the same propagation network. With cycle consistency training, the propagation network is better regularized than in standard forward-only training approaches. Evaluation results on challenging benchmarks such as AbdomenCT-1k and OAI-ZIB demonstrate the effectiveness of our method. To the best of our knowledge, we are the first to explore cycle consistency learning in interactive volume segmentation.},
keywords = {deep learning,segmentation,interactive}
}
@article{DBLP:journals/corr/abs-2303-09234,
author = {Yining Jiao and
Carlton J. Zdanski and
Julia S. Kimbell and
Andrew Prince and
Cameron Worden and
Samuel Kirse and
Christopher Rutter and
Benjamin Shields and
William Dunn and
Jisan Mahmud and
Marc Niethammer},
title = {{NAISR:} {A} 3D Neural Additive Model for Interpretable Shape Representation},
journal = {CoRR},
volume = {abs/2303.09234},
year = {2023},
url = {https://doi.org/10.48550/arXiv.2303.09234},
doi = {10.48550/arXiv.2303.09234},
eprinttype = {arXiv},
eprint = {2303.09234},
timestamp = {Mon, 20 Mar 2023 15:23:19 +0100},
biburl = {https://dblp.org/rec/journals/corr/abs-2303-09234.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Deep implicit functions (DIFs) have emerged as a powerful paradigm for many computer vision tasks such as 3D shape reconstruction, generation, registration, completion, editing, and understanding. However, given a set of 3D shapes with associated covariates there is at present no shape representation method which allows to precisely represent the shapes while capturing the individual dependencies on each covariate. Such a method would be of high utility to researchers to discover knowledge hidden in a population of shapes. We propose a 3D Neural Additive Model for Interpretable Shape Representation (NAISR) which describes individual shapes by deforming a shape atlas in accordance to the effect of disentangled covariates. Our approach captures shape population trends and allows for patient-specific predictions through shape transfer. NAISR is the first approach to combine the benefits of deep implicit shape representations with an atlas deforming according to specified covariates. Although our driving problem is the construction of an airway atlas, NAISR is a general approach for modeling, representing, and investigating shape populations. We evaluate NAISR with respect to shape reconstruction, shape disentanglement, shape evolution, and shape transfer for the pediatric upper airway. Our experiments demonstrate that NAISR achieves competitive shape reconstruction performance while retaining interpretability.},
keywords = {deep learning,shape,interpretability}
}
@inproceedings{DBLP:conf/iccv/GreerKVN21,
author = {Hastings Greer and
Roland Kwitt and
Fran{\c{c}}ois{-}Xavier Vialard and
Marc Niethammer},
title = {{ICON:} Learning Regular Maps Through Inverse Consistency},
booktitle = {2021 {IEEE/CVF} International Conference on Computer Vision, {ICCV}
2021, Montreal, QC, Canada, October 10-17, 2021},
pages = {3376--3385},
publisher = {{IEEE}},
year = {2021},
url = {https://drive.google.com/file/d/1kaqRx9DXqB4ksDZ1ztuCKETVN9s8F8lg},
doi = {10.1109/ICCV48922.2021.00338},
timestamp = {Fri, 11 Mar 2022 10:01:59 +0100},
biburl = {https://dblp.org/rec/conf/iccv/GreerKVN21.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Learning maps between data samples is fundamental. Applications range from representation learning, image translation and generative modeling, to the estimation of spatial deformations. Such maps relate feature vectors, or map between feature spaces. Well-behaved maps should be regular, which can be imposed explicitly or may emanate from the data itself. We explore what induces regularity for spatial transformations, e.g., when computing image registrations. Classical optimization-based models compute maps between pairs of samples and rely on an appropriate regularizer for well-posedness. Recent deep learning approaches have attempted to avoid using such regularizers altogether by relying on the sample population instead. We explore if it is possible to obtain spatial regularity using an inverse consistency loss only and elucidate what explains map regularity in such a context. We find that deep networks combined with an inverse consistency loss and randomized off-grid interpolation yield well behaved, approximately diffeomorphic, spatial transformations. Despite the simplicity of this approach, our experiments present compelling evidence, on both synthetic and real data, that regular maps can be obtained without carefully tuned explicit regularizers, while achieving competitive registration performance.},
keywords = {deep learning,knee,registration,ICCV}
}
@inproceedings{DBLP:conf/nips/ShenFLCEEN21,
author = {Zhengyang Shen and
Jean Feydy and
Peirong Liu and
Ariel Hern{\'{a}}n Curiale and
Rub{\'{e}}n San Jos{\'{e}} Est{\'{e}}par and
Ra{\'{u}}l San Jos{\'{e}} Est{\'{e}}par and
Marc Niethammer},
editor = {Marc'Aurelio Ranzato and
Alina Beygelzimer and
Yann N. Dauphin and
Percy Liang and
Jennifer Wortman Vaughan},
title = {Accurate Point Cloud Registration with Robust Optimal Transport},
booktitle = {Advances in Neural Information Processing Systems 34: Annual Conference
on Neural Information Processing Systems 2021, NeurIPS 2021, December
6-14, 2021, virtual},
pages = {5373--5389},
year = {2021},
url = {https://drive.google.com/file/d/1gquq0Ha8bDUeXMl03ujYSKbIJzop5KVS},
timestamp = {Tue, 03 May 2022 16:20:47 +0200},
biburl = {https://dblp.org/rec/conf/nips/ShenFLCEEN21.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {This work investigates the use of robust optimal transport (OT) for shape matching. Specifically, we show that recent OT solvers improve both optimization-based and deep learning methods for point cloud registration, boosting accuracy at an affordable computational cost. This manuscript starts with a practical overview of modern OT theory. We then provide solutions to the main difficulties in using this framework for shape matching. Finally, we showcase the performance of transport-enhanced registration models on a wide range of challenging tasks: rigid registration for partial shapes; scene flow estimation on the Kitti dataset; and nonparametric registration of lung vascular trees between inspiration and expiration. Our OT-based methods achieve state-of-the-art results on Kitti and for the challenging lung registration task, both in terms of accuracy and scalability. We also release PVT1010, a new public dataset of 1,010 pairs of lung vascular trees with densely sampled points. This dataset provides a challenging use case for point cloud registration algorithms with highly complex shapes and deformations. Our work demonstrates that robust OT enables fast pre-alignment and fine-tuning for a wide range of registration models, thereby providing a new key method for the computer vision toolbox. Our code and dataset are available online at: https://github.com/uncbiag/robot.},
keywords = {registration,point cloud,lung,deep learning,NeurIPS}
}
@article{DBLP:journals/corr/abs-2303-10249,
author = {Boqi Chen and
Marc Niethammer},
title = {{MRIS:} {A} Multi-modal Retrieval Approach for Image Synthesis on
Diverse Modalities},
journal = {CoRR},
volume = {abs/2303.10249},
year = {2023},
url = {https://doi.org/10.48550/arXiv.2303.10249},
doi = {10.48550/arXiv.2303.10249},
eprinttype = {arXiv},
eprint = {2303.10249},
timestamp = {Wed, 22 Mar 2023 14:41:36 +0100},
biburl = {https://dblp.org/rec/journals/corr/abs-2303-10249.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Multiple imaging modalities are often used for disease diagnosis, prediction, or population-based analyses. However, not all modalities might be available due to cost, different study designs, or changes in imaging technology. If the differences between the types of imaging are small, data harmonization approaches can be used; for larger changes, direct image synthesis approaches have been explored. In this paper, we develop an approach based on multi-modal metric learning to synthesize images of diverse modalities. We use metric learning via multi-modal image retrieval, resulting in embeddings that can relate images of different modalities. Given a large image database, the learned image embeddings allow us to use k-nearest neighbor (k-NN) regression for image synthesis. Our driving medical problem is knee osteoarthritis (KOA), but our developed method is general after proper image alignment. We test our approach by synthesizing cartilage thickness maps obtained from 3D magnetic resonance (MR) images using 2D radiographs. Our experiments show that the proposed method outperforms direct image synthesis and that the synthesized thickness maps retain information relevant to downstream tasks such as progression prediction and Kellgren-Lawrence grading (KLG). Our results suggest that retrieval approaches can be used to obtain high-quality and meaningful image synthesis results given large image databases.},
keywords = {deep learning,embedding,retrieval,knee,MICCAI}
}
@article{DBLP:journals/corr/abs-2305-00067,
author = {Nurislam Tursynbek and
Marc Niethammer},
title = {Unsupervised Discovery of 3D Hierarchical Structure with Generative
Diffusion Features},
journal = {CoRR},
volume = {abs/2305.00067},
year = {2023},
url = {https://doi.org/10.48550/arXiv.2305.00067},
doi = {10.48550/arXiv.2305.00067},
eprinttype = {arXiv},
eprint = {2305.00067},
timestamp = {Thu, 04 May 2023 16:57:18 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-2305-00067.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Inspired by recent findings that generative diffusion models learn semantically meaningful representations, we use them to discover the intrinsic hierarchical structure in biomedical 3D images using unsupervised segmentation. We show that features of diffusion models from different stages of a U-Net-based ladder-like architecture capture different hierarchy levels in 3D biomedical images. We design three losses to train a predictive unsupervised segmentation network that encourages the decomposition of 3D volumes into meaningful nested subvolumes that represent a hierarchy. First, we pretrain 3D diffusion models and use the consistency of their features across subvolumes. Second, we use the visual consistency between subvolumes. Third, we use the invariance to photometric augmentations as a regularizer. Our models achieve better performance than prior unsupervised structure discovery approaches on challenging biologically-inspired synthetic datasets and on a real-world brain tumor MRI dataset.},
keywords = {deep learning,segmentation,unsupervised,diffusion,MICCAI}
}
@article{DBLP:journals/corr/abs-2305-00087,
author = {Hastings Greer and
Lin Tian and
Fran{\c{c}}ois{-}Xavier Vialard and
Roland Kwitt and
Sylvain Bouix and
Ra{\'{u}}l San Jos{\'{e}} Est{\'{e}}par and
Richard Rushmore and
Marc Niethammer},
title = {Inverse Consistency by Construction for Multistep Deep Registration},
journal = {CoRR},
volume = {abs/2305.00087},
year = {2023},
url = {https://doi.org/10.48550/arXiv.2305.00087},
doi = {10.48550/arXiv.2305.00087},
eprinttype = {arXiv},
eprint = {2305.00087},
timestamp = {Thu, 04 May 2023 16:57:18 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-2305-00087.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Inverse consistency is a desirable property for image registration. We propose a simple technique to make a neural registration network inverse consistent by construction, as a consequence of its structure, as long as it parameterizes its output transform by a Lie group. We extend this technique to multi-step neural registration by composing many such networks in a way that preserves inverse consistency. This multi-step approach also allows for inverse-consistent coarse to fine registration. We evaluate our technique on synthetic 2-D data and four 3-D medical image registration tasks and obtain excellent registration accuracy while assuring inverse consistency.},
keywords = {deep learning,registration,inverse consistent}
}
@article{DBLP:journals/corr/abs-2305-03125,
author = {Yifeng Shi and
Marc Niethammer},
title = {Multimodal Understanding Through Correlation Maximization and Minimization},
journal = {CoRR},
volume = {abs/2305.03125},
year = {2023},
url = {https://doi.org/10.48550/arXiv.2305.03125},
doi = {10.48550/arXiv.2305.03125},
eprinttype = {arXiv},
eprint = {2305.03125},
timestamp = {Wed, 10 May 2023 16:08:41 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-2305-03125.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Multimodal learning has mainly focused on learning large models on, and fusing feature representations from, different modalities for better performances on downstream tasks. In this work, we take a detour from this trend and study the intrinsic nature of multimodal data by asking the following questions: 1) Can we learn more structured latent representations of general multimodal data?; and 2) can we intuitively understand, both mathematically and visually, what the latent representations capture? To answer 1), we propose a general and lightweight framework, Multimodal Understanding Through Correlation Maximization and Minimization (MUCMM), that can be incorporated into any large pre-trained network. MUCMM learns both the common and individual representations. The common representations capture what is common between the modalities; the individual representations capture the unique aspect of the modalities. To answer 2), we propose novel scores that summarize the learned common and individual structures and visualize the score gradients with respect to the input, visually discerning what the different representations capture. We further provide mathematical intuitions of the computed gradients in a linear setting, and demonstrate the effectiveness of our approach through a variety of experiments.},
keywords = {deep learning,multi modal,interpretability}
}
@article{DBLP:journals/cmbbeiv/Ben-ZikriHFSACN22,
author = {Yehuda K. Ben{-}Zikri and
Mar{\'{\i}}a Helguera and
David Fetzer and
David A. Shrier and
Stephen R. Aylward and
Deepak Chittajallu and
Marc Niethammer and
Nathan D. Cahill and
Cristian A. Linte},
title = {A feature-based affine registration method for capturing background
lung tissue deformation for ground glass nodule tracking},
journal = {Comput. methods Biomech. Biomed. Eng. Imaging Vis.},
volume = {10},
number = {5},
pages = {521--539},
year = {2022},
url = {https://doi.org/10.1080/21681163.2021.1994471},
doi = {10.1080/21681163.2021.1994471},
timestamp = {Tue, 06 Dec 2022 13:15:05 +0100},
biburl = {https://dblp.org/rec/journals/cmbbeiv/Ben-ZikriHFSACN22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Apparent changes in lung nodule size assessed via simple image-based measurements from computed tomography (CT) images may be compromised by the effect of the background lung tissue deformation on the nodule, leading to erroneous nodule tracking. We propose a feature-based affine registration method and study its performance vis-a-vis several other registration methods. We implement and test each registration method using a lung- and a lesion-centred region of interest on 10 patient CT datasets featuring 12 nodules. We evaluate each registration method according to the target registration error (TRE) computed across 30–50 homologous fiducial landmarks selected by expert radiologists. Our results show that the proposed feature-based affine lesion-centred registration yielded a 1.11.2 mm TRE, while a Symmetric Normalisation deformable registration yielded a 1.21.2 mm TRE, with a baseline least-square fit of the validation fiducial landmarks of 1.51.2 mm TRE. The proposed feature-based affine registration is computationally efficient, eliminates the need for nodule segmentation, and reduces the susceptibility of artificial deformations. We also conducted a pilot pre-clinical study that showed the proposed featurebased lesion-centred affine registration effectively compensates for the background lung tissue deformation and serves as a reliable baseline registration method prior to assessing lung nodule changes due to disease.},
keywords = {registration,lung}
}
@article{DBLP:journals/mia/HuangXSLLNNGNZ22,
author = {Chao Huang and
Zhenlin Xu and
Zhengyang Shen and
Tianyou Luo and
Tengfei Li and
Daniel Nissman and
Amanda Nelson and
Yvonne Golightly and
Marc Niethammer and
Hongtu Zhu},
title = {{DADP:} Dynamic abnormality detection and progression for longitudinal
knee magnetic resonance images from the Osteoarthritis Initiative},
journal = {Medical Image Anal.},
volume = {77},
pages = {102343},
year = {2022},
url = {https://doi.org/10.1016/j.media.2021.102343},
doi = {10.1016/j.media.2021.102343},
timestamp = {Wed, 07 Dec 2022 13:34:33 +0100},
biburl = {https://dblp.org/rec/journals/mia/HuangXSLLNNGNZ22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Osteoarthritis (OA) is the most common disabling joint disease. Magnetic resonance (MR) imaging has been commonly used to assess knee joint degeneration due to its distinct advantage in detecting morphologic cartilage changes. Although several statistical methods over conventional radiography have been developed to perform quantitative cartilage analyses, little work has been done capturing the development and progression of cartilage lesions (or abnormal regions) and how they naturally progress. There are two major challenges, including (i) the lack of building spatial-temporal correspondences and correlations in cartilage thickness and (ii) the spatio-temporal heterogeneity in abnormal regions. The goal of this work is to propose a dynamic abnormality detection and progression (DADP) framework for quantitative cartilage analysis, while addressing the two challenges. First, spatial correspondences are established on flattened 2D cartilage thickness maps extracted from 3D knee MR images both across time within each subject and across all subjects. Second, a dynamic functional mixed effects model (DFMEM) is proposed to quantify abnormality progression across time points and subjects, while accounting for the spatio-temporal heterogeneity. We systematically evaluate our DADP using simulations and real data from the Osteoarthritis Initiative (OAI). Our results show that DADP not only effectively detects subject-specific dynamic abnormal regions, but also provides population-level statistical disease mapping and subgroup analysis.},
keywords = {knee,MRI,cartilage,MEDIA}
}
@article{DBLP:journals/sadm/WeiN22,
author = {Susan Wei and
Marc Niethammer},
title = {The fairness-accuracy Pareto front},
journal = {Stat. Anal. Data Min.},
volume = {15},
number = {3},
pages = {287--302},
year = {2022},
url = {https://drive.google.com/file/d/1vni9fw5azk7pFof9CSJbBinSqQwHGxlq},
doi = {10.1002/sam.11560},
timestamp = {Thu, 02 Jun 2022 16:43:02 +0200},
biburl = {https://dblp.org/rec/journals/sadm/WeiN22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Algorithmic fairness seeks to identify and correct sources of bias in machine learning algorithms. Confoundingly, ensuring fairness often comes at the cost of accuracy. We provide formal tools in this work for reconciling this fundamental tension in algorithm fairness. Specifically, we put to use the concept of Pareto optimality from multiobjective optimization and seek the fairness‐accuracy Pareto front of a neural network classifier. We demonstrate that many existing algorithmic fairness methods are performing the so‐called linear scalarization scheme, which has severe limitations in recovering Pareto optimal solutions. We instead apply the Chebyshev scalarization scheme which is provably superior theoretically and no more computationally burdensome at recovering Pareto optimal solutions compared to the linear scheme.},
keywords = {fairness,Pareto front}
}
@inproceedings{DBLP:conf/cvpr/DingN22,
author = {Zhipeng Ding and
Marc Niethammer},
title = {Aladdin: Joint Atlas Building and Diffeomorphic Registration Learning
with Pairwise Alignment},
booktitle = {{IEEE/CVF} Conference on Computer Vision and Pattern Recognition,
{CVPR} 2022, New Orleans, LA, USA, June 18-24, 2022},
pages = {20752--20761},
publisher = {{IEEE}},
year = {2022},
url = {https://drive.google.com/file/d/1gdceCsBTiSute52GcgIHdf3lh9OkhjQa},
doi = {10.1109/CVPR52688.2022.02012},
timestamp = {Wed, 05 Oct 2022 16:31:19 +0200},
biburl = {https://dblp.org/rec/conf/cvpr/DingN22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Atlas building and image registration are important tasks for medical image analysis. Once one or multiple atlases from an image population have been constructed, commonly (1) images are warped into an atlas space to study intra-subject or inter-subject variations or (2) a possibly probabilistic atlas is warped into image space to assign anatomical labels. Atlas estimation and nonparametric transformations are computationally expensive as they usually require numerical optimization. Additionally, previous approaches for atlas building often define similarity measures between a fuzzy atlas and each individual image, which may cause alignment difficulties because a fuzzy atlas does not exhibit clear anatomical structures in contrast to the individual images. This work explores using a convolutional neural network (CNN) to jointly predict the atlas and a stationary velocity field (SVF) parameterization for
diffeomorphic image registration with respect to the atlas. Our approach does not require affine pre-registrations and utilizes pairwise image alignment losses to increase registration accuracy. We evaluate our model on 3D knee magnetic resonance images (MRI) from the OAI-ZIB dataset. Our results show that the proposed framework achieves better performance than other state-of-the-art image registration algorithms, allows for end-to-end training, and for fast inference at test time.},
keywords = {MRI,atlas,knee,CVPR}
}
@inproceedings{DBLP:conf/cvpr/LiuLAN22,
author = {Peirong Liu and
Yueh Z. Lee and
Stephen R. Aylward and
Marc Niethammer},
title = {Deep Decomposition for Stochastic Normal-Abnormal Transport},
booktitle = {{IEEE/CVF} Conference on Computer Vision and Pattern Recognition,
{CVPR} 2022, New Orleans, LA, USA, June 18-24, 2022},
pages = {18769--18779},
publisher = {{IEEE}},
year = {2022},
url = {https://drive.google.com/file/d/1E64NrDAHZyIXs21CC_qAEPlfTsHGASmU},
doi = {10.1109/CVPR52688.2022.01823},
timestamp = {Wed, 05 Oct 2022 16:31:19 +0200},
biburl = {https://dblp.org/rec/conf/cvpr/LiuLAN22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Advection-diffusion equations describe a large family of natural transport processes, e.g., fluid flow, heat transfer,
and wind transport. They are also used for optical flow and perfusion imaging computations. We develop a machine learning model, D2-SONATA, built upon a stochastic advection-diffusion equation, which predicts the velocity and diffusion fields that drive 2D/3D image time-series of transport. In particular, our proposed model incorporates a model of transport atypicality, which isolates abnormal differences between expected normal transport behavior and the observed transport. In a medical context such a normal-abnormal decomposition can be used, for example, to quantify pathologies. Specifically, our model identifies the advection and diffusion contributions from the transport time-series and simultaneously predicts an anomaly value field to provide a decomposition into normal and abnormal advection and diffusion behavior. To achieve improved estimation performance for the velocity and diffusion-tensor fields underlying the advection-diffusion process and for the estimation of the anomaly fields, we create a 2D/3D anomaly-encoded advection-diffusion simulator, which allows for supervised learning. We further apply our model on a brain perfusion dataset from ischemic stroke patients via transfer learning. Extensive comparisons demonstrate that our model successfully distinguishes stroke lesions (abnormal) from normal brain regions, while reconstructing the underlying velocity and diffusion tensor fields.},
keywords = {perfusion,stroke,CVPR}
}
@inproceedings{DBLP:conf/eccv/LiuZPKCNW22,
author = {Qin Liu and
Meng Zheng and
Benjamin Planche and
Srikrishna Karanam and
Terrence Chen and
Marc Niethammer and
Ziyan Wu},
editor = {Shai Avidan and
Gabriel J. Brostow and
Moustapha Ciss{\'{e}} and
Giovanni Maria Farinella and
Tal Hassner},
title = {PseudoClick: Interactive Image Segmentation with Click Imitation},
booktitle = {Computer Vision - {ECCV} 2022 - 17th European Conference, Tel Aviv,
Israel, October 23-27, 2022, Proceedings, Part {VI}},
series = {Lecture Notes in Computer Science},
volume = {13666},
pages = {728--745},
publisher = {Springer},
year = {2022},
url = {https://arxiv.org/pdf/2207.05282.pdf},
doi = {10.1007/978-3-031-20068-7\_42},
timestamp = {Mon, 05 Dec 2022 13:35:31 +0100},
biburl = {https://dblp.org/rec/conf/eccv/LiuZPKCNW22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {The goal of click-based interactive image segmentation is to obtain precise object segmentation masks with limited user interaction, i.e., by a minimal number of user clicks. Existing methods require users to provide all the clicks: by first inspecting the segmentation mask and then providing points on mislabeled regions, iteratively. We ask the question: can our model directly predict where to click, so as to further reduce the user interaction cost? To this end, we propose PseudoClick, a generic framework that enables existing segmentation networks to propose candidate next clicks. These automatically generated clicks, termed pseudo clicks in this work, serve as an imitation of human clicks to refine the segmentation mask. We build PseudoClick on existing segmentation backbones and show how the click prediction mechanism leads to improved performance. We evaluate PseudoClick on 10 public datasets from different domains and modalities, showing that our model not only outperforms existing approaches but also demonstrates strong generalization capability in cross-domain evaluation. We obtain new state-of-theart results on several popular benchmarks, e.g., on the Pascal dataset, our model significantly outperforms existing state-of-the-art by reducing 12.4\% number of clicks to achieve 85\% IoU.},
keyword = {deep learning,segmentation,interactive,ECCV}
}
@inproceedings{DBLP:conf/miccai/LiuXJN22,
author = {Qin Liu and
Zhenlin Xu and
Yining Jiao and
Marc Niethammer},
editor = {Linwei Wang and
Qi Dou and
P. Thomas Fletcher and
Stefanie Speidel and
Shuo Li},
title = {iSegFormer: Interactive Segmentation via Transformers with Application
to 3D Knee {MR} Images},
booktitle = {Medical Image Computing and Computer Assisted Intervention - {MICCAI}
2022 - 25th International Conference, Singapore, September 18-22,
2022, Proceedings, Part {V}},
series = {Lecture Notes in Computer Science},
volume = {13435},
pages = {464--474},
publisher = {Springer},
year = {2022},
url = {https://drive.google.com/file/d/1CwtA_P7m1cwTI1YLRGOEc3LM7oXiJO5m},
doi = {10.1007/978-3-031-16443-9\_45},
timestamp = {Tue, 13 Dec 2022 14:39:06 +0100},
biburl = {https://dblp.org/rec/conf/miccai/LiuXJN22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Interactive image segmentation has been widely applied to obtain high-quality voxel-level labels for medical images. The recent success of Transformers on various vision tasks has paved the road for developing Transformer-based interactive image segmentation approaches. However, these approaches remain unexplored and, in particular, have not been developed for 3D medical image segmentation. To fill this research gap, we investigate Transformer-based interactive image segmentation and its application to 3D medical images. This is a nontrivial task due to two main challenges: 1) limited memory for computationally inefficient Transformers and 2) limited labels for 3D medical images. To tackle the first challenge, we propose iSegFormer, a memory-efficient Transformer that combines a Swin Transformer with a lightweight multilayer perceptron (MLP) decoder. To address the second challenge, we pretrain iSegFormer on large amount of unlabeled datasets and then finetune it with only a limited number of segmented 2D slices. We further propagate the 2D segmentations obtained by iSegFormer to unsegmented slices in 3D images using a pre-existing segmentation propagation model pretrained on videos. We evaluate iSegFormer on the public OAI-ZIB dataset for interactive knee cartilage segmentation. Evaluation results show that iSegFormer outperforms its convolutional neural network (CNN) counterparts on interactive 2D knee cartilage segmentation, with competitive computational efficiency. When propagating the 2D interactive segmentations of 5 slices to other unprocessed slices within the same 3D volume, we achieve 82.2\% Dice score for 3D knee cartilage segmentation. Code is available at https://github.com/uncbiag/iSegFormer.},
keywords = {MICCAI,deep learning,segmentation,interactive}
}
@inproceedings{DBLP:conf/miccai/TianLEN22,
author = {Lin Tian and
Yueh Z. Lee and
Ra{\'{u}}l San Jos{\'{e}} Est{\'{e}}par and
Marc Niethammer},
editor = {Linwei Wang and
Qi Dou and
P. Thomas Fletcher and
Stefanie Speidel and
Shuo Li},
title = {LiftReg: Limited Angle 2D/3D Deformable Registration},
booktitle = {Medical Image Computing and Computer Assisted Intervention - {MICCAI}
2022 - 25th International Conference, Singapore, September 18-22,
2022, Proceedings, Part {VI}},
series = {Lecture Notes in Computer Science},
volume = {13436},
pages = {207--216},
publisher = {Springer},
year = {2022},
url = {https://drive.google.com/file/d/13Dw3RO1ZhF3vtLr9TyJGhJkF7wz8DrjC},
doi = {10.1007/978-3-031-16446-0\_20},
timestamp = {Tue, 13 Dec 2022 14:39:06 +0100},
biburl = {https://dblp.org/rec/conf/miccai/TianLEN22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {We propose LiftReg, a 2D/3D deformable registration approach. LiftReg is a deep registration framework which is trained using sets of digitally reconstructed radiographs (DRR) and computed tomography (CT) image pairs. By using simulated training data, LiftReg can use a high-quality CT-CT image similarity measure, which helps the network to learn a high-quality deformation space. To further improve registration quality and to address the inherent depth ambiguities of very limited angle acquisitions, we propose to use features extracted from the backprojected 2D images and a statistical deformation model. We test our approach on the DirLab lung registration dataset and show that it outperforms an existing learning-based pairwise registration approach.},
keywords = {tomosynthesis,CT,2D/3D,registration,MICCAI,lung}
}
@inproceedings{DBLP:conf/nips/GrafZRNK22,
author = {Florian Graf and
Sebastian Zeng and
Bastian Rieck and
Marc Niethammer and
Roland Kwitt},
title = {On Measuring Excess Capacity in Neural Networks},
booktitle = {NeurIPS},
year = {2022},
url = {https://drive.google.com/file/d/1O9LlGse4ZDWBEjvBKBDxDfTFfFkRVmRo},
timestamp = {Thu, 11 May 2023 17:08:21 +0200},
biburl = {https://dblp.org/rec/conf/nips/GrafZRNK22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {We study the excess capacity of deep networks in the context of supervised classification. That is, given a capacity measure of the underlying hypothesis class – in our case, empirical Rademacher complexity – to what extent can we (a priori) constrain this class while retaining an empirical error on a par with the unconstrained regime? To assess excess capacity in modern architectures (such as residual networks), we extend and unify prior Rademacher complexity bounds to accommodate function composition and addition, as well as the structure of convolutions. The capacitydriving terms in our bounds are the Lipschitz constants of the layers and an (2, 1) group norm distance to the initializations of the convolution weights. Experiments on benchmark datasets of varying task difficulty indicate that (1) there is a substantial amount of excess capacity per task, and (2) capacity can be kept at a surprisingly similar level across tasks. Overall, this suggests a notion of compressibility with respect to weight norms, complementary to classic compression via weight pruning. Source code is available at https://github.com/rkwitt/excess_capacity.},
keywords={deep learning,capacity,generalization,NeurIPS}
}
@inproceedings{DBLP:conf/nips/XuNR22,
author = {Zhenlin Xu and
Marc Niethammer and
Colin Raffel},
title = {Compositional Generalization in Unsupervised Compositional Representation
Learning: {A} Study on Disentanglement and Emergent Language},
booktitle = {NeurIPS},
year = {2022},
url = {https://arxiv.org/abs/2210.00482},
timestamp = {Thu, 11 May 2023 17:08:21 +0200},
biburl = {https://dblp.org/rec/conf/nips/XuNR22.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Deep learning models struggle with compositional generalization, i.e. the ability to recognize or generate novel combinations of observed elementary concepts. In hopes of enabling compositional generalization, various unsupervised learning algorithms have been proposed with inductive biases that aim to induce compositional structure in learned representations (e.g. disentangled representation and emergent language learning). In this work, we evaluate these unsupervised learning algorithms in terms of how well they enable compositional generalization. Specifically, our evaluation protocol focuses on whether or not it is easy to train a simple model on top of the learned representation that generalizes to new combinations of compositional factors. We systematically study three unsupervised representation learning algorithms - β-VAE, β-TCVAE, and emergent language (EL) autoencoders - on two datasets that allow directly testing compositional generalization. We find that directly using the bottleneck representation with simple models and few labels may lead to worse generalization than using representations from layers before or after the learned representation itself. In addition, we find that the previously proposed metrics for evaluating the levels of compositionality are not correlated with actual compositional generalization in our framework. Surprisingly, we find that increasing pressure to produce a disentangled representation produces representations with worse generalization, while representations from EL models show strong compositional generalization. Taken together, our results shed new light on the compositional generalization behavior of different unsupervised learning algorithms with a new setting to rigorously test this behavior, and suggest the potential benefits of delevoping EL learning algorithms for more generalizable representations.},
keywords = {deep learning,compositionality,NeurIPS}
}
@article{DBLP:journals/corr/abs-2206-05897,
author = {Lin Tian and
Hastings Greer and
Fran{\c{c}}ois{-}Xavier Vialard and
Roland Kwitt and
Ra{\'{u}}l San Jos{\'{e}} Est{\'{e}}par and
Marc Niethammer},
title = {GradICON: Approximate Diffeomorphisms via Gradient Inverse Consistency},
journal = {CVPR},
volume = {abs/2206.05897},
year = {2023},
url = {https://drive.google.com/file/d/1j8u5n50knQUxhnHp1OMGEwsl8CX-lODX},
doi = {10.48550/arXiv.2206.05897},
eprinttype = {arXiv},
eprint = {2206.05897},
timestamp = {Tue, 21 Mar 2023 21:04:57 +0100},
biburl = {https://dblp.org/rec/journals/corr/abs-2206-05897.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {We present an approach to learning regular spatial transformations between image pairs in the context of medical image registration. Contrary to optimization-based registration techniques and many modern learning-based methods, we do not directly penalize transformation irregularities but instead promote transformation regularity via an inverse consistency penalty. We use a neural network to predict a map between a source and a target image as well as the map when swapping the source and target images. Different from existing approaches, we compose these two resulting maps and regularize deviations of the Jacobian of this composition from the identity matrix. This regularizer - GradICON - results in much better convergence when training registration models compared to promoting inverse consistency of the composition of maps directly while retaining the desirable implicit regularization effects of the latter. We achieve state-of-the-art registration performance on a variety of real-world medical image datasets using a single set of hyperparameters and a single non-dataset-specific training protocol. Code is available at https://github.com/uncbiag/ICON.},
keywords = {CVPR,registration,gradient inverse consistency,lung,knee,brain}
}
@inproceedings{DBLP:conf/iccv/DingHLN21,
author = {Zhipeng Ding and
Xu Han and
Peirong Liu and
Marc Niethammer},
title = {Local Temperature Scaling for Probability Calibration},
booktitle = {2021 {IEEE/CVF} International Conference on Computer Vision, {ICCV}
2021, Montreal, QC, Canada, October 10-17, 2021},
pages = {6869--6879},
publisher = {{IEEE}},
year = {2021},
url = {https://drive.google.com/file/d/1XDPwmr4iBZaY06OebVhKqwD_ZaARBL8X},
doi = {10.1109/ICCV48922.2021.00681},
timestamp = {Fri, 11 Mar 2022 10:01:59 +0100},
biburl = {https://dblp.org/rec/conf/iccv/DingHLN21.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {For semantic segmentation, label probabilities are often uncalibrated as they are typically only the by-product of a segmentation task. Intersection over Union (IoU) and Dice score are often used as criteria for segmentation success, while metrics related to label probabilities are rarely explored. On the other hand, probability calibration approaches have been studied, which aim at matching probability outputs with experimentally observed errors, but they mainly focus on classification tasks, not on semantic segmentation. Thus, we propose a learning-based calibration method that focuses on multi-label semantic segmentation. Specifically, we adopt a tree-like convolution neural network to predict local temperature values for probability calibration. One advantage of our approach is that it does not change prediction accuracy, hence allowing for calibration as a post-processing step. Experiments on the COCO and LPBA40 datasets demonstrate improved calibration performance over different metrics. We also demonstrate the performance of our method for multi-atlas brain segmentation from magnetic resonance images.},
keywords = {ICCV,probability calibration,segmentation}
}
@article{DBLP:journals/corr/abs-2210-11006,
author = {Qin Liu and
Zhenlin Xu and
Gedas Bertasius and
Marc Niethammer},
title = {SimpleClick: Interactive Image Segmentation with Simple Vision Transformers},
journal = {CoRR},
volume = {abs/2210.11006},
year = {2022},
url = {https://doi.org/10.48550/arXiv.2210.11006},
doi = {10.48550/arXiv.2210.11006},
eprinttype = {arXiv},
eprint = {2210.11006},
timestamp = {Tue, 25 Oct 2022 14:25:08 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-2210-11006.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Click-based interactive image segmentation aims at extracting objects with a limited user clicking. A hierarchical backbone is the de-facto architecture for current methods. Recently, the plain, non-hierarchical Vision Transformer (ViT) has emerged as a competitive backbone for dense prediction tasks. This design allows the original ViT to be a
foundation model that can be finetuned for downstream tasks without redesigning a hierarchical backbone for pretraining. Although this design is simple and has been proven effective, it has not yet been explored for interactive image segmentation. To fill this gap, we propose SimpleClick, the first interactive segmentation method that leverages a plain backbone. Based on the plain backbone, we introduce a symmetric patch embedding layer that encodes clicks into the backbone with minor modifications to the backbone itself. With the plain backbone pretrained as a masked autoencoder (MAE), SimpleClick achieves state-of-theart performance. Remarkably, our method achieves 4.15 NoC@90 on SBD, improving 21.8\% over the previous best result. Extensive evaluation on medical images demonstrates the generalizability of our method. We further develop an extremely tiny ViT backbone for SimpleClick and provide a detailed computational analysis, highlighting its suitability as a practical annotation tool.},
keywords = {deep learning,segmentation,interactive}
}
@inproceedings{DBLP:conf/cvpr/LiuTZALN21,
author = {Peirong Liu and
Lin Tian and
Yubo Zhang and
Stephen R. Aylward and
Yueh Z. Lee and
Marc Niethammer},
title = {Discovering Hidden Physics Behind Transport Dynamics},
booktitle = {{IEEE} Conference on Computer Vision and Pattern Recognition, {CVPR}
2021, virtual, June 19-25, 2021},
pages = {10082--10092},
publisher = {Computer Vision Foundation / {IEEE}},
year = {2021},
url = {https://drive.google.com/file/d/1Egl7mcuvgwpE_Owbb8P5h1KrhdRHVYP7},
doi = {10.1109/CVPR46437.2021.00995},
timestamp = {Tue, 29 Nov 2022 14:53:03 +0100},
biburl = {https://dblp.org/rec/conf/cvpr/LiuTZALN21.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Transport processes are ubiquitous. They are, for example, at the heart of optical flow approaches; or of perfusion imaging, where blood transport is assessed, most com-
monly by injecting a tracer. An advection-diffusion equation is widely used to describe these transport phenomena. Our goal is estimating the underlying physics of advection-diffusion equations, expressed as velocity and diffusion tensor fields. We propose a learning framework (YETI) building on an auto-encoder structure between 2D and 3D image time-series, which incorporates the advection-diffusion model. To help with identifiability, we develop an advection-diffusion simulator which allows pre-training of our model by supervised learning using the velocity and diffusion tensor fields. Instead of directly learning these velocity and diffusion tensor fields, we introduce representations that assure incompressible flow and symmetric positive semi-definite diffusion fields and demonstrate the additional benefits of these representations on improving estimation accuracy. We further use transfer learning to apply YETI on a public brain magnetic resonance (MR) perfusion dataset of stroke patients and show its ability to successfully distinguish stroke lesions from normal brain regions via the estimated velocity and diffusion tensor fields.},
keywords = {perfusion,stroke,CVPR}
}
@inproceedings{DBLP:conf/icml/GrafHNK21,
author = {Florian Graf and
Christoph D. Hofer and
Marc Niethammer and
Roland Kwitt},
editor = {Marina Meila and
Tong Zhang},
title = {Dissecting Supervised Constrastive Learning},
booktitle = {Proceedings of the 38th International Conference on Machine Learning,
{ICML} 2021, 18-24 July 2021, Virtual Event},
series = {Proceedings of Machine Learning Research},
volume = {139},
pages = {3821--3830},
publisher = {{PMLR}},
year = {2021},
url = {http://proceedings.mlr.press/v139/graf21a.html},
timestamp = {Wed, 14 Jul 2021 15:41:58 +0200},
biburl = {https://dblp.org/rec/conf/icml/GrafHNK21.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Minimizing cross-entropy over the softmax scores of a linear map composed with a high-capacity encoder is arguably the most popular choice for training neural networks on supervised learning tasks. However, recent works show that one can directly optimize the encoder instead, to obtain equally (or even more) discriminative representations via a supervised variant of a contrastive objective. In this work, we address the question whether there are fundamental differences in the sought-for representation geometry in the output space of the encoder at minimal loss. Specifically, we prove, under mild assumptions, that both losses attain their minimum once the representations of each class collapse to the vertices of a regular simplex, inscribed in a hypersphere. We provide empirical evidence that this configuration is attained in practice and that reaching a close-to-optimal state typically indicates good generalization performance. Yet, the two losses show remarkably different optimization behavior. The number of iterations required to perfectly fit to data scales superlinearly with the amount of randomly flipped labels for the supervised contrastive loss. This is in contrast to the approximately linear scaling previously reported for networks trained with cross-entropy.},
keywords = {deep learning,contrastive learning,ICML}
}
@inproceedings{DBLP:conf/isbi/DingN21,
author = {Zhipeng Ding and
Marc Niethammer},
title = {Votenet++: Registration Refinement For Multi-Atlas Segmentation},
booktitle = {18th {IEEE} International Symposium on Biomedical Imaging, {ISBI}
2021, Nice, France, April 13-16, 2021},
pages = {275--279},
publisher = {{IEEE}},
year = {2021},
url = {https://drive.google.com/file/d/1MWCl3YhQR8qTt-CPmXKkqtg_aoWq33IZ},
abstract = {Multi-atlas segmentation (MAS) is a popular image segmen- tation technique for medical images. In this work, we improve the performance of MAS by correcting registration errors be- fore label fusion. Specifically, we use a volumetric displace- ment field to refine registrations based on image anatomical appearance and predicted labels. We show the influence of the initial spatial alignment as well as the beneficial effect of using label information for MAS performance. Experiments demonstrate that the proposed refinement approach improves MAS performance on a 3D magnetic resonance dataset of the knee.},
doi = {10.1109/ISBI48211.2021.9434031},
timestamp = {Mon, 31 May 2021 11:40:24 +0200},
biburl = {https://dblp.org/rec/conf/isbi/DingN21.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
keywords = {segmentation,brain,ISBI}
}
@inproceedings{DBLP:conf/iclr/XuLYRN21,
author = {Zhenlin Xu and
Deyi Liu and
Junlin Yang and
Colin Raffel and
Marc Niethammer},
title = {Robust and Generalizable Visual Representation Learning via Random
Convolutions},
booktitle = {9th International Conference on Learning Representations, {ICLR} 2021,
Virtual Event, Austria, May 3-7, 2021},
publisher = {OpenReview.net},
year = {2021},
url = {https://openreview.net/forum?id=BVSM0x3EDK6},
timestamp = {Wed, 23 Jun 2021 17:36:39 +0200},
biburl = {https://dblp.org/rec/conf/iclr/XuLYRN21.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {While successful for various computer vision tasks, deep neural networks have shown to be vulnerable to texture style shifts and small perturbations to which humans are robust. In this work, we show that the robustness of neural networks can be greatly improved through the use of random convolutions as data augmentation. Random convolutions are approximately shape-preserving and may distort local textures. Intuitively, randomized convolutions create an infinite number of new domains with similar global shapes but random local texture. Therefore, we explore using outputs of multi-scale random convolutions as new images or mixing them with the original images during training. When applying a network trained with our approach to unseen domains, our method consistently improves the performance on domain generalization benchmarks and is scalable to ImageNet. In particular, in the challenging scenario of generalizing to the sketch domain in PACS and to ImageNet-Sketch, our method outperforms state-of-art methods by a large margin. More interestingly, our method can benefit downstream tasks by providing a more robust pretrained visual representation.},
keywords = {deep learning,representation learning,robustness,domain generalization,neural networks,data augmentation,ICLR}
}
@inproceedings{DBLP:conf/aaai/ShiON20,
author = {Yifeng Shi and
Junier Oliva and
Marc Niethammer},
title = {Deep Message Passing on Sets},
booktitle = {The Thirty-Fourth {AAAI} Conference on Artificial Intelligence, {AAAI}
2020, The Thirty-Second Innovative Applications of Artificial Intelligence
Conference, {IAAI} 2020, The Tenth {AAAI} Symposium on Educational
Advances in Artificial Intelligence, {EAAI} 2020, New York, NY, USA,
February 7-12, 2020},
pages = {5750--5757},
publisher = {{AAAI} Press},
year = {2020},
url = {https://aaai.org/ojs/index.php/AAAI/article/view/6031},
timestamp = {Thu, 04 Jun 2020 01:00:00 +0200},
biburl = {https://dblp.org/rec/conf/aaai/ShiON20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Modern methods for learning over graph input data have shown the fruitfulness of accounting for relationships among elements in a collection. However, most methods that learn over set input data use only rudimentary approaches to exploit intra-collection relationships. In this work we introduce Deep Message Passing on Sets (DMPS), a novel method that incorporates relational learning for sets. DMPS not only connects learning on graphs with learning on sets via deep kernel learning, but it also bridges message passing on sets and traditional diffusion dynamics commonly used in denoising models. Based on these connections, we develop two new blocks for relational learning on sets: the set-denoising block and the set-residual block. The former is motivated by the connection between message passing on general graphs and diffusion-based denoising models, whereas the latter is inspired by the well-known residual network. In addition to demonstrating the interpretability of our model by learning the true underlying relational structure experimentally, we also show the effectiveness of our approach on both synthetic and real-world datasets by achieving results that are competitive with or outperform the state-of-the-art. For readers who are interested in the detailed derivations of serveral results that we present in this work, please see the supplementary material at: https://arxiv. org/abs/1909.09877.},
keywords = {AAAI,deep learning}
}
@inproceedings{DBLP:conf/miccai/MooreMNGA20,
author = {Brad T. Moore and
Sean Montgomery and
Marc Niethammer and
Hastings Greer and
Stephen R. Aylward},
editor = {Yipeng Hu and
Roxane Licandro and
J. Alison Noble and
Jana Hutter and
Stephen R. Aylward and
Andrew Melbourne and
Esra Abaci Turk and
Jordina Torrents{-}Barrena},
title = {Automatic Optic Nerve Sheath Measurement in Point-of-Care Ultrasound},
booktitle = {Medical Ultrasound, and Preterm, Perinatal and Paediatric Image Analysis
- First International Workshop, {ASMUS} 2020, and 5th International
Workshop, {PIPPI} 2020, Held in Conjunction with {MICCAI} 2020, Lima,
Peru, October 4-8, 2020, Proceedings},
series = {Lecture Notes in Computer Science},
volume = {12437},
pages = {23--32},
publisher = {Springer},
year = {2020},
url = {https://doi.org/10.1007/978-3-030-60334-2\_3},
doi = {10.1007/978-3-030-60334-2\_3},
timestamp = {Wed, 07 Apr 2021 16:01:49 +0200},
biburl = {https://dblp.org/rec/conf/miccai/MooreMNGA20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Intracranial hypertension associated with traumatic brain injury is a life-threatening condition which requires immediate diagnosis and treatment. The measurement of optic nerve sheath diameter (ONSD), using ultrasonography, has been shown to be a promising, non-invasive predictor of intracranial pressure (ICP). Unfortunately, the reproducibility and accuracy of this measure depends on the expertise of the sonologist- a requirement that limits the broad application of ONSD. Previous work on ONSD measurement has focused on computer-automated annotation of expert-acquired ultrasound taken in a clinical setting. Here, we present a system using a handheld point-of-care ultrasound probe whereby the ONSD is automatically measured without requiring an expert sonographer to acquire the images. We report our results on videos from ocular phantoms with varying ONSDs. We show that our approach accurately measures the ONSD despite the lack of an observer keeping the ONSD in focus or in frame.},
keywords = {ultrasound,MICCAI}
}
@inproceedings{DBLP:conf/miccai/HanSXBABDN20,
author = {Xu Han and
Zhengyang Shen and
Zhenlin Xu and
Spyridon Bakas and
Hamed Akbari and
Michel Bilello and
Christos Davatzikos and
Marc Niethammer},
editor = {Mingxia Liu and
Pingkun Yan and
Chunfeng Lian and
Xiaohuan Cao},
title = {A Deep Network for Joint Registration and Reconstruction of Images
with Pathologies},
booktitle = {Machine Learning in Medical Imaging - 11th International Workshop,
{MLMI} 2020, Held in Conjunction with {MICCAI} 2020, Lima, Peru, October
4, 2020, Proceedings},
series = {Lecture Notes in Computer Science},
volume = {12436},
pages = {342--352},
publisher = {Springer},
year = {2020},
url = {https://drive.google.com/file/d/1AaLx1KS3QvxAAHsh6qFrxBoMfV5IELUO},
doi = {10.1007/978-3-030-59861-7\_35},
timestamp = {Tue, 06 Oct 2020 16:05:22 +0200},
biburl = {https://dblp.org/rec/conf/miccai/HanSXBABDN20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Registration of images with pathologies is challenging due to tissue appearance changes and missing correspondences caused by the pathologies. Moreover, mass effects as observed for brain tumors may displace tissue, creating larger deformations over time than what is observed in a healthy brain. Deep learning models have successfully been applied to image registration to offer dramatic speed up and to use surrogate information (e.g., segmentations) during training. However, existing approaches focus on learning registration models using images from healthy patients. They are therefore not designed for the registration of images with strong pathologies for example in the context of brain tumors, and traumatic brain injuries. In this work, we explore a deep learning approach to register images with brain tumors to an atlas. Our model learns an appearance mapping from images with tumors to the atlas, while simultaneously predicting the transformation to atlas space. Using separate decoders, the network disentangles the tumor mass effect from the reconstruction of quasi-normal images. Results on both synthetic and real brain tumor scans show that our approach outperforms cost function masking for registration to the atlas and that reconstructed quasi-normal images can be used for better longitudinal registrations.},
keywords = {registration,brain,pathology,MICCAI}
}
@inproceedings{DBLP:conf/miccai/GerberN20,
author = {Samuel Gerber and
Marc Niethammer},
editor = {Anne L. Martel and
Purang Abolmaesumi and
Danail Stoyanov and
Diana Mateus and
Maria A. Zuluaga and
S. Kevin Zhou and
Daniel Racoceanu and
Leo Joskowicz},
title = {Spatial Component Analysis to Mitigate Multiple Testing in Voxel-Based
Analysis},
booktitle = {Medical Image Computing and Computer Assisted Intervention - {MICCAI}
2020 - 23rd International Conference, Lima, Peru, October 4-8, 2020,
Proceedings, Part {VII}},
series = {Lecture Notes in Computer Science},
volume = {12267},
pages = {667--677},
publisher = {Springer},
year = {2020},
url = {https://doi.org/10.1007/978-3-030-59728-3\_65},
doi = {10.1007/978-3-030-59728-3\_65},
timestamp = {Mon, 26 Apr 2021 14:27:06 +0200},
biburl = {https://dblp.org/rec/conf/miccai/GerberN20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Voxel-based analysis provides a simple, easy to interpret approach to discover regions correlated with a variable of interest such as for example a pathology indicator. Voxel-based analysis methods perform a statistical test at each voxel and are prone to false positives due to multiple testing, or when corrected for multiple testing may miss regions of interest. Component based approaches, such as principal or independent component analysis provide an approach to mitigate multiple testing, by testing for correlations to projections of the data to the components. We propose a spatially regularized component analysis approach to find components for image data sets that are spatially localized and smooth. We show that the proposed approach leads to components that are easier to interpret and can improve predictive performance when used with linear regression models. We develop an efficient optimization approach using the Grassmannian projection kernel and a randomized SVD. The proposed optimization is capable to deal with data sets to large too fit all at once into memory. We demonstrate the approach with an application to study Alzheimer's disease using over 1200 images from the OASIS-3 data set.},
keywords = {MICCAI,statistics,brain},
}
@inproceedings{DBLP:conf/miccai/LiuLAN20,
author = {Peirong Liu and
Yueh Z. Lee and
Stephen R. Aylward and
Marc Niethammer},
editor = {Anne L. Martel and
Purang Abolmaesumi and
Danail Stoyanov and
Diana Mateus and
Maria A. Zuluaga and
S. Kevin Zhou and
Daniel Racoceanu and
Leo Joskowicz},
title = {{PIANO:} Perfusion Imaging via Advection-Diffusion},
booktitle = {Medical Image Computing and Computer Assisted Intervention - {MICCAI}
2020 - 23rd International Conference, Lima, Peru, October 4-8, 2020,
Proceedings, Part {VII}},
series = {Lecture Notes in Computer Science},
volume = {12267},
pages = {688--698},
publisher = {Springer},
year = {2020},
url = {https://drive.google.com/file/d/1goRkNQ27N262Fg-arCzCZG_Cx4-pMqYV},
doi = {10.1007/978-3-030-59728-3\_67},
timestamp = {Mon, 26 Apr 2021 14:27:06 +0200},
biburl = {https://dblp.org/rec/conf/miccai/LiuLAN20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Perfusion imaging (PI) is clinically used to assess strokes and brain tumors. Commonly used PI approaches based on magnetic resonance imaging (MRI) or computed tomography (CT) image the effect of a contrast agent moving through blood vessels and into tissue. Contrast-agent free approaches, for example, based on intravoxel incoherent motion, also exist, but are so far not routinely used clinically. MR or CT perfusion imaging based on contrast agents relies on the estimation of the arterial input function (AIF) to approximately model tissue perfusion, neglecting spatial dependencies. Reliably estimating the AIF is also non-trivial, leading to difficulties with standardizing perfusion measures. In this work we therefore propose a data-assimilation approach (PIANO) which estimates the velocity and diffusion fields of an advection-diffusion model best explaining the contrast dynamics. PIANO accounts for spatial dependencies and neither requires estimating the AIF nor relies on a particular contrast agent bolus shape. Specifically, we propose a convenient parameterization of the estimation problem, a numerical estimation approach, and extensively evaluate PIANO. We demonstrate that PIANO can successfully resolve velocity and diffusion field ambiguities and results in sensitive measures for the assessment of stroke, comparing favorably to conventional measures of perfusion.},
keywords = {MICCAI,perfusion,brain}
}
@inproceedings{DBLP:conf/miccai/TianPLSALN20,
author = {Lin Tian and
Connor Puett and
Peirong Liu and
Zhengyang Shen and
Stephen R. Aylward and
Yueh Z. Lee and
Marc Niethammer},
editor = {Anne L. Martel and
Purang Abolmaesumi and
Danail Stoyanov and
Diana Mateus and
Maria A. Zuluaga and
S. Kevin Zhou and
Daniel Racoceanu and
Leo Joskowicz},
title = {Fluid Registration Between Lung {CT} and Stationary Chest Tomosynthesis
Images},
booktitle = {Medical Image Computing and Computer Assisted Intervention - {MICCAI}
2020 - 23rd International Conference, Lima, Peru, October 4-8, 2020,
Proceedings, Part {III}},
series = {Lecture Notes in Computer Science},
volume = {12263},
pages = {307--317},
publisher = {Springer},
year = {2020},
doi = {10.1007/978-3-030-59716-0\_30},
timestamp = {Mon, 05 Oct 2020 18:46:13 +0200},
biburl = {https://dblp.org/rec/conf/miccai/TianPLSALN20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
url = {https://drive.google.com/file/d/1-gORB0x9qa8hDpnpLSISXGmb9I6j9SG9},
abstract = {Registration is widely used in image-guided therapy and image-guided surgery to estimate spatial correspondences between organs of interest between planning and treatment images. However, while high-quality computed tomography (CT) images are often available at planning time, limited angle acquisitions are frequently used during treatment because of radiation concerns or imaging time constraints. This requires algorithms to register CT images based on limited angle acquisitions. We, therefore, formulate a 3D/2D registration approach which infers a 3D deformation based on measured projections and digitally reconstructed radiographs of the CT. Most 3D/2D registration approaches use simple transformation models or require complex mathematical derivations to formulate the underlying optimization problem. Instead, our approach entirely relies on differentiable operations which can be combined with modern computational toolboxes supporting automatic differentiation. This then allows for rapid prototyping, integration with deep neural networks, and to support a variety of transformation models including fluid flow models. We demonstrate our approach for the registration between CT and stationary chest tomosynthesis (sDCT) images and show how it naturally leads to an iterative image reconstruction approach.},
keywords = {MICCAI,registration,lung,LDDMM}
}
@inproceedings{DBLP:conf/miccai/ShenXON20,
author = {Zhengyang Shen and
Zhenlin Xu and
Sahin Olut and
Marc Niethammer},
editor = {Anne L. Martel and
Purang Abolmaesumi and
Danail Stoyanov and
Diana Mateus and
Maria A. Zuluaga and
S. Kevin Zhou and
Daniel Racoceanu and
Leo Joskowicz},
title = {Anatomical Data Augmentation via Fluid-Based Image Registration},
booktitle = {Medical Image Computing and Computer Assisted Intervention - {MICCAI}
2020 - 23rd International Conference, Lima, Peru, October 4-8, 2020,
Proceedings, Part {III}},
series = {Lecture Notes in Computer Science},
volume = {12263},
pages = {318--328},
publisher = {Springer},
year = {2020},
doi = {10.1007/978-3-030-59716-0\_31},
timestamp = {Mon, 05 Oct 2020 18:46:13 +0200},
biburl = {https://dblp.org/rec/conf/miccai/ShenXON20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
url = {https://drive.google.com/file/d/1WzuwW5Hk8LIGEUCfQOrTHle3B1A1LroY},
abstract = {We introduce a fluid-based image augmentation method for medical image analysis. In contrast to existing methods, our framework generates anatomically meaningful images via interpolation from the geodesic subspace underlying given samples. Our approach consists of three steps: 1) given a source image and a set of target images, we construct a geodesic subspace using the Large Deformation Diffeomorphic Metric Mapping (LDDMM) model; 2) we sample transformations from the resulting geodesic subspace; 3) we obtain deformed images and segmentations via interpolation. Experiments on brain (LPBA) and knee (OAI) data illustrate the performance of our approach on two tasks: 1) data augmentation during training and testing for image segmentation; 2) one-shot learning for single atlas image segmentation. We demonstrate that our approach generates anatomically meaningful data and improves performance on these tasks over competing approaches.},
keywords = {MICCAI,registration,segmentation,brain,knee,LDDMM}
}
@inproceedings{DBLP:conf/eccv/OlutSXGN20,
author = {Sahin Olut and
Zhengyang Shen and
Zhenlin Xu and
Samuel Gerber and
Marc Niethammer},
editor = {Andrea Vedaldi and
Horst Bischof and
Thomas Brox and
Jan{-}Michael Frahm},
title = {Adversarial Data Augmentation via Deformation Statistics},
booktitle = {Computer Vision - {ECCV} 2020 - 16th European Conference, Glasgow,
UK, August 23-28, 2020, Proceedings, Part {XXIX}},
series = {Lecture Notes in Computer Science},
volume = {12374},
pages = {643--659},
publisher = {Springer},
year = {2020},
url = {https://doi.org/10.1007/978-3-030-58526-6\_38},
doi = {10.1007/978-3-030-58526-6\_38},
timestamp = {Wed, 07 Oct 2020 19:50:12 +0200},
biburl = {https://dblp.org/rec/conf/eccv/OlutSXGN20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Deep learning models have been successful in computer vision and medical image analysis. However, training these models frequently requires large labeled image sets whose creation is often very time and labor intensive, for example, in the context of 3D segmentations. Approaches capable of training deep segmentation networks with a limited number of labeled samples are therefore highly desirable. Data augmentation or semi-supervised approaches are commonly used to cope with limited labeled training data. However, the augmentation strategies for many existing approaches are either hand-engineered or require computationally demanding searches. To that end, we explore an augmentation strategy which builds statistical deformation models from unlabeled data via principal component analysis and uses the resulting statistical deformation space to augment the labeled training samples. Specifically, we obtain transformations via deep registration models. This allows for an intuitive control over plausible deformation magnitudes via the statistical model and, if combined with an appropriate deformation model, yields spatially regular transformations. To optimally augment a dataset we use an adversarial strategy integrated into our statistical deformation model. We demonstrate the effectiveness of our approach for the segmentation of knee cartilage from 3D magnetic resonance images. We show favorable performance to state-of-the-art augmentation approaches.},
keywords = {ECCV,registration,segmentation,knee,statistics,deep learning}
}
@inproceedings{DBLP:conf/isbi/DingHN20,
author = {Zhipeng Ding and
Xu Han and
Marc Niethammer},
title = {Votenet+: An Improved Deep Learning Label Fusion Method for Multi-Atlas
Segmentation},
booktitle = {17th {IEEE} International Symposium on Biomedical Imaging, {ISBI}
2020, Iowa City, IA, USA, April 3-7, 2020},
pages = {363--367},
publisher = {{IEEE}},
year = {2020},
url = {https://drive.google.com/file/d/17wMnHNosMbFjYhu4TuNd8zmbXtCA2398},
doi = {10.1109/ISBI45749.2020.9098493},
timestamp = {Sun, 07 Jun 2020 18:48:26 +0200},
biburl = {https://dblp.org/rec/conf/isbi/DingHN20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {In this work, we improve the performance of multi-atlas segmentation (MAS) by integrating the recently proposed VoteNet model with the joint label fusion (JLF) approach. Specifically, we first illustrate that using a deep convolutional neural network to predict atlas probabilities can better distinguish correct atlas labels from incorrect ones than relying on image intensity difference as is typical in JLF. Motivated by this finding, we propose VoteNet+, an improved deep network to locally predict the probability of an atlas label to differ from the label of the target image. Furthermore, we show that JLF is more suitable for the VoteNet framework as a label fusion method than plurality voting. Lastly, we use Platt scaling to calibrate the probabilities of our new model. Results on LPBA40 3D MR brain images show that our proposed method can achieve better performance than VoteNet.},
keywords = {ISBI,deep learning,segmentation,registration,brain}
}
@inproceedings{DBLP:conf/icml/HoferGNK20,
author = {Christoph D. Hofer and
Florian Graf and
Marc Niethammer and
Roland Kwitt},
title = {Topologically Densified Distributions},
booktitle = {Proceedings of the 37th International Conference on Machine Learning,
{ICML} 2020, 13-18 July 2020, Virtual Event},
series = {Proceedings of Machine Learning Research},
volume = {119},
pages = {4304--4313},
publisher = {{PMLR}},
year = {2020},
url = {http://proceedings.mlr.press/v119/hofer20a.html},
timestamp = {Tue, 15 Dec 2020 17:40:19 +0100},
biburl = {https://dblp.org/rec/conf/icml/HoferGNK20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {We study regularization in the context of small sample-size learning with over-parameterized neural networks. Specifically, we shift focus from architectural properties, such as norms on the network weights, to properties of the internal representations before a linear classifier. Specifically, we impose a topological constraint on samples drawn from the probability measure induced in that space. This provably leads to mass concentration effects around the representations of training instances, ie, a property beneficial for generalization. By leveraging previous work to impose topological constraints in a neural network setting, we provide empirical evidence (across various vision benchmarks) to support our claim for better generalization.},
keywords={ICML,deep learning,topology}
}
@article{DBLP:journals/corr/abs-2006-04259,
author = {Yifeng Shi and
Christopher M. Bender and
Junier B. Oliva and
Marc Niethammer},
title = {Deep Goal-Oriented Clustering},
journal = {CoRR},
volume = {abs/2006.04259},
year = {2020},
url = {https://arxiv.org/abs/2006.04259},
archivePrefix = {arXiv},
eprint = {2006.04259},
timestamp = {Fri, 12 Jun 2020 01:00:00 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-2006-04259.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Clustering and prediction are two primary tasks in the fields of unsupervised and supervised learning, respectively. Although much of the recent advances in machine learning have been centered around those two tasks, the interdependent, mutually beneficial relationship between them is rarely explored. One could reasonably expect appropriately clustering the data would aid the downstream prediction task and, conversely, a better prediction performance for the downstream task could potentially inform a more appropriate clustering strategy. In this work, we focus on the latter part of this mutually beneficial relationship. To this end, we introduce Deep Goal-Oriented Clustering (DGC), a probabilistic framework that clusters the data by jointly using supervision via side-information and unsupervised modeling of the inherent data structure in an end-to-end fashion. We show the effectiveness of our model on a range of datasets by achieving prediction accuracies comparable to the state-of-the-art, while, more importantly in our setting, simultaneously learning congruent clustering strategies.},
keywords = {deep learning,clustering}
}
@inproceedings{DBLP:conf/nips/VialardKWN20,
author = {Fran{\c{c}}ois{-}Xavier Vialard and
Roland Kwitt and
Susan Wei and
Marc Niethammer},
editor = {Hugo Larochelle and
Marc'Aurelio Ranzato and
Raia Hadsell and
Maria{-}Florina Balcan and
Hsuan{-}Tien Lin},
title = {A shooting formulation of deep learning},
booktitle = {Advances in Neural Information Processing Systems 33: Annual Conference
on Neural Information Processing Systems 2020, NeurIPS 2020, December
6-12, 2020, virtual},
year = {2020},
url = {https://proceedings.neurips.cc/paper/2020/hash/89562dccfeb1d0394b9ae7e09544dc70-Abstract.html},
timestamp = {Tue, 19 Jan 2021 15:56:51 +0100},
biburl = {https://dblp.org/rec/conf/nips/VialardKWN20.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Continuous-depth neural networks can be viewed as deep limits of discrete neural networks whose dynamics resemble a discretization of an ordinary differential equation (ODE). Although important steps have been taken to realize the advantages of such continuous formulations, most current techniques are not truly continuous-depth as they assume identical layers. Indeed, existing works throw into relief the myriad difficulties presented by an infinite-dimensional parameter space in learning a continuous-depth neural ODE. To this end, we introduce a shooting formulation which shifts the perspective from parameterizing a network layer-by-layer to parameterizing over optimal networks described only by a set of initial conditions. For scalability, we propose a novel particle-ensemble parametrization which fully specifies the optimal weight trajectory of the continuous-depth neural network. Our experiments show that our particle-ensemble shooting formulation can achieve competitive performance, especially on long-range forecasting tasks. Finally, though the current work is inspired by continuous-depth neural networks, the particle-ensemble shooting formulation also applies to discrete-time networks and may lead to a new fertile area of research in deep learning parametrization.},
keywords = {deep learning,NeurIPS}
}
@article{DBLP:journals/ans/StanleyBKNM19,
author = {Natalie Stanley and
Thomas Bonacci and
Roland Kwitt and
Marc Niethammer and
Peter J. Mucha},
title = {Stochastic block models with multiple continuous attributes},
journal = {Applied Network Science},
volume = {4},
number = {1},
pages = {54:1--54:22},
year = {2019},
url = {https://doi.org/10.1007/s41109-019-0170-z},
doi = {10.1007/s41109-019-0170-z},
timestamp = {Tue, 20 Aug 2019 01:00:00 +0200},
biburl = {https://dblp.org/rec/journals/ans/StanleyBKNM19.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {The stochastic block model (SBM) is a probabilistic model for community structure in networks. Typically, only the adjacency matrix is used to perform SBM parameter inference. In this paper, we consider circumstances in which nodes have an associated vector of continuous attributes that are also used to learn the node-to-community assignments and corresponding SBM parameters. Our model assumes that the attributes associated with the nodes in a network’s community can be described by a common multivariate Gaussian model. In this augmented, attributed SBM, the objective is to simultaneously learn the SBM connectivity probabilities with the multivariate Gaussian parameters describing each community. While there are recent examples in the literature that combine connectivity and attribute information to inform community detection, our model is the first augmented stochastic block model to handle multiple continuous attributes. This provides the flexibility in biological data to, for example, augment connectivity information with continuous measurements from multiple experimental modalities. Because the lack of labeled network data often makes community detection results difficult to validate, we highlight the usefulness of our model for two network prediction tasks: link prediction and collaborative filtering. As a result of fitting this attributed stochastic block model, one can predict the attribute vector or connectivity patterns for a new node in the event of the complementary source of information (connectivity or attributes, respectively). We also highlight two biological examples where the attributed stochastic block model provides satisfactory performance in the link prediction and collaborative filtering tasks.},
keywords = {SBM,network,Applied Network Science}
}
@article{DBLP:journals/jmlr/HoferKN19,
author = {Christoph D. Hofer and
Roland Kwitt and
Marc Niethammer},
title = {Learning Representations of Persistence Barcodes},
journal = {J. Mach. Learn. Res.},
volume = {20},
pages = {126:1--126:45},
year = {2019},
url = {http://jmlr.org/papers/v20/18-358.html},
timestamp = {Thu, 18 Jun 2020 01:00:00 +0200},
biburl = {https://dblp.org/rec/journals/jmlr/HoferKN19.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {We consider the problem of supervised learning with summary representations of topological features in data. In particular, we focus on persistent homology, the prevalent tool used in topological data analysis. As the summary representations, referred to as barcodes or persistence diagrams, come in the unusual format of multi sets, equipped with computationally expensive metrics, they can not readily be processed with conventional learning techniques. While different approaches to address this problem have been proposed, either in the context of kernel-based learning, or via carefully designed vectorization techniques, it remains an open problem how to leverage advances in representation learning via deep neural networks. Appropriately handling topological summaries as input to neural networks would address the disadvantage of previous strategies which handle this type of data in a task-agnostic manner. In particular, we propose an approach that is designed to learn a task-specific representation of barcodes. In other words, we aim to learn a representation that adapts to the learning problem while, at the same time, preserving theoretical properties (such as stability). This is done by projecting barcodes into a finite dimensional vector space using a collection of parametrized functionals, so called structure elements, for which we provide a generic construction scheme. A theoretical analysis of this approach reveals sufficient conditions to preserve stability, and also shows that different choices of structure elements lead to great differences with respect to their suitability for numerical optimization. When implemented as a neural network input layer, our approach demonstrates compelling performance on various types of problems, including graph classification and eigenvalue prediction, the classification of 2D/3D object shapes and recognizing activities from EEG signals.},
keywords = {topology,deep learning,JMLR}
}
@article{DBLP:journals/mia/DingFYTKN19,
author = {Zhipeng Ding and
Greg M. Fleishman and
Xiao Yang and
Paul Thompson and
Roland Kwitt and
Marc Niethammer},
title = {Fast predictive simple geodesic regression},
journal = {Medical Image Anal.},
volume = {56},
pages = {193--209},
year = {2019},
url = {https://drive.google.com/file/d/1D2RastKoyhEl9lwOYDD5wG_Fy4LwEFc5},
doi = {10.1016/j.media.2019.06.003},
timestamp = {Sat, 30 May 2020 01:00:00 +0200},
biburl = {https://dblp.org/rec/journals/mia/DingFYTKN19.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Deformable image registration and regression are important tasks in medical image analysis. However, they are computationally expensive, especially when analyzing large-scale datasets that contain thousands of images. Hence, cluster computing is typically used, making the approaches dependent on such computational infrastructure. Even larger computational resources are required as study sizes increase. This limits the use of deformable image registration and regression for clinical applications and as component algorithms for other image analysis approaches. We therefore propose using a fast predictive approach to perform image registrations. In particular, we employ these fast registration predictions to approximate a simplified geodesic regression model to capture longitudinal brain changes. The resulting method is orders of magnitude faster than the standard optimization-based regression model and hence facilitates large-scale analysis on a single graphics processing unit (GPU). We evaluate our results on 3D brain magnetic resonance images (MRI) from the ADNI datasets.},
keywords = {registration,brain,regression,MEDIA}
}
@article{DBLP:journals/tbe/ChittajalluMGCG19,
author = {Deepak Roy Chittajallu and
Matthew McCormick and
Samuel Gerber and
Tomasz J. Czernuszewicz and
Ryan C. Gessner and
Monte S. Willis and
Marc Niethammer and
Roland Kwitt and
Stephen R. Aylward},
title = {Image-Based Methods for Phase Estimation, Gating, and Temporal Superresolution
of Cardiac Ultrasound},
journal = {{IEEE} Trans. Biomed. Engineering},
volume = {66},
number = {1},
pages = {72--79},
year = {2019},
url = {https://doi.org/10.1109/TBME.2018.2823279},
doi = {10.1109/TBME.2018.2823279},
timestamp = {Sat, 30 May 2020 01:00:00 +0200},
biburl = {https://dblp.org/rec/journals/tbe/ChittajalluMGCG19.bib},
bibsource = {dblp computer science bibliography, https://dblp.org},
abstract = {Objective: Ultrasound is an effective tool for rapid noninvasive assessment of cardiac structure and function. Determining the cardiorespiratory phases of each frame in the ultrasound video and capturing the cardiac function at a much higher temporal resolution are essential in many applications. Fulfilling these requirements is particularly challenging in preclinical studies involving small animals with high cardiorespiratory rates, requiring cumbersome and expensive specialized hardware. Methods: We present a novel method for the retrospective estimation of cardiorespiratory phases directly from the ultrasound videos. It transforms the videos into a univariate time series preserving the evidence of periodic cardiorespiratory motion, decouples the signatures of cardiorespiratory motion with a trend extraction technique, and estimates the cardiorespiratory phases using a Hilbert transform approach. We also present a robust nonparametric regression technique for respiratory gating and a novel kernelregression model for reconstructing images at any cardiac phase facilitating temporal superresolution. Results: We validated our methods using two-dimensional echocardiography videos and electrocardiogram (ECG) recordings of six mice. Our cardiac phase estimation method provides accurate phase estimates with a mean-phase-error range of 3\%–6\% against ECG derived phase and outperforms three previously published methods in locating ECGs R-wave peak frames with a mean-frame-error range of 0.73–1.36. Our kernel-regression model accurately reconstructs images at any cardiac phase with a mean-normalizedcorrelation range of 0.81–0.85 over 50 leave-one-outcross-validation rounds. Conclusion and significance: Our methods can enable tracking of cardiorespiratory phases without additional hardware and reconstruction of respiration-free single cardiac-cycle videos at a much higher temporal resolution.},
keywords = {heart,regression,TBME}
}
@inproceedings{DBLP:conf/cvpr/ShenHXN19,
author = {Zhengyang Shen and
Xu Han and
Zhenlin Xu and
Marc Niethammer},
title = {Networks for Joint Affine and Non-Parametric Image Registration},
booktitle = {{IEEE} Conference on Computer Vision and Pattern Recognition, {CVPR}
2019, Long Beach, CA, USA, June 16-20, 2019},
pages = {4224--4233},
publisher = {Computer Vision Foundation / {IEEE}},
year = {2019},