Skip to content

Commit 9574ec2

Browse files
authored
Merge pull request #973 from qlinsey/main
Projects' Proposal Submission
2 parents e076c28 + 4a5bf5a commit 9574ec2

File tree

8 files changed

+194
-5
lines changed

8 files changed

+194
-5
lines changed

content/authors/LinseyPang/_index.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -10,17 +10,17 @@ authors:
1010
superuser: false
1111

1212
# Role/position
13-
role: "Principal Applied Scientist, Salesforce"
13+
role: "Distinguished Scientist, Paypal"
1414

1515

1616
# Organizations/Affiliations
1717
organizations:
18-
- name: Salesforce
19-
url: "https://www.salesforce.com/"
18+
- name: Paypal
19+
url: "https://www.paypal.com/"
2020

2121

2222
# Short bio (displayed in user profile at end of posts)
23-
bio: Linsey's research interests are in Deep Learning + Machine Learning + AI.
23+
bio: Linsey's research interests are in Machine Learning + AI.
2424

2525

2626

@@ -50,6 +50,6 @@ user_groups:
5050
- University of California Mentors
5151
- Summer of Reproducibility Mentors
5252
---
53-
Linsey Pang is a Principal Machine Learning Scientist at Salesforce, working on deep learning recommendation modeling projects. Prior to this, she was a Principal Data Scientist at Walmart Lab. She was leading the Pricing, One demand data science projects. Prior to joining Walmart Lab, she worked as an applied scientist at eBay Inc. She finished her PhD degree in machine learning research at the University of Sydney. She co-organizes multiple workshops and tutorials at CIKM, KDD, and WSDM etc in top data mining conferences.
53+
Linsey Pang is a Distinguished Scientist at Paypal. Previously she was a Principal Machine Learning Scientist at Salesforce. Prior to this, she was a Principal Data Scientist at Walmart Lab.Prior to joining Walmart Lab, she worked as an applied scientist at eBay Inc. She finished her PhD degree in machine learning research at the University of Sydney. She co-organizes multiple workshops and tutorials at CIKM, KDD, and WSDM etc in top data mining conferences.
5454

5555

content/project/osre26/NE

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
556 KB
Loading
Lines changed: 111 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,111 @@
1+
---
2+
title: "MedJEPA: Self-Supervised Medical Image Representation Learning with JEPA"
3+
authors: [bindong,LinseyPang]
4+
author_notes: ["Research Scientist, Lawrence Berkeley Lab","Northeastern"]
5+
tags: ["osre26", "uc", "AI", "LLM", "Health"]
6+
date: 2026-01-19T10:15:56-07:00
7+
lastmod: 2026-01-19T10:15:56-07:00
8+
---
9+
10+
### Project Description
11+
[MedJEPA] Medical image analysis is fundamental to modern healthcare, enabling disease diagnosis, treatment planning, and patient monitoring across diverse clinical applications. In radiology and pathology, deep learning models support automated detection of abnormalities, tumor segmentation, and diagnostic assistance. Medical imaging modalities including X-rays, CT scans, MRI, ultrasound, and histopathology slides generate vast amounts of unlabeled data that could benefit from self-supervised representation learning. Clinical applications include cancer detection and staging, cardiovascular disease assessment, neurological disorder diagnosis, and infectious disease screening. In drug discovery and clinical research, analyzing medical images helps evaluate treatment efficacy, predict patient outcomes, and identify biomarkers for disease progression. Telemedicine and point-of-care diagnostics benefit from AI-powered image analysis that extends expert-level interpretation to underserved regions. However, medical imaging faces unique challenges: limited labeled datasets due to expensive expert annotation, patient privacy concerns restricting data sharing, domain shift across different imaging equipment and protocols, and the need for models that generalize across hospitals and populations.
12+
Traditional medical image analysis relies heavily on supervised learning with manually annotated labels, creating bottlenecks due to the scarcity and cost of expert annotations. Existing self-supervised methods applied to medical imaging often employ complex training procedures with numerous heuristics—momentum encoders, stop-gradients, teacher-student architectures, and carefully tuned augmentation strategies—that may not translate well across different medical imaging modalities and clinical contexts. These approaches struggle with domain-specific challenges such as subtle pathological features, high-resolution images, 3D volumetric data, and the need for interpretable representations that clinicians can trust. To address these challenges, we propose MedicalJEPA: Self-Supervised Medical Image Representation Learning with Joint-Embedding Predictive Architecture, which leverages the theoretically grounded LeJEPA framework for 2D medical images and V-JEPA principles for medical video and volumetric data, creating a unified, scalable, and heuristics-free approach specifically tailored for medical imaging applications.
13+
By utilizing the principled JEPA frameworks with objectives like Sketched Isotropic Gaussian Regularization (SIGReg), MedJEPA eliminates complex training heuristics while learning clinically meaningful representations from unlabeled medical images. Unlike conventional self-supervised methods that require extensive hyperparameter tuning and may not generalize across medical imaging modalities, MedicalJEPA provides a clean, theoretically motivated framework with minimal hyperparameters that adapts to diverse medical imaging contexts—from chest X-rays to histopathology slides to cardiac MRI sequences. The learned representations can support downstream tasks including disease classification, lesion detection, organ segmentation, and survival prediction, while requiring significantly fewer labeled examples for fine-tuning. This approach democratizes access to state-of-the-art medical AI by enabling effective learning from the vast amounts of unlabeled medical imaging data available in hospital archives, addressing the annotation bottleneck that has limited progress in medical AI.
14+
15+
16+
17+
18+
### Project Objectives
19+
Aligned with the vision of the 2026 Open Source Research Experience (OSRE), this project aims to apply Joint-Embedding Predictive Architecture (JEPA) frameworks to medical image representation learning, addressing the critical challenge of learning from limited labeled medical data. Medical imaging generates enormous amounts of unlabeled data, but supervised learning approaches are bottlenecked by the scarcity and cost of expert annotations. Existing self-supervised methods often rely on complex heuristics that don't generalize well across diverse medical imaging modalities, equipment vendors, and clinical protocols.
20+
This project will leverage the theoretically grounded LeJEPA framework for 2D medical images (X-rays, histopathology slides, fundus images) and V-JEPA principles for temporal and volumetric medical data (cardiac MRI sequences, CT scans, surgical videos). The core challenge lies in adapting these heuristics-free, stable frameworks to medical imaging's unique characteristics: subtle pathological features requiring fine-grained representations, high-resolution images demanding efficient processing, domain shift across hospitals and equipment, and the need for interpretable features that support clinical decision-making. The learned representations will be evaluated on diverse downstream clinical tasks including disease classification, lesion detection, organ segmentation, and prognosis prediction, with emphasis on few-shot learning scenarios that reflect real-world annotation constraints. Below is an outline of the methodologies and models that will be developed in this project.
21+
22+
23+
- **Step 1: Medical Data Preparation**:
24+
Develop data processing pipelines for diverse medical imaging modalities, implementing DICOM/NIfTI parsing, standardized preprocessing, and efficient data loading for self-supervised pre-training.
25+
Prepare 2D medical image datasets:
26+
Chest X-rays: ChestX-ray14, MIMIC-CXR, CheXpert for lung disease detection
27+
Histopathology: Camelyon16/17 (breast cancer), PCam (patch-level classification)
28+
Retinal imaging: EyePACS, APTOS (diabetic retinopathy), Messidor
29+
Dermatology: HAM10000, ISIC (skin lesion classification)
30+
Prepare 3D volumetric and temporal medical data:
31+
CT scans: LIDC-IDRI (lung nodules), Medical Segmentation Decathlon datasets
32+
MRI sequences: BraTS (brain tumors), ACDC (cardiac MRI), UK Biobank cardiac videos
33+
Medical video: Surgical procedure videos, endoscopy recordings, ultrasound sequences
34+
Implement medical imaging-specific preprocessing: intensity normalization, resolution standardization, handling of multi-channel medical images (different MRI sequences, RGB histopathology), and privacy-preserving anonymization.
35+
Design masking strategies appropriate for medical imaging: spatial masking for 2D images, volumetric masking for 3D scans, temporal masking for sequences, and anatomy-aware masking that respects organ boundaries.
36+
Create data loaders supporting high-resolution medical images, 3D volumes, and multi-modal inputs (e.g., multiple MRI sequences).
37+
38+
39+
40+
41+
- **Step 2: JEPA Model Implementation for Medical Imaging**:
42+
Implement LeJEPA for 2D medical images:
43+
Adapt joint-embedding predictive architecture for medical image characteristics (high resolution, subtle features, domain-specific patterns)
44+
Apply Sketched Isotropic Gaussian Regularization (SIGReg) to learn clinically meaningful embedding distributions
45+
Maintain single trade-off hyperparameter and heuristics-free training for reproducibility across medical imaging centers
46+
Support various encoder architectures: Vision Transformers for global context, ConvNets for local features, hybrid approaches
47+
Extend to V-JEPA for medical video and volumetric data:
48+
Spatiotemporal encoding for cardiac MRI sequences, surgical videos, and time-series medical imaging
49+
Temporal prediction objectives for understanding disease progression and treatment response
50+
3D volume processing for CT and MRI scans with efficient memory management
51+
Multi-slice and multi-sequence learning for comprehensive medical imaging contexts
52+
Develop medical domain-specific enhancements:
53+
Multi-scale representation learning to capture both fine-grained pathological details and global anatomical context
54+
Interpretability mechanisms: attention visualization, feature attribution, and embedding space analysis for clinical validation
55+
Robustness to domain shift: training strategies that generalize across different scanners, protocols, and institutions
56+
Privacy-preserving training considerations compatible with medical data regulations (HIPAA, GDPR)
57+
Implement efficient training infrastructure:
58+
Support for distributed training across multiple GPUs for large medical imaging datasets
59+
Memory-efficient processing of high-resolution images and 3D volumes
60+
Checkpoint management and model versioning for clinical deployment pipelines
61+
Minimal-code implementation (≈50-100 lines) demonstrating framework simplicity
62+
63+
64+
65+
- **Step 3: Evaluation & Safety Validation**: :
66+
Disease Classification Tasks:
67+
Multi-label chest X-ray classification: 14 pathology classes on ChestX-ray14, MIMIC-CXR
68+
Diabetic retinopathy grading: 5-class classification on EyePACS, APTOS
69+
Skin lesion classification: 7-class classification on HAM10000
70+
Brain tumor classification: glioma grading on BraTS dataset
71+
Evaluate with linear probing, few-shot learning (5-shot, 10-shot), and full fine-tuning
72+
Lesion Detection and Segmentation:
73+
Lung nodule detection on LIDC-IDRI dataset
74+
Tumor segmentation on Medical Segmentation Decathlon tasks
75+
Polyp detection in colonoscopy videos
76+
Cardiac structure segmentation in MRI sequences
77+
Clinical Prediction Tasks:
78+
Survival prediction from histopathology slides
79+
Disease progression prediction from longitudinal imaging
80+
Treatment response assessment from pre/post imaging pairs
81+
Few-Shot and Low-Data Regime Evaluation:
82+
Systematic evaluation with 1%, 5%, 10%, 25%, 50% of labeled training data
83+
Comparison against supervised baselines and ImageNet pre-training
84+
Analysis of annotation efficiency: performance vs. number of labeled examples required
85+
86+
87+
88+
### Project Deliverables
89+
This project will deliver three components: software implementation, clinical evaluation, and practical deployment resources. The software implementing MedicalJEPA will be hosted on GitHub as an open-access repository with modular code supporting multiple medical imaging modalities (2D images, 3D volumes, videos), pre-trained model checkpoints on major medical imaging datasets (chest X-rays, histopathology, MRI), training and evaluation scripts with medical imaging-specific preprocessing pipelines, privacy-preserving training implementations compatible with clinical data regulations, and comprehensive documentation including tutorials for medical AI researchers and clinicians. The evaluation results will include benchmarks on 10+ medical imaging datasets across diverse modalities and clinical tasks, few-shot learning analysis demonstrating annotation efficiency gains, cross-institutional validation studies showing robustness to domain shift, interpretability visualizations enabling clinical validation of learned representations, and detailed comparisons against supervised baselines and existing medical self-supervised methods. .
90+
91+
### NeuroHealth
92+
93+
- **Topics**: Self-Supervised Medical Image Representation Learning with JEPA
94+
- **Skills**: Proficiency in Python, Pytorch, Github, JEPA
95+
- **Difficulty**: Difficult
96+
- **Size**: Large (350 hours)
97+
- **Mentor**: {{% mention bindong %}}, {{% mention LinseyPang %}}
98+
99+
### References:
100+
101+
- LeJEPA: Provable and Scalable Self-Supervised Learning Without the Heuristics - Randall Balestriero and Yann LeCun, arXiv 2024
102+
- Revisiting Feature Prediction for Learning Visual Representations from Video (V-JEPA) - Adrien Bardes et al., arXiv 2024
103+
- Self-Supervised Learning from Images with a Joint-Embedding Predictive Architecture - Mahmoud Assran et al., CVPR 2023 (I-JEPA)
104+
- ChestX-ray14: Hospital-Scale Chest X-Ray Database - https://nihcc.app.box.com/v/ChestXray-NIHCC
105+
- Medical Segmentation Decathlon - http://medicaldecathlon.com/
106+
- MIMIC-CXR Database - https://physionet.org/content/mimic-cxr/
107+
- The Cancer Imaging Archive (TCIA) - https://www.cancerimagingarchive.net/
108+
- UK Biobank Imaging Study - https://www.ukbiobank.ac.uk/enable-your-research/about-our-data/imaging-data
109+
110+
111+
Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
321 KB
Loading

0 commit comments

Comments
 (0)