Course Schedule
Course overview and objectives; transformative impact of deep learning across clinical medicine; applications in medical image diagnosis (radiology, pathology, ophthalmology); protein structure prediction with AlphaFold; AI-driven drug discovery pipelines; unique challenges of biomedical data (class imbalance, annotation scarcity, privacy constraints, regulatory requirements)
Perceptrons and multilayer feedforward networks; forward and backward propagation of gradients; activation functions (ReLU, sigmoid, tanh); optimization algorithms (SGD, Adam, RMSProp); learning rate scheduling; regularization techniques (dropout, batch normalization, weight decay); bias-variance tradeoff and overfitting in medical settings
Convolution operations, stride, padding, and receptive fields; pooling layers for spatial downsampling; LeNet, AlexNet, VGG, and ResNet architectures; skip connections and residual learning; feature maps and hierarchical representation learning; image classification pipelines for biomedical imagery
Transfer learning from ImageNet to medical domains; feature extraction vs. fine-tuning strategies; pretrained architectures (ResNet, DenseNet, EfficientNet); MedMNIST and clinical imaging datasets; end-to-end classification pipelines; evaluation metrics (accuracy, AUC-ROC, F1-score, confusion matrix); handling class imbalance with weighted loss and oversampling
Advanced fine-tuning strategies (layer freezing, differential learning rates, gradual unfreezing); ensemble learning theory; soft voting (probability averaging) and hard voting (majority vote) ensembles; bagging and model diversity; combining heterogeneous architectures; performance gains and computational trade-offs of ensembles on medical imaging benchmarks
Semantic vs. instance segmentation; U-Net encoder-decoder architecture with skip connections; Attention U-Net and 3D volumetric extensions; combined Dice and cross-entropy loss; evaluation metrics (Dice coefficient, IoU, Hausdorff distance); clinical applications in tumor delineation, organ segmentation, and surgical planning; data augmentation strategies for limited medical annotations
Clinical text preprocessing and tokenization; named entity recognition for diseases, medications, and procedures; word embeddings (Word2Vec, GloVe) and contextual representations; transformer architecture (self-attention, multi-head attention, positional encoding); BERT and BioBERT for clinical NLP; information extraction from electronic health records (EHRs); ICD coding and clinical concept normalization
Recurrent neural networks and LSTMs for sequential clinical data; sequence-to-sequence models with encoder-decoder attention; character-level and subword language models; clinical note summarization and discharge summary generation; question answering over medical literature; ROUGE and BLEU evaluation metrics; GPT-style models and instruction tuning for healthcare text tasks
Autoencoder architecture and latent space properties, variational autoencoders and probabilistic encoding, reparameterization trick, ELBO loss, GAN generator/discriminator training dynamics, Wasserstein GANs, mode collapse and vanishing gradients, medical imaging applications (MRI reconstruction, MR-to-CT synthesis, cycle-consistent GANs)
Deep dive into GAN architectures for clinical imaging: U-Net generator with skip connections, PatchGAN discriminator, multi-component loss (pixel-wise MSE, frequency-domain MSE, VGG perceptual loss, adversarial loss), MR-to-CT synthesis using cycle-consistent GANs for unpaired image translation, MRI acceleration with DAGAN (deep learning augmented GAN) using k-space undersampling and refinement learning
Denoising Diffusion Probabilistic Models (DDPMs): forward diffusion process, reverse denoising process, noise scheduling, sinusoidal time embeddings; U-Net with ResNet blocks and self-attention for the denoising network; CycleGAN with residual learning for multi-modality synthesis (MRI-to-CT, 3T-to-7T MRI); text-to-image generation (DALL-E 2, Imagen); diffusion models for medical imaging tasks including segmentation, reconstruction, anomaly detection, and histopathology image generation; genotype-conditioned synthesis for glioma classification; evaluation metrics (FID, Inception Score, Improved Precision/Recall)
Assignments
Homework 1: Transfer Learning and Ensemble Methods
Assigned: Week 5 (February 7) · Due: Week 8 · Worth: 15% of final grade
Train pretrained models on BloodMNIST dataset, implement ensemble methods (averaging and voting), and compare performance across different architectures.
Homework 2: Medical Image Segmentation and NLP
Assigned: Week 8 (February 28) · Due: Week 11 · Worth: 15% of final grade
Implement a 3D Attention U-Net for volumetric medical image segmentation, train a character-level LSTM for medical text generation, and build an RNN with additive attention for clinical note classification.
Homework 3: Variational Autoencoders for Medical Image Generation
Assigned: Week 11 (March 28) · Due: Week 14 · Worth: 15% of final grade
Implement a Convolutional VAE from scratch — encoder, reparameterization trick, decoder, and ELBO loss — and apply it to PathMNIST colon pathology slides. Evaluate generation quality via reconstructions, random sampling, and latent space interpolation.