|
| 1 | +\documentclass[11pt]{article} |
| 2 | + |
| 3 | +\usepackage[a4paper,margin=1in]{geometry} |
| 4 | +\usepackage{amsmath,amssymb} |
| 5 | +\usepackage{hyperref} |
| 6 | + |
| 7 | +\title{Semester long Project: Integrating Discriminative and Generative Models} |
| 8 | +\author{FYS5429/9429} |
| 9 | +\date{Spring 2026} |
| 10 | + |
| 11 | +\begin{document} |
| 12 | + |
| 13 | +\maketitle |
| 14 | + |
| 15 | +\section{Introduction and Motivation} |
| 16 | + |
| 17 | +Modern machine learning spans two complementary paradigms: |
| 18 | + |
| 19 | +\begin{itemize} |
| 20 | +\item \textbf{Discriminative models}, which learn mappings from inputs to outputs (e.g.\ classification and regression). |
| 21 | +\item \textbf{Generative models}, which aim to learn the underlying data distribution. |
| 22 | +\end{itemize} |
| 23 | + |
| 24 | +While discriminative models such as neural networks, convolutional |
| 25 | +neural networks (CNNs), and recurrent neural networks (RNNs) are |
| 26 | +highly effective for prediction tasks, generative models such as |
| 27 | +variational autoencoders (VAEs), Boltzmann machines, and diffusion |
| 28 | +models provide deeper insight into the structure and distribution of |
| 29 | +data. |
| 30 | + |
| 31 | +The purpose of this project is to explore the interplay between these paradigms through theoretical analysis, implementation, and empirical evaluation. |
| 32 | + |
| 33 | +\section{Project Objectives} |
| 34 | + |
| 35 | +The main objectives of the project are: |
| 36 | + |
| 37 | +\begin{itemize} |
| 38 | +\item Develop a solid understanding of both discriminative and generative modeling approaches. |
| 39 | +\item Implement and train selected models on real datasets. |
| 40 | +\item Compare performance across models and tasks. |
| 41 | +\item Investigate how generative models can enhance discriminative tasks. |
| 42 | +\item Analyze strengths, limitations, and computational complexity. |
| 43 | +\end{itemize} |
| 44 | + |
| 45 | +\section{Project Tasks} |
| 46 | + |
| 47 | +The project consists of three main components. |
| 48 | + |
| 49 | +\subsection{Part I: Discriminative Modeling} |
| 50 | + |
| 51 | +You should implement and analyze at least two of the following models: |
| 52 | + |
| 53 | +\begin{itemize} |
| 54 | +\item Fully connected neural networks |
| 55 | +\item Convolutional neural networks (CNNs) |
| 56 | +\item Recurrent neural networks (RNNs) for temporal data |
| 57 | +\item Autoencoders (as feature extractors) |
| 58 | +\end{itemize} |
| 59 | + |
| 60 | +Tasks include (you should select whether you wish to study |
| 61 | +classification or regression problems): |
| 62 | + |
| 63 | +\begin{itemize} |
| 64 | +\item Training on classification datasets (e.g.\ MNIST, Fashion-MNIST) |
| 65 | +\item Optional regression tasks (e.g.\ time series prediction) |
| 66 | +\item Evaluation using appropriate metrics (accuracy, MSE, etc.) |
| 67 | +\end{itemize} |
| 68 | + |
| 69 | +\subsection{Part II: Generative Modeling} |
| 70 | + |
| 71 | +You should implement at least one generative model: |
| 72 | + |
| 73 | +\begin{itemize} |
| 74 | +\item Variational Autoencoder (VAE) |
| 75 | +\item Restricted Boltzmann Machine (RBM) or Boltzmann machine |
| 76 | +\item Diffusion model (optional advanced component) |
| 77 | +\end{itemize} |
| 78 | + |
| 79 | +Tasks include: |
| 80 | + |
| 81 | +\begin{itemize} |
| 82 | +\item Learning latent representations |
| 83 | +\item Generating new samples |
| 84 | +\item Evaluating generative quality (visual inspection, likelihood proxies) |
| 85 | +\end{itemize} |
| 86 | + |
| 87 | +\subsection{Part III: Integration and Comparison} |
| 88 | + |
| 89 | +The key part of the project is to combine the two paradigms. |
| 90 | + |
| 91 | +Possible directions include: |
| 92 | + |
| 93 | +\begin{itemize} |
| 94 | +\item Using latent variables from a VAE as input to a classifier |
| 95 | +\item Data augmentation using generative models |
| 96 | +\item Comparing classification performance with and without generative preprocessing |
| 97 | +\item Hybrid models combining reconstruction and prediction losses |
| 98 | +\end{itemize} |
| 99 | + |
| 100 | +\section{Datasets} |
| 101 | + |
| 102 | +You may choose among the following datasets (please feel free to add): |
| 103 | + |
| 104 | +\begin{itemize} |
| 105 | +\item \textbf{MNIST / Fashion-MNIST}: image classification |
| 106 | +\item \textbf{CIFAR-10}: more complex image classification |
| 107 | +\item \textbf{Time series datasets}: e.g.\ stock prices, physical signals |
| 108 | +\item \textbf{Synthetic datasets}: for controlled experiments |
| 109 | +\end{itemize} |
| 110 | + |
| 111 | +You are encouraged to explore both: |
| 112 | + |
| 113 | +\begin{itemize} |
| 114 | +\item \textbf{Static data} (images, tabular) |
| 115 | +\item \textbf{Temporal data} (time series, sequences) |
| 116 | +\end{itemize} |
| 117 | + |
| 118 | +\section{Methodology} |
| 119 | + |
| 120 | +You should: |
| 121 | + |
| 122 | +\begin{itemize} |
| 123 | +\item Clearly define model architectures |
| 124 | +\item Perform hyperparameter tuning |
| 125 | +\item Use appropriate training procedures (e.g.\ Adam optimizer) |
| 126 | +\item Ensure reproducibility of results |
| 127 | +\end{itemize} |
| 128 | + |
| 129 | +The report should include: |
| 130 | + |
| 131 | +\begin{itemize} |
| 132 | +\item Mathematical formulation of models |
| 133 | +\item Description of training procedures |
| 134 | +\item Analysis of results |
| 135 | +\item Visualization of learned representations |
| 136 | +\end{itemize} |
| 137 | + |
| 138 | + |
| 139 | +\section{Possible Extensions} |
| 140 | + |
| 141 | +If you wish to go beyond the above you may explore: |
| 142 | + |
| 143 | +\begin{itemize} |
| 144 | +\item Diffusion models for data generation |
| 145 | +\item Energy-based models and connections to statistical physics |
| 146 | +\item Bayesian interpretations of generative models |
| 147 | +\item Application to scientific data (e.g.\ physics simulations) |
| 148 | +\end{itemize} |
| 149 | + |
| 150 | +\section{Conclusion} |
| 151 | + |
| 152 | +This project provides an opportunity to bridge two central paradigms |
| 153 | +in machine learning. By combining discriminative and generative |
| 154 | +approaches, the hope is that you will gain a deeper understanding of both |
| 155 | +predictive modeling and data representation, preparing them for |
| 156 | +research and applications in modern machine learning. |
| 157 | + |
| 158 | +\section{Introduction to numerical projects} |
| 159 | + |
| 160 | + |
| 161 | +Here follows a brief recipe and recommendation on how to write a report for the |
| 162 | +project. |
| 163 | + |
| 164 | +\begin{itemize} |
| 165 | + \item Give a short description of the nature of the problem and the eventual numerical methods you have used. |
| 166 | + |
| 167 | + \item Describe the algorithm you have used and/or developed. Here you may find it convenient to use pseudocoding. In many cases you can describe the algorithm in the program itself. |
| 168 | + |
| 169 | + \item Include the source code of your program. Comment your program properly. |
| 170 | + |
| 171 | + \item If possible, try to find analytic solutions, or known limits in order to test your program when developing the code. |
| 172 | + |
| 173 | + \item Include your results either in figure form or in a table. Remember to label your results. All tables and figures should have relevant captions and labels on the axes. |
| 174 | + |
| 175 | + \item Try to evaluate the reliabilty and numerical stability/precision of your results. If possible, include a qualitative and/or quantitative discussion of the numerical stability, eventual loss of precision etc. |
| 176 | + |
| 177 | + \item Try to give an interpretation of you results in your answers to the problems. |
| 178 | + |
| 179 | + \item Critique: if possible include your comments and reflections about the exercise, whether you felt you learnt something, ideas for improvements and other thoughts you've made when solving the exercise. We wish to keep this course at the interactive level and your comments can help us improve it. |
| 180 | + |
| 181 | + \item Try to establish a practice where you log your work at the computerlab. You may find such a logbook very handy at later stages in your work, especially when you don't properly remember what a previous test version of your program did. Here you could also record the time spent on solving the exercise, various algorithms you may have tested or other topics which you feel worthy of mentioning. |
| 182 | +\end{itemize} |
| 183 | + |
| 184 | +\noindent |
| 185 | +\subsection*{Format for electronic delivery of report and programs} |
| 186 | + |
| 187 | +The preferred format for the report is a PDF file. You can also use DOC or postscript formats or as an ipython notebook file. As programming language we prefer that you choose between C/C++, Fortran2008 or Python. The following prescription should be followed when preparing the report: |
| 188 | + |
| 189 | +\begin{itemize} |
| 190 | + \item Send us an email in order to hand in your projects with a link to your GitHub/Gitlab repository. |
| 191 | + |
| 192 | + \item In your GitHub/GitLab or similar repository, please include a folder which contains selected results. These can be in the form of output from your code for a selected set of runs and input parameters. |
| 193 | +\end{itemize} |
| 194 | + |
| 195 | +\noindent |
| 196 | +Finally, |
| 197 | +we encourage you to collaborate. Optimal working groups consist of |
| 198 | +2-3 students. You can then hand in a common report. |
| 199 | + |
| 200 | + |
| 201 | + |
| 202 | + |
| 203 | + |
| 204 | +\end{document} |
| 205 | + |
| 206 | + |
| 207 | + |
| 208 | +% ------------------- end of main content --------------- |
| 209 | + |
| 210 | +\end{document} |
| 211 | + |
0 commit comments