In the context of high throughput RNA sequencing, we aim at solving the inverse problem of subtype mixing within a sample.
We formulate this problem as a matrix factorization problem, where the data matrix is factorized into two matrices, one representing the subtypes and the other the mixing proportions.
Let
We assume that we have a matrix
$W^,H^ = \arg\min_{H \in \mathbb{R}^{n \times K}_{+} ; W \in \mathbb{R}^{m \times K}} \text{dst}_X(WH) + \lambda g_M(W)$
where we have:
To do so we use alterned gradient descent.
For more details and experiments, please refer to the notebook.