-
Notifications
You must be signed in to change notification settings - Fork 12
[DOC] Very high dim #425
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
[DOC] Very high dim #425
Conversation
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files@@ Coverage Diff @@
## main #425 +/- ##
=======================================
Coverage 99.28% 99.28%
=======================================
Files 23 23
Lines 1253 1253
=======================================
Hits 1244 1244
Misses 9 9 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
3eab43f to
ebfc3b3
Compare
jpaillard
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Would it make sense to add a figure from an example, like the fMRI or the 2D simulation?
|
|
||
| # compute estimated support | ||
| selected_cdl = np.logical_or(pval_corr < alpha, one_minus_pval_corr < alpha) | ||
| print(f'Clustered Desparsified Lasso selected {np.sum(selected_cdl)} features among {np.sum(beta > 0)} ') |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This code is not executed in the rendered doc. Is there a way to run the code snippets?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
For the user guide, I don't think that it's a good idea to have code.
In my opinion, it's better to have an example associated with it and make reference to it.
|
|
||
| As hinted in :footcite:t:`meinshausen2009pvalues` an efficient way to deal with such configuration is to take the per-group average of the features: this leads to a *reduced design*. After inference, all the feature in a given group obtain the p-value of the group representative. When the inference engine is Desparsified Lasso, the resulting method is called Clustered Desparsified lasso, or **CluDL**. | ||
|
|
||
| Using the same example as previously, we start by defining a clustering method that will perform the grouping. For image data, Ward clustering is a good default model:: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should we add a bit more details explaining that the clustering is constrained by pixel's neighborhood?
docs/src/high_dimension.rst
Outdated
| * The :ref:`Grouping documentation <grouping>` | ||
|
|
||
|
|
||
| As hinted in :footcite:t:`meinshausen2009pvalues` an efficient way to deal with such configuration is to take the per-group average of the features: this leads to a *reduced design*. After inference, all the feature in a given group obtain the p-value of the group representative. When the inference engine is Desparsified Lasso, the resulting method is called Clustered Desparsified lasso, or **CluDL**. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| As hinted in :footcite:t:`meinshausen2009pvalues` an efficient way to deal with such configuration is to take the per-group average of the features: this leads to a *reduced design*. After inference, all the feature in a given group obtain the p-value of the group representative. When the inference engine is Desparsified Lasso, the resulting method is called Clustered Desparsified lasso, or **CluDL**. | |
| As hinted in :footcite:t:`meinshausen2009pvalues` an efficient way to deal with such configuration is to take the per-group average of the features: this leads to a *reduced design*. After inference, all the feature in a given group obtain the p-value of the group representative. When the inference engine is Desparsified Lasso, the resulting method is called Clustered Desparsified lasso, or :py:func:`CluDL <hidimstat.clustered_inference>`. |
I think this links to the API
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There are other models to cite it in this section; it should be interesting to add the link, for example Desparsified Lasso and EnCluDL.
lionelkusch
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you try to stick to the maximum length of 80 characters?
Co-authored-by: Joseph Paillard <[email protected]>
Co-authored-by: Joseph Paillard <[email protected]>
Co-authored-by: Joseph Paillard <[email protected]>
Co-authored-by: Joseph Paillard <[email protected]>
Co-authored-by: Joseph Paillard <[email protected]>
Co-authored-by: Joseph Paillard <[email protected]>
One setp toward #306 : populating the high dimension documentation module.