-
Notifications
You must be signed in to change notification settings - Fork 99
Adding a registry to have the hashes of datasets #1072
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
for more information, see https://pre-commit.ci
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files@@ Coverage Diff @@
## main #1072 +/- ##
==========================================
+ Coverage 65.90% 65.92% +0.02%
==========================================
Files 45 46 +1
Lines 6772 6776 +4
Branches 1138 1138
==========================================
+ Hits 4463 4467 +4
Misses 1875 1875
Partials 434 434
🚀 New features to boost your workflow:
|
|
ok is this something I can do or is it do we need an account etc for it? |
|
We do have an account but I'm a) not sure how to upload datasets there and b) also not sure how to properly download from there. One of the many things we should homogenise across packages. But Lukas and Gregor would know, so waiting for them to chime in |
|
It's pretty simple. We have an AWS account and I can share the credentials for our datasets IAM with you on zulip. Then you create a folder on S3 for squidpy and upload the datasets there. There's a field on the UI for each dataset where you get a raw URL that is publicly exposed so you can just copy it here. If you need more guidance just let us know. Edit: Selman got credentials. |
|
ok to confirm I am trying to download these files atm: I am also going to need to restructure the dataset classes etc. |
|
I will continue this here #1076. I decided starting from scratch for this was more reasonable since I refactored many things for the datasets |
hi this is a continuation of #1069 .