You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/leaderboard.md
+51-56Lines changed: 51 additions & 56 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,8 +2,6 @@
2
2
3
3
# Tasks
4
4
5
-
There are three tasks within the MillionTrees package.
6
-
7
5
## Zero-shot
8
6
9
7
The first task is to create a zero-shot detection system to generalize across geography and aquisition conditions. Selected datasets are held out from training completely and used for evaluation in new conditions. This is a challenging task with no local training data.
@@ -16,7 +14,6 @@ The second task is to create the best global detector for individual trees given
16
14
17
15
Off the shelf tools often limit users for a single annotation type. We have 'point' models, 'box' models and 'polygon' models. To create truly global models for biological inference, we need models that can use all available data, not just one annotation geometry. In particular, polygon annotations are very time consuming to create, but are often desirable for downstream usecases. We opted against polygon training sources, for example polygons to points, as this is an unrealistic, or atleast, very uncommon downstream use case.
18
16
19
-
20
17
### Boxes to Polygons
21
18
22
19
All box sources are used to train and predict all polygon sources. There is no local data from the test localities in train.
@@ -29,69 +26,27 @@ All point sources are used to train and predict all polygon sources
29
26
30
27
All point sources are used to train and predict all box sources.
31
28
32
-
33
-
# Submissions
34
-
35
-
## Submit to the leaderboard
36
-
37
-
Once you have trained a model and evaluated its performance, you can submit your results to the MillionTrees leaderboard. Here's how:
38
-
39
-
1. Create a public repository with your code and model training scripts. Make sure to include:
40
-
- Clear instructions for reproducing your results
41
-
- Requirements file listing all dependencies
42
-
- Training configuration files/parameters
43
-
- Code for data preprocessing and augmentation
44
-
- Model architecture definition
45
-
- Evaluation code
46
-
47
-
2. Generate predictions on the test split:
48
-
```python
49
-
test_dataset = dataset.get_subset("test") # Use test split
0 commit comments