@@ -76,17 +76,20 @@ The aggregation happens separately for all positive predictions (i.e., $p_c^{m_i
7676($p_c^{m_i} < 0.5$). If the aggregated value is larger for the positive predictions than for the negative predictions,
7777the ensemble makes a positive prediction for class $c$:
7878
79+ <img width =" 2297 " height =" 114 " alt =" image " src =" https://github.com/user-attachments/assets/2f0263ae-83ac-41ea-938a-c71b46082c22 " />
80+ <!-- For some reason, this formula does not render in GitHub markdown. Therefore, I rendered it locally and added it as an image. The rendered formula is:
7981$$
8082\text{ensemble}(c) = \begin{cases}
81831 & \text{if } \sum_{i: p_c^{m_i} \geq 0.5} [\text{confidence}_c^{m_i} \cdot \text{model_weight}_{m_i} \cdot \text{trust}_c^{m_i}] > \sum_{i: p_c^{m_i} < 0.5} [\text{confidence}_c^{m_i} \cdot \text{model_weight}_{m_i} \cdot \text{trust}_c^{m_i}] \\
82840 & \text{otherwise}
8385\end{cases}
8486$$
87+ -->
8588
8689Here, confidence is the model's (self-reported) confidence in its prediction, calculated as
87- $$
90+ $
8891\text{confidence}_ c^{m_i} = 2|p_c^{m_i} - 0.5|
89- $$
92+ $
9093For example, if a model makes a positive prediction with $p_c^{m_i} = 0.55$, the confidence is $2|0.55 - 0.5| = 0.1$.
9194One could say that the model is not very confident in its prediction and very close to switching to a negative prediction.
9295If another model is very sure about its negative prediction with $p_c^{m_j} = 0.1$, the confidence is $2|0.1 - 0.5| = 0.8$.
@@ -114,4 +117,4 @@ both, we select one of them randomly and set the other to 0.
114117- (3) Since the second step might have introduced new inconsistencies into the hierarchy, we repeat the first step, but
115118with a small change. For a pair of classes $A \subseteq B$ with predictions $1$ and $0$, instead of setting $B$ to $1$,
116119we now set $A$ to $0$. This has the advantage that we cannot introduce new disjointness-inconsistencies and don't have
117- to repeat step 2.
120+ to repeat step 2.
0 commit comments