Synthetic domains for transfer lerning #634
Replies: 3 comments 7 replies
-
|
Would this also include cases where the parameter space is not the same between source and target? E.g. Parameters (p1, p2, p3, p4, p5) contribute to source outcome and (p1, p2, p6, p7) to target outcome? If this is of interest, I am happy to discuss more. |
Beta Was this translation helpful? Give feedback.
-
|
Hi @Hrovatin, Thank you for compiling this list! Are we focused on identifying failure cases for our models, even in unrealistic scenarios, or are we aiming to evaluate their robustness against simpler cases? Specifically, what would "failure" mean for us? Is it converging to the wrong optimum, or to the correct one but slower than a non-transfer learning approach? For now, I would focus on scenarios without constraints. Since we are working with synthetic benchmarks, we could start with simple, low-dimensional synthetic functions to identify failure cases. From there, we can think about what we might encounter in real world usecases and derive higher dimensional examples. I would be most interested in the following relations: Shifted Outcomes: I would expect this to be challenging for hierarchical and additive approaches. I'm particularly interested in how the Inverted Source: I would like to see if the Different Optima Ranking: We could explore this using the Forrester function, which is a well-known benchmark for multifidelity Bayesian optimization. Second priority would be: Noisy Source: While it should be manageable with ample source data, I suggest we test this on a simple 1D example for completeness. Wider Source Optima: While this may slow down exploitation, it shouldn't lead us to converge on the wrong optimum. The same reasoning applies to the different outcome scale. If we're comfortable starting with 1D examples first, I can develop a new Forrester benchmark that addresses the above cases. |
Beta Was this translation helpful? Give feedback.
-
|
Comment from the team: Maybe not all of these differences are equally severe for BO outcome. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Collection of source-target relationships to consider when designing synthetic TL domains for benchmarking.
Focus of this discussion is on different types of unreliable source information, rather than other types of missmatch between source and target (such as non-overlapping parameter spaces suggested by Sara W. below).
Types of useful transfer information
Below is collection of different types of useful and harmful (inconsistent) information between source and target. They are not mutually exclusive. The plots show parameter on x-axis, outcome on y-axis, target in red, and source in black.
Some of the inconsistency types may become more relevant in multi-parameter setting with parameter interactions or constraints. But these can be often explained with other types of inconsistencies if the searchspace is regarded as a whole and not parameter wise, as explained by the examples below. For multi-parameter setting plots optima is colored yellow while worse outcome is blue-violet and two parameters are on axes.
All shown examples focus on optima being minimum.
Noisy source
Source follows same underlying function but is noisier. This should be easy to optimize if lots of source data is present - the main challenge occurs if little source data is present as then the learned relationships will be more biased by the noise effects.
Shifted outcomes
Modelling this would be easy if there was option to add interaction term between task parameter and the non-task parameter (x-axis on the plot).
Inverted source
This is what the current TaskKernel was designed to account for.
Different optima ranking
This may lead to prolonged exploration of optima more prominent in source/missing global target optima.
Partially shared optima
Some local optima are shared, but there are also some source or target specific local optima.
This could be regarded as special case of different optima ranking.
Missing target optima
While it would be hard to find global source optima, it may still help to find local optima due to most of the parameter space having similar outcome.
This could be regarded as special case of different optima ranking.
Wider source optima
This may be problematic in the case of flat and wide source optima.
Narrower source optima
This may be less interesting as target optima can then be easily found.
However, this may become relevant in multi-parameter setting with constraints, where having wider optima may lead to different regions of the allowed space being selected. But this can be probably simplified into missing target optima case by regarding only the allowed parameter space region.
Different outcome scale
Would not expect to have negative effect on TL.
However, this may again become relevant in multi-paramater setting where optima of two parameters are mutually exclusive due to some negative interaction. Depending on the contribution of each parameter to the outcome a different optima may be selected. The effect of this could be also explained with other inconsistency types (such as too wide optima if the first function on below plots is source and the second is target).
Y = - (X1 + X2 - X1*X2)

Stronger contribution of X1 while X2 and the interaction term contribute equally
Y = - (2*X1 + X2 - X1*X2)

Parameters with useful information
For some of the inconsistencies it may be relevant to be only present in part of parameter range (e.g. noise only at boundary).
In multi-feature (parameter) settings, it may happen that:
Beta Was this translation helpful? Give feedback.
All reactions