Mulit-Task Normalization problem #2387
-
Hi I am trying to do a Multi-Fidelity optimization using the MultiTaskGP, but I have issues when normalizing the training data. My training data is x-training data
y-training data:
These are exact observations form the multifidelity forrester function defined in Analytical Benchmark Problems for Multifidelity Optimization Methods. when i do the input transform with the Normalize and outcome_tranform with Standardize the model fit is not correct:
When i plot the results with:
I get this result: but when I omit the input and outcome transform I get a much better fit: Why is it so different? Has it something to do with the strong priors? Thanks to anyone reading my post! I'm grateful for any help. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 2 replies
-
I don't think you should be normalizing the task values. This will raise an exception in newer versions of BoTorch. When I tried to run this, it errored out at
I'm not an expert on multi-task models, so I'll see if I can get back to you on that. |
Beta Was this translation helpful? Give feedback.
I don't think you should be normalizing the task values. This will raise an exception in newer versions of BoTorch. When I tried to run this, it errored out at
pred2 = model.posterior(eval_x2)
withI'm not an expert on multi-task models, so I'll see if I can get back to you on that.