-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Questions on Usage (Output of Acquisition Function Values, Design Methods for Search Space) #333
Comments
You could take the candidates after their generation and and fed them into the
Regarding your second question: I am not sure that I completely understand your question, if you have hand designed descriptors, you could setup the search space using Maybe also this tutorial notebook could be helpful for you: https://github.com/experimental-design/bofire/blob/main/tutorials/benchmarks/009-Bayesian_optimization_over_molecules.ipynb cc: @simonsung06 |
@jduerholt Also, thank you for the information about CategoricalDescriptorInput and CategoricalMolecularInput. I apologize for the additional questions, but I have two points of inquiry regarding the use of the above:
|
Nothing to apologize! Regarding your questions:
Best, Johannes |
Thank you very much once again. Regarding categorical descriptors, I managed to resolve it by creating a class that decomposes them when using Cross-validation. Thank you for your support! As for the mixed rules, there remains discussion in the Baysian optimization community, but for now, creating composite descriptors and implementing recursive feature selection might be better. The support index that includes mixed instant and recursive feature selection has been very helpful. I'm thinking of developing a method to generate simple composite variable combinations from categorical descriptors. I have an additional question: is it possible to further classify the descriptors registered under categorical descriptors into continuous, discrete, and categorical? Also, I understand that when experimental data is added via 'tell' in a strategy, it undergoes training. Is it possible to set up cross-validation or LOO (Leave-One-Out). Thank you very much for your assistance. |
regarding the class that you wrote for cross validation, can you share it? I would be interested. Regarding the descriptors, currently we support there only ordina ones (meaning continuous and discrete) and there is no furhter classification. But of course you can setup categorical molecular features and use for example mordred descriptors on the fly ... Regarding bofire/bofire/surrogates/trainable.py Line 57 in 33a2053
Within the BO loop, you can define
This works for every surrogate that implements a so calles
Was this helpful? Best, Johannes |
@jduerholt
Subsequently, the Input_features are updated, and the ExperimentData is also updated. input_features = Inputs(features = decomposition_input_features(Molecule))
Thank you also for your advice regarding cross-validation. Adopting your second suggestion made the process simpler |
Regarding your last line, you also have to set the The helper function regarding the cross val is smart! Good idea. |
Thank you always
Is there a way to obtain the evaluation value of the acquisition function, not just the mean and variance of the predicted results obtained with the .ask function? I want to use it as a clue to prioritize candidates.
Also, if the search space is, for instance, the types of molecules of reagents that can be purchased, and the descriptors generated from the molecular structure of the reagents are the features, the combination of features is fixed. In such a case, how should I design the search space using bofire?"
The text was updated successfully, but these errors were encountered: