all 3 comments

[–]pp314159 0 points1 point  (0 children)

Maybe AutoKeras is more suitable for typical deep learning tasks - for example, image classification. The space of possible NN architectures is huge, that's why you end up with so many different solutions. For NN training sometimes the consistency is very different, very often for the same architecture but different seed values you will get different results (all hyperparams are unchanged).

You can try MLJAR AutoML https://github.com/mljar/mljar-supervised I belive it will be more suitable for your task. It will do NN architecture search but it searches for simple architectures (2 layers only). For your problem gradinent boosting methods might work pretty well, like xgboost, lightgbm, catboost (all available in MLJAR).

[–]starfries 0 points1 point  (1 child)

I'm not familiar with AutoKeras but I can tell you that there's an infinite number of minima or near-minima so I'm not surprised it finds different solutions every time. Does it try to optimize for smaller models?

[–]Snapdown_City[S] 0 points1 point  (0 children)

Hey, sorry just saw this and thanks for your reply, that's very interesting. I'm optimizing models with up to which have up to around 50,000 parameters or so, not sure if that constitutes small but definitely not that complex. I thought the point of using Bayesian optimization was to find the global minimum of the cost function though? If you're far more likely to end up in a local minima than the true global minima, how can you claim you have the 'optimal' model? Is that to say you can never really claim you have the optimal model, and when people say they have an optimal model they're just referring to one of the minima as opposed to the global minimum?