Like to train an SVR to combine the predictions of the top 10% MLPRegressors using the same data that were used for training of the MLPRegressors? Wouldn't that lead to overfitting?
On 7 January 2017 at 21:20, Sebastian Raschka <se.raschka@gmail.com> wrote:Hi, Thomas,
sorry, I overread the regression part …
This would be a bit trickier, I am not sure what a good strategy for averaging regression outputs would be. However, if you just want to compute the average, you could do sth like
np.mean(np.asarray([r.predict(X) for r in list_or_your_mlps]))
However, it may be better to use stacking, and use the output of r.predict(X) as meta features to train a model based on these?Like to train an SVR to combine the predictions of the top 10% MLPRegressors using the same data that were used for training of the MLPRegressors? Wouldn't that lead to overfitting?
Best,
Sebastian
> On Jan 7, 2017, at 1:49 PM, Thomas Evangelidis <tevang3@gmail.com> wrote:
>
> Hi Sebastian,
>
> Thanks, I will try it in another classification problem I have. However, this time I am using regressors not classifiers.
>
> On Jan 7, 2017 19:28, "Sebastian Raschka" <se.raschka@gmail.com> wrote:
> Hi, Thomas,
>
> the VotingClassifier can combine different models per majority voting amongst their predictions. Unfortunately, it refits the classifiers though (after cloning them). I think we implemented it this way to make it compatible to GridSearch and so forth. However, I have a version of the estimator that you can initialize with “refit=False” to avoid refitting if it helps. http://rasbt.github.io/mlxtend/user_guide/classifier/ EnsembleVoteClassifier/# example-5-using-pre-fitted- classifiers
>
> Best,
> Sebastian
>
>
>
> > On Jan 7, 2017, at 11:15 AM, Thomas Evangelidis <tevang3@gmail.com> wrote:
> >
> > Greetings,
> >
> > I have trained many MLPRegressors using different random_state value and estimated the R^2 using cross-validation. Now I want to combine the top 10% of them in how to get more accurate predictions. Is there a meta-estimator that can get as input a few precomputed MLPRegressors and give consensus predictions? Can the BaggingRegressor do this job using MLPRegressors as input?
> >
> > Thanks in advance for any hint.
> > Thomas
> >
> >
> > --
> > ============================================================ ==========
> > Thomas Evangelidis
> > Research Specialist
> > CEITEC - Central European Institute of Technology
> > Masaryk University
> > Kamenice 5/A35/1S081,
> > 62500 Brno, Czech Republic
> >
> > email: tevang@pharm.uoa.gr
> > tevang3@gmail.com
> >
> > website: https://sites.google.com/site/thomasevangelidishomepage/
> >
> >
> > _______________________________________________
> > scikit-learn mailing list
> > scikit-learn@python.org
> > https://mail.python.org/mailman/listinfo/scikit-learn
>
> _______________________________________________
> scikit-learn mailing list
> scikit-learn@python.org
> https://mail.python.org/mailman/listinfo/scikit-learn
> _______________________________________________
> scikit-learn mailing list
> scikit-learn@python.org
> https://mail.python.org/mailman/listinfo/scikit-learn
_______________________________________________
scikit-learn mailing list
scikit-learn@python.org
https://mail.python.org/mailman/listinfo/scikit-learn
--======================================================================
Thomas Evangelidis
Research Specialist
CEITEC - Central European Institute of Technology
Masaryk University
Kamenice 5/A35/1S081,
62500 Brno, Czech Republic
_______________________________________________
scikit-learn mailing list
scikit-learn@python.org
https://mail.python.org/mailman/listinfo/scikit-learn