-
Notifications
You must be signed in to change notification settings - Fork 156
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add fit_transform methods? #546
Comments
Thanks for that. Yes, sk-learn has this, and so you are presumably not the first to wonder about this. Users can sometimes get confused when there are lot of methods and many ways to do the same thing. In balancing convenience with simplicity, I feel some reluctance to do this. I wonder what others think of the suggestion? |
I kind of like the approach of overloading
|
If we are going to do this, then I agree this might be the best option. It seem the method table would allow this. Still, like to ponder this a bit more before committing. |
I do not see the motivation for removing the fit with As @CameronBieganek mentions sklearn has the option to
I would argue otherwise. When you are doing Cross-Validation fitting a preprocess can be very handy. For example, one of feature values of a categorical variable might rarely appear. It might happen that the folds in which I do not have any counter argument for having a |
Sometimes it's handy to have a combined
fit_transform
method. For example, right now I'm doing my one-hot encoding once at the beginning of my analysis (not in a pipeline). This is safe to do, since it's a static transformer that doesn't learn from the data. So for this use case, it would be cool if I could do something like this:Or perhaps it would make sense to just overload
transform
?The text was updated successfully, but these errors were encountered: