Cross validation of Pipeline/estimators using MLDataset / xarray.Dataset#221
Cross validation of Pipeline/estimators using MLDataset / xarray.Dataset#221PeterDSteinberg wants to merge 27 commits intomasterfrom
Conversation
|
Currently status of tests (for a simple Pipeline of only one unsupervised estimator step) - these are mostly failing due to test harness not putting together all the requisite arguments for the cross validators (such as not giving it a grouping variable): |
|
I'm going to add more tests using |
| for cls in get_module_classes(m).values(): | ||
| if cls.__name__ in _seen: | ||
| continue | ||
| if not m in cls.__module__: |
There was a problem hiding this comment.
This is just checking that we are getting StandardScaler or similar from the sklearn module where it is actually defined, not some other one where it is imported for internal usage.
…ther methods are wrapped
| import pytest | ||
|
|
||
|
|
||
| def new_pipeline(*args, flatten_first=True): |
There was a problem hiding this comment.
This was not Python 2.7 compatible.
| _from_numpy_arrs = _from_numpy_arrs | ||
|
|
||
| def _call_sk_method(self, sk_method, X=None, y=None, **kw): | ||
| def _call_sk_method(self, sk_method, X=None, y=None, do_split=True, **kw): |
There was a problem hiding this comment.
I am currently working on simplifying this function - checking what is actually needed.
|
Added the several label encoding classes to The new test module When hyperparameterizing a The problem above with can be changed to: I'll open issues and link them here:
I'm running this PR with: To run the tests: Test summary |
|
Notes:
|
|
Replaced by #228 |
Work in progress to fix #204