Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

WIP: Added implementation of feature_permutation #91

Closed
wants to merge 1 commit into from

Conversation

andersbogsnes
Copy link
Owner

Working on issue #59

@codecov
Copy link

codecov bot commented Oct 29, 2018

Codecov Report

Merging #91 into develop will decrease coverage by 2.09%.
The diff coverage is 14.28%.

Impacted file tree graph

@@            Coverage Diff             @@
##           develop      #91     +/-   ##
==========================================
- Coverage    97.41%   95.31%   -2.1%     
==========================================
  Files            7        7             
  Lines          541      555     +14     
  Branches        69       71      +2     
==========================================
+ Hits           527      529      +2     
- Misses           7       19     +12     
  Partials         7        7
Impacted Files Coverage Δ
src/ml_tooling/metrics.py 77.35% <14.28%> (-22.65%) ⬇️

Continue to review full report at Codecov.

Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 3653b8b...a81de62. Read the comment docs.

@thomasfrederikhoeck
Copy link
Contributor

thomasfrederikhoeck commented Oct 31, 2018

This implementation can be multi-threaded. I can give it at shot at a later point.

Leaving this link for later:

TeamHG-Memex/eli5#244

@thomasfrederikhoeck
Copy link
Contributor

As this is similar to ELI5's PermutationImportance I suggest the following features that it has:

  • Multiple permutation of the same feature to improve estimate of importance.

  • Option for Cross-Calidation if model is refitted. This can be used to show how important the features are for generalization.

metric=None):
"""
Calculates feature importance by randomly permuting features and comparing result to baseline
:param y:
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Isn't y training/test target?

Copy link
Owner Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yeah, changed the implementation without updating docs - still have some work to do on this feature 😄

@andersbogsnes
Copy link
Owner Author

I'll have a look at their implementation - see if there's some cleverness we can use

@andersbogsnes andersbogsnes changed the title Added implementation of feature_permutation WIP: Added implementation of feature_permutation Oct 31, 2018
@andersbogsnes andersbogsnes deleted the feature_permutation branch March 30, 2019 17:30
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants