Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What qualities should metrics have? #2

Open
rbiswas4 opened this issue Dec 5, 2017 · 2 comments
Open

What qualities should metrics have? #2

rbiswas4 opened this issue Dec 5, 2017 · 2 comments
Labels

Comments

@rbiswas4
Copy link
Collaborator

rbiswas4 commented Dec 5, 2017

While thinking about classifiction metrics, we have all felt that it would be nice to have metrics that can re
ward better classification methods. To give a concrete example, we would prefer a binary classification method that assigns a probability p, (1-p) for an event to belong to the two classes, rather than a method which chooses to assign only 0, 1 or 1, 0 values. Obviously, it is important that this be that case only if the classifier does this reliably, while it would be preferable to choose a discrete probability classifier if this is not the case. This might be an example of a consistency condition that we need. Can we figure out if we can have such a consistency condition, and ensure that metrics respect that?

@aimalz
Copy link
Owner

aimalz commented Dec 14, 2017

This is exactly what I was hoping to test in the notebooks. Please do write a demo if you can think of a concrete example.

@aimalz
Copy link
Owner

aimalz commented Dec 18, 2017

After a chat with @rbiswas4, I think we can identify some properties of classifiers we'd like our metric to favor and disfavor.

  1. We do not want to encourage classification of only on the most common class(es) by ignoring rarer classes, i.e. all test set objects should not be of equal value.

  2. We may want to reward classifiers that respect hierarchical classes, i.e. there could be a smaller penalty for misclassifying a SN Ib as SN II than for misclassifying a SN Ia as some type of AGN.

  3. We may not want to favor classifiers that perform well only on the most perfect data (brightest/lowest noise/best possible sampling), i.e. a classifier with better performance on lower quality data could be rewarded more than a classifier that only functions on the highest quality data. Conversely, we might want a higher penalty for misclassifying in the presence of higher quality data than for lower quality data, i.e. the penalty for misclassification on lower quality data could be lower than for higher quality data.

These matters should influence the weightings that may need to be used for multi-class metrics. Do folks have more major considerations we want to bake into the metrics?

@aimalz aimalz added the question label Apr 3, 2018
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants