Supervised Data Stream Classification on dionis
Supervised Data Stream Classification on dionis
Issue
#Downvotes for this reason
By
Metric:
area under roc curve
average cost
binominal test
build cpu time
build memory
c index
chi-squared
class complexity
class complexity gain
confusion matrix
correlation coefficient
cortana quality
coverage
f measure
information gain
jaccard
kappa
kb relative information score
kohavi wolpert bias squared
kohavi wolpert error
kohavi wolpert sigma squared
kohavi wolpert variance
kononenko bratko information score
matthews correlation coefficient
mean absolute error
mean class complexity
mean class complexity gain
mean f measure
mean kononenko bratko information score
mean precision
mean prior absolute error
mean prior class complexity
mean recall
mean weighted area under roc curve
mean weighted f measure
mean weighted precision
weighted recall
number of instances
os information
positives
precision
predictive accuracy
prior class complexity
prior entropy
probability
quality
ram hours
recall
relative absolute error
root mean prior squared error
root mean squared error
root relative squared error
run cpu time
run memory
run virtual memory
scimark benchmark
single point area under roc curve
total cost
unclassified instance count
usercpu time millis
usercpu time millis testing
usercpu time millis training
webb bias
webb error
webb variance
joint entropy
pattern team auroc10
wall clock time millis
wall clock time millis training
wall clock time millis testing
unweighted recall
0 Runs
Metric:
area under roc curve
average cost
binominal test
build cpu time
build memory
c index
chi-squared
class complexity
class complexity gain
confusion matrix
correlation coefficient
cortana quality
coverage
f measure
information gain
jaccard
kappa
kb relative information score
kohavi wolpert bias squared
kohavi wolpert error
kohavi wolpert sigma squared
kohavi wolpert variance
kononenko bratko information score
matthews correlation coefficient
mean absolute error
mean class complexity
mean class complexity gain
mean f measure
mean kononenko bratko information score
mean precision
mean prior absolute error
mean prior class complexity
mean recall
mean weighted area under roc curve
mean weighted f measure
mean weighted precision
weighted recall
number of instances
os information
positives
precision
predictive accuracy
prior class complexity
prior entropy
probability
quality
ram hours
recall
relative absolute error
root mean prior squared error
root mean squared error
root relative squared error
run cpu time
run memory
run virtual memory
scimark benchmark
single point area under roc curve
total cost
unclassified instance count
usercpu time millis
usercpu time millis testing
usercpu time millis training
webb bias
webb error
webb variance
joint entropy
pattern team auroc10
wall clock time millis
wall clock time millis training
wall clock time millis testing
unweighted recall
Timeline
Plotting contribution timeline
Leaderboard
Rank
Name
Top Score
Entries
Highest rank
Note: The leaderboard ignores resubmissions of previous solutions, as well as parameter variations that do not improve performance.
Challenge
Given a dataset with a nominal target, various data samples of increasing size are defined. A model is build for each individual data sample; from this a learning curve can be drawn.
Given inputs
Expected outputs
evaluations
A list of user-defined evaluations of the task as key-value pairs.
KeyValue (optional)
predictions
The desired output format
Predictions (optional)
How to submit runs
Using your favorite machine learning environment
Download this task directly in your environment and automatically upload your results
OpenML bootcamp
From your own software
Use one of our APIs to download data from OpenML and upload your results
OpenML APIs