Skip to content

chunchun1116/ml-boosting

 
 

Repository files navigation

Boosting Classifier

This classifier was part of a project for my Fall 2011 machine learning class.

To run, put spambase.data from the UCI spambase dataset into the same directory as homework4.py.

$ python2 homework4.py --beststump

This will create decision stump for each feature & datapoint in folds 2..10 and boost with the optimal one until a convergence criterion is met. Data for a ROC curve is written on exit.

Options:

homework4.py
    --folds    NUM    #  How many folds to make from the dataset (default: 10).
    --testfold NUM    #  Which fold to use for testing           (default: 0).
    --rounds   NUM    #  How many rounds to run                  (default: until convergence).
    --beststump       #  Choose the optimal decision stump       (default: random choice).

See my analysis for a discussion of the results.

-- PLR


Requires Python 2.7

About

Adaboost of trivial classifiers for the UCI spambase dataset.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%