PyRACER is an unofficial Python implementation of the RACER classification algorithm described by Basiri et. al, 2019. RACER is designed specifically for discrete datasets and therefore uses the entropy-based MDLP discretization algorithm by Fayyad and Irani, 1993 for binary tasks and an optimal binning strategy for the multiclass case. The code is also heavily documented for ease of use.
Please consider citing this work if you intend to use it in an academic setting.
A new release will be made available on PyPI every time new features are added or bugs are fixed so you can simply use pip
to install the package:
$ pip install pyracer
If you would like to develop the package for your own use case however, you may clone this repository and then simply install the requirements. Reading the documentation prior to this is strongly advised, however, as you may find native support for your specific task in the private methods already available.
$ git clone https://github.com/Adversarian/RACER/
$ cd RACER
$ pip install -r requirements-dev.txt
Otherwise, you may also try to monkey-patch the class in your own code if you find that solution more appealing.
PyRACER is designed to be consistent with Scikit-learn estimator API which makes it very easy to use.
The following example demonstrates the use of RACER on the Zoo dataset. Take a look at examples for more use cases.
from RACER import RACER, RACERPreprocessor
from sklearn.model_selection import train_test_split
import pandas as pd
# dataset from https://archive.ics.uci.edu/ml/machine-learning-databases/zoo/
df = pd.read_csv(
"datasets/zoo.data",
names=[
"animal_name",
"hair",
"feathers",
"eggs",
"milk",
"airborne",
"aquatic",
"predator",
"toothed",
"backbone",
"breathes",
"venomous",
"fins",
"legs",
"tail",
"domestic",
"catsize",
"type",
],
)
X = df.drop(columns=['animal_name', 'type']).astype('category')
Y = df[['type']].astype('category')
RACER requires a preprocessing step to be performed on the data prior to splitting into test and train portions. This step discretizes continous features and then converts each feature into a dummy encoded variable. Note that since different discretization methods are used for multiclass and binary classification tasks you need to either specify the task using the target
keyword argument or leave it to default to "auto"
which attempts to infer your task when you call fit_transform(X,y)
from the number of unique values in y
.
RACERPreprocessor now also supports separate fit
and transform
functions but it is still recommended to use fit_transform
or perform fit
on the entire dataset prior to splitting. This ensures that new unseen values are not left out of the transformation at test time.
X, Y = RACERPreprocessor(target="multiclass").fit_transform(X, Y)
X_train, X_test, Y_train, Y_test = train_test_split(X,Y, random_state=1, test_size=0.3)
RACER provides a benchmark
keyword argument that can be used to time the fit
method. Moreover, the hyperparameter alpha
can be set using its respective keyword argument. (Note that beta
is uniquely determined as 1.0 - alpha
and is therefore not exposed through a keyword argument)
racer = RACER(alpha=0.95, benchmark=True)
racer.fit(X_train, Y_train)
Now you may access the public methods available within the racer
object such as score
and display_rules
. For example:
>>> racer.score(X_test, Y_test)
... 0.8709677419354839
>>> racer.display_rules()
... Algorithm Parameters:
... - Alpha: 0.95
... - Time to fit: 0.008133015999987947s
...
... Final Rules (8 total): (if --> then (label) | fitness)
... [111011011111111101011011111000111111] --> [1000000] (0) | 0.9685714285714285
... [100101101111111001011010010000011111] --> [0100000] (1) | 0.9607142857142856
... [101001101001110101101101100000011111] --> [0001000] (3) | 0.9571428571428571
... [101011101011011010111110101011111011] --> [0000001] (6) | 0.9542857142857143
... [111001101110101010011110000010101010] --> [0000010] (5) | 0.9535714285714285
... [101011101011111101111110101000011011] --> [0010000] (2) | 0.9528571428571428
... [101001101001110101011110001000101010] --> [0000100] (4) | 0.9521428571428571
... [101001101010101010011010100000101010] --> [0000001] (6) | 0.9507142857142856
Add another example notebook featuring Scikit-learn's built-in datasets.Replacepandas.get_dummies()
with Scikit-learn'sOneHotEncoder
for better consistency.- Unify discretization algorithms for all tasks.
- Better docs!
Found a problem within the implementation or an inconsistency with the original algorithm? Or maybe you would like to request a feature? Please feel free to submit a PR or create a new issue.
@Article{Basiri2019,
author="Basiri, Javad
and Taghiyareh, Fattaneh
and Faili, Heshaam",
title="RACER: accurate and efficient classification based on rule aggregation approach",
journal="Neural Computing and Applications",
year="2019",
month="Mar",
day="01",
volume="31",
number="3",
pages="895--908",
issn="1433-3058",
doi="10.1007/s00521-017-3117-2",
url="https://doi.org/10.1007/s00521-017-3117-2"
}