Official repository of the paper "$f$-Divergence Based Classification: Beyond the Use of Cross-Entropy " published at ICML 2024.
Novel class of objective functions for supervised classification problems, based on the
$f$ -divergence. We estimate the posterior probability (continuous or discrete) and maximize it to solve classification tasks.
AWGN | PAM Attenuation | PAM Triangular |
---|---|---|
For the image classification tasks, the file main.py
runs the experiments. The code runs iterating over multiple random seeds, network architectures and objective functions. They can be set by modifying the lists:
list_cost_func_v = [5]
random_seeds = [0]
net_architectures = ["ResNet18"]
dataset_type = "cifar10"
where the IDs of the objective functions are:
- 2: GAN
- 3: CE
- 5: SL
- 7: KL with softplus as last activation function
- 9: RKL
- 10: HD
- 12: P
while the available network architectures are:
- ResNet18
- PreActResNet18
- MobileNetV2
- VGG
- SimpleDLA
- DenseNet121
For the decoding tasks, the file main_communications.py
runs the experiments.
If you use the code for your research, please cite our paper:
@InProceedings{pmlr-v235-novello24a,
title = {$f$-Divergence Based Classification: Beyond the Use of Cross-Entropy},
author = {Novello, Nicola and Tonello, Andrea M},
booktitle = {Proceedings of the 41st International Conference on Machine Learning},
pages = {38448--38473},
year = {2024},
volume = {235},
series = {Proceedings of Machine Learning Research},
month = {21--27 Jul},
publisher = {PMLR},
}
The implementation is based on / inspired by:
- https://github.com/kuangliu/pytorch-cifar
- https://github.com/tonellolab/MIND-neural-decoder