<stdin> |

My Thoughts, Trials and Adventures

MusicGenreClassification - Combining multiple features to make a better model

Posted at — Oct 31, 2020 | Last Modified on — May 11, 2023


This post is about the extension of my work in one of my projects MusicGenreClassification - a collection of methods to classify given audio clips into any of the 10 categories of GTZAN dataset that it was trained on.


in MusicGenreClassification_FeatureEnsemble.ipynb, i combined multiple features of a audio clip to make a dataset. refer to librosa documentaion for more details.

running a Dense Network on this data, this proved to be the best model giving 5-7% more accurate results compared to previous models based on MEL.the best loss achieved with this model is 0.9987499713897705

The accuracy graph of train and validation sets:

accuracy graph