亚洲免费av电影一区二区三区,日韩爱爱视频,51精品视频一区二区三区,91视频爱爱,日韩欧美在线播放视频,中文字幕少妇AV,亚洲电影中文字幕,久久久久亚洲av成人网址,久久综合视频网站,国产在线不卡免费播放

        ?

        Arabic Music Genre Classification Using Deep Convolutional Neural Networks(CNNs)

        2022-11-11 10:47:52LaialiAlmazaydehSalehAtiewiArarAlTawilandKhaledElleithy
        Computers Materials&Continua 2022年9期

        Laiali Almazaydeh,Saleh Atiewi,Arar Al Tawil and Khaled Elleithy

        1Department of Software Engineering,Al-Hussein Bin Talal University,Ma’an,71111,Jordan

        2Department of Computer Science,Al-Hussein Bin Talal University,Ma’an,71111,Jordan

        3King Abdullah II School of Information Technology,The University of Jordan,Amman,Jordan

        4Department of Computer Science and Engineering,University of Bridgeport,Bridgeport,CT,06604,USA

        Abstract: Genres are one of the key features that categorize music based on specific series of patterns.However, the Arabic music content on the web is poorly defined into its genres,making the automatic classification of Arabic audio genres challenging.For this reason, in this research, our objective is first to construct a well-annotated dataset of five of the most well-known Arabic music genres,which are:Eastern Takht,Rai,Muwashshah,the poem,and Mawwal, and finally present a comprehensive empirical comparison of deep Convolutional Neural Networks(CNNs)architectures on Arabic music genres classification.In this work,to utilize CNNs to develop a practical classification system, the audio data is transformed into a visual representation(spectrogram)using Short Time Fast Fourier Transformation (STFT), then several audio features are extracted using Mel Frequency Cepstral Coefficients(MFCC).Performance evaluation of classifiers is measured with the accuracy score,time to build,and Matthew’s correlation coefficient(MCC).The concluded results demonstrated that AlexNet is considered among the topperforming five CNNs classifiers studied: LeNet5, AlexNet, VGG, ResNet-50,and LSTM-CNN,with an overall accuracy of 96%.

        Keywords:CNN;MFCC;spectrogram;STFT;arabic music genres

        1 Introduction

        Music Information Retrieval(MIR)is an interdisciplinary field that aims to extract meaningful information due to the rapid growth of the music volume produced daily.This kind of process is required to have a major benefit in various applications.MIR can be used in copyright monitoring,music management,and Music genre classification.

        Music Genre Classification(MGC)has become one of the most significant MIR techniques as it is vital for digital music platforms such as Tidal,sound cloud,and Apple Music.Automatic approaches for performing MGC can extract meaningful information directly from the audio content.Such information is used to enhance the output of various application systems such as playlist generation,music recommendation,and search optimization.

        Music genre refers to the categorization of music descriptions based on the interaction between culture, artists, and market forces.It assists in sorting music into collections by showing similarities between musicians or compositions[1].Due to the very elusive properties of auditory musical data,it isn’t easy to distinguish between distinct genres[2].

        Several studies proposed MGC methods for the western music genre with comparative results[3-9].However, till the last decade, studies rarely offered such an automatic MGC method of nonwestern music.As mentioned by Downie et al.[10], MGC is considered one of the most significant challenges in the International Society of Music Information Retrieval(ISMIR)to extend its musical horizons to non-western music.

        The Arabic music content on the web is poorly categorized,and its genres are not well defined.Most Arabic music content needs both exact description“Labeling”and classification“genres.”For this reason,we introduce a method for automatic classification between five of the most well-known Arabic music genres:Eastern Takht,Rai,Muwashshah,the poem,and Mawwal.

        The implementation is carried out to classify the five different music genres using Deep learning techniques,explicitly using Convolutional Neural Networks.Deep learning can build Powerful Artificial Intelligence(AI)applications capable of solving extraordinary complex tasks.All AI applications have the same type of neural network architecture,which is Convolutional Neural Networks(CNN).CNN is an extensively used model in image information retrieval applications, and it has a high capacity for extracting relevant characteristics from changes of musical patterns with no prior knowledge[2].

        There have been various CNN architectures[11]since LeNet by LeCun et al.in 1998[12],and the first deep learning network applied in the competition of object recognition with the advancement of GPUs,also the AlexNet network by Krizhevsky et al.[13],and VGGNet[14]are an advanced CNN architecture.

        Notably, the various CNN architectures conduct similarly and have a few cases where they are better.Therefore,this paper presents a comprehensive empirical comparison of CNN architectures on Arabic music genres classification.

        Considering the mentioned below reviewed related works in western MGC,our contribution can be summarized as follows:

        ? Covering the absence of Arabic Genre Music dataset by gathering,structuring,and annotating a large corpus of Arabic audio clips,since most of the literature works towards western genre music using the GTZAN dataset,which although it is considered a benchmark dataset for MGC but according to Strum in[15]it has some shortcomings such as mislabeling,distortions,and replicas.

        ? Presenting a comprehensive systematic empirical comparison of various deep learning models using fine-tuned CNNs architecture.

        ? Evaluating the performance of the various implemented CNNs classifiers on the constructed corpus.

        Developing such an Arabic music genre classifier with many beneficial applications such as search optimization,playlist generation,and music management.

        This paper is organized as follows: Section 2 offers the related research.Section 3 describes the dataset and CNN architectures used.Section 4 demonstrates the experimental results and evaluation.Section 5 summarizes how the research objectives are being achieved and future works.

        2 Related Works

        Compared to rarely proposed methods to classify non-western genre music, most developed methods were primarily oriented toward western MGC.Some popular western music genres are rap,folk,jazz,country,pop,and rock.

        The work in[16]was carried out to distinguish between three different folk music:German,Irish,and Austrian.The dataset was tested and compared using different Hidden Markov Models(HMM)structures to explore statistical differences among the various folk.The classification performance is averaging 77%.However, the work carried out in [17] using Support Vector Machines (SVM)as a statistical ML technique achieved a higher performance rate in MGC than HMM.

        A recent study [18] examined the selection of frequency-domain features and low-level features using a genetic algorithm.Comparative analysis is performed with different classification algorithms,such as Na?ve Bayes(NB)and K-nearest neighbor(KNN),and SVM.This study was experimented on samples from the GTZAN dataset[7,19]to differentiate between a collection of 10 genres.Optimal classification accuracy of 80.1%was obtained with SVM compared to NB and KNN.

        Similarly,SVM was used as a basis for classification in[20],and the classification was evaluated on GTZAN dataset and a proposed Brazilian Music Dataset(BMD).The authors used features that belong to six sets of descriptors:time-domain,spectral,tonal,sound effect,high-level,and rhythm.In addition to the following common features:tonality,loudness,dissonance,sharpness,inharmonicity,tempo, key, and beat histograms.Higher accuracy of 86.11% was achieved with BMD compared to GTZAN.

        Many classifiers are also employed for automatic MGC using different feature extraction methods.These classifiers are:Gaussian mixture model(GMM)[21],radial basis function(RBF)[22],AdaBoost[23],and semi-supervised method[24].However,the results of several ML classification methods have also been reviewed in a comprehensive survey in[25],which shows that seeking the perfect classifier is still required.

        In recent years,deep learning approaches for the MGC have significantly impacted the classification results.According to Choi et al.[26],adopting deep learning models in the context of MGC will be beneficial for different reasons.The first one is that it provides classification with learned features obtained from different hierarchical layers of the neural networks rather than derived features obtained from structured data.The second one is the hierarchical topology properties of deep learning models that can be useful for musical analysis at any time and frequency range.In this regard,Convolution Neural Networks(CNNs)have been employed by many researchers for MGC[27-33].For instance,based on the sample level, Allamy et al.[27] proposed that 1D CNN architecture consists of nine residual blocks and two Convolution layers to classify the GTZAN dataset containing 1000 audio clips.The proposed 1D CNN achieved 80.93%accuracy.However,as future work,the authors believe that better results could be achieved on large music datasets such as MSD dataset[28],LMD dataset[29], and free music archive dataset [30].On the other hand, the work carried out in [31] using the preprocessed spectrogram as input to the CNN consisting of five convolutional layers achieved a higher accuracy rate of 84%on the GTZAN dataset.On the same dataset,Senac et al.[32]adopted CNN for their work with the set of eight music features along three main music dimensions: timbre, tonality,and dynamics(previously used in[33]).As a result,the trained CNN model achieved an accuracy rate of up to 91%.

        3 Materials and Methods

        This section discusses the methodology illustrated in Fig.1 through steps such as data construction,audio pre-processing,feature extraction,training,and testing.

        Figure 1:Overall framework

        3.1 Dataset Construction

        The large corpus of the Arabic Music Genre (AMG)dataset is built by extracting numerous audio clips from YouTube.The dataset consists of five different known genre classes, as presented in Tab.1.The AMG consists of 1266 audio tracks,each music piece of 30 s long,stored as a 799 MB wav audio file.

        Table 1: Arabic music genre

        Table 1:Continued

        The AMG dataset is well-annotated and well-structured of Arabic audio clips.This dataset is available freely for the research community on Arabic MGC on Kaggle under the title “Ar-MGC:Arabic Music Genre Classification Dataset”[34].

        The dataset has the following folders:

        ? Genres original—A collection of 5 genres with 1266 audio files,all are having a length of 30 s.

        ? Images original—A visual representation for each audio file.

        ? 1 JSON file—Contains features of the audio files.This JSON file contains three sections:

        Mapping->contains a name of the music genre.

        Labels->which represented as numbers for each genre[0-4].

        MFCC->which represented as a vector with size 13, and contains the features extracted by MFCC.

        3.2 WAV File

        The methodology implementation starts with WAVs files to convert them into spectrograms.WAVs were initially developed by Microsoft and IBM back in 1991.It is a waveform audio file format,and it is used to store uncompressed recorded sound with high fidelity.

        Figs.2 and 3 show waveforms of two AMG;Muwashshah and the Poem as time on the x-axis and amplitude on the y-axis.In this case, the visual representations for both genres clearly illustrate the difficulty in distinguishing each genre from the other one,so the Genre prediction by visual inspection is not evident.

        Figure 2:Waveform of muwashshah genre

        Figure 3:Waveform of the poem genre

        3.3 STFT

        STFT stands for Short Time Fourier Transform.It shows the changes in the frequency content over time by applying a series of windows to the signal using a DFT algorithm,and then all resulting DFTs are placed together in a single graph called a spectrogram,as shown in Fig.4.In this work,the parameters used to generate the spectrogram are as follows:

        ? Sampling rate(sr)=22050

        ? Window size(n_fft)=2048

        ? Hop_size=512

        Figure 4:Spectrogram of the poem genre

        3.4 MFCCs

        MFCCs stands for Mel-Frequency Cepstral Coefficients,introduced in the 1990 s by Davis et al.[35].It is one of the most popular state-of-the-art feature extraction methods because of its faster extraction technique than other methods such as Perceptual Linear Prediction (PLP)and Linear Prediction Coefficients (LPC).The MFCC summarizes the frequency distribution throughout the window size,allowing for analysis of the sound’s time and frequency characteristics.

        In this work, first, the STFT of the audio signal is taken with n_fft= 2048 and hop_size=512,next, the logarithm of the powers is taken at domain frequencies, then to map the powers onto the Mel-Scale using triangular overlapping windows, and then the final step to apply Discrete Cosine Transform,thereby obtaining the MFCCs,whereas 12656 is many audio features are extracted using the LibROSA package[36].

        3.5 CNNs

        CNNs are a type of deep neural network that is commonly used in computer vision applications.Furthermore, the uses of CNNs might be extended for any audio analysis applications because the architecture of 2D CNNs can process audio data after MFCC transformation, as it will deal with spectrograms as input.Similarly,as the architecture of 2D CNNs used with image processing tasks,it can also process spectrograms as images due to variations of musical patterns.A spectrogram is a 2D visual representation of the spectrum of audio frequencies over time, which is the convolved input using filters that identify essential characteristics to match the output.Therefore,CNN learns mostly hierarchical features through convolutions.Convolution produces a dot product between the filter and the pixels convolving the entire image as the output of the first convolution layer.Then this features map is sent into the next layer to produce many more features maps until it reaches the end of the network with extremely detailed general information about the image contents.The numbers within the filters are known as weights, and these represent the parameters trained during the training phase.CNNs are composed of convolution layers to learn hierarchical features and an activation function, and a pooling layer between each convolution layer.These activation functions employ the backpropagation approach to calculate the error and then propagate this error across the network,altering the weights of the filters based on this error.Indeed,the most used activation function is known as the Rectified Linear Unit(ReLU)function.Typically,to simplify the network and reduce the number of parameters,pooling layers are another building block of a CNN,and max pooling is the most common operation used in pooling[37].

        As explained above,that is the basic form of CNN;accordingly,there have been many different CNN architectures since the pioneering one, which is LeNet5 [12] by LeCun et al., in 1998, next, is AlexNet[38]by Krizhevsky et al.,in 2012,the deep learning network applied in the most popular object recognition competition,called ImageNet LSVRC-2010 contest to classify the 1.2 million images into the 1000 different classes.Next,with the progress of the GPUs,is VGG[39]by Simonyan et al.in 2015.Later, ResNet [40] by He et al.in 2016.Nowadays, most state-of-the-art architectures [41] perform similarly and have some specific use cases where they are better.In the following,we detail the most used CNN architectures[42]implemented in this paper,in addition to LSTM[43].

        3.6 Lenet5

        LeNet is the most common class of neural network architectures,as it is one of the earliest deep learning architectures.LeNet5 constitutes of five alternating convolutional and pooling,followed by two fully connected layers in the end.It employs convolution to preserve the spatial orientation of features and average pooling for downsampling of the feature space and ReLU and softmax activation between layers.The summary of the LeNet5 model is shown in Tab.2.The summary shows the total number of layers, the input size of each layer, the used activation functions employed, and more parameters.

        Table 2: The LeNet5 architecture and the model summary

        3.7 AlexNet

        AlexNet was one of the first deep neural networks in the 21stcentury.It is a deeper version of the LeNet5,which won the most popular object recognition competition,called ImageNet LSVRC-2010 contest,to classify more than one million images into the 1000 different classes.Advancement in computational hardware, GPU, and huge dataset availability aided the network’s success in the competition.

        The architecture of AlexNet consists of eight learned layers.It is five convolution layers followed by three fully connected layers.Each convolutional layer has a convolutional filter followed by a nonlinear activation function(ReLU).Between the first and second convolutional layers,max pooling and normalization operations add shift invariance and numerically stabilize learning,respectively.The third, fourth and fifth convolutional layers are connected to one another without any intervening pooling or normalization layers.Tab.3 shows the summary of the AlexNet model.The summary shows the total number of layers,the input size of each layer,the used activation functions,and more parameters.

        Table 3: The AlexNet architecture and the model summary

        Table 3:Continued

        3.8 VGG

        The VGG model was developed in 2014 by the visual geometry group at Oxford to handle another crucial component of convolution architecture design which is depth.VGG would have 11 to 19 layers compared to AlexNet’s eight layers.To this aim,other architectural parameters were fixed.At the same time,depth was gradually increased by adding more convolutional layers,which was possible due to relatively small convolution filters in all layers’levels.The spectrograms are passed through a stack of convolution layers.The filters have a very small receptive field of 3 cross 3, which is the smallest size required to capture the notions of left-right, up-down, and center.This results in a significant parameter decrease.Spatial pooling is also carried out by five max pooling layers that follow some of the convolutional layers.Tab.4 summarizes the VGG model.The summary displays the overall number of layers,the input size of each layer,the activation functions employed,and additional parameters.

        Table 4: The VGG architecture and the model summary

        3.9 ResNet-50

        Typically,previous models to the ResNet have depths of 16 and 30 layers,whereas ResNet could be trained very deep up to a hundred and fifty-two layers in the network.This is 8 times more than that of the VGG nets.ResNet-50 is a CNN that is 50 layers deep using residual connections.As a result,instead of learning unreferenced functions,the layers are reformulated as learning residual functions concerning the layer inputs.Fig.5.Shows a simple flow chart for RestNet with the residual connection,which is the building block that utilizes skip connections to bypass some layers.There are two primary reasons for adding skip connections: to prevent vanishing gradients and alleviate the Degradation(accuracy saturation)problem.In addition, an extra weight matrix may be employed; these models are known as HighwayNets.

        Figure 5:The ResNet flowchart

        3.10 LSTM-CNN

        The overall architecture of LSTM-CNN model combines convolutional block and Long Short-Term Memory (LSTM)block to capture spatial features and temporal features within the data,respectively.In the first convolutional block, the input data will be convoluted with a filter map by sliding the kernel window.There is a pooling layer,which is a way of compressing the convoluted data further.Here, the max pooling is used by considering the max from the vector.The second block,LSTM block[43],acts as a sequence to vector,and it is ideal for capturing long-term patterns within the time series data.Finally,the final output layer acts as a fully connected artificial neural network,and SoftMax activation function is used to generate the output.The summary of the LSTM-CNN model is shown in Tab.5.The summary shows the total number of layers,the input size of each layer,the activation functions employed,and more parameters.

        Table 5: The LSTM-CNN architecture and the model summary

        4 Results

        One of the main objectives of this study is to evaluate the performance of all trained models described in Section 3 on our new created dataset; the AMG dataset, which consists of 1266 audio tracks,each track piece of 30 s long,stored as wav audio files at a sampling rate of 22050 Hz.Samples were classified into the following five Arabic musical genres: Eastern Takht, Rai, Muwashshah, the poem,and Mawwal.AMG dataset was split into75%for the training set and 25%for the testing set.

        In order to build our approach and evaluate it,the experiment was performed under a GPU server(Kaggle notebook)using Python 3[44].MFCC features is computed from audio signals were sent to the deep CNN models for training.

        In general, MIR field provides two main parameters for evaluating the system performance:accuracy and Matthew’s correlation coefficient (MCC), in addition to the time required to run the model.

        The percentage of correctly classified audio samples is referred to as accuracy[45].It includes all four possibilities:True Positives(TP),True Negatives(TN),False Positives(FP),and False Negatives(FN).This formula defines the accuracy.

        Matthew’s correlation coefficient determines the correlation between true class and predicted class.The higher the correlation between true and predicted values, the better the prediction, and mathematically is defined as

        An empirical experiment was performed using different epochs and learning rate parameters to compare CNN model performance.Learning rate controls how quickly or slowly a CNN model learns a problem,and epoch refers to one cycle through the full training dataset.Indeed,it is an art in ML to determine the number of sufficient epochs for a network;hence,three epochs’values were used(30,50,70),and three learning rates were used(0.001,0.0001,0.00005).The experimental results show a perfectly configured learning rate,and epochs are 0.00005 and 70,respectively.

        The performance of the different CNN models with the learning rate 0.00005 and the different epochs values is summarized in Tab.6.The best result shows that AlexNet model obtained a 96%test accuracy on the dataset and 95%on MCC within 6.4 s.Figs.6-8 illustrate the overall performance in terms of Accuracy,MCC and time,respectively.Furthermore,the confusion matrix of AlexNet model shows superior performance with Eastern Takht genre classification reaching an accuracy rate of 99%,as illustrated in Fig.9.

        Figure 6:Performance evaluation of all classifiers on the dataset in terms of accuracy

        Figure 7:Performance evaluation of all classifiers on the dataset in terms of MCC

        Figure 8:Performance evaluation of all classifiers on the dataset in terms of time

        Figure 9:Confusion matrix of AlexNet model

        5 Conclusion and Future Works

        Using the constructed dataset titled “Ar-MGC: Arabic Music Genre Classification Dataset”,we performed a complete empirical comparison of deep CNNs architectures in this study.In the methodology, the audio data was transformed into a spectrogram using STFT.MFCC was applied to extract the audio features,and finally,a classification task was carried out using CNNs.

        Comparing the reviewed related works mainly implemented in western MGC using various ML algorithms.AlexNet model obtained higher accuracy on automatic classification between five of the most well-known Arabic music genres:Eastern Takht,Rai,Muwashshah,the poem,and Mawwal.

        Many CNNs architectures were explored with their design parameters and evaluation.The results of the experimental evaluation are encouraging to incorporate this work into the mood analysis system on music preferences as music has the potential to impact our brains.This psychological investigation involves EEG analysis.

        Funding Statement:The authors received no specific funding for this study.

        Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

        欧美黑人群一交| 成人性生交大片免费看l| 美女主播福利一区二区| 中文字幕乱码熟妇五十中出| 亚洲综合色成在线播放| 91在线区啪国自产网页| 最新日本女优中文字幕视频| 97在线视频人妻无码| 狠狠色噜噜狠狠狠狠色综合久| 亚洲国产成人久久综合一区77| 久久免费看视频少妇高潮| 精品人妻系列无码人妻漫画| 熟女熟妇伦av网站| 国产亚洲欧洲AⅤ综合一区| 熟女高潮av一区二区| 曰韩无码av一区二区免费| 亚洲av综合色区无码一二三区| 国产精品美女黄色av| 精品国产亚洲av高清日韩专区| 私人vps一夜爽毛片免费| 欧美国产日韩a在线视频| 亚洲无码毛片免费视频在线观看 | 国产精品无码人妻在线| 真实国产乱啪福利露脸| 亚洲av中文aⅴ无码av不卡| 丝袜美足在线视频国产在线看| 久久久久久亚洲av成人无码国产| 亚洲欧美国产日韩制服bt| 大量老肥熟女老女人自拍| 色婷婷av一区二区三区久久| 在线观看免费人成视频色9| 国产69口爆吞精在线视频喝尿| 婷婷开心五月亚洲综合| 国产精品免费av片在线观看| 伊人99re| 国产在线看不卡一区二区| 国产成人av无码精品| 中文字幕熟妇人妻在线视频| 亚州五十路伊人网| 就爱射视频在线视频在线 | 国产午夜福利短视频|