DCNN for Pig Vocalization and Non-Vocalization Classification: Evaluate Model Robustness with New Data

Author:

Pann Vandet1ORCID,Kwon Kyeong-seok1ORCID,Kim Byeonghyeon1ORCID,Jang Dong-Hwa1,Kim Jong-Bok1ORCID

Affiliation:

1. Animal Environment Division, National Institute of Animal Science, Rural Development Administration, Wanju 55365, Republic of Korea

Abstract

Since pig vocalization is an important indicator of monitoring pig conditions, pig vocalization detection and recognition using deep learning play a crucial role in the management and welfare of modern pig livestock farming. However, collecting pig sound data for deep learning model training takes time and effort. Acknowledging the challenges of collecting pig sound data for model training, this study introduces a deep convolutional neural network (DCNN) architecture for pig vocalization and non-vocalization classification with a real pig farm dataset. Various audio feature extraction methods were evaluated individually to compare the performance differences, including Mel-frequency cepstral coefficients (MFCC), Mel-spectrogram, Chroma, and Tonnetz. This study proposes a novel feature extraction method called Mixed-MMCT to improve the classification accuracy by integrating MFCC, Mel-spectrogram, Chroma, and Tonnetz features. These feature extraction methods were applied to extract relevant features from the pig sound dataset for input into a deep learning network. For the experiment, three datasets were collected from three actual pig farms: Nias, Gimje, and Jeongeup. Each dataset consists of 4000 WAV files (2000 pig vocalization and 2000 pig non-vocalization) with a duration of three seconds. Various audio data augmentation techniques are utilized in the training set to improve the model performance and generalization, including pitch-shifting, time-shifting, time-stretching, and background-noising. In this study, the performance of the predictive deep learning model was assessed using the k-fold cross-validation (k = 5) technique on each dataset. By conducting rigorous experiments, Mixed-MMCT showed superior accuracy on Nias, Gimje, and Jeongeup, with rates of 99.50%, 99.56%, and 99.67%, respectively. Robustness experiments were performed to prove the effectiveness of the model by using two farm datasets as a training set and a farm as a testing set. The average performance of the Mixed-MMCT in terms of accuracy, precision, recall, and F1-score reached rates of 95.67%, 96.25%, 95.68%, and 95.96%, respectively. All results demonstrate that the proposed Mixed-MMCT feature extraction method outperforms other methods regarding pig vocalization and non-vocalization classification in real pig livestock farming.

Funder

Rural Development Administration, Republic of Korea

Publisher

MDPI AG

Reference52 articles.

1. Domestic pig sound classification based on TransformerCNN;Liao;Appl. Intell.,2023

2. Pork market crisis in Romania: Pig livestock, pork production, consumption, import, export, trade balance and price;Popescu;Sci. Pap. Ser. Manag. Econ. Eng. Agric. Rural Dev.,2020

3. Liang, Y., Cheng, Y., Xu, Y., Hua, G., Zheng, Z., Li, H., and Han, L. (2022). Consumer preferences for animal welfare in China: Optimization of pork production-marketing chains. Animals, 12.

4. Hou, Y., Li, Q., Wang, Z., Liu, T., He, Y., Li, H., Ren, Z., Guo, X., Yang, G., and Liu, Y. (2024). Study on a Pig Vocalization Classification Method Based on Multi-Feature Fusion. Sensors, 24.

5. Dohlman, E., Hansen, J., and Boussios, D. (2022). USDA Agricultural Projections to 2031, United States Department of Agriculture.

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3