Next Article in Journal
Phenotypic and Genetic Analyses of In Vitro Embryo Production Traits in Chinese Holstein Cattle
Next Article in Special Issue
Behavioral Adaptations of Nursing Brangus Cows to Virtual Fencing: Insights from a Training Deployment Phase
Previous Article in Journal
iTRAQ-Based Phosphoproteomic Analysis Exposes Molecular Changes in the Small Intestinal Epithelia of Cats after Toxoplasma gondii Infection
Previous Article in Special Issue
A Framework for Transparency in Precision Livestock Farming
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Sound Location to Monitor Farrowing in Sows

by
Elaine van Erp-van der Kooij
1,*,
Lois F. de Graaf
1,
Dennis A. de Kruijff
1,
Daphne Pellegrom
1,
Renilda de Rooij
1,
Nian I. T. Welters
2 and
Jeroen van Poppel
1
1
Department of Animal Husbandry, HAS Green Academy, University of Applied Sciences, P.O. Box 90108, 5200 MA ‘s-Hertogenbosch, The Netherlands
2
Department of Applied Biology, HAS Green Academy, University of Applied Sciences, P.O. Box 90108, 5200 MA ‘s-Hertogenbosch, The Netherlands
*
Author to whom correspondence should be addressed.
Animals 2023, 13(22), 3538; https://doi.org/10.3390/ani13223538
Submission received: 2 October 2023 / Revised: 11 November 2023 / Accepted: 14 November 2023 / Published: 16 November 2023
(This article belongs to the Special Issue 2nd U.S. Precision Livestock Farming Conference)

Abstract

:

Simple Summary

Automated monitoring can help farmers care during the farrowing of sows and piglets. Five sows were monitored in two field studies. A sound camera with small microphones showing sounds as coloured spots and a security camera was used to record the farrowing of sows and piglets. First, sound spots were compared with audible sounds, analysing video data. This gave many false positives, including visible sound spots but no audible sounds. Of piglet births, 23 of 50 piglet births were visible, but none were audible. The sow’s behaviour changed when farrowing started. One piglet was silently crushed. Secondly, data were analysed at a slower speed, and sound spots were compared with sounds and behaviour separately. This resulted in better results, but again, many sound spots showed without audible sound. When adding up audible sounds and visible sow behaviour and comparing sound spots with the combination of sound and behaviour, results were much improved, with an accuracy of 91.2%, an error of 8.8%, a sensitivity of 99.6%, and a specificity of 69.7%. We conclude that sound cameras are promising tools, detecting sounds more accurately than humans. The most promising application for the sound camera is detecting the onset of farrowing.

Abstract

Precision Livestock Farming systems can help pig farmers prevent health and welfare issues around farrowing. Five sows were monitored in two field studies. A Sorama L642V sound camera, visualising sound sources as coloured spots using a 64-microphone array, and a Bascom XD10-4 security camera with a built-in microphone were used in a farrowing unit. Firstly, sound spots were compared with audible sounds, using the Observer XT (Noldus Information Technology), analysing video data at normal speed. This gave many false positives, including visible sound spots without audible sounds. In total, 23 of 50 piglet births were visible, but none were audible. The sow’s behaviour changed when farrowing started. One piglet was silently crushed. Secondly, data were analysed at a 10-fold slower speed when comparing sound spots with audible sounds and sow behaviour. This improved results, but accuracy and specificity were still low. When combining audible sound with visible sow behaviour and comparing sound spots with combined sound and behaviour, the accuracy was 91.2%, the error was 8.8%, the sensitivity was 99.6%, and the specificity was 69.7%. We conclude that sound cameras are promising tools, detecting sound more accurately than the human ear. There is potential to use sound cameras to detect the onset of farrowing, but more research is needed to detect piglet births or crushing.

1. Introduction

It is estimated that by 2050, the human world population could reach >9 billion, consuming 50–60% more food than at present. The majority of people still prefer animal proteins over plant-based food, and the demand for livestock products continues to grow, as does global food insecurity [1]. Sustainable intensification is one of the solutions [2]. With the intensification of food production and the industrialisation of animal production systems comes the fear of decreased animal welfare [3]. While meat production increases, society expects that animals used for meat are treated humanely and individually. Precision Livestock Farming (PLF) can improve or monitor animal welfare on farms if properly implemented [4,5]. PLF can be defined as managing livestock production using the principles of process engineering. Smart sensors are used to measure and monitor animal health and welfare [6,7,8,9,10]. Several sensors have been developed for the livestock sector. For pigs, the main focus is on the health and productivity of pigs, with sensors such as cameras, microphones, thermometers, and accelerometers being developed and applied [1,11,12,13,14]. A new development is the application of a sound camera, providing sound source localisation through an array of 64 microphones and visualising sound sources as coloured spots. Sound cameras are presently used for crowd control under outdoor conditions [15] and have been introduced in ecology [16] and agriculture, specifically in poultry and pigs [17,18,19]. Potentially, these cameras can be of use in monitoring welfare in pig farms since automated behaviour monitoring via sound and vision could help farmers to prevent health and welfare issues [20,21,22,23,24].
The farrowing phase in pig production and breeding is a crucial moment that has a great impact on the economics of the farm but also on the welfare of the sow and piglets. Piglet mortality is a big problem in modern sow farming [25]. For sow farms, relevant welfare issues around farrowing are a prolonged birthing process, possibly resulting in stillborn piglets and the crushing of piglets after farrowing [26,27,28]. Globally, approximately one piglet per litter is stillborn [29]. More stillbirths occur in sows that are confined in a crate during farrowing than in sows in open pens [30]. Stillbirths are often due to asphyxiation during the farrowing process. Piglets born later in the farrowing process have a higher chance of asphyxiation [29]. Asphyxiation is related to dystocia in sows due to prolonged farrowing or weak uterine contractions. This requires the stockman to assist the sow during farrowing. Improved farrowing management and human supervision during farrowing might decrease piglet mortality [25]. The automated monitoring of sows and alerting the farmer when the farrowing process stagnates can aid the farmer in optimising the health and welfare of his animals [8]. Therefore, in this study, we focus on the automatic monitoring of the farrowing process. Sows show specific behaviours and behavioural patterns before, during, and after farrowing. If we can monitor these behaviours, we might detect the onset of farrowing and follow the process of farrowing. Before farrowing, sows show a natural pattern of nest-building behaviour, such as foraging, rooting, and pawing, which are motivated by the desire to build a shelter to protect their offspring against predators and cold weather. Domestic sows, when given nest-building materials, still perform nest-building behaviours [31]. In farming systems with farrowing crates, nest-building behaviour is reflected in ‘playing’ with the available material, which can be a jute sack or a handful of straw. During farrowing, sows show pain-related behaviour such as tail flicks, straining, and pushing the back leg forward [30]. Approximately 50% of postnatal deaths in piglets are caused by the crushing of the sow [28]. Piglet crushing happens mostly within the first four days after farrowing. There is a large individual variation in piglet crushing between sows; sows show a more protective mothering style and crush fewer piglets. Especially during posture changes of the sows, the young piglets are at risk. Piglets vocalise during trapping events but also during other stress-related events. However, detecting these sounds might be used in a monitoring device [32].
In this study, we have used a sound camera together with a security camera to monitor sounds, vision, and sound location around farrowing as the first step in developing a sound-based early warning system for a stagnating birthing process and the prevention of piglet crushing.

2. Materials and Methods

2.1. Sound Camera

The L642V is a camera with an array of 64 microphones. The device uses delay-and-sum beamforming to localise different sound sources and visualises these on an acoustic map. Delay-and-sum beamforming signal processing can be divided into four steps. Since the sound of each source travels to every microphone along a different path (Step 1), the signals captured by the microphones are similar in wave form but differ in their delay and phase. Delay and phase are proportional to the travelled distances. The delays can be determined from the sound speed, the distance between the microphones, and the sound sources (Step 2). The Beamformer targets the point of one of the sound sources, shifting the signal of each microphone via the difference in runtime depending on the focus point. Therefore, the signal components of this one sound source are in phase, and of the other sound sources are out of phase (Step 3). Finally, the signals of the microphones are summed together and normalised by the number of microphones (Step 4). If a certain target point does not contain a source, the signals partly cancel each other out due to destructive interference. At target points with a source, the signals align, and add up due to positive interference. The maximum amplitude is calculated from the time signal and the sound source is visualised on the acoustic map. Due to the positive interference, target points with a source have a higher magnitude than those that do not, and thus, source locations can be found. In this study, we denoised sound by selecting a frequency band for which the sounds were visualised. In a pilot study, we tested different frequency bands and found that in a range from 39,000 to 49,000 Hz, the sounds of the sow and her environment were best visible, with the lowest influence of background noise. In the present study, each camera was manually tuned to a specific optimal frequency band of 2000 Hz within these limits, according to the noise in the farrowing unit in that period. The camera has a built-in spatial filter, which means it only shows sound sources within the selected area, which, in our study, was the farrowing pen for each sow. Finally, the camera also denoised automatically by not visualising a sound when no clear source could be found.

2.2. Study Design

The study was performed at a commercial pig farm with two farrowing units for 64 sows each. Sows were housed in farrowing crates within a farrowing pen, with access to a jute sack in the pen but no straw and a solid concrete floor. Sows were monitored around farrowing, staying in a farrowing pen of 2.80 × 1.75 m with a farrowing crate of 2.1 × 1.0 m. Two Bascom XD10-4 security cameras that showed sound and vision were placed above the pens to record audible sounds and the behaviour of the sow and piglets, with each camera viewing two pens. Three Sorama L642V sound cameras were placed directly above three farrowing pens, with each camera viewing one pen (Figure 1).
Data from the cameras could not directly be recorded due to the safety settings. We, therefore, streamed the data to three laptops in the office of the farm. The screen of the laptops showed the image of the security camera and the sound camera side by side, as well as a clock, to synchronise the images if necessary (Figure 2). We recorded data using the screen recorder software Open Broadcast Software version 27.2.1 (OBS-studio), resulting in video files in the mp3 format. Laptops were remotely controlled using TeamViewer.
Approximately 45 min of video data from each of the four sows were analysed after the first field study. Birthing events and crushing events were recorded from the video footage using the Observer XT (Noldus Information Technology). Four hours before farrowing and during farrowing, sow behaviours were recorded that were possibly associated with the birthing process. We used a simplified ethogram with three behavioural categories: lying (inactive), playing with the jute sack/rooting (snout on the concrete floor), and sitting/standing (inactive). Audible sounds were recorded from the recorded video of the security camera. Visible sounds were recorded from the sound camera. When a sound spot was visible at roughly the right location within ±1 s from the audible sound, the spot was considered correct and positive. Audible sounds with no corresponding sound spot were considered false negatives. Sound spots with no audible sounds were considered false positives. Finally, when no sound was visible or audible for 2 s, this was considered correct negative. In Table 1, the connected audible and visible sounds and sound locations are shown.
In Table 2, the connected audible and visible sounds and behaviours are shown.
Data from three sows were analysed after the second field study. From the security cameras, sow behaviours, and audible sounds were recorded using the Observer XT, and from the sound camera, sound spots were recorded for a short period during farrowing. When a sound spot was visible at roughly the right location within ±1 s from the audible sound, the spot was considered correct and positive (CP). Audible sounds with no corresponding sound spot were considered false negatives (FN). Sound spots with no audible sounds were considered false positives (FP). Finally, when no sound was visible or audible for 2 s, this was considered correct negative (CN). For validation purposes, accuracy, error%, sensitivity and specificity were calculated as follows:
a c c u r a c y = ( C P + C N t o t a l   ) × 100 %
e r r o r   % = ( F P + F N t o t a l   ) × 100 %
s e n s i t i v i t y = C P F N + C P × 100 %
s p e c i f i c i t y = ( C N ( F N + C P )   ) × 100 %

3. Results

Data from five sows in two field studies were gathered and analysed for visible sounds (sound spots) and audible sounds, and, in field test 2, visible sow behaviour as well. There was a minor time lag in the recording of the visible sounds of approximately 1.5 s, which was corrected by adding 1.5 s to the recorded times of the sound spots. We adjusted the frequency settings for each camera per sow manually by testing which frequency band showed the best visualising of sound sources with the least noise at that time and place. This resulted in frequency bands of 39,570–41,570 Hz (camera 1), 46,060–48,060 Hz (camera 2) and 41,730–43,730 Hz (camera 3) in the first study and an adjustment to 45,630–47,630 Hz for camera 1 in the second study.

3.1. Field Study 1

In the first field study, we compared the audible and visible sounds of the sows before farrowing and recorded 13,351 sound spots and 981 audible sounds in 177 min of video data. We found a low agreement between the sound and vision data (Table 3).

3.2. Field Study 2

In the second field study, we analysed 6 min (360 s) of video from two sows during farrowing. Video data were analysed at a 10-fold slower speed, and audible sounds, sow behaviour, and sound spots were recorded. This resulted in a somewhat higher but still unsatisfactory agreement between the sound and vision data when sound spots were compared to either audible sounds or visible behaviour (Table 4). However, when comparing sound spots with the combination of audible sounds and visible behaviour, results were much improved, with an accuracy of 91.2, an error% of 8.8, and a specificity of 69.7 (Table 4). In this analysis, we added up sounds and behaviours to compare with sound spots, considering an event as a correct positive if either sound or behaviour (or both) were shown at the same time and location as the sound spot.

3.3. Birthing Events and Piglet Crushing

For the detection of birthing events, data from four sows were used. A total of 50 piglets were born from these sows. For 23 of 50 piglet births, a sound spot was visible at the correct time; piglet births were usually not audible, but most were visible on the safety camera. This resulted in an accuracy of 71.4%. Sound spots were visible in the area behind the sow for some time after the piglet was born, but the movement of the piglets was not audible.
One piglet was crushed shortly after birth, but no sound was heard, and no sound spots were visible for this event, which was only visible on the video footage of the Bascom camera.
When we analysed behaviours around parturition, we found that the sows showed specific behaviours at specific times: playing with the jute sack in the farrowing pen was seen more before parturition and especially during the last hours before the piglets were born. Once farrowing started, this behaviour stopped almost completely (Table 5). During farrowing, typical movements of the legs of the sows were seen, seemingly associated with the birthing events.

4. Discussion

For this study, we used data from five sows to study the application of a sound camera in pig farming. Further research is needed using more sows and more repetitions to validate these findings. However, in this study, we were able to optimise the methods for analysing the results of sound spots, sound, and behaviour, and we found promising results for the application of the sound camera to monitor farrowing sows.
In this study, the human observer was the gold standard for audible sound. When comparing manual to automated scoring, there were some problems with finding the gold standard. Clinical research has shown that manual scores are usually qualitative or semiquantitative and subjective, even when conducted by a seasoned observer, while automated image analysis is quantitative, reproducible, and objective. Manual image analysis has some drawbacks [33] that can easily be extrapolated to manual sound analysis. The sources of bias include the illusion of size (size being influenced by the context), distinguishing colours, and lateral inhibition (increased response to edges). For sound analysis, these can translate into an illusion of loudness (being influenced by the loudness of other sounds), a distinguishing pitch (depending on the pitch of surrounding sounds), and an increased response to short and sharply defined sounds. General sources of bias include inattentional blindness (i.e., not paying attention) and confirmation bias (i.e., hearing what you expect or want to hear). Labelling audible sounds from videos recorded with a safety camera probably resulted in many false negatives for audible sounds and inaccuracies in labelling since the human observer either hears the sound and reacts too late or does not hear the sound at all, while the sound camera does receive the sound. Furthermore, the labelling of the sound spots was probably not accurate enough since we labelled at a normal speed. This resulted in many ‘cloud of sound spots’ events, with a cluster of sound spots occurring at once. Playing the videos at a 10-fold slower speed showed that the sound spot clouds were actually a series of sound spots that started with one spot in the correct place, followed by a cluster of spots in the area. Therefore, we adjusted the analysis for the second field study.
Most sows farrowed at night, with low visibility on the cameras, which increased the number of false positives (i.e., sounds visible in a different spot than audible) due to human error. In all tests, we counted many more sound spots than visible behaviours or audible sounds. This may very well be due to human error. A reliability analysis for labelling sound spots between the observers showed an agreement of 82%, but a Kappa value of 0.17 (indicating slight agreement) was obtained. The high number of sound spots and almost no silent periods lead to a high agreement by a chance of 0.78. This lowered the Kappa value [34,35]. In addition, the manual labelling of data as the gold standard is a point of discussion.
In the second field study, adding visible behaviour gave much better results, correctly classifying sound spots for visible but inaudible behaviours. Furthermore, we recorded sound from the Bascon camera and listened to the recordings, which is an indirect way of working with data and probably lowered the audibility for human observers. It seems that the sound camera is much more accurate and precise at detecting sounds than we, as human observers, are. Therefore, we advise combining audible sounds and visible behaviours when validating sound location sensors such as the sound camera we used.
We detected less than half of the piglet birth events with the sound camera. These sounds were low-pitched and soft. Filtering the noise from the sound camera and using the high-frequency bands to visualise sounds might have increased the number of false negatives. One piglet was crushed during the study without an audible sound or a sound spot from the sound camera. We expected to capture high-pitched sounds that accompanied crushing events, but if it happened fast, the piglet had no time to scream. From only one crushing event, we could not validate whether the sound camera could detect piglet crushing. In a study where sound was used to detect piglet distress, it was found that many more piglet stress calls were associated with other stress-related events than associated with trapping events. Although adding context-based event filters increased the results, sound might not be the preferred method for detecting crushing events [32]. Crushing events are mostly associated with posture changes in the sow, such as rolling over during resting. Sows that crush fewer piglets show fewer posture changes [28]. If these posture changes could be detected with a sound camera, this might be used to prevent the crushing of piglets.
The automatic monitoring of sow behaviours can be performed using cameras or activity meters [20,36,37,38]. To investigate whether the sound camera can be used to detect specific behaviours associated with the onset of farrowing, behaviours were recorded around farrowing. We found that playing with the jute sack and rooting was seen almost only before farrowing. This behaviour is associated with nest-building behaviour [31]. Typical leg movements were seen predominantly during farrowing and seemed to be associated with birthing events. These behaviours resembled the leg behaviours that were reported in previous research, which were classified as pain-related behaviours during farrowing [30,39]. Nest building behaviour was only seen before farrowing, and leg movements during farrowing; after farrowing, these behaviours stopped completely. These behaviours were associated with sound spots in the corresponding, typical locations as follows: playing with the jute sack or rooting on the floor was seen as sound spots near the head of the sow, and leg movements were seen as sound spots near the front or hind legs, detected with the sound camera. This is interesting because the cross-over from showing nest-building behaviour (playing with the jute sack and rooting on the floor) to lying down and showing only some typical leg movements seems to mark the onset of farrowing. If so, we could use the detection of these behaviours with the camera as a signal for the farmer that farrowing has started. Furthermore, the absence of both these behaviours, with sows lying down and being fully inactive, marks the end of the farrowing process. This could be combined with the fact that sound spots behind the sow showed the movement of piglets that were just being born. As long as new spots appeared behind the sow, the farrowing process was still in progress. The combination of these findings could be used to monitor the farrowing process and its duration. However, more research is needed to validate these assumptions.

5. Conclusions

Sound cameras are potentially interesting to apply in pig farming since they can detect sounds and sound locations better than the human observer. The behaviour of the sow and the movement of piglets in the pen could be reliably detected. However, we could not reliably detect piglet births and crushing events in this study. When analysing sound and visual data, it is important that a slower speed must be used to record the order of events and sound spots and that sound data are connected not only to audible sounds but also to behaviours that are inaudible to the human ear. The most interesting application for sound cameras seems to be detecting the onset of farrowing by recording sounds from the prepartum sow as she is preparing for the farrowing process and monitoring the farrowing process by detecting sounds of newly born piglets in the area behind the sow. Further research is needed to test this application, using more sows and more repetitions.

Author Contributions

Conceptualisation, E.v.E.-v.d.K.; methodology, E.v.E.-v.d.K. and J.v.P.; software, E.v.E.-v.d.K.; validation, E.v.E.-v.d.K.; formal analysis, E.v.E.-v.d.K.; investigation, E.v.E.-v.d.K., J.v.P., L.F.d.G., D.A.d.K., D.P., R.d.R., and N.I.T.W.; resources, E.v.E.-v.d.K.; data curation, L.F.d.G., D.A.d.K., D.P., R.d.R., and N.I.T.W.; writing—original draft preparation, E.v.E.-v.d.K.; writing—review and editing, E.v.E.-v.d.K. and J.v.P.; visualisation, D.P., R.d.R., and N.I.T.W.; supervision, E.v.E.-v.d.K. and J.v.P.; project administration, J.v.P.; funding acquisition, E.v.E.-v.d.K. All authors have read and agreed to the published version of the manuscript.

Funding

This project was funded by Holland High Tech with PPS-premium for research and development in the top sector HTSM, project HTSM 21.0382.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Animals were observed on a commercial farm with cameras only and no handling took place, nor were the animals disturbed in any way due to the study. The farmer signed an informed consent statement, allowing us to use the data of their sows.

Data Availability Statement

Data in this study are not publicly archived due to privacy reasons since the study was performed at a commercial farm. Datasets are available from the author upon request.

Acknowledgments

We thank Twan Dirks for the opportunity to perform this study at his farm ‘Beter Varken’ and use the data of his sows. We are thankful for the loan of one sound camera from Noldus Information Technology, and the technical support from Ben Loke of Noldus Information Technology and Koen Diesveld of Sorama.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Benjamin, M.; Yik, S. Precision Livestock Farming in Swine Welfare: A Review for Swine Practitioners. Animals 2019, 9, 133. [Google Scholar] [CrossRef] [PubMed]
  2. Charles, H.; Godfray, H.; Garnett, T. Food Security and Sustainable Intensification. Philos. Trans. R. Soc. B Biol. Sci. 2014, 369, 6–11. [Google Scholar] [CrossRef]
  3. Broom, D.M. Animal Welfare: An Aspect of Care, Sustainability, and Food Quality Required by the Public. J. Vet. Med. Educ. 2010, 37, 83–88. [Google Scholar] [CrossRef] [PubMed]
  4. Banhazi, T.M.; Lehr, H.; Black, J.L.; Crabtree, H.; Schofield, P.; Tscharke, M.; Berckmans, D. Precision Livestock Farming: An International Review of Scientific and Commercial Aspects. Int. J. Agric. Biol. Eng. 2012, 5, 1–11. [Google Scholar] [CrossRef]
  5. van Erp-van der Kooij, E.; Rutter, S.M. Using Precision Farming to Improve Animal Welfare. CAB Rev. Perspect. Agric. Vet. Sci. Nutr. Nat. Resour. 2020, 15, 1–10. [Google Scholar] [CrossRef]
  6. Wathes, C.M.; Kristensen, H.H.; Aerts, J.M.; Berckmans, D. Is Precision Livestock Farming an Engineer’s Daydream or Nightmare, an Animal’s Friend or Foe, and a Farmer’s Panacea or Pitfall? Comput. Electron. Agric. 2008, 64, 2–10. [Google Scholar] [CrossRef]
  7. Neethirajan, S. Recent Advances in Wearable Sensors for Animal Health Management. Sens. Bio-Sens. Res. 2017, 12, 15–29. [Google Scholar] [CrossRef]
  8. Berckmans, D. General Introduction to Precision Livestock Farming. Anim. Front. 2017, 7, 6–11. [Google Scholar] [CrossRef]
  9. Li, N.; Ren, Z.; Li, D.; Zeng, L. Review: Automated Techniques for Monitoring the Behaviour and Welfare of Broilers and Laying Hens: Towards the Goal of Precision Livestock Farming. Animal 2020, 14, 617–625. [Google Scholar] [CrossRef]
  10. Halachmi, I.; Guarino, M.; Bewley, J.; Pastell, M. Smart Animal Agriculture: Application of Real-Time Sensors to Improve Animal Well-Being and Production. Annu. Rev. Anim. Biosci. 2019, 7, 403–425. [Google Scholar] [CrossRef]
  11. Chapa, J.M.; Maschat, K.; Iwersen, M.; Baumgartner, J.; Drillich, M. Accelerometer Systems as Tools for Health and Welfare Assessment in Cattle and Pigs—A Review. Behav. Process. 2020, 181, 104262. [Google Scholar] [CrossRef] [PubMed]
  12. Berckmans, D. Precision Livestock Farming Technologies for Welfare Management in Intensive Livestock Systems. OIE Rev. Sci. Tech. 2014, 33, 189–196. [Google Scholar] [CrossRef] [PubMed]
  13. Wallenbeck, A.; Keeling, L.J. Using Data from Electronic Feeders on Visit Frequency and Feed Consumption to Indicate Tail Biting Outbreaks in Commercial Pig Production. J. Anim. Sci. 2013, 91, 2879–2884. [Google Scholar] [CrossRef] [PubMed]
  14. D’Eath, R.B.; Jack, M.; Futro, A.; Talbot, D.; Zhu, Q.; Barclay, D.; Baxter, E.M. Automatic Early Warning of Tail Biting in Pigs: 3D Cameras Can Detect Lowered Tail Posture before an Outbreak. PLoS ONE 2018, 13, e0194524. [Google Scholar] [CrossRef]
  15. Nguyen, Q.; Shen, G.; Choi, J.S. Sound Detection and Localization in Windy Conditions for Intelligent Outdoor Security Cameras. Circuits Syst. Signal Process. 2016, 35, 233–251. [Google Scholar] [CrossRef]
  16. Mennill, D.J.; Battiston, M.; Wilson, D.R.; Foote, J.R.; Doucet, S.M. Field Test of an Affordable, Portable, Wireless Microphone Array for Spatial Monitoring of Animal Ecology and Behaviour. Methods Ecol. Evol. 2012, 3, 704–712. [Google Scholar] [CrossRef]
  17. Exadaktylos, V.; Silva, M.; Ferrari, S.; Guarino, M.; Berckmans, D. Sound Localisation in Practice: An Application in Localisation of Sick Animals in Commercial Piggeries. In Advances in Sound Localization; IntechOpen: London, UK, 2011. [Google Scholar]
  18. Du, X.; Lao, F.; Teng, G. A Sound Source Localisation Analytical Method for Monitoring the Abnormal Night Vocalisations of Poultry. Sensors 2018, 18, 2906. [Google Scholar] [CrossRef]
  19. Adrion, F.; Keller, M.; Bozzolini, G.B.; Umstatter, C. Setup, Test and Validation of a UHF RFID System for Monitoring Feeding Behaviour of Dairy Cows. Sensors 2020, 20, 7035. [Google Scholar] [CrossRef]
  20. Oczak, M.; Bayer, F.; Vetter, S.; Maschat, K.; Baumgartner, J. Comparison of the Automated Monitoring of the Sow Activity in Farrowing Pens Using Video and Accelerometer Data. Comput. Electron. Agric. 2022, 192, 106517. [Google Scholar] [CrossRef]
  21. Shao, B.; Xin, H. A Real-Time Computer Vision Assessment and Control of Thermal Comfort for Group-Housed Pigs. Comput. Electron. Agric. 2008, 62, 15–21. [Google Scholar] [CrossRef]
  22. Nilsson, M.; Herlin, A.H.; Ardö, H.; Guzhva, O.; Aström, K.; Bergsten, C. Development of Automatic Surveillance of Animal Behaviour and Welfare Using Image Analysis and Machine Learned Segmentation Technique. Animal 2015, 9, 1859–1865. [Google Scholar] [CrossRef]
  23. Chen, C.; Zhu, W.; Liu, D.; Steibel, J.; Siegford, J.; Wurtz, K.; Han, J.; Norton, T. Detection of Aggressive Behaviours in Pigs Using a RealSence Depth Sensor. Comput. Electron. Agric. 2019, 166, 105003. [Google Scholar] [CrossRef]
  24. Wurtz, K.; Camerlink, I.; D’Eath, R.B.; Fernández, A.P.; Norton, T.; Steibel, J.; Siegford, J. Recording Behaviour of Indoor-Housed Farm Animals Automatically Using Machine Vision Technology: A Systematic Review. PLoS ONE 2019, 14, e0226669. [Google Scholar] [CrossRef]
  25. Okinda, C.; Lu, M.; Nyalala, I.; Li, J.; Shen, M. Asphyxia Occurrence Detection in Sows during the Farrowing Phase by Inter-Birth Interval Evaluation. Comput. Electron. Agric. 2018, 152, 221–232. [Google Scholar] [CrossRef]
  26. Skovbo, D.K.F.; Hales, J.; Kristensen, A.R.; Moustsen, V.A. Comparison of Management Strategies for Confinement of Sows around Farrowing in Sow Welfare And Piglet Protection Pens. Livest. Sci. 2022, 263, 105026. [Google Scholar] [CrossRef]
  27. Singh, C.; Verdon, M.; Cronin, G.M.; Hemsworth, P.H. The Behaviour and Welfare of Sows and Piglets in Farrowing Crates or Lactation Pens. Animal 2017, 11, 1210–1221. [Google Scholar] [CrossRef]
  28. Andersen, I.L.; Berg, S.; Bøe, K.E. Crushing of Piglets by the Mother Sow (Sus Scrofa)—Purely Accidental or a Poor Mother? Appl. Anim. Behav. Sci. 2005, 93, 229–243. [Google Scholar] [CrossRef]
  29. Leenhouwers, J.I.; Wissink, P.; van der Lende, T.; Paridaans, H.; Knol, E.F. Stillbirth in the Pig in Relation to Genetic Merit for Farrowing Survival1. J. Anim. Sci. 2003, 81, 2419–2424. [Google Scholar] [CrossRef]
  30. Nowland, T.L.; van Wettere, W.H.E.J.; Plush, K.J. Allowing Sows to Farrow Unconfined Has Positive Implications for Sow and Piglet Welfare. Appl. Anim. Behav. Sci. 2019, 221, 104872. [Google Scholar] [CrossRef]
  31. Yun, J.; Swan, K.M.; Vienola, K.; Farmer, C.; Oliviero, C.; Peltoniemi, O.; Valros, A. Nest-Building in Sows: Effects of Farrowing Housing on Hormonal Modulation of Maternal Characteristics. Appl. Anim. Behav. Sci. 2013, 148, 77–84. [Google Scholar] [CrossRef]
  32. Manteuffel, C.; Hartung, E.; Schmidt, M.; Hoffmann, G.; Schön, P.C. Online Detection and Localisation of Piglet Crushing Using Vocalisation Analysis and Context Data. Comput. Electron. Agric. 2017, 135, 108–114. [Google Scholar] [CrossRef]
  33. Aeffner, F.; Wilson, K.; Martin, N.T.; Black, J.C.; Hendriks, C.L.L.; Bolon, B.; Rudmann, D.G.; Gianani, R.; Koegler, S.R.; Krueger, J.; et al. The Gold Standard Paradox in Digital Image Analysis: Manual versus Automated Scoring as Ground Truth. Arch. Pathol. Lab. Med. 2017, 141, 1267–1275. [Google Scholar] [CrossRef] [PubMed]
  34. Banerjee, M.; Fielding, J. Focus on Quantitative Methods Interpreting Kappa Values for Two-Observer Nursing Diagnosis Data. Res. Nurs. Health 1997, 20, 465–470. [Google Scholar] [CrossRef]
  35. Byrt, T.; Bishop, J.; Carlin, J.B. Bias, Prevalence and Kappa. J. Clin. Epidemiol. 1993, 46, 423–429. [Google Scholar] [CrossRef]
  36. Oczak, M.; Maschat, K.; Baumgartner, J. Dynamics of Sows’ Activity Housed in Farrowing Pens with Possibility of Temporary Crating Might Indicate the Time When Sows Should Be Confined in a Crate before the Onset of Farrowing. Animals 2020, 10, 6. [Google Scholar] [CrossRef] [PubMed]
  37. Oczak, M.; Maschat, K.; Berckmans, D.; Vranken, E.; Baumgartner, J. Classification of Nest-Building Behaviour in Sows on the Basis of Accelerometer Data. Biosyst. Eng. 2015, 140, 632–640. [Google Scholar] [CrossRef]
  38. Küster, S.; Kardel, M.; Ammer, S.; Brünger, J.; Koch, R.; Traulsen, I. Usage of Computer Vision Analysis for Automatic Detection of Activity Changes in Sows during Final Gestation. Comput. Electron. Agric. 2020, 169, 105177. [Google Scholar] [CrossRef]
  39. Mainau, E.; Manteca, X. Pain and Discomfort Caused by Parturition in Cows and Sows. Appl. Anim. Behav. Sci. 2011, 135, 241–251. [Google Scholar] [CrossRef]
Figure 1. Experimental set-up for three sows with two Bascom XD10-4 security cameras and three L642V sound cameras.
Figure 1. Experimental set-up for three sows with two Bascom XD10-4 security cameras and three L642V sound cameras.
Animals 13 03538 g001
Figure 2. The screen capture that was used for the analysis of behaviour and sound in the farrowing pen.
Figure 2. The screen capture that was used for the analysis of behaviour and sound in the farrowing pen.
Animals 13 03538 g002
Table 1. Sound spots and corresponding audible sounds that were considered correct for field study 1.
Table 1. Sound spots and corresponding audible sounds that were considered correct for field study 1.
Sound SpotAudible Sound
Head of the sowSow in crate
Rump of the sowSow in crate
FenceMetal fence
Head of the sowTrough
TroughTrough
Outside penNeighbour sow
Cloud of spotsNeighbour sow
Table 2. Sound spots and corresponding audible sounds and corresponding behaviours that were considered correct for field study 2.
Table 2. Sound spots and corresponding audible sounds and corresponding behaviours that were considered correct for field study 2.
Sound SpotAudible SoundSound SpotBehaviour
Head of the sow
Rump of the sow
Fence
Head of the sow
Trough
Outside pen
Cloud of spots
Front of the pen
Metal fence
Metal fence
Trough
Trough
Neighbour sow
Neighbour sow
Head of the sow
Head of the sow
Trough
Rump of the sow
Fence
Fence
Outside pen
Cloud of spots
None
Playing with sack
Eating
Eating
Standing
Standing
Moving leg lying
None
None
Lying down
Table 3. Results of visible and audible sounds before farrowing in the first field study (N = 4 sows, 177 min); sound spots without audible sound are false positives (FP), sound spots with audible sound are correct positive (CP); audible sounds without sound spot are false negative (FN) and 2 s of no audible sound and no sound spot indicates a correct negative (CN). If a sound spot was in the wrong location but at the correct time for an audible sound, it was considered either as a false negative (1st column) or as a correct positive (2nd column).
Table 3. Results of visible and audible sounds before farrowing in the first field study (N = 4 sows, 177 min); sound spots without audible sound are false positives (FP), sound spots with audible sound are correct positive (CP); audible sounds without sound spot are false negative (FN) and 2 s of no audible sound and no sound spot indicates a correct negative (CN). If a sound spot was in the wrong location but at the correct time for an audible sound, it was considered either as a false negative (1st column) or as a correct positive (2nd column).
Wrong Location Considered False NegativeWrong Location Considered Correct Positive
FP10,75110,751
FN182376
CP15093256
CN15821582
Accuracy19.730.9
Error%80.369.1
Sensitivity45.397.7
Specificity12.812.8
Table 4. Sound spots compared with audible sound, visible behaviour, or sound and behaviour combined.
Table 4. Sound spots compared with audible sound, visible behaviour, or sound and behaviour combined.
SoundBehaviourSound and Behaviour
# Sows422
Accuracy50.671.491.2
Error%49.428.68.8
% Correct positive35.457.571.3
Precision41.866.889.2
Sensitivity99.799.999.6
Specificity23.632.769.7
Table 5. Behaviour of sows (N = 4) four hours before farrowing, during farrowing and four hours after farrowing: percentage of time inactive, either lying or sitting/standing, or being active, playing with the jute sack or rooting on the floor.
Table 5. Behaviour of sows (N = 4) four hours before farrowing, during farrowing and four hours after farrowing: percentage of time inactive, either lying or sitting/standing, or being active, playing with the jute sack or rooting on the floor.
BehavioursBefore FarrowingDuring FarrowingAfter Farrowing
Lying71.7%93.5%100.0%
Sitting/standing3.3%6.2%0%
Jute sack/rooting floor25%0.3%0%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

van Erp-van der Kooij, E.; de Graaf, L.F.; de Kruijff, D.A.; Pellegrom, D.; de Rooij, R.; Welters, N.I.T.; van Poppel, J. Using Sound Location to Monitor Farrowing in Sows. Animals 2023, 13, 3538. https://doi.org/10.3390/ani13223538

AMA Style

van Erp-van der Kooij E, de Graaf LF, de Kruijff DA, Pellegrom D, de Rooij R, Welters NIT, van Poppel J. Using Sound Location to Monitor Farrowing in Sows. Animals. 2023; 13(22):3538. https://doi.org/10.3390/ani13223538

Chicago/Turabian Style

van Erp-van der Kooij, Elaine, Lois F. de Graaf, Dennis A. de Kruijff, Daphne Pellegrom, Renilda de Rooij, Nian I. T. Welters, and Jeroen van Poppel. 2023. "Using Sound Location to Monitor Farrowing in Sows" Animals 13, no. 22: 3538. https://doi.org/10.3390/ani13223538

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop