Site icon aviNews International, poultry information

Machine Vision Technologies for Monitoring Poultry Welfare

Escrito por: Lilong Chai

BACKGROUND ON POULTRY WELFARE

Poultry production plays a critical role in feeding the increasing world’s population with affordable protein (i.e., chicken and eggs). The United States is currently the world’s largest broiler producer and 2nd largest egg producer due to continuous innovation in animal breeding, nutrition management, environmental control, and disease prevention, etc.

However, US poultry and egg farms are facing several production challenges such as animal welfare concerns.

To address those issues, researchers at the University of Georgia (UGA, Dr. Lilong Chai’s precision poultry farming lab) developed several precision farming technologies for monitoring welfare and behaviors of broilers and cage-free layers.

BROILERS’ FLOOR DISTRIBUTION AND BEHAVIORS

The spatial distribution of broiler chickens is an indication of a healthy flock or not. Routine inspections of broiler chickens’ floor distribution are done manually in commercial houses daily or multiple times a day, which is labor intensive, time consuming, and subject to farm staff’s errors.

Figure 1. A top view of a pen and zone definition. The redbox (1) drinking zone; (2) feeding zone; and (3) resting/exercise zone.

Figure 2. Data collection and analysis.

Figure 3. Behavior classification results of the Densenet-264 model in broilers dataset.

Figure 4. Broilers detection on the floor.

CAGE FREE LAYERS’ PECKING, MISLAYING, AND DISTRIBUTION

Major restaurants and grocery chains in the United States have pledged to buy cage-free (CF) eggs only by 2025 or 2030.

Machine vision methods were developed and tested in tracking chickens’ floor and spatial distribution (Figure 5 and Figure 6), and identifying pecking behaviors of hens and potential damages (Figure 7 and Figure 8) in research cage-free facilities at UGA. The YOLOv5x-pecking model was tested with a precision of 88.3% in tracking pecking.

Figure 5. Number of chickens identified under low density and moderate density by our model: low density (a) vs. moderate density (b).

Figure 6. Number of chickens identified under horizontal angle and vertical angle by our model: horizontal angle (a) and vertical angle (b).

Figure 7. Pecking behavior and damages in layers.

Figure 8. Performance of YOLOv5-pecking deep learning model in pecking detection: a – pecking in a rest zone, b – pecking in a feeding zone, c –pecking in a drinking zone; d – two birds are pecking one bird (i.e., the same bird in c was pecked by the two birds at the same time).

In addition, about 5400 images were collected and used to train another deep learning model (i.e., YOLOv5m-FELB – floor egg laying behavior), which reached 90% of precision (Figure 9).

Figure 9. The floor egg laying behaviors detected in test data using the YOLOv5s model for different hen proportions: a) individual hen detection; and b) group hens detection.

Besides, the method could also be used to detect or scan floor eggs (Figure 10).

Figure 10. Floor egg scanning with machine vision.

SUMMARY

Different machine vision or deep learning methods were developed at the University of Georgia’s poultry science department to monitor broiler and cage-free layers’ welfare and behaviors.

Those findings provide references for developing precision poultry farming systems on commercial broiler and egg farms to address
poultry production, welfare, and health associated issues. Dr. Lilong Chai’s projects were sponsored by USDA-NIFA, USDA ARS, Egg Industry Center, Georgia Research Alliance, UGA, Oracle, and poultry companies, etc.

Bibliography available upon request

 

PDF
Exit mobile version