Ijraset Journal For Research in Applied Science and Engineering Technology
Authors: Harshada Aher, Sakshi Deore, Harsh Karwande, Bhagyesh Kale, Swamiraj Jadhav
DOI Link: https://doi.org/10.22214/ijraset.2024.65263
Certificate: View Certificate
Water pollution, particularly from non-biodegradable waste such as plastics, is one of the most pressing environmental issues facing aquatic ecosystems today. This paper presents an innovative solution for detecting and classifying waste in water bodies using machine learning techniques, specifically You Only Look Once (YOLO) and Convolutional Neural Networks (CNNs). The Water Trash Detection System (WTDS) can process both static images and real-time video streams to efficiently detect various types of aquatic trash, including plastic bottles, bags, and fishing debris. The system generates detailed reports on detected waste, providing environmental organizations and researchers with actionable data for monitoring and cleaning polluted water bodies. The combination of YOLO for real-time object detection and CNNs for waste classification allows the WTDS to deliver high accuracy and real-time processing, which is essential for effective environmental management.
I. INTRODUCTION
Water pollution has emerged as one of the most urgent environmental crises, largely driven by the vast accumulation of non-biodegradable waste, primarily plastics, in the world’s oceans, rivers, and lakes. Each year, millions of tons of plastic waste infiltrate aquatic ecosystems, posing severe risks not only to marine life but also to biodiversity, human health, and the stability of entire ecosystems. As plastic particles break down over time, they release toxic chemicals and microplastics, which are ingested by marine species, contaminating the food chain and reaching humans through seafood. Additionally, larger pieces of plastic waste create physical barriers and entrap aquatic species, resulting in loss of biodiversity and disruption of natural habitats.
To address this escalating issue, researchers and environmental organizations have developed various technologies, ranging from large-scale physical clean-up projects to advanced monitoring and tracking systems. Despite these efforts, the task of waste detection and collection in aquatic environments remains daunting. Physical clean-up operations are resource-intensive and often limited to specific areas, leaving vast water bodies unchecked. Manual detection of waste in water bodies is also challenging due to limited resources, vast spatial coverage, and time constraints, which make consistent and widespread monitoring difficult.
In response to these challenges, automated solutions leveraging advanced technology are gaining attention. The Water Trash Detection System (WTDS) presents an innovative approach to waste detection using machine learning and computer vision. This system harnesses the power of You Only Look Once (YOLO), a state-of-the-art object detection framework, and Convolutional Neural Networks (CNNs) to efficiently detect, classify, and locate various types of waste in water bodies. YOLO is specifically designed for real-time object detection, making it highly suitable for analyzing both static images and video streams. By utilizing YOLO for object detection, the WTDS can quickly and accurately identify waste items like plastic bottles, bags, and fishing nets, even in diverse aquatic conditions.
Beyond detection, the WTDS employs CNNs for classifying waste materials into specific categories, enabling a detailed understanding of the types of pollution present in any given water body. This classification aids in generating actionable data for environmental organizations and policymakers, empowering them to implement targeted clean-up strategies. Furthermore, WTDS supports live video stream analysis, allowing for real-time monitoring of waste in rivers, lakes, and coastal areas. This functionality makes it versatile and capable of continuous, large-scale environmental monitoring without the need for constant human intervention.
In essence, the WTDS bridges the gap between limited manual monitoring and the need for extensive waste detection across vast aquatic ecosystems. By automating the detection and classification of aquatic waste, the system not only enhances the efficiency and accuracy of pollution tracking but also provides valuable insights that can drive impactful environmental conservation efforts.
III. RELATED WORK
A. Machine Learning for Environmental Monitoring:
B. YOLO for Real-Time Waste Detection:
C. CNN for Waste Classification:
D. Real-Time Video Stream Processing:
E. Challenges in Aquatic Waste Detection:
III. DATASET
The primary focus of this study is on the classification of recyclable materials, as efficient waste management and recycling systems are crucial components of a sustainable economy. Replacing human labor with intelligent systems in waste sorting facilities is essential for enhancing both the safety and cost-effectiveness of recycling operations.
In line with this objective, our system is designed to identify common recyclable materials, including glass, paper, cardboard, plastic, metal, and general waste. We use the Trash Net dataset, which includes six categories of recyclable items. The dataset comprises images with a consistent white background, capturing various items of waste and recycling materials collected around Stanford. Each image showcases objects in diverse poses and lighting conditions, providing a range of variations within the dataset. The images are resized to 512 x 384 pixels, with the total dataset size reaching approximately 3.5GB. Table I below lists the number of images for each category, and Figure 1 provides sample images from the dataset.
|
|
|
|
|
|||
(a) Plastic |
(b) Glass |
(c) Paper |
(d) Metal |
|
|||
|
|
|
|
||||
(e) Metal |
(f) Cardboard |
(g) Trash |
(h) Cloth |
||||
Figure. Sample Images of Dataset
IV. METHODOLOGY
A. YOLO for Object Detection
B. CNN for Object Classification
Once objects are detected by YOLO, the system employs a Convolutional Neural Network (CNN) to classify the detected objects into specific waste categories. CNNs are particularly effective for image classification due to their ability to learn spatial hierarchies of features automatically. The WTDS uses a CNN architecture pre-trained on a large dataset of aquatic waste images, which includes various types of waste such as plastics, fishing gear, and metal debris.
The classification process involves feeding the detected objects into the CNN, which then predicts the specific type of waste. The system’s accuracy improves with each additional training cycle and dataset expansion.
C. System Architecture
The system is designed around two main components:
System Architecture
V. SYSTEM DESIGN AND IMPLEMENTATION
A. User Interface
The user interface of the WTDS is designed to be intuitive and easy to navigate. Users can upload images or provide real-time video streams from webcams or IP cameras. The results are displayed in a dashboard, which includes a list of detected waste items, their locations, and type classifications.
The system also provides a history of previously detected waste, allowing users to track pollution trends over time. This data can be exported for further analysis or used to generate reports for environmental organizations.
B. Data Collection and Training
A key aspect of the system's performance is the dataset used to train the YOLO and CNN models. To build a robust and diverse dataset, images of various aquatic environments—rivers, lakes, and coastal areas—were collected, along with different types of waste such as plastic bottles, bags, fishing nets, and organic debris.
The training process involved manually labeling these images to create ground truth data, which was used to train both the YOLO and CNN models. The system was fine-tuned with additional data to improve the accuracy of both detection and classification.
C. Real-Time Video Processing
In addition to analyzing static images, the WTDS supports real-time video stream processing. The system processes video frames at a rate of up to 30 frames per second (FPS), making it suitable for live monitoring of water bodies. By analyzing each frame as it is captured, the WTDS can detect and classify waste objects in real time, providing immediate feedback to environmental agencies or researchers.
VI. EQUATIONS
A. YOLO Object Detection Equation
The YOLO (You Only Look Once) model works by predicting bounding boxes and class probabilities directly from the image. YOLO divides the image into N × NN × NN × N grid. All cell phone predictions BBB tie box and confidence score.The output is a set of class probabilities and bounding box coordinates.
The equation for YOLO's bounding box prediction can be written as:
y^?i?=(x,y,w,h,C,P1?,P2?,…,Pk?)
Where:
x,yx, yx,y = Coordinates of the bounding box center (relative to the grid cell)
w,h, = width and height of the bounding box (relative to the entire image)
C = confidence level (specify whether the box contains an object)
P1,P2,…,Pk = Class probabilities for each object type (e.g., plastic bottle, plastic bag, etc.)
Confidence Score: The confidence score C is calculated as:
C=Pobject?⋅IOU
Where:
B. CNN Loss Function
For the CNN (Convolutional Neural Network) component used for waste classification, a standard cross-entropy loss function is often used. The cross-entropy loss LCE for a multi-class classification problem is:
LCE?=−i=1∑N?yi?log(pi?)
Where:
This loss function helps minimize the difference between the true labels and the predicted labels, improving the classification accuracy of waste items.
C. Model Evaluation Metrics (Precision, Recall, and F1-Score)
To evaluate the performance of the search and classification model, precision, recall and F1 score
are commonly used metrics.
These can be defined as:
Where:
Where:
F1-Score (harmonic mean of precision and recall):
These metrics are critical for assessing the overall performance of the detection and classification system in terms of both accuracy and the ability to detect all objects of interest
VII. CHALLENGES
Exploring these challenges could enrich the research on this system and provide practical insights for environmental monitoring.
VIII. FUTURE SCOPE
The Water Trash Detection System (WTDS) has significant potential for future improvements and research in several key areas:
In this paper, we presented the Water Trash Detection System, an innovative solution for identifying and classifying waste in aquatic environments using YOLO and CNN technologies. The system is designed to process both static images and real-time video streams, offering efficient, scalable, and accurate detection of water pollution. By automating the detection and classification of waste, such as plastic bottles, bags, and other debris, the WTDS aids in environmental conservation efforts, providing actionable insights for both researchers and organizations working to combat water pollution. While the system demonstrates strong performance in surface-level waste detection, there remain several challenges that require further research, such as detecting submerged waste and adapting to dynamic environmental conditions like varying water turbidity, waves, and lighting. Future work will focus on integrating multi-sensor data, enhancing real-time processing capabilities, and refining waste classification to cover a broader range of materials. Additionally, improving the user experience through intuitive interfaces and mobile apps will broaden the system’s accessibility, encouraging wider adoption by environmental agencies and local communities. Ultimately, the WTDS represents a significant step toward the automation of water pollution monitoring, offering a valuable tool for global environmental conservation efforts. Continued advancements in deep learning, sensor technologies, and real-time processing will further enhance its potential to address the growing challenge of aquatic waste, contributing to cleaner, healthier water bodies worldwide.
[1] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, \"Koj tsuas pom nws ib zaug: Joint real-time object detection,\" Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), 2016, pp. 779–788. [Online]. Available: https://arxiv.org/abs/1506.02640 [2] D. P. Kingma and J. Ba, “Adam: A stochastic optimization approach,” presented at the International Conference on Learning Representations (ICLR), 2015. [online] URL: https://arxiv.org/abs/1412.6980 [3] Y. Tian, S. Wu, and P. S. Yu, “Deep learning for environmental monitoring: A review and future perspectives,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, vol. 49, no. 4, pp. 790-804, 2019. DOI: 10.1109/TSMC.2018.2852571 [4] Li Xiaowei Wu, A.M.B.C. Tock and R.S.S.C. Ang, “Detecting and classifying marine debris using deep learning and drone imagery,” Environmental Monitoring and Assessment, vol. 192, no. 20:12 - Wednesday 12, 2020. Thacker, i. H. and Rashed, A. (2020). \"Deep learning-based identification of debris in aquatic environments.\" Sensors, 20(12), 3471. [5] L. Zhou, T. Xie, and M. Zhou, “Plastic waste detection in marine environment using deep learning from aerial imagery,” International Journal of Environmental Research and Public Health, vol. 17, no. 17, pp. 6264-6279, 2020. DOI: 10.3390/ijerph17176264 [6] R. Prakash, G. J. V. K. Reddy, and B. S. Pradeep, “Marine debris detection using drone imagery and convolutional neural networks,” Marine Pollution Bulletin, vol. 154, pp. 111059, 2020. DOI: 10.1016/j.marpolbul.2020.111059 [7] Y. Yuan, Y. Cheng, Y. Yang and W. Zhang, “Instant detection of floating debris using YOLOv3 and deep learning,” IEEE Access, vol. 9 Ib., 10951-10961, 2021. [8] H. Shen, F. Yang, X. Zhang, and H. Zhang, \"Degradation of marine debris using deep diffusion methods: A comparative study\", Marine Pollution Bulletin, vol. 151 Ib., p. 110789, 2020. [9] A. Bui, T. Nguyen, and T. Pham, “Plastic waste detection and classification using drone-based images and deep learning techniques,” Journal of Environmental Management, vol. 246, pp. 830-839, 2019. DOI: 10.1016/j.jenvman.2019.05.059 [10] A. Bui, T. Nguyen ve T. Pham, \"Waste detection and classification using dronebased imagery and deep learning,\" Journal of Environmental Management, Vol. 246, pp. 830-839, 2019. [11] P. Vidyasagar, V. B. Rao, and R. C. Gupta, \"Instant waste detection and monitoring of water bodies using dronebased imagery and artificial intelligence,\" Sensors, vol. 20. No. 21 Ib., 6150, 2020. [12] Thacker, I. H. and Rashed, A. (2020). \"Deep learning based identification of debris in aquatic environments.\" Sensors, 20(12), 3471. [13] Kumar, S. and Kumar, R. (2021). \"Image processing and deep learning techniques for water pollution detection.\" Journal of Water and Climate Change, 12(4), 1201-1215. DOI: 10.2166/wcc.2020.168. [14] Mohamed, S. and Abdo, A. (2023). \"A smart water monitoring system leveraging deep learning and Internet of Things.\" Journal of Environmental Management, 320, 115870. [15] Choudhury, S.R., and Shirin, A. (2020). \"Application of Convolutional Neural Networks to Marine Debris Analysis: Marine Pollution Bulletin, 150, 110684. [16] Khan, M. A., & Arif, M. (2021). \"Real-Time Detection of Floating Waste in Water Using Deep Learning.\" Journal of Environmental Engineering and Science, 6(1), 67-75. DOI: 10.1080/24150072.2021.1873456.
Copyright © 2024 Harshada Aher, Sakshi Deore, Harsh Karwande, Bhagyesh Kale, Swamiraj Jadhav. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Paper Id : IJRASET65263
Publish Date : 2024-11-14
ISSN : 2321-9653
Publisher Name : IJRASET
DOI Link : Click Here