Next Article in Journal
Genetic Algorithm-Based Online-Partitioning BranchyNet for Accelerating Edge Inference
Previous Article in Journal
A Financial Management Platform Based on the Integration of Blockchain and Supply Chain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization of Trash Identification on the House Compound Using a Convolutional Neural Network (CNN) and Sensor System

1
Institute of IR4.0, Universiti Kebangsaan Malaysia, Bangi 43600, Malaysia
2
Faculty of Computer Science, Universitas Putra Indonesia YPTK Padang, Padang 25221, Indonesia
*
Author to whom correspondence should be addressed.
Submission received: 10 November 2022 / Revised: 17 December 2022 / Accepted: 22 December 2022 / Published: 29 January 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
This study aims to optimize the object identification process, especially identifying trash in the house compound. Most object identification methods cannot distinguish whether the object is a real image (3D) or a photographic image on paper (2D). This is a problem if the detected object is moved from one place to another. If the object is 2D, the robot gripper only clamps empty objects. In this study, the Sequential_Camera_LiDAR (SCL) method is proposed. This method combines a Convolutional Neural Network (CNN) with LiDAR (Light Detection and Ranging), with an accuracy of ±2 mm. After testing 11 types of trash on four CNN architectures (AlexNet, VGG16, GoogleNet, and ResNet18), the accuracy results are 80.5%, 95.6%, 98.3%, and 97.5%. This result is perfect for object identification. However, it needs to be optimized using a LiDAR sensor to determine the object in 3D or 2D. Trash will be ignored if the fast scanning process with the LiDAR sensor detects non-real (2D) trash. If Real (3D), the trash object will be scanned in detail to determine the robot gripper position in lifting the trash object. The time efficiency generated by fast scanning is between 13.33% to 59.26% depending on the object’s size. The larger the object, the greater the time efficiency. In conclusion, optimization using the combination of a CNN and a LiDAR sensor can identify trash objects correctly and determine whether the object is real (3D) or not (2D), so a decision may be made to move the trash object from the detection location.

1. Introduction

A clean house compound is a dream for everyone. Trash can be caused by leaves falling from trees or plastic trash such as plastic bottles and snack packets. Usually, trash on the house compound is cleaned by the owner or the person assigned. If homeowners are busy and do not have time to clean the house compound, then the house will look unclean because trash is scattered in the house compound.
Currently, the state considers essential urban services, such as water, sanitation, and solid waste management, to be the responsibility of local or national governments [1,2,3,4]. Research on the identification and classification of trash has been performed [5,6,7]. However, it was not optimal regarding the amount of trash detected and accurate trash detection. At the same time, optimization plays an essential role in computer vision because many computer vision algorithms employ an optimization step at some point in their proceeding [8]. A study was conducted by Fuchikawa in 2005 on trash collection using an OSR robot (Outdoor Service Robot) [9]. Unfortunately, this research was only aimed at collecting plastic trash in PET bottles. Another study was conducted by an Italian research group led by Mazzolai [10]. They named the robot DustCart. This robot groups junk types based on user input. After the user inputs the type of trash, the robot opens the trashed store according to the type of trash input. However, robot cameras are for navigation, avoiding obstacles, and not classifying trash types.
Improvements in the reliability and processing speed of the vision system and experiments with other trash must be conducted [9]. Improvement was achieved at Stanford University through a study on trash classified using the CNN (Convolution Neural Network) method [11]. However, the amount of trash that can be identified is only up to six items of trash [12,13,14], while there are more than six items in the house compound.
Identification of trash is an essential step before separation, and this can be performed efficiently with the help of different machine-learning and image processing algorithms. A Convolutional Neural Network (CNN) is preferred for the classification of images [15]. However, most methods of identifying objects cannot distinguish whether the object is a real (3D) image or an image in the form of a photograph on paper (2D) [12,16,17,18,19,20]. Figure 1 shows a test object (mouse) in 2D and 3D. The object is detected using a Convolutional Neural Network (CNN). The test results show that 2D and 3D (real) objects can be detected well, with a prediction accuracy of 82%. In Figure 1, the 2D object is located on the right and the 3D object is located on the left.
Then, each object is detected, 2D objects and 3D objects, using the Convolutional Neural Network (CNN). An Intel Core i7 (8th gen) laptop, with 16 GB RAM, a NVIDIA GeForce RTX2060 graphics card and MATLAB 2020a software was used. For 2D objects, the samples are photos of objects on paper. The test results for 2D objects can be seen in Figure 2a. The test results were obtained with a prediction accuracy of 98%.
Furthermore, testing was carried out for 3D objects (real). This object was placed next to the 2D object. The test results can be seen in Figure 2b. The test results were obtained with a prediction accuracy of 91%. From these three test results, it can be said that the Convolutional Neural Network (CNN) can detect objects well. This will be a problem if the object is moved from one place to another using the robot gripper. How would it be if the robot wanted to move the trash object while the trash object was just a photograph of trash on paper? A robot must identify/recognize real (3D) trash objects.
In any application, we can use a particular classifier and try to optimize its performance. The usual approach is to try several different classifiers and choose the one that performs the best on a separate validation set [21]. In this case, we propose a Convolutional Neural Network (CNN) combined with LiDAR sensors to identify whether the object is real (3D) or not (2D). A CNN was chosen because, at present, a CNN is the best machine-learning method in object identification. Therefore, this study proposes optimizing trash identification in the house compound using a Convolutional Neural Network (CNN).
The proposed research has similarities with sensor fusion. In sensor fusion, several sensors are combined to take measurements. The goal is that the resulting data are more accurate than using only one sensor. In this study, a camera and a LiDAR sensor were used. Many studies use a combination of LiDAR sensors and cameras. Table 1 shows a comparison of their use [22].

2. Materials and Methods

In this study, data collection is performed using own trash dataset, the TrashNet dataset [11], and the Leafsnap dataset [32]. These data are grouped into 11 types of data, namely, cardboard, fabric, food packaging, fruit, glass, leaf, metal, paper, plastic, rubber, and wood. Figure 3 shows sample images of the trash dataset. The TrashNet dataset was created by Gary Thung and Mindy Yang. This is a small dataset and contains 1989 images. The LeafSnap dataset was created by Kumar et al. [33]. The dataset consists of 7500 field-augmented images [34]. Own trash datasets were captured by a mobile device (iPhone). The image data are augmented by a laptop using MATLAB R2020a software. This own dataset is used to complement the TrashNet dataset and LeafSnap dataset to obtain 11 types of data.
The data were tested with four pre-trained Convolutional Neural Network (CNN) architectures: AlexNet, VGG16, GoogleNet, and ResNet18. The training parameters can be seen in Table 2.
The training parameters used are those with standard values already existing in the pre-trained CNN architecture. After the data are collected, the next step is to capture the trash object. Figure 4 depicts the block diagram of the proposed system.
After the IP camera captures the image data, the MATLAB program resizes the image. It is helpful for further processing on CNN Image Detection. The CNN architecture used can classify up to 1000 types of objects. However, modifications are made to the feature learner to speed up classifying trash objects. The original value is 1000 fully connected layers changed to 11 fully connected layers. This value is changed on every architecture—AlexNet, VGG16, GoogleNet, and ResNet18. Figure 5 depicts transfer learning using the CNN architecture. In the Figure, it can be seen that the first step of the trash dataset is loading.
Furthermore, this dataset is divided into training and validation datasets with a 70:30 ratio. Then, this dataset image is resized according to the size of the input layer architecture used. For AlexNet, the image size is 227 × 227. For VGG16 GoogleNet and ResNet18, the image size is 224 × 224, respectively. The next step is to modify the feature learner layer and classification layer. Fully connected, it is changed from 1000 to 11, so the output from the classification layer will automatically be 11 classes.
In MATLAB, the modification process can be performed by changing the architectural design using Deep Network Designer. Figure 6 shows the process of modifying the GoogleNet architecture.
Furthermore, the MATLAB program ensures that an image is a trash object. If the object is not trash, it is ignored. If it is trash, the laptop (PC) sends a command to Arduino to activate the actuator (servo1 and servo2). This command is based on the detected object bounding box value. The actuator moves the LiDAR sensor in X and Z coordinates based on the value. After that, the data from the LiDAR sensor are sent by Arduino to the laptop (PC). The proposed flowchart can be seen in Figure 7.
The Mapping Bounding box is used to convert the movement of the servo motor to match the object’s size to be detected, as shown in Figure 8 [35]. After obtaining the x and z, it is calculated/mapped to the degree of motor servo movement. In the scanning process, servo motor 1 (x) and servo motor 2 (z) move at the boundary in the bounding box.
In the designed system, the LiDAR sensor is mounted on top of a servo motor 1. The scanning process is then carried out using a LiDAR sensor. The scanning process is carried out in 2 stages, namely, the fast and detailed scanning processes. This is to achieve an efficient scanning time. If the detected object is flat, then the object is a 2D object. If it is 2D, the detailed scanning process does not need to be performed. However, if the object is not flat, the detected object is a 3D object. If the object is 3D, a detailed process needs to be performed. It aims to determine with certainty the geometric shape of the object. With the object’s geometry data, the robot gripper can easily lift trash objects.

2.1. TF40 LiDAR

TF40 is an millimeter-level accuracy LiDAR with a range of up to 40 m [36,37,38]. Accuracy in mm is essential because the trash object is relatively small, 15 cm × 20 cm. It is adjusted to the width of the robot gripper. TF40 has the following features: high accuracy, tiny, small light spot, visible laser, and easier aiming. Table 3 is the main parameter of TF40, and Figure 9 shows the physical form of the TF40 LiDAR sensor and its dimensions.

2.2. Fast Scanning Image Using LiDAR

The scanning process is fast, from the top left to the bottom right. The fast scanning process is carried out five times according to the path in Figure 10.
The height of the bounding box (z) can be different depending on the size of the detected trash object. However, the value of the degree of movement of the servo motor 2 (z) in this fast scanning system is obtained from the height of the bounding box divided by five, as shown in the following formula:
Servo   motor   2   ( z ) degree   = Height   of   bounding   box 5
After obtaining the value of x and z based on the bounding box, it is calculated (mapped) to the degree of motor servo movement. Point x0 line z0 in Figure 10 is the starting point of the fast scanning image process. If the position of the servo motor 1 and servo motor 2 is not at that point, then the program contained in the Arduino will move them to that position.
After that, servo motor 1 (x) and servo motor 2 (z) move, as shown in Figure 11 in line z0. In the initial conditions, the servo motor will move from point x0 line z0 to point xn line z0. Then, the servo motor will move down along the servo motor 2 (z) degrees. For example, if the height of z is 100 degrees, then the servo motor 2 (z) will move down 20 degrees (100 degrees/5 = 20 degrees). The next step is servo motor 1 (x) will move along z1 in the direction of movement opposite to line z0. The movement of the servo motor is continued by following lines z1, z2, z3, and z4 and arriving at line z5 at point x0.
In Figure 11, the position and movement of the LiDAR sensor during the scanning process are shown. The following algorithm determines whether the detected object has flat, concave, or convex sides.
  • If sensor_value = hypotenuse, then the line/point of the image is a flat plane;
  • If sensor_value > hypotenuse, then the lines/dots of the image become concave;
  • If sensor_value < hypotenuse, then the lines/dots of the image become convex.
In the fast scanning process, the data read is time data (Time Stamp) and distance data (Distance). Figure 12 is an example of a graph of the LiDAR TF40 sensor reading.
Figure 13 is a flowchart of a fast scanning image. This flowchart is in the form of a procedure that is called after the MATLAB program detects the trash object.
This fast scanning image flowchart is used on Arduino. In the early stages, initialization of the variables and constants that will be used is carried out. Next, the program will read the value of the bounding box pixel (width, height). After this value is known, bounding box mapping is performed. The results of this mapping are in the form of x and z values, which are the basic values for the degree of movement of the servo motor 1 and servo motor 2. The next step is to read the servo motor’s position (1) and servo motor 2. If the position is not yet in the initial position, then the servo motor 1 and servo motor 2 will be driven in that initial position. Then, the zj value is divided by 5. If the zj value is a fractional number, then this value will be rounded off with the int(zj) command. The servo motor can only move in integer degrees (not floating point degrees). The next step is to map the movement of the servo motor 1 and servo motor 2.
The servo motor uses this mapping to move from point x0 line z0 to point x0 line z5. Simultaneously with the movement of the servo motor along the z0, z1, z2, z3, z4, and z5 lines, LiDAR sensor data are also read and stored in Arduino memory. After the servo motor 1 and servo motor 2 arrive at point x0 line z5, the LiDAR sensor data are sent to the MATLAB program. In MATLAB, the LiDAR sensor data are processed. The result can be 2D or 3D. By the flowchart of the proposed study in Figure 7, if it is 2D, it is ignored. If it is in 3D, then the next step is to do a detailed scanning image process.

2.3. Detail Scanning Image Using LiDAR

Detail scanning is helpful for the process of taking objects using a robot gripper. If the gripper is not positioned correctly, the lifting process may fail. The scanning process is almost the same as fast scanning, but the degree of movement of the servo motor has been determined from the start, which is 2 degrees. The details can be seen in Figure 14.
The difference in the process path between a fast scanning image and a detailed scanning image lies only in the number of lines (z). In fast scanning images, the number of lines is limited to 5 lines. Meanwhile, the minimum number of lines for detailed scanning images is five. If the spacing between rows is limited to 2 degrees, then the degree of movement of the servo motor 2 is at least 10 degrees.
Figure 15 is a flowchart of a detail scanning image. This flowchart is in the form of a procedure that is called after the MATLAB program detects the trash object. This flowchart is similar to the flowchart in Figure 13. The difference is that the scanning process will stop if xi = x0, and zj = zn. Another difference is that the degree of movement of the servo motor 2 in fast scanning is determined by the z value divided by 5. Meanwhile, in the detail scanning image, the degree of movement is equal to 2 degrees.
In this study, four parameters were used to evaluate the accuracy of the model, namely accuracy (Ac), precision (Pr), recall (Re), F1 score (F1) [39]. The formula can be seen in the following equation.
Accuracy   = T P + T N T P + F P + T N + F N
Precision   = T P T P + F P
Recall   = T P T P + F N
F 1   Score   = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l
where TP is short for the number of true positives. TN is short for the number of true negatives. False positive is defined as FP. FN presents false negatives.

3. Results

Image data were tested for 11 types of trash using four CNN architectures, namely AlexNet, VGG16, GoogleNet, and ResNet18. The number of images for each type of trash is 150 pieces. The data are augmented into 1200 images, so the total image is 13,200. The data are divided into 70% training data and 30% test data.
Figure 16 is the result of training progress for GoogleNet. The validation accuracy values for each CNN architecture can be seen in Table 4.
Table 4 shows that the CNN GoogleNet architecture has the highest validation accuracy value, while the CNN AlexNet architecture has the lowest validation value. However, in terms of training time, the fastest training time was achieved by AlexNet and the slowest training time was the VGG16 architecture.

3.1. Confusion Matrix for Trash Classification Testing

This trash object identification system is tested for its performance using a confusion matrix. The four pre-trained CNN architectures (AlexNet, VGG16, GoogleNet, and ResNet18) were tested according to their respective architectures. Figure 17 shows the results of the confusion matrix with VGG16.
Table 5 is the result of comparing the accuracy of the confusion matrix. The table shows that the AlexNet architecture has the lowest accuracy value, namely 80.5%, while the GoogleNet architecture has the highest value, namely 98.3%.

3.2. Trash Identification Test Using a Real-Time Camera

After the training process is complete, the identification test of the trash object is carried out. The test is carried out in real-time using a mobile device (iPhone) camera. The IP address used during testing is as follows.
Camera = ipcam (‘http://192.168.43.28:8080/video’ accessed on 18 October 2022);
This address can be seen on the screen display of mobile device (iPhone) as shown in Figure 18. Tests of the 11 types of trash can be seen in Figure 19, Figure 20 and Figure 21. This test uses the AlexNet architecture.
At the time of testing, there was also an error in identifying the trash object, as shown in Figure 22. The trash object was identified as glass, even though the trash object was plastic.
The tests performed in Figure 19, Figure 20 and Figure 21 are used to test the performance of pre-trained CNN architectures (AlexNet, VGG16, GoogleNet, and ResNet18) in identifying trash in real time. The number tested was 411 pieces of trash. Then, record the total trash detected for each type of trash. This test used trash that had never been used in training. Table 6, Table 7, Table 8 and Table 9 show the results of the accuracy obtained by each pre-trained CNN architecture, while Table 10 shows the average accuracy that can be achieved by each pre-trained CNN architecture.

3.3. Result of a Fast Scanning Image

A fast scanning image is used to quickly ensure that the identified image is an image in 2D or 3D. The results of the fast scanning image can be seen in Figure 23.
In Figure 23, the results of the fast scanning image divide the value of the z-axis into 5. This value comes from the servo motor 2.

3.4. Result of Detail Scanning Image

After the fast scanning process states that an observed object is an object in 3D, the detailed scanning image process is carried out. The function of this detail scanning image is to ensure the robot gripper position when lifting trash.
In Figure 24, it can be seen that the detailed scanning image results are better than the fast scanning image results. This is because the amount of data on the y-axis and z-axis are more than the amount of data on the y- and z-axis in fast scanning images.

3.5. Result of Time Speed Comparison between Fast Scanning Image and Detail Scanning Image

One of the optimizations in the trash identification process is to make time efficient in identifying trash objects. Therefore, it is necessary to examine the difference between fast and detailed scanning images, as depicted in Figure 25. The efficiency of time consumption used in the trash identification process will be obtained from the results of this test.

4. Discussion

In the initial experiment, the total images entered were 690 images with details: cardboard 50, fabric 15, food packaging 50, fruit 50, glass 50, leaf 165, metal 65, paper 100, plastic 120, rubber 15, and wood 15. In the results of training progress, accuracy ranges from 65.32% to 74.51%. Accuracy results on the confusion matrix ranged from 84.2% to 92.1%.
Furthermore, to increase the low accuracy value, the number of each data image in each class needs to be added, especially on images less than 20. Images of each class are added to a minimum of 150 images per class. Then, each image is augmented seven times so that each class has 1200 images. There is an increase in the training progress, with accuracy ranging from 77.54% to 86.38%. The accuracy in the confusion matrix increased from 80.5% to 98.3%. Thus, data images must be added to increase the accuracy of identifying the trash object.
Based on Table 4 and Table 5, it can be seen that there is a linear correlation between the resulting accuracy values. The lowest accuracy value in Table 4 is also the lowest accuracy value in Table 5. Likewise, the highest accuracy value in Table 4 is also the highest accuracy value in Table 5. Thus, without testing the value of the confusion matrix, we can also predict the accuracy results on the confusion matrix by looking at the value of validation accuracy.
In the real-time trash identification test, there is a decrease in accuracy from each pre-trained CNN. The lowest accuracy is generated by AlexNet, which is equal to 79.410%, and the highest accuracy is generated by GoogleNet with accuracy of 96.513%. This is caused by the lighting factor when detecting the trash object. However, this accuracy value is already good for the identification process and can be continued for the next process, namely the 2D or 3D determination of the identified trash.
In Table 4, it can be seen that the accuracy of GoogleNet is higher than ResNet18, but ResNet18 has a much shorter training time compared to GoogleNet. This can also be taken into consideration in choosing the pre-trained CNN to be used. Moreover, if the system is made based on embedded systems.
Figure 11 illustrates the position of the LiDAR sensor during the scanning process. Based on the figure, there are three formulas used, namely:
  • If the object is straight in front of the LiDAR sensor, then the y-coordinate:
y = d i s t a n c e   v a l u e   t h a t   m e a s u r e d   b y   L i D A R   s e n s o r
2.
If the object is on the front left side/the front right side of the LiDAR sensor, the y-coordinate value can be calculated by the formula:
y = C o s θ × d i s t a n c e   m e a s u r e d   b y   L i D A R   s e n s o r
3.
The formula can calculate the x-coordinate value:
x = Degree   of   servo   Motor   1 90
The value of 90 is due to servo motor 1 being set at 90 degrees. After all the x values are read and stored in the matrix variable, then the x value is added to the maximum x value with the formula:
x = x + maximum   value   of   x
The formula aims to make the x-coordinate values all positive.
After the x-coordinate and y-coordinate values are obtained, then the z-value is obtained by the formula:
z = Degree   of   servo   Motor   2 90
After all the z values are read and stored in the matrix variable, then the z value is added to the maximum z value with the formula:
z = z + maximum   value   of   z
With a LiDAR accuracy of ±2 mm, the results of detail scanning images displayed using MATLAB are similar to the original object. However, because what is scanned is only the surface in front of the LiDAR sensor, the detailed scan only shows the object’s surface. At first, the data generated by the LiDAR sensor were still in the form of numbers containing the y-coordinate data after combining the data with the x coordinates (servo motor 1) and z coordinates (servo motor 2).
Scanning images in detail is relatively time-consuming because the frame rate capability of the LiDAR sensor is only 5 Hz. If every time it finds an object that is considered trash, the robot must scan a detailed image, so the decision to lift the object will take longer. For this reason, the system is equipped with fast scanning. With this fast scanning, if the object considered trash is not in 3D shape, then the lifting process by the robot gripper is ignored. This time efficiency can be performed, and the robot can search for the next trash object.
In this study, the Sequential_Camera_LiDAR (SCL) method is proposed. This method is relatively simple, combining cameras and LiDAR sensors. A CNN processes the camera’s output, followed by a fast scanning process by the LiDAR sensor. A comparison of the proposed method with the previous method is in Table 1. However, this proposed study can be used in a house compound. The proposed method differs from the previous method in Table 1 (Early Fusion, Sequential Fusion, Late Fusion). These three methods are applied to vehicles that must have a relatively high speed compared to the speed of the robot. In addition, the vehicle’s response to obstacles in front of the camera is relatively faster than the robot’s movement in the house compound. So the proposed method is suitable for use in housing, but it is necessary to adjust the existing parameters for use on highways.
Indeed, in the real world, there is a possibility that 2D paper is also trash. However, the robot in this study has the disadvantage of not being able to lift 2D objects, such as 2D paper. It is due to the geometry of the robot gripper itself. It will be a gap for further research. In future studies, the robot gripper must be equipped with a suction system to lift 2D paper-shaped objects by being sucked on.

5. Conclusions

The optimization of the trash object detection system has been successfully carried out by using a fast scanning system based on bounding boxes. The time efficiency ranges from 13.33% to 59.26%, depending on the size of the detected object. The larger the object, the greater the time efficiency. The bigger the object, the more time it takes. Testing is limited to objects of 15 cm × 20 cm. To overcome this, in future research, we will try to use stereo LiDAR so that the scan time can be faster.
In testing the identification of trash objects using several CNN architectures—AlexNet, VGG16, GoogleNet, and ResNet18—all have a trash object identification accuracy of 80.5%, 95.6%, 98.3%, and 97.5%, respectively. This system uses a LiDAR sensor to ensure that the object is real or not real. The results of LiDAR scanning in graphic form can be adequately produced because the LiDAR sensor has a reading accuracy of 2 mm. The graph is the basis for determining the robot gripper position in lifting the trash object. Future research will focus on the trash collection process based on the resulting graph.

Author Contributions

Supervision, R.S. and N.M.A.; draft preparation, E.N., R.S. and N.M.A.; review and editing, E.N., R.S. and N.M.A. All authors have read and agreed to the published version of the manuscript.

Funding

This paper was supported by the grant: Institute of IR4.0 UKM, GP-K007341.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Baud, I.; Post, J.; Furedy, C. (Eds.) Solid Waste Management and Recycling; Springer: Dordrecht, The Netherlands, 2004; Volume 76. [Google Scholar] [CrossRef]
  2. Kshirsagar, P.R.; Kumar, N.; Almulihi, A.H.; Alassery, F.; Khan, A.I.; Islam, S.; Rothe, J.P.; Jagannadham, D.B.V.; Dekeba, K. Artificial Intelligence-Based Robotic Technique for Reusable Waste Materials. Comput. Intell. Neurosci. 2022, 2022, 2073482. [Google Scholar] [CrossRef]
  3. Owusu-Ansah, P.; Obiri-Yeboah, A.A.; Kwesi Nyantakyi, E.; Kwame Woangbah, S.; Idris Kofi Yeboah, S.I. Ghanaian inclination towards household waste segregation for sustainable waste management. Sci. Afr. 2022, 17, e01335. [Google Scholar] [CrossRef]
  4. Sheng, T.J.; Islam, M.S.; Misran, N.; Baharuddin, M.H.; Arshad, H.; Islam, M.R. An Internet of Things Based Smart Waste Management System Using LoRa and Tensorflow Deep Learning Model. IEEE Access 2020, 8, 148793–148811. [Google Scholar] [CrossRef]
  5. Raza, S.M.; Hassan, S.M.G.; Hassan, S.A.; Shin, S.Y. Real-Time Trash Detection for Modern Societies using CCTV to Identifying Trash by utilizing Deep Convolutional Neural Network. arXiv 2021, arXiv:2109.09611. [Google Scholar]
  6. Alsubaei, F.S.; Al-Wesabi, F.N.; Hilal, A.M. Deep Learning-Based Small Object Detection and Classification Model for Garbage Waste Management in Smart Cities and IoT Environment. Appl. Sci. 2022, 12, 2281. [Google Scholar] [CrossRef]
  7. Longo, E.; Sahin, F.A.; Redondi, A.E.C.; Bolzan, P.; Bianchini, M.; Maffei, S. A 5G-Enabled Smart Waste Management System for University Campus. Sensors 2021, 21, 8278. [Google Scholar] [CrossRef]
  8. Treiber, M.A. Introduction. In Optimization for Computer Vision; Springer: London, UK, 2013; pp. 1–16. [Google Scholar] [CrossRef]
  9. Fuchikawa, Y.; Nishida, T.; Kurogi, S.; Kondo, T.; Ohkawa, F.; Suehiro, T.; Kihara, Y. Development of a Vision System for an Outdoor Service Robot to Collect Trash on Streets. In the Proceedings of the Eighth IASTED International Conference on Computer Graphics and Imaging, CGIM 2005, Honolulu, HI, USA, 15–17 August 2005; pp. 100–105. [Google Scholar]
  10. Salvini, P.; Teti, G.; Spadoni, E.; Laschi, C.; Mazzolai, B.; Dario, P. The Robot DustCart. IEEE Robot. Autom. Mag. 2011, 18, 59–67. [Google Scholar] [CrossRef]
  11. Yang, M.; Thung, G. Classification of Trash for Recyclability Status. CS229Project Rep. 2016, 2016, 1–6. [Google Scholar]
  12. Mao, W.L.; Chen, W.C.; Wang, C.T.; Lin, Y.H. Recycling waste classification using optimized convolutional neural network. Resour. Conserv. Recycl. 2021, 164, 105132. [Google Scholar] [CrossRef]
  13. Rahman, M.W.; Islam, R.; Hasan, A.; Bithi, N.I.; Hasan, M.M.; Rahman, M.M. Intelligent waste management system using deep learning with IoT. J. King Saud Univ. Comput. Inf. Sci. 2022, 34, 2072–2087. [Google Scholar] [CrossRef]
  14. Adedeji, O.; Wang, Z. Intelligent waste classification system using deep learning convolutional neural network. Procedia Manuf. 2019, 35, 607–612. [Google Scholar] [CrossRef]
  15. Hulyalkar, S.; Deshpande, R.; Makode, K. Implementation of Smartbin Using Convolutional Neural Networks. Int. Res. J. Eng. Technol. 2018, 5, 3352–3358. [Google Scholar]
  16. Zhao, Z.-Q.; Zheng, P.; Xu, S.-T.; Wu, X. Object Detection With Deep Learning: A Review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [Green Version]
  17. Funch, O.I.; Marhaug, R.; Kohtala, S.; Steinert, M. Detecting glass and metal in consumer trash bags during waste collection using convolutional neural networks. Waste Manag. 2021, 119, 30–38. [Google Scholar] [CrossRef]
  18. Ren, C.; Jung, H.; Lee, S.; Jeong, D. Coastal waste detection based on deep convolutional neural networks. Sensors 2021, 21, 7269. [Google Scholar] [CrossRef]
  19. Liu, C.; Xie, N.; Yang, X.; Chen, R.; Chang, X.; Zhong, R.Y.; Peng, S.; Liu, X. A Domestic Trash Detection Model Based on Improved YOLOX. Sensors 2022, 22, 6974. [Google Scholar] [CrossRef]
  20. Wu, X.; Sahoo, D.; Hoi, S.C.H. Recent advances in deep learning for object detection. Neurocomputing 2020, 396, 39–64. [Google Scholar] [CrossRef] [Green Version]
  21. Dougherty, G. Classification. In Pattern Recognition and Classification: An Introduction; Springer: New York, NY, USA, 2013; pp. 9–26. [Google Scholar] [CrossRef]
  22. Khan, M.A.U.; Nazir, D.; Pagani, A.; Mokayed, H.; Liwicki, M.; Stricker, D.; Afzal, M.Z. A Comprehensive Survey of Depth Completion Approaches. Sensors 2022, 22, 6969. [Google Scholar] [CrossRef]
  23. Qiu, J.; Cui, Z.; Zhang, Y.; Zhang, X.; Liu, S.; Zeng, B.; Pollefeys, M. DeepLiDAR: Deep Surface Normal Guided Depth Prediction for Outdoor Scene From Sparse LiDAR Data and Single Color Image. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3308–3317. [Google Scholar] [CrossRef] [Green Version]
  24. Cheng, X.; Wang, P.; Guan, C.; Yang, R. CSPN++: Learning Context and Resource Aware Convolutional Spatial Propagation Networks for Depth Completion. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 10615–10622. [Google Scholar] [CrossRef]
  25. Van Gansbeke, W.; Neven, D.; De Brabandere, B.; Van Gool, L. Sparse and Noisy LiDAR Completion with RGB Guidance and Uncertainty. In Proceedings of the 2019 16th International Conference on Machine Vision Applications (MVA), Tokyo, Japan, 27–31 May 2019; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
  26. Hu, M.; Wang, S.; Li, B.; Ning, S.; Fan, L.; Gong, X. PENet: Towards Precise and Efficient Image Guided Depth Completion. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May 2021–5 June 2021; pp. 13656–13662. [Google Scholar] [CrossRef]
  27. Yan, Z.; Wang, K.; Li, X.; Zhang, Z.; Li, J.; Yang, J. RigNet: Repetitive Image Guided Network for Depth Completion. In Computer Vision-ECCV 2022; Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T., Eds.; Springer Nature Switzerland: Cham, Germany, 2022; pp. 214–230. [Google Scholar] [CrossRef]
  28. Nazir, D.; Liwicki, M.; Stricker, D.; Afzal, M.Z. SemAttNet: Towards Attention-based Semantic Aware Guided Depth Completion. IEEE Access 2022, 10, 120781–120791. [Google Scholar] [CrossRef]
  29. Liu, L.; Song, X.; Lyu, X.; Diao, J.; Wang, M.; Liu, Y.; Zhang, L. FCFR-Net: Feature Fusion based Coarse-to-Fine Residual Learning for Depth Completion. Proc. Conf. AAAI Artif. Intell. 2021, 25, 2136–2144. [Google Scholar] [CrossRef]
  30. Eldesokey, A.; Felsberg, M.; Holmquist, K.; Persson, M. Uncertainty-Aware CNNs for Depth Completion: Uncertainty from Beginning to End. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 12011–12020. [Google Scholar] [CrossRef]
  31. Jaritz, M.; de Charette, R.; Wirbel, E.; Perrotton, X.; Nashashibi, F. Sparse and Dense Data with CNNs: Depth Completion and Semantic Segmentation. In Proceedings of the 2018 IEEE International Conference on 3D Vision (3DV), Verona, Italy, 5–8 September 2018; pp. 52–60. [Google Scholar] [CrossRef] [Green Version]
  32. Barré, P.; Stöver, B.C.; Müller, K.F.; Steinhage, V. LeafNet: A computer vision system for automatic plant species identification. Ecol. Inform. 2017, 40, 50–56. [Google Scholar] [CrossRef]
  33. Kumar, N.; Belhumeur, P.N.; Biswas, A.; Jacobs, D.W.; Kress, W.J.; Lopez, I.C.; Soares, J.V. Leafsnap: A Computer Vision System for Automatic Plant Species Identification. In Computer Vision-ECCV 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 502–516. [Google Scholar] [CrossRef]
  34. Adetiba, E.; Ajayi, O.T.; Kala, J.R.; Badejo, J.A.; Ajala, S.; Abayomi, A. LeafsnapNet: An Experimentally Evolved Deep Learning Model for Recognition of Plant Species based on Leafsnap Image Dataset. J. Comput. Sci. 2021, 17, 349–363. [Google Scholar] [CrossRef]
  35. Chen, J.; Zhang, H.; Lu, Y.; Zhang, Q. The Research on Control and Dynamic Property of Autonomous Vehicle Adaptive Lidar System. In Proceedings of the 2020 International Conferences on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData) and IEEE Congress on Cybermatics (Cybermatics), Rhodes, Greece, 2–6 November 2020; pp. 462–468. [Google Scholar] [CrossRef]
  36. Benewake, T. “TF40 Datasheet”. 2022. Available online: https://my.mouser.com/datasheet/2/1099/Benewake_10152020_TF40-1954048.pdf (accessed on 20 June 2022).
  37. Iordan, D.; Popescu, G. The accuracy of LiDAR measurements for the different land cover categories. Environ. Eng. 2015, 4, 158–164. [Google Scholar]
  38. Wang, Y.; Che, J.; Zhang, L.; Ma, M. Research of garbage salvage system based on deep learning. In Proceedings of the International Conference on Computer Application and Information Security (ICCAIS 2021), Wuhan, China, 24 May 2022. [Google Scholar] [CrossRef]
  39. Fan, Z.; Li, C.; Chen, Y.; Mascio, P.D.; Chen, X.; Zhu, G.; Loprencipe, G. Ensemble of Deep Convolutional Neural Networks for Automatic Pavement Crack Detection and Measurement. Coatings 2020, 10, 152. [Google Scholar] [CrossRef]
Figure 1. A sample detection object (mouse) in 2D and 3D.
Figure 1. A sample detection object (mouse) in 2D and 3D.
Sensors 23 01499 g001
Figure 2. A sample of the detection object (mouse). (a) A photo of a mouse on paper (2D); (b) a real mouse photo (3D).
Figure 2. A sample of the detection object (mouse). (a) A photo of a mouse on paper (2D); (b) a real mouse photo (3D).
Sensors 23 01499 g002
Figure 3. (a) Sample images of own trash dataset. (b) Sample images of the TrashNet dataset. (c) Sample images of the LeafSnap dataset.
Figure 3. (a) Sample images of own trash dataset. (b) Sample images of the TrashNet dataset. (c) Sample images of the LeafSnap dataset.
Sensors 23 01499 g003
Figure 4. Block diagram.
Figure 4. Block diagram.
Sensors 23 01499 g004
Figure 5. Simple transfer learning using CNN architecture (AlexNet, VGGG16, GooogleNet, and ResNet18).
Figure 5. Simple transfer learning using CNN architecture (AlexNet, VGGG16, GooogleNet, and ResNet18).
Sensors 23 01499 g005
Figure 6. Modifying the GoogleNet architecture using Deep Network Designer.
Figure 6. Modifying the GoogleNet architecture using Deep Network Designer.
Sensors 23 01499 g006
Figure 7. Flowchart of the proposed study.
Figure 7. Flowchart of the proposed study.
Sensors 23 01499 g007
Figure 8. Convert the bounding box to x, y, and z coordinates.
Figure 8. Convert the bounding box to x, y, and z coordinates.
Sensors 23 01499 g008
Figure 9. TF40 LiDAR.
Figure 9. TF40 LiDAR.
Sensors 23 01499 g009
Figure 10. Fast scanning image process path.
Figure 10. Fast scanning image process path.
Sensors 23 01499 g010
Figure 11. The position of the LiDAR sensor during the scanning process.
Figure 11. The position of the LiDAR sensor during the scanning process.
Sensors 23 01499 g011
Figure 12. Example of LiDAR sensor data.
Figure 12. Example of LiDAR sensor data.
Sensors 23 01499 g012
Figure 13. Flowchart of fast scanning image.
Figure 13. Flowchart of fast scanning image.
Sensors 23 01499 g013
Figure 14. Detail scanning image process path.
Figure 14. Detail scanning image process path.
Sensors 23 01499 g014
Figure 15. Flowchart of detail scanning image.
Figure 15. Flowchart of detail scanning image.
Sensors 23 01499 g015
Figure 16. Result of training progress trash object using GoogleNet.
Figure 16. Result of training progress trash object using GoogleNet.
Sensors 23 01499 g016
Figure 17. Result of confusion matrix trash using VGG16.
Figure 17. Result of confusion matrix trash using VGG16.
Sensors 23 01499 g017
Figure 18. IP address of the IP camera displays on the mobile device (iPhone).
Figure 18. IP address of the IP camera displays on the mobile device (iPhone).
Sensors 23 01499 g018
Figure 19. Results of trash identification using a real-time camera. (a) Cardboard; (b) food packaging.
Figure 19. Results of trash identification using a real-time camera. (a) Cardboard; (b) food packaging.
Sensors 23 01499 g019
Figure 20. Results of trash identification using a real-time camera. (a) metal; (b) paper.
Figure 20. Results of trash identification using a real-time camera. (a) metal; (b) paper.
Sensors 23 01499 g020
Figure 21. Results of trash identification using a real-time camera. (a) glass; (b) plastic.
Figure 21. Results of trash identification using a real-time camera. (a) glass; (b) plastic.
Sensors 23 01499 g021
Figure 22. Result of the error trash identification.
Figure 22. Result of the error trash identification.
Sensors 23 01499 g022
Figure 23. Result of fast scanning image of the trash object (box).
Figure 23. Result of fast scanning image of the trash object (box).
Sensors 23 01499 g023
Figure 24. Result of detail scanning image of the trash object (box).
Figure 24. Result of detail scanning image of the trash object (box).
Sensors 23 01499 g024
Figure 25. Result of fast scanning image and detail scanning image of the trash object.
Figure 25. Result of fast scanning image and detail scanning image of the trash object.
Sensors 23 01499 g025
Table 1. Comparison between the Sensor Fusion Method and the Proposed Study.
Table 1. Comparison between the Sensor Fusion Method and the Proposed Study.
ReferencesSensorsMethodComment
[23,24,25,26,27,28,29]Camera RGB, LiDAREarly FusionData from RGB camera images and LiDAR sensors are directly input for deep learning and processed together. The result is complete depth.
[26,28,30]Camera RGB, LiDARSequential FusionThe data from the RGB camera image are first processed by deep learning. The result is RGB depth. Then, these results and the LiDAR sensor data directly become the input for subsequent deep learning. Both are processed together, and the result is a complete depth.
[26,28,31]Camera RGB, LiDARLate FusionThe data from the RGB camera image are first processed by deep learning. The result is RGB depth. In addition,
LiDAR sensor data are also processed by deep learning. The result is LiDAR depth. Furthermore, the respective outputs are processed together, resulting in complete depth.
ProposedCamera RGB, LiDARSequential_Camera_LiDAR (SCL): CNN + Fast Scanning Image + Detail Scanning ImageImage data from the RGB camera are processed using deep learning to detect the presence of trash objects. If what is detected is a trash object, then a Fast Scanning image is performed to ensure that the object is in 3D form. The scanning process uses a LiDAR sensor. If the object is 3D, then a detail scanning Image is performed to determine the correct position in lifting the trash object using a robot gripper.
Table 2. Training Parameter Used for Identification of Trash.
Table 2. Training Parameter Used for Identification of Trash.
ParameterAlexNetVGG16GoogleNetResnet18
MiniBatch size5555
Learning rate0.00030.00030.00030.0003
Max epoch6666
Data augmentedYesYesYesYes
Table 3. Main Parameters of TF40.
Table 3. Main Parameters of TF40.
Parameter NameStandard Version
Product performanceRange0.04–40 m 90% reflectivity, 0.04–20 m 10% reflectivity
Accuracy±2 mm
Distance resolution1 mm
Frame rate5 Hz
Optical parametersLight sourceLD
Wavelength635 nm
Laser classCLASS 2 (EN 60825)
Detection angle<1 mrad
Electrical parametersSupply voltage3.3 V
Average current≤180 mA
Power consumption≤0.6 W
Communication voltage levelLVTTL (3.3 V)
Table 4. Training Progress Result.
Table 4. Training Progress Result.
CNN ArchitectureTraining Time (s)Validation Accuracy (%)
AlexNet22977.54
VGG1689479.09
GoogleNet58186.38
ResNet1833580.75
Table 5. Comparison of the Confusion Matrix.
Table 5. Comparison of the Confusion Matrix.
CNN ArchitectureAccuracy (%)
AlexNet80.5
VGG1695.6
GoogleNet98.3
ResNet1897.5
Table 6. The Accuracy of 11 Types of Trash Using a Real-Time Camera with Pre-Trained AlexNet.
Table 6. The Accuracy of 11 Types of Trash Using a Real-Time Camera with Pre-Trained AlexNet.
Types of TrashNumber of Real Trash TestsTotal Trash is Correctly DetectedAccuracy (%)
Cardboard453782.22222222
Fabric171376.47058824
Food Packaging635282.53968254
Fruit211780.95238095
Glass342779.41176471
Leaf564682.14285714
Metal211676.19047619
Paper615183.60655738
Plastic524280.76923077
Rubber141071.42857143
Wood272177.77777778
Overall41133279.41019176
Table 7. The Accuracy of 11 Types of Trash Using a Real-Time Camera with Pre-Trained VGG16.
Table 7. The Accuracy of 11 Types of Trash Using a Real-Time Camera with Pre-Trained VGG16.
Types of TrashNumber of Real Trash TestsTotal Trash is Correctly DetectedAccuracy (%)
Cardboard454088.88888889
Fabric171482.35294118
Food Packaging635790.47619048
Fruit211990.47619048
Glass343088.23529412
Leaf564987.50000000
Metal211885.71428571
Paper615488.52459016
Plastic524586.53846154
Rubber141178.57142857
Wood272385.18518519
Overall41136086.58758694
Table 8. The Accuracy of 11 Types of Trash Using a Real-Time Camera with Pre-Trained GoogleNet.
Table 8. The Accuracy of 11 Types of Trash Using a Real-Time Camera with Pre-Trained GoogleNet.
Types of TrashNumber of Real Trash TestsTotal Trash is Correctly DetectedAccuracy (%)
Cardboard454497.77777778
Fabric171694.11764706
Food Packaging636298.41269841
Fruit212095.23809524
Glass343397.05882353
Leaf565598.21428571
Metal212095.23809524
Paper616098.36065574
Plastic525198.07692308
Rubber141392.85714286
Wood272696.29629630
Overall41140096.51349463
Table 9. The Accuracy of 11 Types of Trash Using a Real-Time Camera with Pre-Trained ResNet18.
Table 9. The Accuracy of 11 Types of Trash Using a Real-Time Camera with Pre-Trained ResNet18.
Types of TrashNumber of Real Trash TestsTotal Trash is Correctly DetectedAccuracy (%)
Cardboard454395.55555556
Fabric171588.23529412
Food Packaging636196.82539683
Fruit212095.23809524
Glass343397.05882353
Leaf565496.42857143
Metal212095.23809524
Paper615996.72131148
Plastic525096.15384615
Rubber141392.85714286
Wood272696.2962963
Overall41139495.14622079
Table 10. The Average Accuracy of Pre-Trained CNN in Real-Time Trash Identification.
Table 10. The Average Accuracy of Pre-Trained CNN in Real-Time Trash Identification.
CNN ArchitectureAverage Accuracy (%)
AlexNet79.410
VGG1686.588
GoogleNet96.513
ResNet1895.146
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Naf’an, E.; Sulaiman, R.; Ali, N.M. Optimization of Trash Identification on the House Compound Using a Convolutional Neural Network (CNN) and Sensor System. Sensors 2023, 23, 1499. https://doi.org/10.3390/s23031499

AMA Style

Naf’an E, Sulaiman R, Ali NM. Optimization of Trash Identification on the House Compound Using a Convolutional Neural Network (CNN) and Sensor System. Sensors. 2023; 23(3):1499. https://doi.org/10.3390/s23031499

Chicago/Turabian Style

Naf’an, Emil, Riza Sulaiman, and Nazlena Mohamad Ali. 2023. "Optimization of Trash Identification on the House Compound Using a Convolutional Neural Network (CNN) and Sensor System" Sensors 23, no. 3: 1499. https://doi.org/10.3390/s23031499

APA Style

Naf’an, E., Sulaiman, R., & Ali, N. M. (2023). Optimization of Trash Identification on the House Compound Using a Convolutional Neural Network (CNN) and Sensor System. Sensors, 23(3), 1499. https://doi.org/10.3390/s23031499

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop