A Smart Recycling Bin Using Waste Image Classification at The Edge
A Smart Recycling Bin Using Waste Image Classification at The Edge
[email protected] , [email protected]
Electrical and Electronic Engineering
University College London
Abstract
arXiv:2210.00448v1 [cs.CV] 2 Oct 2022
Rapid economic growth gives rise to the urgent demand for a more efficient waste recycling system.
This work thereby developed an innovative recycling bin that automatically separates urban waste
to increase the recycling rate. We collected 1800 recycling waste images and combined them with
an existing public dataset to train classification models for two embedded systems, Jetson Nano and
K210, targeting different markets. The model reached an accuracy of 95.98% on Jetson Nano and
96.64% on K210. A bin program was designed to collect feedback from users. On Jetson Nano, the
overall power consumption of the application was reduced by 30% from the previous work to 4.7
W, while the second system, K210, only needed 0.89 W of power to operate. In summary, our work
demonstrated a fully functional prototype of an energy-saving, high-accuracy smart recycling bin,
which can be commercialized in the future to improve urban waste recycling.
I. Introduction replace the traditional trash bins at the locations without plug
sockets. Therefore, it can hardly increase the household and
Waste generation has increased dramatically in the 21st century street recycling rate. As a result, this paper proposes low-cost
due to the growth in the global population. According to battery-powered real-time recycling waste segmentation bin
the World Bank Group’s report [1], 2.01 billion tonnes of systems to fill this gap.
municipal solid waste are generated worldwide every year, In this research, we trained a light MobileNet image classifica-
with only about 19% adequately recycled [1]. Recycling not tion model for Jetson Nano with TrashNet [3], an open-source
only helps to preserve raw material but, more importantly, waste dataset, and 1800 new training samples collected by
reduces the landfills required, which is an undesirable way us. The model demonstrated great performance with a high
of waste disposal due to its high demand for space and the test accuracy of 95.98% and a small parameter size of 3.0 M.
danger of introducing contaminants into the ground or even Jetson Nano only consumed 4.7 W when it ran the model. We
groundwater system. A major step of recycling is to separate also found a cheaper and less energy-hungry device, K210
the waste into specific categories according to its material. [4] to further reduce power consumption. The K210 board
Failure to do so will significantly harm the effectiveness of only consumed 0.89 W at inference time and the model on
recycling. Traditionally, workers at the recycling company sort it achieved a high test accuracy of 96.64%, which made it a
the waste into corresponding categories by hand, which is an more practical solution for trash bin applications. The code
inefficient method and requires unnecessary labor. Therefore, used in this study is available on Github [5].
people are now in dire need of a more advanced and automated
waste separation system. This paper is organized as follows. In Section II, the related
works performed in waste classification and AI trash bins
Image classification machine learning (ML) algorithms have are introduced. Section III introduces the system design and
been used to build automatic waste classification systems and explains the background theories that support this study.
assist waste management. This research aimed to improve the The methodology of data collection and model training is
existing classification system and pursue the commercialization introduced in section IV. Section V analyses the test results
of an artificial intelligence (AI) street bin. Specifically, we and evaluates the system performance on Jetson Nano and
focused on reducing the power consumption of the controller K210. Finally, Section VI concludes the achievements and
board to extend AI bin’s battery life, increasing its waste provides suggestions for future research.
classification accuracy and reducing its price.
The only intelligent recycling bin on the market, Bin-e [2],
uses image classification algorithms to separate the trash into
II. Related works
four categories: plastic, paper, metal and glass. It achieves
a waste segmentation accuracy of 92% [2]. However, the The United Kingdom (UK) government has planned to increase
price USD 5200 [2] is expensive for a rubbish bin. Also, the household recycling rate in England to 50% by 2020,
Bin-e’s operation requires a 230V power supply, so it cannot but only 44% of municipal waste was reused and recycled
in 2020 [6]. The inconvenience and lack of knowledge are
Preprint – A SMART RECYCLING BIN USING WASTE IMAGE CLASSIFICATION AT THE EDGE 2
two crucial factors that prevent people from recycling [7]. The classification speed is quantified by inference time, which
Automatic waste segmentation bins were designed to overcome refers to the time taken for one classification to be completed.
these problems by helping people classify and send the waste Our application looks for a smaller model that can interact
into the corresponding containers, making waste disposal more with the user in real-time, with an inference time that must be
convenient. Table I outlines the approaches and the primary smaller than 0.1 s, which is the human visual reaction time. The
hardware components involved in related studies to construct CNN model, EfficientNet B0, used in [17] is the starting point
the waste segmentation systems. of this research and it achieved an accuracy of 95.38% and
an inference time of 0.07 s on Jetson Nano. While the model
In early research, a microcontroller is connected to a variety of
is fast enough to be used in real-time applications, the 96%
sensors to determine the composition of the waste. For example,
high memory usage needs optimization for bin applications.
inductive and capacitive sensors can detect the metal element
[8][9][10] and moisture sensors separate wet waste from dry Power consumption is another factor that needs to be consid-
waste [9][10]. The microcontroller makes decisions based ered in the final product. Unfortunately, rare research has paid
on its readings [8][9]. However, although the sensor-based attention to it. For instance, none of the trash bin cases listed in
classification method can detect the composition precisely, it Table I has measured the power consumption of the proposed
lacks the ability to classify waste into more specific groups. The system. Still, it is evident that the CNN-based approaches have
sensors cannot discriminate between plastic, paper, glass, and higher power consumption than the sensor-based one and the
other unrecyclable dry waste, which are important categories Pynq-Zl and Raspberry Pi will typically have lower power
in recycling. consumption than Jetson Nano used in [12] and [17]. The
Raspberry Pi 4 has a typical power consumption of between
The development of machine learning and image classification
2.7 W and 6.4 W. It reduces the power consumption but
enables the bin to sort the waste based on visual input
results in undesirable low performance [14]. As a result, the
like a human. The convolutional neural network (CNN) is
CNN architecture, MobileNet V2, has a low average per-class
a branch of image classification algorithms that performs
precision of 91.76% and a long inference time of 0.358 s on
mainly convolution operations on the pixels [15]. It is the most
it.
popular choice due to its high accuracy and power efficiency
compared to other methods and is used in all the four papers The limitation of the current research is that the energy-
[11][12][13][14]. It gives the bin ability to differentiate between saving systems will have unacceptable low performance in
spoons and cups, which is tremendous progress compared to the the classification task for commercialization. As a result, this
sensor-based approaches. Traditionally, the CNN models run on study designed two high-accuracy waste classification systems
a cloud raising the data transmission latency and user privacy with lower power consumption than previous studies. The first
security problems. To solve these problems, recent research system reduced the power consumption of the application on
moved the computation to an edge embedded system. However, Jetson Nano by using a lighter model, MobileNet, to reduce
edge computing has the drawback of limiting computation power consumption while maintaining the accuracy.
resources, so the model size is important in selecting the CNN
The second system is developed based on K210, a less
model structure.
expensive and more energy-saving embedded device. K210 is
To develop, evaluate and select the proper CNN structures, most an unpopular choice and is used in fewer than 100 research.
research in this field used the TrashNet dataset developed by Most research implemented YOLO object detection models on
Mindy Yang and Gary Thung [3] in 2017, the first open-access it [18][19] and demonstrated its outstanding power efficiency.
recycling waste dataset. This high-quality dataset contains K210 has not been used for recycling waste classification when
2527 waste photos from 6 groups: metal, glass, plastic, paper, this paper was written, so this paper proposed and evaluated
cardboard, and trash. It provides a foundation for later research an innovative approach.
and our study also used this dataset.
Efforts have been made to increase the segmentation accuracy III. System design
of CNN models based on TrashNet. The state-of-art accuracy
of 97.86% on TrashNet was reached in [16]. However, the A. The AI recycling bin design
CNN architecture, GoogLeNet, used in this research will cause
out-of-memory (OOM) on Jetson Nano [17]. It demonstrated The AI bin consists of five recycling waste containers and a
the potential of CNN classification, but the model size and detection box. The waste will be sorted after it is placed in
computation cost must be cut down to implement the machine the detection box. The whole system can be controlled by a
learning algorithm on embedded systems. A lighter CNN Jetson Nano or a K210 board.
model, WasteNet, was constructed by White et al. and achieved
an accuracy of 97% on the TrashNet dataset [12]. The paper The bin design using Jetson Nano is summarized in Figure
claimed that the model could be loaded to Jetson Nano, but 1. Jetson Nano will interact with the users and collects
it did not provide details regarding the edge implementation, feedback through the touch screen in front of it, displaying
such as the classification speed. Nevertheless, from the model the instructions for using the bin and the camera inputs. The
structure, we could estimate that the classification speed for Raspberry Pi camera at the top takes photos of the waste in
one image would be too slow for real-time classification. It the detection box. The images will be fed to the classification
can only be used in a bin application that classifies the waste algorithm in Jetson Nano, which classifies the photos into
objects based on one photo. seven groups. The waste classification models in the previous
study [17] have five output classes: "paper", "metal", "plastic",
Preprint – A SMART RECYCLING BIN USING WASTE IMAGE CLASSIFICATION AT THE EDGE 3
TABLE I
Related works in AI bins
TABLE II
Training parameters used in [17]
Pre-Processing Method Value
Random flip Horizontal and vertical
Random rotation Up to 180◦
Random translation Up to 10%
Random zoom Up to 75%
Training Parameter Value
Learning rate scheduler Constant learning rate scheduler
Fig. 1. System block design of WasteNet. Train/Validation/Test split ratio 72/18/10
Optimizer Adam optimiser
Training epochs 50
"cardboard" and "glass". They will be reproduced in the next Learning rate 4.3e-05
section as the benchmark models of our study. Our model will Fine-tuning training epochs 8
be trained with two new classes, "empty" and "hand". "Empty" Fine-tuning learning rate 4e-06
Loss functions Sparse categorical cross entropy
means there is nothing in the photo, while the "hand" group
Classifier activation function Softmax
detects a human’s hand to avoid trapping the user’s hand by Include top layers in base model False
the door. When these two groups are detected, the bin waits Batch size 16
and continues detecting.
B. The Benchmark models and validation, and 238 images were used for testing. The
train/validation/test split is 72/18/10.
The waste classification models in [17] consist of an input
layer, a pre-processing augmentation layer, an EfficientNet B0 To begin with, the base model is initialized with pre-trained
base model layer, a global average pooling 2D layer and a weights on ImageNet. The two models are trained separately
dense layer. The first model has an input layer size of 512x384 by setting the corresponding input sizes. The learning rate
pixels, while the second model input has 384x288 pixels. Table was set to 4.3e-05 at the first 50 epochs and was reduced to
II shows the configuration of the augmentation layer and the 4e-06 at the eight epochs afterwards. Eventually, both models
model training hyperparameters for reproduction. achieved the same test accuracy of 95.38%. Figure 2 shows
the confusion matrix of the test results and indicates that most
The models were trained with the TrashNet dataset, which
mistakes are made in classifying between paper and cardboard.
consists of six categories of images: glass, paper, cardboard,
plastic, metal, and trash. Only the first five groups of data
were used to build the model that classified recycling into
five categories. In total, 2152 images were used for training
Preprint – A SMART RECYCLING BIN USING WASTE IMAGE CLASSIFICATION AT THE EDGE 4
TABLE III
Model classification accuracy on ImageNet
CNN architecture Number of Top-1 accuracy on
parameters (M) ImageNet (%)
EfficientNet B0 [23] 5.3 76.3
MobileNet V3 Large 5.4 75.2
[22]
MobileNet V1 [20] 4.2 70.6
ShuffleNet V2 [24] 7.4 74.9
Dense-169 [25] 14.3 76.2
ResNet50 [26] 25. 6 74.9
TABLE IV
Comparison of Jetson Nano and K210
Nvidia Jetson Sipeed M1w dock
Nano 4GB[27] K210 kit[28]
Clock 1.43GHz 0.4GHz
RAM 4GB 8 MB
Power consumption 5-10W 0.3W-3W (typically
1W)
Programming language Python Micropython
Cost GBP 133 GBP 23
TABLE V
Number of waste items in each category
Category Number of images in Number of images in
TrashNet our dataset
Cardboard 403 268
Glass 501 130
Paper 594 89
Plastic 482 298
Metal 410 140
Hand 0 920
Empty 0 26
We kept the same data augmentation setup and used the same
training hyperparameters applied in the previous training, as
shown in Table VIII. All training data is resized into 224x224
Fig. 6. Waste classification model structure pixels before training. The model used the same test/validation
split 70/15 as our previous models on Jetson Nano. We first
Preprint – A SMART RECYCLING BIN USING WASTE IMAGE CLASSIFICATION AT THE EDGE 7
TABLE IX
Accuracy of three models on 70/15/15 split
TABLE X
Accuracies of models trained by the final training data
Model Base model Training Top-1 Accuracy Per-class precision Top-1 Accuracy Per-class precision
Architecture input resolution accuracy on 5-class test of 5-class test set on new 7-class of 7-class test set
(pixels) (%) set (%) (%) test set (%) (%)
EfficientNet B0 512x384 99.59 95.23 94.79 95.98 95.97
MobileNet V3 Large 512x384 99.46 94.68 94.53 95.38 95.68
MobileNet V3 Large 224x224 99.46 95.79 95.70 95.98 96.41
TABLE XII
Inference speed on Jetson Nano at 5 W power mode
Model Base model In- Inference Memory Power
Architecture put Resolution Per (%) consumption
(pixels) second (W)
EfficientNetB0 384x288 13 94.93 6.682
MobileNet 224x224 40 90.90 4.698
V3 Large
Future works can continue to develop the trash bin with the to improve the recycling performance of urban areas,”
two embedded systems. While Jetson Nano targets the highly Journal of Cleaner Production, vol. 245, p. 118605,
accurate bin used in households and offices, K210 can be used 2020. [Online]. Available: https://www.sciencedirect.com/
in cheaper along-road self-powered bins. For Jetson Nano, science/article/pii/S0959652619334754
the research can investigate the hyperparameters for continual [8] R. Rajkamal, V. Anitha, P. G. Nayaki, K. Ramya, and
training and implement the algorithm on Jetson Nano with E. Kayalvizhi, “A novel approach for waste segregation
limited computation resources that restricts the training time at source level for effective generation of electricity
and samples. For K210, future works can build the circuit board - GREENBIN,” in 2014 International Conference on
for powering K210 with a solar panel. As the energy generated Science Engineering and Management Research, ICSEMR
by the solar panel fluctuates depending on the sunshine, a 2014. Institute of Electrical and Electronics Engineers
power management system is required to ensure constant Inc., February 2014.
voltage output to K210. Moreover, the performance of the
two applications could be tested with a larger unseen real- [9] B. R. S. Kumar, N. Varalakshmi, S. S. Lokeshwari,
world test set with a size greater than 10% of the training data. K. Rohit, Manjunath, and D. N. Sahana, “Eco-friendly
This will give a better estimation of their performance. iot based waste segregation and management,” in 2017
International Conference on Electrical, Electronics, Com-
In conclusion, this paper has provided a state-of-the-art solution munication, Computer, and Optimization Techniques
that reduced the power consumption of Jetson Nano by 30% (ICEECCOT), 2017, pp. 297–299.
and increased the inference speed by 200%. K210 only used
20% of the power of Jetson Nano while maintaining accuracy. [10] W. Pereira, S. Parulekar, S. Phaltankar, and V. Kamble,
The use of this bin can increase the recycling rate up to 95%, “Smart Bin (Waste Segregation and Optimisation),” in
significantly improving the recycling rate in the UK. 2019 Amity International Conference on Artificial Intelli-
gence (AICAI), 2019, pp. 274–279.
[11] D. Ziouzios and M. Dasygenis, “A Smart Recycling Bin
VII. Acknowledgements for Waste Classification,” in 2019 Panhellenic Conference
on Electronics Telecommunications (PACET), 2019, pp.
The authors would like to acknowledge their gratitude to Mindy 1–4.
Yang and Gary Thung who open-sourced the TrashNet dataset.
The authors would also like to show gratitude to the Institute [12] G. White, C. Cabrera, A. Palade, F. Li, and
of Making staff for technical assistance with constructing the S. Clarke, “WasteNet: Waste Classification at the
detection box. Edge for Smart Bins,” June 2020. [Online]. Available:
http://arxiv.org/abs/2006.05873
[13] F. N. Jimeno, B. J. A. Briz, M. R. P. Artiaga, R. E.
References Angelia, and N. B. Limsangan, “Development of Smart
Waste Bin Segregation using Image Processing,” in
[1] S. Kaza, L. C. Yao, P. Bhada-Tata, and F. Van Woerden, 2021 IEEE 13th International Conference on Humanoid,
What a Waste 2.0 : A Global Snapshot of Solid Waste Nanotechnology, Information Technology, Communication
Management to 2050. Urban Development. World and Control, Environment, and Management (HNICEM).
Bank Publications, 2018. [Online]. Available: https: IEEE, November 2021, pp. 1–6. [Online]. Available:
//openknowledge.worldbank.org/handle/10986/30317 https://ieeexplore.ieee.org/document/9732038/
[2] “ Bin-e.” [Online]. Available: https://www.bine.world/ [14] N. C. A. Sallang, M. T. Islam, M. S. Islam, and
[3] Gary Thung and Mindy Yang, “garythung/trashnet: H. Arshad, “A CNN-Based Smart Waste Management
Dataset of images of trash; Torch-based CNN for System Using TensorFlow Lite and LoRa-GPS Shield
garbage image classification.” [Online]. Available: in Internet of Things Environment,” IEEE Access,
https://github.com/garythung/trashnet vol. 9, pp. 153 560–153 574, 2021. [Online]. Available:
https://ieeexplore.ieee.org/document/9615047/
[4] Canaan, “Kendryte K210.” [Online]. Available: https:
//canaan.io/product/kendryteai [15] R. Yamashita, M. Nishio, R. K. G. Do, and K. Togashi,
“Convolutional neural networks: an overview and appli-
[5] Xueying Li, “Paper GitHub Repository.”
cation in radiology,” pp. 611–629, August 2018.
[Online]. Available: https://github.com/molvcan/AN-
SMART-RECYCLING-BIN-USING-WASTE-IMAGE- [16] U. Ozkaya and L. Seyfi, “Fine-tuning models comparisons
CLASSIFICATION-AT-THE-EDGE on garbage classification for recyclability,” CoRR,
vol. abs/1908.04393, 2019. [Online]. Available: http:
[6] “Progress report on recycling and recovery targets
//arxiv.org/abs/1908.04393
for England 2020,” United Kindom Government,
Tech. Rep., January 2022. [Online]. Available: [17] M. Abdulmahmood and R. Grammenos, “Improving the
https://www.gov.uk/government/publications/progress- Deployment of Recycling Classification through Efficient
report-on-recycling-and-recovery-targets-for-england- Hyper-Parameter Analysis,” October 2021. [Online].
2020/progress-report-on-recycling-and-recovery-targets- Available: http://arxiv.org/abs/2110.11043
for-england-2020 [18] Henderi, A. S. Rafika, H. L. H. Spits Warnar, and M. A.
[7] D. Knickmeyer, “Social factors influencing household Saputra, “An Application of Mask Detector for Prevent
waste separation: A literature review on good practices Covid-19 in Public Services Area,” in Journal of Physics:
Preprint – A SMART RECYCLING BIN USING WASTE IMAGE CLASSIFICATION AT THE EDGE 12
Conference Series, vol. 1641, no. 1. IOP Publishing [35] “tensorflow/models/research/slim/nets/mobilenet/.” [On-
Ltd, November 2020. line]. Available: https://github.com/tensorflow/models/
[19] S. Li, K. Huang, L. Feng, Z. Lv, and Y. Wu, tree/master/research/slim/nets/mobilenet
“ADRC-based position control of QCSP with vision [36] Kendryte, “kendryte / kflash.py.” [Online]. Available:
feedback,” in 2021 China Automation Congress (CAC). https://github.com/kendryte/kflash.py
IEEE, October 2021, pp. 5062–5068. [Online]. Available: [37] Y. Yu and R. Grammenos, “Towards artificially
https://ieeexplore.ieee.org/document/9727408/ intelligent recycling: Improving image processing for
[20] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, waste classification,” Tech. Rep. [Online]. Available:
W. Wang, T. Weyand, M. Andreetto, and H. Adam, https://arxiv.org/abs/2108.06274
“MobileNets: Efficient Convolutional Neural Networks [38] R. A. Aral, S. R. Keskin, M. Kaya, and M. Haciömeroǧlu,
for Mobile Vision Applications,” April 2017. [Online]. “Classification of TrashNet Dataset Based on Deep Learn-
Available: http://arxiv.org/abs/1704.04861 ing Models,” in Proceedings - 2018 IEEE International
[21] M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, and Conference on Big Data, Big Data 2018. Institute of
L.-C. Chen, “MobileNetV2: Inverted Residuals and Electrical and Electronics Engineers Inc., January 2019,
Linear Bottlenecks,” January 2018. [Online]. Available: pp. 2058–2062.
http://arxiv.org/abs/1801.04381
[39] C. Bircanoglu, M. Atay, F. Beser, O. Genc, and M. A.
[22] A. Howard, M. Sandler, G. Chu, L.-C. Chen, B. Chen, Kizrak, “RecycleNet: Intelligent Waste Sorting Using
M. Tan, W. Wang, Y. Zhu, R. Pang, V. Vasudevan, Q. V. Deep Neural Networks,” in 2018 IEEE (SMC) Interna-
Le, and H. Adam, “Searching for MobileNetV3,” May tional Conference on Innovations in Intelligent Systems
2019. [Online]. Available: http://arxiv.org/abs/1905.02244 and Applications, INISTA 2018. Institute of Electrical
[23] M. Tan and Q. V. Le, “EfficientNet: Rethinking Model and Electronics Engineers Inc., September 2018.
Scaling for Convolutional Neural Networks,” May 2019. [40] M. H. Hesamian, S. Mashohor, M. I. Saripan, and
[Online]. Available: http://arxiv.org/abs/1905.11946 W. A. Wan Adnan, “Effect of image resolution on
[24] N. Ma, X. Zhang, H.-T. Zheng, and J. Sun, “ShuffleNet intensity based scene illumination classification using
V2: Practical Guidelines for Efficient CNN Architecture neural network,” Imaging Science Journal, vol. 63, no. 8,
Design,” Tech. Rep. pp. 433–439, November 2015.
[25] G. Huang, Z. Liu, L. van der Maaten, and [41] Xueying Li, “Demonstration video.” [Online]. Available:
K. Q. Weinberger, “Densely Connected Convolutional https://www.youtube.com/watch?v=_ri6iLEXBhk
Networks,” August 2016. [Online]. Available: [42] European Commission, “PHOTOVOLTAIC GEOGRAPH-
http://arxiv.org/abs/1608.06993 ICAL INFORMATION SYSTEM.” [Online]. Available:
[26] K. He, X. Zhang, S. Ren, and J. Sun, “Deep Residual https://re.jrc.ec.europa.eu/pvg_tools/en/
Learning for Image Recognition,” December 2015. [43] “Lixada 10W Portable Solar Panel High Power
[Online]. Available: http://arxiv.org/abs/1512.03385 Paper Shaped Mini Monocrystalline Silicon Panel
[27] Nvidia, “Jetson Nano Developer Kit.” [Online]. Charger USB Port for Cell Phone Camping
Available: https://developer.nvidia.com/embedded/jetson- Riding Climbing Travel Outdoor Activity.” [Online].
nano-developer-kit Available: https://www.amazon.co.uk/Lixada-Portable-
[28] “Sipeed maixdock datasheet v1.0 key features,” Tech. Monocrystalline-Climbing-Activity/dp/B071D9MB7J/
Rep., 2019. [Online]. Available: www.sipeed.com ref=sr_1_1?geniuslink=true&keywords=lixada+solar+
panel+charger+usb+port+portable+high+power+paper+
[29] TensorFlow, “tensor- shaped+monocrystalline+silicon+for+cell+phone+
flow.org/lite/performance/model_optimization.” [Online]. camping+hiking+travel&qid=1647222985&sr=8-1
Available: https://www.tensorflow.org/lite/performance/
model_optimization [44] Saur News Bureau, “Here is how you can
calculate the annual solar energy output of a
[30] Kendryte, “kendryte / nncase.” [Online]. Available: photovoltaic system,” March 2016. [Online]. Available:
https://github.com/kendryte/nncase https://www.saurenergy.com/solar-energy-blog/here-is-
[31] NVDIA, “NVIDIA TensorRT.” [Online]. Available: how-you-can-calculate-the-annual-solar-energy-output-
https://developer.nvidia.com/tensorrt of-a-photovoltaic-system
[32] L. WM Intellectual Property Holdings,
“Recycling-Myths,” 2020. [Online]. Available:
https://www.wm.com/content/dam/wm/documents/
RecyclingResources/Posters-Guides-Tools/Recycling-
Myths.pdf
[33] Sipeed, “sipeed / maix_train.” [Online]. Available:
https://github.com/sipeed/maix_train
[34] “tensorflow/models/research/slim/nets/mobilenet_v1.md.”
[Online]. Available: https://github.com/tensorflow/models/
blob/master/research/slim/nets/mobilenet_v1.md