Loading…
Overlapping bubble detection and tracking method based on convolutional Neural network and Kalman Filter
•An identification method of overlapping bubbles in high void fraction conditions by Convolutional Neural Network was proposed.•A trajectory tracking technology for overlapping bubbles was achieved based on Kalman Filter.•The model was trained only by synthetic images generated by GAN.•Overlapping b...
Saved in:
Published in: | Chemical engineering science 2022-12, Vol.263, p.118059, Article 118059 |
---|---|
Main Authors: | , , , , , |
Format: | Article |
Language: | English |
Subjects: | |
Citations: | Items that this one cites Items that cite this one |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | •An identification method of overlapping bubbles in high void fraction conditions by Convolutional Neural Network was proposed.•A trajectory tracking technology for overlapping bubbles was achieved based on Kalman Filter.•The model was trained only by synthetic images generated by GAN.•Overlapping bubbles can be detected and tracked accurately under low illumination and strong noise conditions.
Gas-liquid bubbly flow is widely applied in chemical process engineering. Geometric and dynamic parameters of bubbles play an essential role in the numerical prediction of mass and heat transfer processes. However, the critical obstacle in bubble detection is the inability of bubble segmentation and reconstruction when the overlapping issue of multiple bubbles is serious under high void fraction conditions. A new detection and tracking technique for overlapping bubbles was proposed in this paper to identify the overlapped bubbles. First, a novel convolutional neural network is used to detect bubbles. Afterward, the relationship between the detected bubbles in two frames is correlated using the Kalman Filter and neural network. The algorithm achieves 85 % accuracy under high overlap rate conditions in a 10 mm narrow rectangular channel with around 0.1 s for an image. In addition, a comparison test was conducted to evaluate the present technique's accuracy and robustness compared with conventional methods. |
---|---|
ISSN: | 0009-2509 1873-4405 |
DOI: | 10.1016/j.ces.2022.118059 |