亚洲免费av电影一区二区三区,日韩爱爱视频,51精品视频一区二区三区,91视频爱爱,日韩欧美在线播放视频,中文字幕少妇AV,亚洲电影中文字幕,久久久久亚洲av成人网址,久久综合视频网站,国产在线不卡免费播放

        ?

        A three-dimensional measurement method for binocular endoscopes based on deep learning*

        2022-04-24 12:57:18HaoYUChangjiangZHOUWeiZHANGLiqiangWANGQingYANGBoYUAN

        Hao YU ,Changjiang ZHOU ,Wei ZHANG ,Liqiang WANG, ,Qing YANG, ,Bo YUAN??

        1State Key Laboratory of Modern Optical Instrumentation,College of Optical Science and Engineering,Zhejiang University,Hangzhou 310027,China

        2Research Center for Intelligent Sensing,Zhejiang Lab,Hangzhou 311100,China

        Abstract:In the practice of clinical endoscopy,the precise estimation of the lesion size is quite significant for diagnosis.In this paper,we propose a three-dimensional (3D) measurement method for binocular endoscopes based on deep learning,which can overcome the poor robustness of the traditional binocular matching algorithm in texture-less areas.A simulated binocular image dataset is created from the target 3D data obtained by a 3D scanner and the binocular camera is simulated by 3D rendering software to train a disparity estimation model for 3D measurement.The experimental results demonstrate that,compared with the traditional binocular matching algorithm,the proposed method improves the accuracy and disparity map generation speed by 48.9%and 90.5%,respectively.This can provide more accurate and reliable lesion size and improve the efficiency of endoscopic diagnosis.

        Key words:Binocular endoscope;Three-dimensional measurement;Deep learning;Disparity estimation

        1 Introduction

        It is tough to perceive spatial positioning with conventional endoscopes that provide only two-dimensional (2D) images,which will eventually affect the accuracy and safety of diagnosis and treatment.The binocular endoscope has the ability of three-dimensional (3D) imaging and measurement,which can provide depth information to assist endoscopists in operating the endoscope more accurately,efficiently,and safely (Ogino-Nishimura et al.,2015;Cai et al.,2018;Nomura et al.,2019;Omori et al.,2020).In clinical gastrointestinal endoscopy,the precise size of lesions such as polyps is of great significance for diagnosis (Furukawa et al.,2014;Dimas et al.,2020).When operating 2D endoscopes,endoscopists mainly use subjective vision or specific measurement tools,such as endoscopic measurement rulers,to estimate the polyp size.Consequently,the endoscopist’s subjective assessment often deviates greatly from the accurate value due to polyp morphology and optical distortion (Ahmad et al.,2016;Anderson et al.,2016;Shaw and Shaukat,2016;Sakata et al.,2018) and because endoscopic measurement rulers produce inaccurate results and are inefficient (Anderson et al.,2016).Therefore,it is necessary to develop an objective,accurate,and efficient 3D measurement method to improve endoscopic surgery results (Furukawa et al.,2014).

        At present,the binocular endoscope realizes 3D measurement mainly by determining the disparity between the left and right images using the binocular matching algorithm (Wang D et al.,2018) and calculating 3D coordinates of each pixel by triangulation combined with the binocular camera parameters.Traditional binocular matching algorithms,such as the semi-global matching (SGM) algorithm(Hirschmüller,2008),have a better matching result on images with rich texture information.However,most endoscopic images have inadequate areas that are texture-less or have a high gloss.The robustness of the SGM algorithm in such an area is poor,and its generated disparity map often has noise and mismatching errors,which will eventually affect the accuracy and stability of 3D measurement.To overcome the shortcomings of traditional binocular matching algorithms,researchers began to explore the binocular disparity estimation model based on deep learning.Some proposed models (?bontar and LeCun,2014;Kendall et al.,2017;Zhang et al.,2019) can effectively improve the accuracy of disparity predictions in texture-less and high-gloss areas,and produce excellent results in the field of autonomous driving.However,because their application scenarios are quite different from the endoscopic environment,they are ineffective when directly applied in binocular endoscopic images.In addition,binocular endoscopic images with ground truth disparity are difficult to obtain,which limits the transfer learning ability of these models.

        In an effort to solve the above-mentioned problems,we propose a 3D measurement method for binocular endoscopes based on deep learning.A simulated binocular image dataset is created for a gastroscope scenario to train a disparity estimation model for 3D measurement.The experimental results demonstrate that the proposed method is superior to the traditional binocular matching algorithm in terms of stability,accuracy,and real-time performance.

        The contributions of this work are summarized as follows:(1) Given that the ground truth disparity of real binocular endoscopic images is difficult to obtain,we propose a new method for creating simulated binocular images that are very similar to real endoscopic images in terms of content and structure.(2)An end-to-end disparity estimation network,mainly based on Google’s StereoNet (Khamis et al.,2018)with minor adjustments,is trained on virtual binocular images and used for estimating the disparities of real binocular endoscopic images.The 3D measurement result calculated from the estimated disparities is more stable and reliable.

        2 Three-dimensional measurement method based on deep learning

        The process of the proposed 3D measurement method based on deep learning is shown in Fig.1a.The left and right images taken by the binocular camera of the 3D endoscope are used as the input of the disparity estimation model,which will output the disparity map corresponding to the left image.Then,3D coordinates of each pixel in the disparity map in the camera coordinate system can be acquired by triangulation combined with the binocular camera parameters.Finally,the target size is calculated using the 3D coordinates.The training process of the disparity estimation model is shown in Fig.1b.A 3D scanner is used to scan the stomach model to obtain a 3D model file,which is then imported into the 3D rendering software.By using the simulated binocular system established in the 3D rendering software,a simulation dataset is created to train the deep convolutional network for disparity estimation.

        Fig.1 The overall process of our 3D measurement method(a) and the training process of disparity estimation model(b)

        2.1 Simulation dataset

        Training a deep learning network requires a lot of data with labels.However,it is difficult to obtain endoscopic binocular images with ground truth disparity,which limits the transfer training,and application of deep learning models in the actual endoscopic environment.Therefore,a method for creating a simulated binocular endoscopic dataset is proposed in this paper.

        In previous research (Wang XZ et al.,2020),Blender,an open-source 3D animation production software,was used for building a simulated gastrointestinal model to generate a binocular endoscopic dataset.However,because the gastrointestinal model established by them is a pure virtual model with some disadvantages such as limited texture,the dataset was not sufficiently similar to the real gastroscopic image.By contrast,the simulated binocular dataset created in this study is more similar to the images taken by actual binocular endoscopes.The specific generation process is as follows:First,a 3D scanner is used to scan the actual stomach model to obtain its 3D model file,which is then imported into Blender.After that,a simulated binocular camera system similar to the actual 3D gastroscopic binocular camera is established in Blender.Next,the relative position of the simulated binocular camera to the scanned stomach model is set close to the working state of the actual 3D gastroscope.Finally,the binocular image dataset with ground truth disparity is generated through the Blender rendering.Figs.2c and 2d show the left and right images taken by the actual 3D gastroscope,respectively.Figs.2e and 2f show the left and right images from the simulation dataset,respectively.It can be seen that the simulated images are very similar to the real images in terms of the gastric rugae texture and the disparity relationship.In particular,because only the depth of the scanned stomach model relative to the binocular camera can be obtained from Blender,the disparitydneeds to be calculated as

        wherefis the focal length of the camera,bis the baseline of the binocular camera,andzis the distance from the scanned model to the plane of the camera’s optical center.The simulation dataset results are shown in Fig.3,in which the image resolution is 1280×800,and the disparity image corresponds to the left image.

        Fig.2 Comparison of real and scanned models:(a) actual stomach model;(b) image obtained using the 3D scanned stomach model;(c) left image obtained using the actual stomach model;(d) right image obtained using the actual stomach model;(e) left image obtained using the scanned model;(f) right image obtained using the scanned model

        Fig.3 Examples of the simulation dataset:(a) left camera image;(b) right camera image;(c) ground truth disparity

        2.2 Disparity estimation network

        Our end-to-end disparity estimation network is mainly based on the StereoNet proposed by Google researchers (Khamis et al.,2018),which has the advantages of good performance in the weak area,low structure complexity,and high computational efficiency.The pipeline mainly includes the feature extraction network,cost volume filter network,and details refinement network (Fig.4).Here,we do not use the Siamese network in the cost volume filter and refinement stage,but only generate the disparities corresponding to the left image.

        At the feature extraction stage,we use the Siamese network,which first employs three 5×5 convolution layers with a stride of 2 to down-sample the input images and then employs 6 residual blocks composed of 3×3 convolution,batch regularization,and leaky ReLu activations,to extract the features.

        A rough cost volume is obtained by subtracting feature vectors of the binocular images.Next,it is filtered through four 3D convolutions with a size of 3×3×3,batch normalization,leaky ReLu activations,and one final 3×3×3 convolution layer.Then the coarse disparity map is generated from the disparity with the minimum cost at each pixel using differentiable arg min.

        At the disparity optimization stage,the bilinear interpolation method is used to up-sample the coarse disparity map,and then the color information of the input RGB image is used for refining the high-frequency details.The details refinement network takes the concatenated color and disparity as input and generates the 32-dimensional tensor through a 3×3 2D convolution.Then the tensor is passed through 6 residual blocks in which the dilation convolution is used and the dilation is set as (1,2,4,8,1,1).By applying a final 3×3 2D convolution layer,we finally obtain the refined disparity map.

        To adapt to the endoscopic images and improve the accuracy of disparity prediction,we adopt the invariant loss function proposed by other researchers(Wang XZ et al.,2020).The loss function is defined as follows:

        wherenis the total number of pixels in the disparity map,diis the predicted disparity,anddi′ is the ground truth value at pixeli.

        2.3 Evaluation indicators

        To quantitatively evaluate the effect of the disparity map generated from simulated binocular images,end-point-error (EPE) and K-pixel-error (KPE)are adopted as metrics (Wang XZ et al.,2020).EPE is the average Euclidean distance between the estimated disparity and ground truth,and KPE is the percentage of the EPE that exceedsKpixels.EPE and KPE are calculated as

        wherefiis counted only when the EPE of pixeliexceedsKpixels.However,in the evaluation of the disparity map generated from real binocular endoscopic images,EPE and KPE are not suitable for evaluating the results because of the lack of ground truth disparity.Therefore,we propose an indicator to evaluate the stability of 3D measurement achieved using the disparity map,which is the standard deviation of the multiple measured lengths for the same 3D curve.The specific process is as follows:First,the gastric rugae with clear texture in the left image taken by a 3D gastroscope are selected as the curves to be measured,and the curves are fitted with a constant number of points.Then the 3D coordinates of these fitting points in the camera coordinate system are calculated from the generated disparity map using the following equations:

        Fig.4 The network architecture

        wherexiandyiare thex-andy-coordinate of pixeliin the pixel coordinate system,respectively,fis the focal length of the camera,bis the baseline of the binocular camera,diis the disparity of pixeli,cxandcyare thexandy-coordinate of the principal point of the camera in the pixel coordinate system,respectively,andX,Y,andZare the 3D coordinates in the camera coordinate system.Finally,the curve length is acquired from the 3D coordinates of the fitting points.

        3 Experiments and results analysis

        3.1 Experiment description

        The binocular endoscopic images were all taken with a self-developed 3D gastroscope prototype(Fig.5).The prototype consisted of the 3D gastroscope body,the LED light source,and the signal processing circuit.At the tip of the gastroscope were two cameras,two lights,one forceps channel,and one air-water channel.The calibrated focal length of the binocular camera was 1059.6 pixels,the calibrated baseline was 5.9 mm,and the calibratedx-andy-coordinate of the principal point of the camera in the pixel coordinate system was 633.6 pixels and 367.1 pixels,respectively.

        In this study,the Einscan-pro-2x-plus 3D scanner from Hangzhou SHINING 3D Technology Co.,Ltd.was used to scan the stomach model.The simulation datasets generated by Blender contained the left image,the right image,and the ground truth disparity corresponding to the left image.Two thousand sets were selected as the training set and the remaining 100 sets as the testing set.The deep learning model used the RMSprop optimizer,and the learning rate was 1e?3.The training process lasted 40 epochs,and the maximum disparity value of the model was set as 192.

        A computer with an Intel Core i9-9900K CPU,an NVIDIA GeForce RTX 2080Ti GPU,and 32 GB of memory was used in our experiment.For comparison,we implemented the traditional semi-global block matching (SGBM) algorithm,fine-tuned for optimal performance.The deep learning algorithm and SGBM algorithm were realized by the PyTorch 1.9.1 deep learning framework in Ubuntu 19.04 and Microsoft Visual Studio 2017 in Windows 10,respectively.

        Fig.5 The 3D prototype endoscope (a) and the tip of the 3D endoscope (b)

        3.2 Experimental results and analysis

        First,the simulated binocular images were used to evaluate the effectiveness of the disparity map,and the results are shown in Fig.6.It can be seen from Fig.6 that the disparity map produced by the SGBM algorithm had a lot of noise in inadequate areas such as shadows,while the disparity map produced by the proposed method was closer to the ground truth disparity map.KPE,EPE,and the single image runtime of each method were calculated as shown in Table 1.Both the KPE and EPE of the proposed method have been greatly improved compared to the SGBM algorithm.The accuracy measured by EPE increased by 48.9%.In particular,whenKequaled 50,the KPE of the disparity map produced by the SGBM algorithm and our method were 2.22% and 0.46%,respectively,which shows that the probability that extremely serious errors occurred in our method is much lower than that with the SGBM algorithm.The runtime of the proposed method was 34 ms,which is 90.5% less than that of the SGBM algorithm.Hence,the proposed method can be used for real-time processing.

        Fig.6 Disparity results of simulated binocular images:(a) input left image;(b) disparity map produced using the SGBM algorithm;(c) disparity map produced by the proposed method;(d) ground truth disparity map

        Table 1 Disparity map performance comparison among different methods

        Similarly,the real binocular endoscopic images were used to evaluate the effectiveness of the disparity map,and the results are shown in Fig.7.Due to the more complicated imaging conditions in the actual environment,in the disparity map produced by the SGBM algorithm,a large amount of noise is densely distributed in the shadow and overexposed areas.By contrast,the disparity map generated by the proposed method had less noise and was smoother.In addition,the stability of 3D measurement by the proposed method and SGBM was compared.Three gastric rugae of the stomach model were selected as the curves to be measured (Fig.8).According to their disparity maps,the length of each curve was measured 10 times,and the corresponding average and standard deviation were calculated as shown in Table 2,where the incorrect measurement results obtained using the SGBM algorithm were highlighted and eliminated when calculating the average and standard deviation.From Table 2,the measurement error rate of the SGBM algorithm was approximately 50%,and the standard deviation was extremely large,which indicates that the noise generated by the SGBM algorithm will seriously affect the stability and reliability of 3D measurement.In contrast,the proposed method had no serious errors and the standard deviation was smaller,which demonstrates that the 3D measurement results of the proposed method are more stable and reliable.

        Fig.7 Disparity results of real endoscopic images:(a) input image (left);(b) disparity map produced by the SGBM algorithm;(c) disparity map produced by our method

        Fig.8 Three-dimensional measurement curves in the experiment

        The proposed method has the potential for further optimization and there are two possible major sources of the remaining errors in our approach:First,due to the unsatisfactory imaging conditions in the actual environment,there is a high probability that the real endoscopic images will contain texture-less areas such as overexposure and shadows,resulting in loss of image features in those local areas.Second,the discrepancy in shallow features,such as color and exposure,between the real endoscopic images and the virtual ones in the training sets affects the accuracy of network prediction.

        Table 2 Three-dimensional measurement performance comparison among different methods

        4 Conclusions

        In this paper,a simulated binocular dataset was generated by establishing a simulated gastroscope scenario through Blender to train a disparity estimation model for 3D measurement.The effectiveness of the disparity maps generated from simulated images was evaluated,and the results demonstrated that compared with the SGBM algorithm,the proposed method improved EPE by 48.9% and reduced the runtime by 90.5%.In addition,the proposed method did not create serious errors and had better 3D measurement stability.The method developed in this study further promotes the potential application of 3D measurement in the endoscopic environment and can provide accurate and reliable sizes to help endoscopists make accurate diagnoses.

        Contributors

        Hao YU and Bo YUAN designed the research.Hao YU,Changjiang ZHOU,and Wei ZHANG processed the data.Hao YU drafted the paper.Liqiang WANG and Qing YANG helped organize the paper.Hao YU revised and finalized the paper.

        Compliance with ethics guidelines

        Hao YU,Changjiang ZHOU,Wei ZHANG,Liqiang WANG,Qing YANG,and Bo YUAN declare that they have no conflict of interest.

        国产午夜亚洲精品不卡福利| 无码熟妇人妻av影音先锋| 97精品一区二区视频在线观看| 国产欧美日韩视频一区二区三区| 日韩欧美精品有码在线观看| 亚洲乱码av中文一区二区第八页 | 日本免费人成视频播放| 欧美成人高清手机在线视频| 国产我不卡在线观看免费| 丁香五月亚洲综合在线| 大香伊蕉国产av| 无码av免费精品一区二区三区| 日本一级三级在线观看| 丰满熟妇人妻av无码区| 亚洲av无码第一区二区三区 | 热re99久久精品国产99热| 亚洲高清在线不卡中文字幕网| 精品视频一区二区在线观看| 国产禁区一区二区三区| 国产精品免费看久久久8| 四虎精品国产一区二区三区| 免费黄网站一区二区三区| 一 级做人爱全视频在线看| 欧美gv在线观看| 三级全黄的视频在线观看| 国产免费操美女逼视频| 久久只精品99品免费久23| y111111少妇影院无码| 国产亚洲午夜高清国产拍精品不卡 | 久久国产成人精品国产成人亚洲| 国产mv在线天堂mv免费观看| 久九九久视频精品网站| 黄片视频大全在线免费播放| 中文字幕欧美人妻精品一区| 免费看奶头视频的网站| 日本一区人妻蜜桃臀中文字幕| 中国少妇×xxxx性裸交| 欧美喷潮久久久xxxxx| 日本第一区二区三区视频| 久久婷婷综合缴情亚洲狠狠| 在线亚洲+欧美+日本专区|