亚洲免费av电影一区二区三区,日韩爱爱视频,51精品视频一区二区三区,91视频爱爱,日韩欧美在线播放视频,中文字幕少妇AV,亚洲电影中文字幕,久久久久亚洲av成人网址,久久综合视频网站,国产在线不卡免费播放

        ?

        Medical image segmentation based on neural network

        2014-09-12 02:03:00WEIFeiLIUShoupeng

        WEI Fei, LIU Shou-peng

        (School of Health Management, Binzhou Medical University,Yantai 264003, China)

        0 Introduction

        The rapid development in the field of medical imaging has greatly promoted the progress of modern medicine. At present, computed tomography (CT), magnetic resonance imaging (MRI), ultrasound and other medical imaging technology has been widely used in clinical diagnosis and treatment. Segmentation is the basis for subsequent processing and three-dimensional visualization, surgical simulation and ultimate identification of diseased tissues. The accuracy of segmentation is crucial in correctly diagnosing patient’s condition; therefore image segmentation plays an important role in the field of medicine[1].

        1 The structure of neural network for image segmentation

        1.1 Back propagation neural network[2]

        Back propagation learning algorithm, or BP algorithm, was introduced by D.E.Rumelhard and W.S.McClelland in 1986.Back propagation neural network is a supervised learning model. It composes of feed-forward networks and backward propagation of errors, and is the most popular neural network modeling tool. Back propagation neural network method is able to derive the underlying data relationship via an arbitrary set of input data. The key to this learning algorithm is via steepest descent, in which the algorithm automatically adjusts its network weight and threshold in order to minimize network delta. Back propagation neural network’s topology consists of input, hidden layer and output layer.

        1.2 Parameter setting

        Let’s assume there are 2 layers of neural network, the corresponding value of its input, hidden layer and output layers areX,n,y.

        1) Number of neurons in input layer:X

        The role of the neural network is equivalent to threshold method, ie, if a given input is greater than the threshold, it will be in the foreground, else it goes to the background. Its value is the actual input vectorX=(x1,x2,…,xn)T.

        2) Number of neurons in output layer:y

        Node in this layer represents the output variable. In a multi-input, single-output system, the number of nodes in the layer defaults to 1, the initial value iswiunder rough membership function degree value. The output of this layer node

        (1)

        3) Number of neurons in hidden layer:n

        Dividennumber of input(x1,x2,…,xn)into different categories randomly. Assign a weight to reach input, the weight is to be between the value of [0,1]. Define the neuronal function of this layer as Gauss function

        (2)

        where,i=1,2,…,n,j=1,2,…,r,ras a discrete number of segmentation,mijas the center of the mean,σijdecided to its width.

        4) Each layer activation function

        Each layer of nodes representing a rule, assuming there arek(k≤n) rule, the layer nodes action functions as

        (3)

        Fig.1 is the structure of neural network.

        Fig.1 The structure of neural network

        2 Optimization

        Taking into account that the nature of neural networks is an m-dimensional input vectorsX=(X1,…,Xm) transformation to theq-dimensional output vectorsO=(O1,…,Oq) is the non-linear mapping[3]. During experiment, given an input sample vector, its output is the actual weight and set over’s independent variable function

        a=F(W,b),

        (4)

        where,Wandbare the weight matrix and set over the matrix,ais the actual output of the network.

        Assumetis the corresponding desired output,SNis the total number of sample, then the minimal mean square deviation can be expressed as (MSE is mean square error)

        (5)

        Constantly adjust the weights between nodes and set over, the result will eventually approximate the desired output.

        Relying purely on the network’s own algorithm to achieve network convergence tends to lead to local optimization. Therefore particle swarm optimization algorithm is used and implemented as shon in Fig.2 of the 3 mappings[4].

        Particle swarm optimization Neural Network

        Fig.2Mappingbetweenneuralnetworkandparticleswarmoptimization

        1) Mapping between neural network’s weights and particle dimension space

        Dimension of each particle in the particle swarm corresponds to a weight in the neural network. Vice versa, the weight and set over in the neural network equal to each particle in the particle swarm optimization.

        2) Mapping between neural network MSE and particle swarm optimization fitness function

        MSE of the neural network is a particle swarm optimization fitness function. It should be minimized via the powerful search performance provided by particle swarm optimization.

        3) Neural network learning and particle search

        The learning process of neural network is about continuously updating the weight and delta to minimize the MSE. The search process of particle swarm optimization is the dimensional change of speed and position of the particles. Taking into account that each particle corresponds to a neural network’s weight and set over, neural network learning process is equivalent to the search for the most optimal location of particles.

        3 Image segmentation

        3.1 Image area description

        Image regional boundaries are represented via regional content and area boundaries. Regional content is often differentiated via colors, texture and geometric meoments, while the regional boundary’s often differentiated by circular degree, rectangle degree etc.

        Fig.3 depicted the multiple images in the simple region. Image 2 is derived from image 1 through pan and zoom, image 4 is derived from image 2 via rotations and translations, while image 6 is derived from image 5 through rotation and scaling.

        Fig.3 Image area

        On the basis of a neural network as the classifier’s thoughts on the divided region, the image region matching into image area between matching, this can effectively reduce the complexity of image matching, improves the efficiency of the algorithm.

        Extraction of regional characteristics and measurement results as shown in Tab.1.

        Tab.1 Image region characteristic data sheet

        Where,A: acreage,P:circumference,C: circularity,F: contour complexity,S: roundness,R: rectangularity,Gavg: gray level, ф: area moment,W: texture.

        3.2 Data discretization

        Based on the maximum and minimum truncation point discretization algorithm simply puts the data into 3 categories, does not require any type of information, the algorithm is as follows: maximum and minimum truncation point discretization algorithm.

        Input:nsamples ofMfeature value data ( see Tab.1), the output: decision tableT=(U,C∪D,V,f).

        Step 1 The attribute value setVa=(C0a,C1a,…,CKa) in increasing order ( the same attribute value to take only one ) and divided into interval equivalence classes. ∪[Cia,Ci+1a], whicha∈C∪D,0≤i

        Step 2 Using the midpoint method to find out the interval [Cia,Ci+1a] truncatedCicomposed of truncated point setVa=(C0,C1,…,CK-1).

        Step 3 Minimum and maximum cut-off pointC0,CK-1.

        Step 4 Category tag.

        (6)

        For decision table, the tables in the same row are merged, to get Tab.2. In Tab.2, various features as condition attributes, add category as decision attribute.

        Tab.2 Decision table

        3.3 Attribute reduction[5]

        UsingA.Skowrondiscernibilitymatrixmethodofdecisiontablereductionstep2,asfollows:

        Step1CalculatethedecisionTab.2correspondingdiscernibilitymatrixM(C,D).

        Step2UsingthediscernibilitymatrixpropertiesthatattributeSisnuclear,deletethediscernibilitymatrixM(C,D)allcontaintheattributeSitem.

        Step3CalculationofthenumberofoccurrencesofeachattributeNA=NP=NG=Nφ=2, NR=1,sothereductionofattributesset{S,φ}and{S, G}.

        Asaresultofdecisiontablereduction,thispaperchooses{S,φ}.Theintuitivemeaningisthroughtheimageareaofthesphericalandregionalmomentinvariantfeaturesofdifferenceimagearea.

        3.4 Rule acquisition

        Accordingtotheobtainedreduction,Tab.2canbesimplifiedtoTab.3,inthesamebankmerger.

        Tab.3 The depicted multiple images in the simple region

        WhenAis an equivalence relation between objects in the domainU, thenU/Arepresents all equivalence classes of objects based on family relationshipUAcomposition.

        The decision rules are as follows:

        Rule of 1:S1ф1→Class1,

        Rule of 2:S1ф0→Class2,

        Rule of 3:S2ф1→Class3,

        Rule of 4:S0ф2→Class4.

        Apparently consistent decision table, for each of which a regulation is consistent.

        3.5 Principle

        Corresponding neural network model based on the above data processing methods, that is: the number of nodes in the first layer-4, the number of nodes in the second layer-4, the number of nodes in the third layer-4, fourth layer nodes is 1. The initial value of the connection weights between the third layer and fourth layer should be set as membership function degree value. Input of each neuron unit is the regional value, apply back propagation algorithm iterations, the output values are candidates for final decision results. Image segmentation is achieved through polymerization.

        4 Experimental results and analysis

        4.1 Experimental procedure

        Divide the 86 medical images collected from Internet into two categories. For the first category, manually segment the image using photoshop to achieve most optimal result, and this will be used as the sample input for neural network. The second category is the test sample[6-11].

        The initial structure of the neural network is set to 9-20-1, and then use the methods described in 3.3 to build the decision tree as shown in Fig.4. By looking at the decision tree, we eventually found seven major nodes {1,2,4,6,7,11,18}, therefore, the network results eventually identified as 9-7-1.

        Fig.4 The decision tree is used to determine the number of neurons in hidden layer

        4.2 Network experimental result

        (6)

        (7)

        (8)

        Among them,ηfor the learning rate,βfor the modified step coefficient,αfor inertia coefficient(0≤α≤1).

        The use of matlab language programming neural networkimage segmentation. The following image segmentation is shown in Fig.5.

        (a)Original image (b) Segmented image

        Fig.5(a) is the original image. Figure in the two larger cells are white blood cells, and the rest are small red blood cells. Fig.5 (b) shows the back propagation algorithm segmentation results. Experimental result indicates that segmentation provided a clearly image and highlight target area.

        This approach significantly reduces training time, improves accuracy, and is superior to conventional segmented images when it comes to meet real-time medical image processing requirements. It presented a whole new set of ideas that’s very effective.

        5 Conclusions

        This paper presents back propagation neural network based image segmentation approach. The experiments show that this method greatly reducing the training time and improve the accuracy, but will also be superior to conventional image segmentation, image processing to meet the real-time requirements. The method has great potential in the field of image segmentation, and its impact is still to be further investigated.

        :

        [1] Pawlak Z. Rough sets[J].International Journal of Information and Computer Science,1982,11:341-256.

        [2] Liu Q. Rough set and rough reasoning[M].Beijing:Science Press,2001.

        [3] Zeng H L. Rough set theory and its application[M].Chongqing: Chongqing University Press,1998.

        [4] Zeng H L, Zeng Q. The neural network based on rough set theory[J].Journal of Sichuan College of Chemical Light,2000,13(1):1-5.

        [5] Xu Z X, Ding Y L. A method based on rough neural networks of rough set theory[J].Nanjing University of Aeronautics and Astronautics Journal,2001,33(4):355-358.

        [6] Li N Y. Rough set theory and its application in image segmentation[J].Sanming Journal,2005,22(4):382-385.

        [7] Jelonek J. Rough set reduction of attributes and their domains for neural net-works[J].Computational Intelligence, 1995,11(2):339-347.

        [8] Zhang Y D, Wu L N. Optimizing weights of neural network using BCO[J].PIER,2008,83:185-198.

        [9] Zhang Y D, Wu L N. A novel pattern recognition method via PCNN and tsallis entropy[J].Sensors,2008,8(11):7518-7529.

        [10] Lin X M, Lv S S, Zhu D, et al. A new particle swarm optimization algorithm for medical image segmentation based on neural network[J].Journal of Changchun University of Technology: Nature Science Edition,2008,29(2):158-161.

        [11] Hong W C. Chaotic particle swarm optimization algorithm in a support vector regression electric load forecasting model[J].Energy Conversion and Management,2009,50:105-117.

        中文字幕在线亚洲日韩6页| 美腿丝袜视频在线观看| 激情人妻中出中文字幕一区| 日韩一区二区三区天堂| 国语对白免费观看123 | 久久天天躁夜夜躁狠狠85麻豆| 日本午夜精品理论片a级app发布| 伊人久久五月丁香综合中文亚洲 | 99伊人久久精品亚洲午夜| 国产老熟妇精品观看| 午夜毛片不卡免费观看视频| 激情另类小说区图片区视频区| 极品 在线 视频 大陆 国产| 欧美xxxxx精品| 国产女主播福利一区二区| 亚洲97成人在线视频| 久久精品国产精油按摩| 热99re久久精品这里都是免费| 久久精品国产一区二区涩涩| 亚洲天堂精品一区入口| 亚洲成在人网站av天堂| 午夜tv视频免费国产区4| 91精品亚洲一区二区三区| 精品女厕偷拍视频一区二区| 猫咪av成人永久网站在线观看| 亚洲国产成人精品无码区在线观看| 另类专区欧美在线亚洲免费| 久久少妇呻吟视频久久久| 亚洲av免费手机在线观看| 中文天堂国产最新| 激情另类小说区图片区视频区 | 国内最真实的xxxx人伦| 人妻少妇精品中文字幕av蜜桃| 日本理论片一区二区三区| 国产一区二区三区口爆在线| 亚洲中文字幕无码天然素人在线| 亚洲av无码专区亚洲av桃| 农村国产毛片一区二区三区女| 粉嫩的极品女神尤物在线| 亚洲 中文 欧美 日韩 在线| 国产网站视频|