Free Essay

Segmentation Using Neural Networks

In:

Submitted By prakrish2001
Words 2010
Pages 9
SEGMENTATION WITH NEURAL NETWORK

B.Prasanna Rahul Radhakrishnan
Valliammai Engineering College Valliammai Engineering College prakrish_2001@yahoo.com krish_rahul_1812@yahoo.com

Abstract: Our paper work is on Segmentation by Neural networks. Neural networks computation offers a wide range of different algorithms for both unsupervised clustering (UC) and supervised classification (SC). In this paper we approached an algorithmic method that aims to combine UC and SC, where the information obtained during UC is not discarded, but is used as an initial step toward subsequent SC. Thus, the power of both image analysis strategies can be combined in an integrative computational procedure. This is achieved by applying “Hyper-BF network”. Here we worked a different procedures for the training, preprocessing and vector quantization in the application to medical image segmentation and also present the segmentation results for multispectral 3D MRI data sets of the human brain with respect to the tissue classes “ Gray matter”, “ White matter” and “ Cerebrospinal fluid”. We correlate manual and semi automatic methods with the results.

Keywords: Image analysis, Hebbian learning rule, Euclidean metric, multi spectral image segmentation, contour tracing.

Introduction: Segmentation can be defined as the identification of meaningful image components. It is a fundamental task in image processing providing the basis for any kind of further highlevel image analysis. In medical image processing, a wide range of application is based on segmentation. A possible realization of high-level image analysis principle is the acquisition and processing of multisprectral image data sets, which forms the basis of the segmentation approach. A good survey is provided by the list of citations published in [1] that may serve as a good starting point for further reference. Different segmentation methods range from simple histogram-based thresholding or region growing algorithms, to more sophisticated techniques such as active contours or watershed transformation. Appropriate preprocessing steps comprise the anatomically correct registration of the data sets and masking a region of interest in which the segmentation should be performed. Data analysis may be performed by two different strategies. The first one tries to identify characteristic properties of the multidimensional data distribution of unlabeled feature vectors. This approach is called as unsupervised clustering (UC). The second strategy involves labeled data, i.e., the learning algorithm requires both the feature vector itself and a target function defining its interpretation with regard to segmentation class membership. We call it supervised classification (SC).

Structure and function of the Hyber-BF network: The general architecture of a
Hyper-BF network is shown in the figure It consists of three layers of neurons:
Input layer, hidden layer, and outer layer

[pic] n” input neurons with activations xi , i €{1,……,n}, the activation pattern of the input layer is represented by an n-dimensional vector x in the feature space R. This activation is propagated to the N neurons of the hidden layer by directed connections with synaptic weights Wji. The synaptic weights Wj € R, j € {1,……, N}, Are computed as a set of prototypical vectors that represent the data set in the feature space. The activation aj of the hidden layer neuron j is chosen as a function of distance d=||x-wj|| of the data vector x with respect to the virtual position wj of the hidden layer neuron j. d hereby defines an arbitrary metric in the feature space, e.g., the Euclidean metric. The term virtual position is based on the idea that the activation aj of the hidden layer neuron should take its maximum value xmax, which can be looked at as a specification of the neuron j with respect to the position xmax. It is obviously reasonable to choose aj as a monotonically declining function of d, i.e., the activation aj of the hidden layer neuron should decline with increasing distance between x and the virtual position wj . A simple choice is an isotropically decreasing function aj,i.e., he declining behavior does not depend on the direction of the difference vector (x-wj). From this results a symmetry with respect to rotation, i.e., a radial decline of aj(x) in the neighborhood of wj: Therefore, we refer to the activation function aj(x) as a Hyper-BF network.With respect to favorable properties regarding function approximation, F.Girosi and T.poggio [2] .The Gaussian activation functions.
[pic]
Moody and Darken [3] propose a global normalization of the hidden layer activation by
[pic]

Which results in a hidden layer activation of
[pic]

The radial symmetry of the activation function is obviously lost by the normalization.
In [2] such a system is called “Hyper-BF Network. In the final step, a linear signal propagation of the hidden layer activation is performed to the m neurons of an output layer by weighted summation.
[pic]

Or in terms of single components,

[pic] In neural network computing this mapping corresponds to a structure called the perceptron (Rosenblatt [4]).

Training procedure: The memorized “knowledge” of the network is represented by the parameter P={(wj,sj,pj)} ,i.e., the weights wj and sj as well as the range pj of the Gaussians in the hidden layer .Here, The two aspects of should be considered : On one hand , the network should reproduce the given training data as well as possible ; on the hand , it should be able to provide some reasonable generalization with respect to unknown test data. An obvious training strategy is the minimization of the error.

[pic]

An attempt to simultaneous optimization of all the free parameters by gradient descend on the cost function. A proposal by Moody and Darken [3] is to split up the parameter space p={(wj,sj,pj)} into three smaller partitions {wj},{sj},{pj} that may be optimized separately. The resulting dimensionality reduction involves an increased convergence rate for the optimization process within the different partitions. This concept of separate optimization within each of the three partitions of p is the basis of the Hyper-BF architecture.

Optimization: Initially the virtue positions wj of the hidden layer neurons are calculated in a procedure called vector quantization (VQ).This is based on an unsupervised clustering Algorithm VQ algorithm can be classified according to the cooperativity function aj .One can distinguish between so called hard clustering method that assign each feature vector x to exactly one cluster represented by the Codebook vector wj and soft clustering methods of fuzzy membership of feature vectors in several clusters. A simple method for hard clustering is the algorithm proposed by Linde, Buzo, and Gray (LBG) [5]. The procedure can be monitored by various control parameters such as the free energy, entropy and reconstruction error, which enable an easy detection of cluster splitting. This version is well suited for implementation on parallel computers .The output weights can be calculated in a procedure called supervised learning. They can be determined by two different ways: global training procedure and local training procedure based on Hebbian learning rule.

[pic]
Application to medical image segmentation: Hyper-BF neural network provides the basis for the application to automatic segmentation of magnetic resonance imaging data sets of the human brain. The four acquisition sequences are T1 weighted, T2 weighted, proton density weighted and inversion recovery sequences. Segmentation aims at classifying each voxel of the multispectra data set as belonging to a specific tissue type, thus obtaining information about structure and volume of the tissue classes. A classical problem with numerous clinical applications is the segmentation of brain imaging data.
With respect to the tissue classes gray matter, white matter, and cerebrospinal fluid. A threefold classification of brain tissue may be sufficient for numerous clinical applications, it should be emphasized that the concept presented in
This paper can be extended to an arbitrary number of tissue classes.
Especially one may think of introducing additional classes for the identification of pathological tissue.

Image Data
[pic]
The image data sets were obtained on 1.5T whole body MRI scanner (siemens, Magneton vision). Four persons were examined in a prospective study employing a standardized MRI sequence protocol: T1 weighted MP-RAGE, T2 weighted and proton density (PD) weighted spin echo, and Inversion-Recovery (IR) sequences. The MRI acquisition parameter of each sequence are listed in table.

[pic]

PREPROCESSING The concept of voxel- based multispectral image segmentation requires anatomically correct alignment of the data sets acquired within different image acquisition procedures. This is known as image registration. All the extracerebral structure those are not required for the tissue classification task should be excluded from the data set. This is pre-segmentation. There is a wide scope of methods for pre-segmentation ranging from manual contour tracing to semiautomatic or fully automatic procedures. The latter exist in numerous implementations and are available on a commercial basis. They are frequently based on filter operations or region-growing algorithms

[pic]

Classification Vector quantization can determine a set of prototypical codebook vectors wj representing the feature space, a set of feature vectors x in a grey-level feature space G. This provides the basis for segmentation of the imaging data set with respect to different tissue classes. Two alternative approaches are Interactive assignment of codebook vectors and supervised classification .The comparison of above approaches are discussed in figure.

[pic]

Results
The paper represents the results for the segmentation of 3D MRI data sets of the human brain. Multispectral MRI data were acquired in four healthy human volunteers. Initially, the four data sets were preprocessed according to the procedure explained before.
After preprocessing, the data sets were segmented by applying two different strategies:
Manual interactive assignment of codebook vectors to tissue classes. At first the result for the vector quantization of the gray value feature space with manual interactive assignment of codebook vector of tissue classes will be discussed .The average manual processing time for the assignment of codebook vectors to tissue classes are 6 minutes per data set, i.e., the procedure requires only a little human intervention. However this approach yields suboptimal results, especially in case where a code book vector cannot be attributed to a single tissue class without ambiguity. Sometimes a codebook vector is positioned at the border between two tissue classes, which makes a clear assignment difficult or even impossible.
Semiautomatic classification by Hyper-BF network
In the following, the results for classification by a Hyper-BF network after preceding vector quantization of the gray level feature space are discussed. The training data comprise 1%of the whole data set each. This approach yields best segmentation result than unsupervised clustering with respect to subjective evaluation by a neuroradiologist. The improvement of segmentation quality was accompanied by a considerable increase of human intervention. Manual processing time for interactive labeling of the training data was approximately 30 minutes per data set. Thus supervised classification requires more human intervention than unsupervised clustering.

Discussion In this paper we have presented a neural network approach to automatic segmentation of MRI data sets of the human brain by neural network computation. The Hyper-BF network computation. The Hyper-BF architecture enables an effective combination of unsupervised and supervised learning procedures. Although the results mentioned above are very encouraging, several critical issues still need to be discussed with regard to the data, the aim of segmentation and the validation of the results. In summary, a wide range of further research topics has to be covered in future studies based on the results presented in this article. Such investigations can help to introduce automatic neural network segmentation as a cost-effective and reliable tool for routine medical image processing according to the growing importance of quantitative image analysis techniques for clinical decision making.

References
1.A.J.Worth . http://demonmac.mgh.harvard.edu/cma/seg_f/seg_refs.html. 2.F.Girosi and T.Poggio, Networks and the best approximation property. A.I.Memo 1164,Massachusetts Institute of Technology, 10,1989.
3.J.Moody and C.Darken. Fast learning in networks of locally tuned processing units.
4. F.Rosenblatt.Principles of neurometrics. Spartans books,1962.
5.Y.Linde,A.Buzo,and R.M.Gray . An algorithm for vector quantizer design.
6.D.O.Hebb.The organization of behaviour.

[pic]

Similar Documents

Premium Essay

Market Segmentation

...reserved. Printed in Great Britain 0160-7383/$30.00 doi:10.1016/j.annals.2004.05.001 MARKET SEGMENTATION A Neural Network Application Jonathan Z. Bloom University of Stellenbosch, South Africa Abstract: The objective of the research is to consider a self-organizing neural network for segmenting the international tourist market to Cape Town, South Africa. A backpropagation neural network is used to complement the segmentation by generating additional knowledge based on input–output relationship and sensitivity analyses. The findings of the self-organizing neural network indicate three clusters, which are visually confirmed by developing a comparative model based on the test data set. The research also demonstrated that Cape Metropolitan Tourism could deploy the neural network models and track the changing behavior of tourists within and between segments. Marketing implications for the Cape are also highlighted. Keywords: segmentation, SOM neural network, input–output analysis, sensitivity analysis, deployment. Ó 2005 Elsevier Ltd. All rights reserved. ´ ´ Resume: Segmentation du marche: une application du reseau neuronal. Le but de la ´ ´ recherche est de considerer un reseau neuronal auto-organisateur pour segmenter le marche ´ ´ ´ touristique international a Cape Town, en Afrique du Sud. On utilise un reseau neuronal de ` ´ retropropogation pour completer la segmentation en generant des connaissances comple´ ´ ´ ´ ´ mentaires basees sur une relation input–output et des...

Words: 7968 - Pages: 32

Premium Essay

Diabetic Retinopathy Literature Review

...Estimation of diabetic retinopathy with artery/vein classification in retinal images using Artificial Neural Network Leshmi Satheesh M.Tech Student,Dept. of Electronics & Communication Mohandas College of Engineering, Kerala University Trivandrum-695541, Kerala, India Email: leshmi24@gmail.com Abstract—Diabetic retinopathy is the single largest explanation for sight loss and visual impairment in eighteen to sixty five year olds . Damage of blood vessels in the eye and the formation of lesions in the retina are the earliest signs of diabetic retinopathy. Efficient image processing and analysis algorithms have to be developed for the automated screening programs to work robustly and effectively. For the detection of vascular changes...

Words: 3295 - Pages: 14

Premium Essay

Brain Hemorrhage Case Study

...Abstract— Brain hemorrhage is a type of stroke, which occurs due to artery bursting in the brain, causing bleeding in the surrounded tissues. Diagnosing brain hemorrhage, which is mainly through the examination of a CT scan enables the accurate prediction of disease and the extraction of reliable and robust measurement for patients in order to describe the morphological changes in the brain as the recovery progresses. The aim of this project is to help radiologist as well as medical students in diagnosis of brain hemorrhage in more refined manner by feeding CT images & identify the type of brain hemorrhage using watershed algorithm along with artificial neural network. Keywords - Computed tomography (CT), Artificial Neural Networks (ANN), Back Propagation Network (BPN), Gray Level Co-occurrence Matrix (GLCM). I. INTRODUCTION This Brain hemorrhage is a type of stroke, which occurs due to artery bursting in the brain, causing bleeding in the surrounded tissues. The symptoms of brain hemorrhage are a sudden severe headache, Weakness in an arm or leg, nausea or vomiting, changes in vision, difficulty in speaking or understanding speech, difficulty in...

Words: 1573 - Pages: 7

Free Essay

Optical Character Recognition

...ISSN: 2277-3754 ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 2, Issue 7, January 2013 Recognition for Handwritten English Letters: A Review Nisha Sharma, Tushar Patnaik, Bhupendra Kumar Abstract -- Character recognition is one of the most interesting and challenging research areas in the field of Image processing. English character recognition has been extensively studied in the last half century. Nowadays different methodologies are in widespread use for character recognition. Document verification, digital library, reading bank deposit slips, reading postal addresses, extracting information from cheques, data entry, applications for credit cards, health insurance, loans, tax forms etc. are application areas of digital document processing. This paper gives an overview of research work carried out for recognition of hand written English letters. In Hand written text there is no constraint on the writing style. Hand written letters are difficult to recognize due to diverse human handwriting style, variation in angle, size and shape of letters. Various approaches of hand written character recognition are discussed here along with their performance. Fig 1.Major Steps of an OCR System Index Terms— Offline Hand written Character Recognition, Pre-Processing, Feature Extraction, Classification, Post Processing. I. INTRODUCTION Optical Character Recognition (OCR) is one of the most fascinating and challenging areas of pattern...

Words: 2982 - Pages: 12

Premium Essay

Dental X-Ray Analysis

...Dental X-Ray Image Analysis by Using Image Processing Techniques Qno.1. Identify the tools and technologies used in the paper. Elaborate or give details of all the tools used in the paper. In this paper author focused on to detect dental Caries. It is diseases that should be occurs in to teeth, so author used different digital image processing tools and techniques such as, Dental X-Rays, Image Processing, Image Segmentation and Binarization these all of defined and elaborate Below. 1. Dental X-Rays 2. Image Processing a) Image segmentation b) Binarization dental x-ray image is used by the dentist for the diagnosis and treatment purpose such as root canal treatment, detection of different factor similar to the...

Words: 3714 - Pages: 15

Free Essay

Neural Network

...– MGT 501 Neural Network Technique Outline * Overview ………………………………………………………….……… 4 * Definition …………………………………………………4 * The Basics of Neural Networks……………………………………………5 * Major Components of an Artificial Neuron………………………………..5 * Applications of Neural Networks ……………….9 * Advantages and Disadvantages of Neural Networks……………………...12 * Example……………………………………………………………………14 * Conclusion …………………………………………………………………14 Overview One of the most crucial and dominant subjects in management studies is finding more effective tools for complicated managerial problems, and due to the advancement of computer and communication technology, tools used in management decisions have undergone a massive change. Artificial Neural Networks (ANNs) is an example, knowing that it has become a critical component of business intelligence. The below article describes the basics of neural networks as well as some work done on the application of ANNs in management sciences. Definition of a Neural Network? The simplest definition of a neural network, particularly referred to as an 'artificial' neural network (ANN), is provided by the inventor of one of the first neurocomputers, Dr. Robert Hecht-Nielsen who defines a neural network as follows: "...a computing system made up of a number of simple, highly interconnected processing elements, which process information by their dynamic state response to external inputs."Neural Network Primer: Part...

Words: 3829 - Pages: 16

Free Essay

Stereoscopic Building Reconstruction Using High-Resolution Satellite Image Data

...Stereoscopic Building Reconstruction Using High-Resolution Satellite Image Data Anonymous submission Abstract—This paper presents a novel approach for the generation of 3D building model from satellite image data. The main idea of 3D modeling is based on the grouping of 3D line segments. The divergence-based centroid neural network is employed in the grouping process. Prior to the grouping process, 3D line segments are extracted with the aid of the elevation information obtained by using area-based stereo matching of satellite image data. High-resolution IKONOS stereo images are utilized for the experiments. The experimental result proved the applicability and efficiency of the approach in dealing with 3D building modeling from high-resolution satellite imagery. Index Terms—building model, satellite image, 3D modeling, line segment, stereo I. I NTRODUCTION Extraction of 3D building model is one of the important problems in the generation of an urban model. The process aims to detect and describe the 3D rooftop model from complex scene of satellite imagery. The automated extraction of the 3D rooftop model can be considered as an essential process in dealing with 3D modeling in the urban area. There has been a significant body of research in 3D reconstruction from high-resolution satellite imagery. Even though a natural terrain can be successfully reconstructed in a precise manner by using correlation-based stereoscopic processing of satellite images [1], 3D building reconstruction...

Words: 2888 - Pages: 12

Free Essay

Electronic Voting

...Project – an Automated Make-up color selection system. Supervisor – Dr. H.L.Premarathne Field(s) of concern – Artificial Neural Networks, Fuzzy Logic, Image Processing, Data Classification, make-up Background: Women typically like to be in the centre of attraction of other the people. In order to be elegant looking and to get the attention of others, ladies often use make-up. Make-up is a favorite topic of women, and is a primary concern, not only when attending functions such as weddings, parties, but in day-to-day life when going for work too. The success of make-up relies on how well one can select the colors that matches her skin color, eye color, shape of the face and other relevant features. Make-up is also an art; hence one should have a good artistic eye to select the make-up which suits her. Inappropriate applying of make-up will cause a person to be in the centre of sarcasm and annoyance, instead of being in the centre of attraction. This is why; ladies often take the service of a beautician. A beautician is a professional who’s trained and who has expertise knowledge on beauty therapy and make-up. With experience, a beautician can match the make-up colors to suit a person, according to her appearance and personality. However, one does not need the help of a beautician, if that person can choose the appropriate make-up colors for herself. Introduction: Selection of colors for a make-up is vital for a Beautician as well as for any lady who rely on make-up...

Words: 1003 - Pages: 5

Premium Essay

Data Minig

...automated, future-ori- ented analyses made possible by data mining move beyond the analyses of past events typically provided by history-oriented tools such as decision support systems. Data mining tools answer business questions that in the past were too time-consuming to pursue. Yet, it is the answers to these questions make customer relationship management possible. Various techniques exist among data mining software, each with their own advantages and challenges for different types of applications. A particular dichotomy exists between neural networks and chi-square automated interaction detection (CHAID). While differing approaches abound in the realm of data mining, the use of some type of data mining is necessary to accomplish the goals of today’s customer relationship management philosophy.  2002 Elsevier Science Ltd. All rights reserved. Keywords: Customer relationship management (CRM); Relationship marketing; Data mining; Neural net- works; Chi-square automated interaction...

Words: 8031 - Pages: 33

Premium Essay

Nt1310 Unit 1 Exercise 1

...battery can be inserted in the Gnd and Vin pin headers of the POWER connector From an external supply of 6 to 20 volts the board will operate. If supply with less than 7V, the board may be unstable. If more than 12V of supply is used, the voltage regulator may overheat and damage the board. 7 to 12 volts is the recommended range. The power pins are as follows: • VIN. The input voltage to the Arduino board when it's using an external power source. You can supply voltage through this pin, or, if supplying voltage thru the power jack, access it through this pin. • 5V. The regulated power supply used to...

Words: 1473 - Pages: 6

Free Essay

A Survey of Ocr Applications

...International Journal of Machine Learning and Computing, Vol. 2, No. 3, June 2012 A Survey of OCR Applications Amarjot Singh, Ketan Bacchuwar, and Akshay Bhasin Abstract—Optical Character Recognition or OCR is the electronic translation of handwritten, typewritten or printed text into machine translated images. It is widely used to recognize and search text from electronic documents or to publish the text on a website. The paper presents a survey of applications of OCR in different fields and further presents the experimentation for three important applications such as Captcha, Institutional Repository and Optical Music Character Recognition. We make use of an enhanced image segmentation algorithm based on histogram equalization using genetic algorithms for optical character recognition. The paper will act as a good literature survey for researchers starting to work in the field of optical character recognition. Index Terms— Genetic algorithm, bimodal images, Captcha, institutional repositories and digital libraries, optical music recognition, optical character recognition. I. INTRODUCTION Highlight in 1950’s [1], applied throughout the spectrum of industries resulting into revolutionizing the document management process. Optical Character Recognition or OCR has enabled scanned documents to become more than just image files, turning into fully searchable documents with text content recognized by computers. Optical Character Recognition extracts the relevant information...

Words: 3379 - Pages: 14

Free Essay

Illicit Image Filtering and Classification Techniques

...Professor Department of E&TC Engineering Sinhgad College of Engineering Pune, Maharashtra (India) Abstract— Adult images have become a threat to social networking, which are having a negative impact, on society. We often hear the stories of blackmailing for adult images which is a serious crime. This should be stopped somewhere, further we will see various techniques to do so. We present a method to classify images into different categories of pornographic content to create a system for filtering illicit images from network traffic. Even though different systems for this application were offered in the past, which are based on simple skin color features and have quite poor performance. Latest advances in the image recognition field in particular for the classification of objects have shown that bag-of-visualwords-approaches are a good method for many image classification problems. Keywords—Bag-of-Visual Words (BoVW), Explicit Content, Neural Network, Region-of-Interest (ROI), Skin Detection I. INTRODUCTION II. The inevitable fact that technology is becoming more intertwined in the daily life of the individual will lead to an increase in court cases and in crime. Due to such image uploads as people have different mentality and culture which causes several serious issues and may revolve into crime. So here we are presenting some of the techniques of filtering and classification of pornographic images. The internet has made people access more information...

Words: 2230 - Pages: 9

Premium Essay

Data Mininng

...An Overview of Data Mining Techniques Page 1 of 48 An Overview of Data Mining Techniques Excerpted from the book Building Data Mining Applications for CRM by Alex Berson, Stephen Smith, and Kurt Thearling Introduction This overview provides a description of some of the most common data mining algorithms in use today. We have broken the discussion into two sections, each with a specific theme: Classical Techniques: Statistics, Neighborhoods and Clustering Next Generation Techniques: Trees, Networks and Rules Each section will describe a number of data mining algorithms at a high level, focusing on the "big picture" so that the reader will be able to understand how each algorithm fits into the landscape of data mining techniques. Overall, six broad classes of data mining algorithms are covered. Although there are a number of other algorithms and many variations of the techniques described, one of the algorithms from this group of six is almost always used in real world deployments of data mining systems. I. Classical Techniques: Statistics, Neighborhoods and Clustering 1.1. The Classics These two sections have been broken up based on when the data mining technique was developed and when it became technically mature enough to be used for business, especially for aiding in the optimization of customer relationship management systems. Thus this section contains descriptions of techniques that have classically been used for decades the next section represents techniques...

Words: 23868 - Pages: 96

Free Essay

Bolta Mosam

...continuous speech recognition system :- In this paper, they describe BYBLOS, the BBN continuous speech recognition system. The system, designed for large vocabulary applications, integrates acoustic, phonetic, lexical, and linguistic knowledge sources to achieve high recognition performance. The basic approach it makes is the extensive use of robust context-dependent models of phonetic coarticulation using Hidden Markov Models (HMM). It describes the components of the BYBLOS system, including: signal processing frontend, dictionary, phonetic model training system, word model generator, grammar and decoder. In recognition experiments, it demonstrates consistently high word recognition performance on continuous speech across: speakers, task domains, and grammars of varying complexity. In speaker-dependent mode, where 15 minutes of speech is required for training to a speaker, 98.5% word accuracy has been achieved in continuous speech for a 350-word task, using grammars with perplexity ranging from 30 to 60. With only 15 seconds of training speech we demonstrate performance of 97% using a grammar. http://ieeexplore.ieee.org/xpl/freeabs_all.jsp?arnumber=1169748 Audio-visual modeling for bimodal speech recognition:- Audio-visual speech recognition is a novel extension of acoustic speech recognition and has received a lot of attention in the last few decades. The main motivation behind bimodal speech recognition is the bimodal characteristics of speech perception and production systems...

Words: 2115 - Pages: 9

Premium Essay

Datamining

...MSc. Information System Management Kyaw Khine Soe (3026039) Data Mining and Business Analytics Boston Housing Dataset Analysis. Table of Contents Introduction 3 Problem Statement 3 The associated data of Boston 5 Data pre-processing / Data preparation 8 Clustering Analysis 11 Cluster segment profile 17 Regression Analysis 18 Predictive analysis using neural network node 19 Decision tree node 21 Regression node analysis 23 Model Comparison 24 The recommendation and conclusion 26 Bibliography 27 Introduction This report included part of assignment for the Data Mining and Business Analytics. This report based on the Boston Housing Dataset to describe prediction, cluster analysis, neural networks and decision tree nodes. Boston Housing is a real estate related dataset from Boston Massachusetts. This is small dataset with 506 rows can show prediction of housing price and regressing using decision trees and neural networks over this dataset. This report shows analysis of the property price over the size, age of property, environment factor such as crime rate, near the river dummy, distanced to employment centers and pollution. Problem Statement In relation to housing intelligence, real estate are usually concerned with following common business concerns: 1. Which area are high rates of crime? How crimes rates effected on housing price? How can reduce the crime? 2. Which area is most/lease house price base on rooms in house/ area and...

Words: 2101 - Pages: 9