Free Essay

Support Vector Machines

In:

Submitted By Suryansh
Words 2848
Pages 12
Support Vector Machines

Operations Management Project Report by-

Suryansh Kapoor

PGPM (2011 – 2013)

11P171

Section – ‘C’

Supervised by- Prof. Manoj Srivastava

Abstract

In today’s highly competitive world markets, high reliability plays increasingly important role in the modern manufacturing industry. Accurate reliability predictions enable companies to make informed decisions when choosing among competing designs or architecture proposals. This is all the more important in case of specialized fields where operations management is a necessary requirement. Therefore, predicting machine reliability is necessary in order to execute predictive maintenance, which has reported benefits include reduced downtime, lower maintenance costs, and reduction of unexpected catastrophic failures. Here, the role of Support Vector Machines or SVMs comes in to predict the reliability of the necessary equipment. SVMs are cited by various sources in the field of medical researches⁶ and other non-mining fields¹ to be better than other classifying methods like Monte-Carlo simulation etc. because SVM models have nonlinear mapping capabilities, and so can more easily capture reliability data patterns than can other models. The SVM model minimizes structural risk rather than minimizing training errors improves the generalization ability of the models.

Contents

1. Objective

2. Literature Review * Introduction of Reliability

* Introduction of Support Vector Machine

3. Case Study

4. Conclusion

5. References

Objective –

The objective of this project is to give a general and temporary overview of the SVM algorithms and the concepts and methodology associated with it, as given by various sources²̕³ in different fields. Basically, the specific methodology adopted by one source´ is given to be understood and its subsequent application to the field of mining engineering is the primary objective.

Literature review –

Introduction to reliability -

Reliability may be defined in several ways:

* The idea that something is fit for a purpose with respect to time;

* The capacity of a device or system to perform as designed;

* The resistance to failure of a device or system;

* The ability of a device or system to perform a required function under stated conditions for a specified period of time

* The probability that a functional unit will perform its required function for a specified interval under stated conditions.

* The ability of something to "fail well" (fail without catastrophic consequences)

Reliability analysis relies heavily on statistics, probability theory, and reliability theory. Many engineering techniques are used in reliability analysis, such as reliability prediction, Weibull analysis, thermal management, reliability testing and SVMs. Because of the large number of reliability techniques, their expense, and the varying degrees of reliability required for different situations, most projects develop a reliability program plan to specify the reliability tasks that will be performed for that specific system.

The purpose of reliability testing is to discover potential problems with the design as early as possible and, ultimately, provide confidence that the system meets its reliability requirements.

Reliability testing may be performed at several levels. Complex systems may be tested at component, circuit board, unit, assembly, subsystem and system levels. (The test level nomenclature varies among applications.) For example, performing environmental stress screening tests at lower levels, such as piece parts or small assemblies, catches problems before they cause failures at higher levels. Testing proceeds during each level of integration through full-up system testing, developmental testing, and operational testing, thereby reducing program risk. System reliability is calculated at each test level. Reliability growth techniques and failure reporting, analysis and corrective active systems (FRACAS) are often employed to improve reliability as testing progresses.

Introduction to Support Vector Machine⁴ –

Support vector machines (SVMs) are a set of related supervised learning methods that analyze data and recognize patterns, used for classification and regression analysis. The original SVM algorithm was invented by Vladimir Vapnik and the current standard incarnation (soft margin) was proposed by Corinna Cortes and Vladimir Vapnik. The standard SVM is a non-probabilistic binary linear classifier, i.e. it predicts, for each given input, which of two possible classes the input is a member of. Since an SVM is a classifier, then given a set of training examples, each marked as belonging to one of two categories, an SVM training algorithm builds a model that predicts whether a new example falls into one category or the other. Intuitively, an SVM model is a representation of the examples as points in space, mapped so that the examples of the separate categories are divided by a clear gap that is as wide as possible. New examples are then mapped into that same space and predicted to belong to a category based on which side of the gap they fall on.

More formally, a support vector machine constructs a hyperplane or set of hyperplanes in a high or infinite dimensional space, which can be used for classification, regression or other tasks. Intuitively, a good separation is achieved by the hyperplane that has the largest distance to the nearest training data points of any class (so-called functional margin), since in general the larger the margin the lower the generalization error of the classifier.

Whereas the original problem may be stated in a finite dimensional space, it often happens that in that space the sets to be discriminated are not linearly separable. For this reason it was proposed that the original finite dimensional space be mapped into a much higher dimensional space presumably making the separation easier in that space. SVM schemes use a mapping into a larger space so that cross products may be computed easily in terms of the variables in the original space making the computational load reasonable. The cross products in the larger space are defined in terms of a kernel function K(x,y) which can be selected to suit the problem. The hyperplanes in the large space are defined as the set of points whose cross product with a vector in that space is constant. The vectors defining the hyperplanes can be chosen to be linear combinations with parameters αi of images of feature vectors which occur in the data base. With this choice of a hyperplane the points x in the feature space which are mapped into the hyperplane are defined by the relation:

∑ αiK(xi,x) = constant i

Formalisation´ –

We are given some training data , a set of n points of the form

where the ci is either 1 or −1, indicating the class to which the point belongs. Each is a p-dimensional real vector. We want

to find the maximum-margin hyperplane that divides the points having ci = 1 from those having ci = − 1. Any hyperplane can be written as the set of points | satisfying w.x – b = 0 where | denotes the dot product. The vector | is a normal vector: it is | perpendicular to the hyperplane. The parameter | determines |

the offset of the hyperplane from the origin along the normal vector .

Maximum-margin hyperplane and margins for an SVM trained with samples from two classes. Samples on the margin are called the support vectors.

We want to choose the and b to maximize the margin, or distance between the parallel hyperplanes that are as far apart as

possible while still separating the data. These hyperplanes can be described by the equations

and

Note that if the training data are linearly separable, we can select the two hyperplanes of the margin in a way that there are no points between them and then try to maximize their distance. By using geometry, we find the distance between these two

hyperplanes is , so we want to minimize . As we also have to prevent data points falling into the margin, we add the following constraint: for each i either

of the first class

or

of the second.

This can be rewritten as:

We can put this together to get the optimization problem:

Minimize (in )

subject to (for any )

Primal form

The optimization problem presented in the preceding section is difficult to solve because it depends on ||w||, the norm of w, which involves a square root. Fortunately it is possible to alter the equation by substituting ||w|| with (the factor of 1/2 being used for mathematical convenience) without changing the solution (the minimum of the original and the modified equation have the same w and b). This is a quadratic programming (QP) optimization problem. More clearly:

Minimize (in )

subject to (for any )

One could be tempted to express the previous problem by means of non-negative Lagrange multipliers αi as

but this would be wrong. The reason is the following: suppose we can find a family of hyperplanes which divide the points; then all . Hence we could find the minimum by sending all αi to , and this minimum would be reached for

all the members of the family, not only for the best one which can be chosen solving the original problem.

Nevertheless the previous constrained problem can be expressed as

that is we look for a saddle point. In doing so all the points which can be separated as do not matter since we must set the corresponding αi to zero.

This problem can now be solved by standard quadratic programming techniques and programs. The solution can be expressed by terms of linear combination of the training vectors as

Only a few αi will be greater than zero. The corresponding are exactly the support vectors, which lie on the margin and satisfy

. From this one can derive that the support vectors also satisfy

which allows one to define the offset b. In practice, it is more robust to average over all NSV support vectors:

Dual form

Writing the classification rule in its unconstrained dual form reveals that the maximum margin hyperplane and therefore the classification task is only a function of the support vectors, the training data that lie on the margin.

Using the fact, that and substituting , one can show that the dual of the SVM reduces to the following optimization problem:

Maximize (in αi )

subject to (for any )

and to the constraint from the minimization in b

Here the kernel is defined by .

The α terms constitute a dual representation for the weight vector in terms of the training set:

Biased and unbiased hyperplanes

For simplicity reasons, sometimes it is required that the hyperplane passes through the origin of the coordinate system. Such hyperplanes are called unbiased, whereas general hyperplanes not necessarily passing through the origin are called biased. An unbiased hyperplane can be enforced by setting b = 0 in the primal optimization problem. The corresponding dual is identical to the dual given above without the equality constraint

Non-linear classification

The original optimal hyperplane algorithm proposed by Vladimir Vapnik in 1963 was a linear classifier. However, in 1992, Bernhard Boser, Isabelle Guyon and Vapnik suggested a way to create non-linear classifiers by applying the kernel trick to maximum-margin hyperplanes. The resulting algorithm is formally similar, except that every dot product is replaced by a non-linear kernel function. This allows the algorithm to fit the maximum-margin hyperplane in a transformed feature space. The transformation may be non-linear and the transformed space high dimensional; thus though the classifier is a hyperplane in the high-dimensional feature space, it may be non-linear in the original input space.

If the kernel used is a Gaussian radial basis function, the corresponding feature space is a Hilbert space of infinite dimension. Maximum margin classifiers are well regularized, so the infinite dimension does not spoil the results. Some common kernels include,

* Polynomial (homogeneous)

* Polynomial (inhomogeneous):

* Gaussian or Radial Basis Function:

, for γ > 0. Sometimes parameterized using γ = 1 / 2σ2  Hyperbolic tangent: , for some
(not every) κ > 0 and c < 0

The kernel is related to the transform by the equation

. The value w is also in the transformed space, with Dot products with w for

classification can again be computed by the kernel trick, i.e.

. However, there does not in general exist a value w' such that

Parameter selection

The effectiveness of SVM depends on the selection of kernel, kernel's parameters and soft margin parameter C.

A common choice is a Gaussian kernel, which has a single parameter γ. Best combination of C and γ is often selected by a grid-search with exponentially growing sequences of C and γ, for example, ; . Each pair of parameters is checked using cross validation and

the parameters with best cross validation accuracy are picked. The final model, which is used for testing and classifying new data, is then trained on the whole training set using the selected parameters.

An Application -

The Load Haul Dump (LHD) is a principal equipment in the Mining Industry. Failure of a single part of LHD can cause downtime over the entire process of mineral transportation. It is well known that failure of equipment is indicated by some indicators. Therefore predicting LHD reliability is necessary in order to execute predictive maintenance.

In the given LHD system, model for each piece of machine should be consistent with the period in which it resides in. The life of machine is dividing into three parts: infant mortality, useful life and wear out period. Reliability is different in different part of the equipment life.

A Case Study –

Reliability prediction was done based on information on LHD in a mine.

The mine was situated in Central Coal Fields Limited. Mechanized bord and pillar mining was the method of operation.

The Details of specific district are –

Size of panel: 150 x 250 m

Size of Pillar: 25 x 25 m
Average Height of Gallery: 2.4m

Average Width of Gallery: 4.2m
No. of heading in the panel: 6

Average gradient of the seam: 1 in 20
Gassiness: Degree 1

The General conditions of the working environment of the LHD were as below –

Ambient Temperature: 35°-40° C

Humidity: 95-98%
Cross-Gradient: 1 in 25

Gradient: 1 in 20
Steering angle: 90°

Side Clearance: 1-1.5 m

Bottom Clearance: 0.3-0.4 m
Make of Water: Heavy

Airborne dust: Less as the seam is watery

Concepts used –

Based on the above SVM models we came to a system of equations to predict reliability of different parts of the LHD.

The system failure rate λ (t) = β/Θ (t/Θ) ^ (β-1) Where β = Shape Parameter

Θ = Scale Parameter t = time taken
Also we have data on the Mean Time between Failure

MTTF (η) =1/λ

Reliability R (t) = e ^ (-(t/Θ) ^β)
Also, R (t) = e ^ (-λt)

The following table shows the values taken and the reliability of the machine over a future period of 250 working hours –

Type of | β | Θ | η | λ (t) | R(t) | Failure | | | | | | Motor | 0.990 | 402.77 | 402.77 | 0.0025 | . | Hydraulic | 1.509 | 367.86 | 330.00 | 0.003 | 0.47 | System | | | | | | Brake | 1.034 | 303.92 | 299.80 | 0.0033 | . | Transmission | 1.034 | 256.44 | 252.90 | 0.0038 | . | Tyres | 1.005 | 298.37 | 297.70 | 0.00335 | . | Bucket | 1.017 | 172.37 | 171.15 | 0.0058 | . | Mechanism | | | | | | Trailing | 0999 | 329.66 | 329.66 | 0.003 | . | Cable | | | | | | Radiator | 1.059 | 296.64 | 289.90 | 0.0045 | . |

We have got the Reliability values of the of different parts of the LHD and forecasting taking the value of t as 250. The Probability of Failure is 1 - R (t)

Results Interpretation -

The following Graph shows the probability of failure of individual parts of the LHD over the next 250 working hours:-

| | Probability of Failure | | 0.8 | | | | | | | | | | 0.7 | | | | | | | | | | 0.6 | | | | | | | | | | 0.5 | | | | | | | | | | 0.4 | | | | | | | | | | 0.3 | | | | | | | | Probability of Failure | | 0.2 | | | | | | | | | | 0.1 | | | | | | | | | | 0 | | | | | | | | | |

Hence, we can infer from the graph that the bucket actuating mechanism is most probable to failure and is therefore least reliable in the LHD. Similarly, the Motor is least probable to fail and is most reliable.

Conclusion –

The method of reliability prediction based on SVM has been presented in this project. SVM models have non-linear mapping capabilities, and so can more easily capture reliability data patterns than can others models.

Effective prediction of equipment failure data helps machine maintenance staff and equipment designers to arrange the repair schedule in order to increase throughputs and productivity. An important outcome of this work is reliability prediction of LHD machine parts in its life period. This can help us understand the various parameters and subsequently improve upon the design of future machines to better our productivities across various fields in operations management.

References –

1. Feng Ding , Zhengjia He, Yanyang Zi , Xuefeng Chen , Jiyong Tan, Hongrui

Cao, Huaxin Chen, ―Application of Support Vector Machine for Equipment Reliability Forecasting‖, The IEEE International Conference, Industrial

Informatics, DCC, Daejeon, Korea, July 13- 16, 2008

2. Ping-Feng Pai , Wei-Chiang Hong, and Yu-Shen Lee, ―Determining Parameters of Support Vector Machines by Genetic Algorithms—Applications to Reliability Prediction‖, International Journal of Operations Research Vol. 2, No. 1, 1-7 (2005)

3. L.P. Wang, Support Vector Machines: Theory and Application. Springer, Berlin, 2005.

4. Hansen G, Sparrow E, Kokate JY, Leland KJ, Iaizzo PA. Wound status evaluation using color image processing. IEEE Transactions on Medical Imaging. 1997;16:78–86. doi: 10.1109/42.552057.

5. http://www.ncbi.nlm.nih.gov/pmc/articles/PMC394338/

PDF to Word

Similar Documents

Free Essay

Offiline Arabic Handwritten Character Recognizer Based on Feature Extraction and Support Vector Machine

...Offline handwritten Arabic character recognizer based on Feature extraction and Support vector machine Thahira banu, Assistant professor in MCA department Sankara College of Science and comerce, Coimbatore-35. thahirshanth@gmail.com. ABSTRACT: Since the problem of Arabic text recognition is a large and complex one, it makes sense to try a simple method to see what performance can be achieved. The characters are written by many people using a great variety of sizes, writing styles, instruments, and with a widely varying amount of care. Some of the characters or words are poorly formed and are hard to classify, even for a human. Of the 280 sample characters used for training, 280 have been used for test purposes. The captured image of a character is normalized and set to eight feature values as parameter values of a vector. Training has given for a character by SVM (Support Vector machine) algorithm. It attempts to work with a subset of the features in a character that a human would typically see for the identification of Arabic characters. 1. Introduction One of the most classical applications of the Artificial Neural Network is the Character Recognition System. Cost effective and less time consuming, businesses, post offices, banks, security systems, and even the field of robotics employ this system as the base of their operations. Handwriting recognition can be defined as the task of transforming text represented in the spatial form...

Words: 1773 - Pages: 8

Free Essay

Support Vecor Machine

...Nonparallel Support Vector Machines for Pattern Classification Lokesh Sharma sharma.123lokesh@gmail.com Anand Mishra anand.lnmiit@gmail.com Vaibhav Kumar Soni vbhvsoni22@gmail.com Sudhanshu Bansal sudhanshu.bansal@lnmiit.ac.in Prasant Rathore prasant.rathore@lnmiit.ac.in The LNM Institute of Information Technology, Jaipur (INDIA) Abstract—We introduce a nonparallel classifier knows as nonparllel support vector machine(NPSVM) for the purpose of binary classification. Proposed NPSVM is totally different from the existing non parallel classifier, such as the generalized eigenvalue proximal support vector machine (GEPSVM) and the twin support vector machine (TWSVM). NPSVM has several incomparable advantages:1) Two primal problems are constructed implementing the structural risk minimization principle; 2) The dual problems of these two primal problems have the same advantages as that of the standard SVMs, so that the kernel trick can be applied directly; 3)The dual problems have the same elegant formulation with that of standard SVMs and can certainly be solved efficiently by sequential minimization optimization algorithm, while existing GEPSVM or TWSVMs are not suitable for large scale problems; 4) It has the inherent sparseness as standard SVMs; 5) Existing TWSVMs are only the special cases of the NPSVM when the parameters of which are appropriately chosen. Experimental results on lots of datasets show the effectiveness of our method in both sparseness and classification...

Words: 1328 - Pages: 6

Free Essay

Natural L Anguage P Rocessing (a Lmost ) from S Cratch

...Journal of Machine Learning Research 12 (2011) 2493-2537 Submitted 1/10; Revised 11/10; Published 8/11 Natural Language Processing (Almost) from Scratch Ronan Collobert∗ Jason Weston† L´ on Bottou‡ e Michael Karlen Koray Kavukcuoglu§ Pavel Kuksa¶ RONAN @ COLLOBERT. COM JWESTON @ GOOGLE . COM LEON @ BOTTOU . ORG MICHAEL . KARLEN @ GMAIL . COM KORAY @ CS . NYU . EDU PKUKSA @ CS . RUTGERS . EDU NEC Laboratories America 4 Independence Way Princeton, NJ 08540 Editor: Michael Collins Abstract We propose a unified neural network architecture and learning algorithm that can be applied to various natural language processing tasks including part-of-speech tagging, chunking, named entity recognition, and semantic role labeling. This versatility is achieved by trying to avoid task-specific engineering and therefore disregarding a lot of prior knowledge. Instead of exploiting man-made input features carefully optimized for each task, our system learns internal representations on the basis of vast amounts of mostly unlabeled training data. This work is then used as a basis for building a freely available tagging system with good performance and minimal computational requirements. Keywords: natural language processing, neural networks 1. Introduction Will a computer program ever be able to convert a piece of English text into a programmer friendly data structure that describes the meaning of the natural language text? Unfortunately, no consensus has...

Words: 4649 - Pages: 19

Free Essay

Bio and Electrocardiogram

...Shivanesan S M. 1, Pradheep M. 1, Sharath K. 1, Aravind Prasad. 1, Manoj M. 1 Ganesan M. 2 Abstract- Electrocardiogram is the recording of the electrical potential of heart versus time. The analysis of ECG signal has great importance in the detection of cardiac abnormalities. In this paper we have dealt about the removal of noises in ECG signals and arrhythmia classification of the signal. The inputs for our analysis is taken from MIT-BIH database (Massachusetts Institute of Technology Beth Israel Hospital database). The denoising is done through wavelet transform and thresholding. Confirmatory tools such as Poincare plot and Detrended Fluctuation Analysis (DFA) are used to find out the healthiness of the signal. Then Support Vector Machine (SVM) is used to find out what type of arrhythmia is present in the signal. Keywords- Classification, DFA Electrocardiogram, MIT-BIH database, Poincare, SVM , Wavelets. I. INTRODUCTION In today’s environment there has been lot of threats due to heart disease and no proper diagnosis With the recent developments in technology, physicians have powerful tools to observe the working of the heart muscle and thus to establish their diagnosis. Among cardiovascular examinations, electrocardiogram (ECG) analysis is the most commonly used and very effective too. This is due to the fact that ECG presents useful information about the rhythm and the electrical activity of the heart. Thus, it is used for the diagnosis of cardiac ...

Words: 3286 - Pages: 14

Free Essay

Writer Adaptation for Handwriting Recognition in Hindi Language – a Survey

...International Journal of Science and Research (IJSR) ISSN (Online): 2319-7064 Writer Adaptation for Handwriting Recognition in Hindi Language – A Survey Nupur Gulalkari1, Shreya Prabhu2, Anjali Pachpute3, Rekha Sugandhi4, Kapil Mehrotra5 1,2 , 3, 4 5 Computer Department, MIT College of Engineering, Pune, India GIST, Center for Development of Advanced Computing, Pune, India Abstract: With the advancement in technology, there is an increased use of pen-based touch screen devices and PDAs. These devices come with an alternative for the traditional alphanumeric or QWERTY keyboard which is input in the form of user’s handwriting. The handwriting is then converted into normal text form. However, these devices require prior training to be done by the user. There is a high demand for robust and accurate recognition systems in the practical applications of handwriting recognition. The real challenge lies with the selection of a classifier which gives accurate results in real-time, while making the system self-adaptive simultaneously. Thus, in this paper various classifiers have been studied so as to find the most appropriate classifier for anonline handwriting recognition system for handwriting in Hindi language that provides a way by which the touch screen device adapts itself to its user handwriting without prior training is studied. Keywords: Active-DTW, Markov Model, Self -adaptation, SVM, Writer adaptation 1. Introduction Hindi is the fourth most spoken language...

Words: 2640 - Pages: 11

Free Essay

Svmdd

... Consider ridge regression We want to learn = =1 Obtain w as = argmin 11 . . 1 ⋮ ⋮ = ⋮ ⋮ 1 = , = 1 ⋮ ⋮ 2 ( −( ) )2 + =1 1 ⋮ ⋮ =1 (for r-th training example) = argmin − 2 + 2 Notation: X is a matrix, x is a vector Solve by setting derivatives to zero, to get = ( + )−1 (Px1) (PxN)(NxP) (PxP) For a new example (PxN) (Nx1) = = ( + )−1 Getting to Dual form = ( + )−1  + = 1 where =  = 1 − = 1 − = − gives the dual solution, from which we can obtain w = or = =1 (here, xi is the i-th example) 11 . . 1 ⋮ ⋮ = ⋮ ⋮ 1 1 ⋮ ⋮ 1 ⋮ ⋮ Substituting w = in = we get = −  + =  = + − 1 − , We can compute as: = ( + )−1 where K = i.e. = , 11 . . ⋮ ⋮ ⋮ 1 1 ⋮ ⋮ ⋮ 11 ⋮ ⋮ ..... 1 1 ⋮ ⋮ =(xi.xj) (dot product) K: matrix of inner products of N vectors (Gram Matrix) K: matrix of similarities of example pairs (since dot product gives similarity between vectors) (1 , 1 ) . . . . . ⋮ K= ⋮ ( , 1 ) (1 , ) ⋮ ⋮ ( , ) Now, = = = , = =1 =1 (since w = ) , So in the dual form: Compute = ( + )−1 where K = , i.e. = , Evaluate on a new data point xnew as y = f = =1 , ...

Words: 2145 - Pages: 9

Free Essay

Fsr Technology

...HUMAN GAIT ANALYSIS Abstract: Sensible Shoes is a hands-free and eyes-free foot-computer interface that supports on-the-go interaction with surrounding environments. We recognize different low-level activities by measuring the user’s continuous weight distribution over the feet with twelve Force Sensing Resistor (FSR) sensors embedded in the insoles of shoes. Using the sensor data as inputs, a Support Vector Machine (SVM) classifier identifies up to eighteen mobile activities and a four-directional foot control gesture at approximately 98% accuracy. By understanding user’s present activities and foot gestures, this system offers a nonintrusive and always-available input method. We present the design and implementation of our system and several proof-of-concept applications. Overview: A person’s weight is not allocated symmetrically over the plantar. As the sole is not flat but arched, the weight mainly centers on the hallex, the first metatarse and the calcaneus. When sitting, the weight of a person’s upper body rest mostly on the chair and the weight on the feet is relatively small. When standing, the whole body’s weight is put evenly on both feet. Leaning left or right changes the weight distribution over the feet. When walking, the weight distribution changes with the pace; the weight on the front and rear part of the foot alternately increases and decreases because not all parts of the sole contact the ground at once. The changes in weight distribution on the feet...

Words: 652 - Pages: 3

Free Essay

Ewfdwefwefwf

...Fourth Week- July 2013 Sept 2013 First week- Aug 2013 Fourth Week- Aug 2013 Sept 2013 TAE - VII Guest Lecture/Industrial Visit  Early computing was entirely mechanical:      abacus (about 500 BC) mechanical adder/subtracter (Pascal, 1642) difference engine design (Babbage, 1827) binary mechanical computer (Zuse, 1941) electromechanical decimal machine (Aiken, 1944)  Mechanical and electromechanical machines have limited speed and reliability because of the many moving parts. Modern machines use electronics for most information transmission.  Computing is normally thought of as being divided into generations.  Each successive generation is marked by sharp changes in hardware and software technologies.  With some exceptions, most of the advances introduced in one generation are carried through to later generations.  We are currently in the fifth generation.  Technology        and Architecture Vacuum tubes and relay memories CPU driven by a program counter (PC) and accumulator Machines had only fixed-point arithmetic  Software and Applications Machine and assembly language Single user at a time No subroutine linkage mechanisms Programmed I/O required continuous use of CPU  Representative IAS, IBM 701 systems: ENIAC, Princeton  Technology       and Architecture Discrete transistors and core memories I/O processors, multiplexed memory access Floating-point arithmetic available Register Transfer...

Words: 2199 - Pages: 9

Free Essay

Assignment

...Active Learning with Support Vector Machines Kim Steenstrup Pedersen Department of Computer Science University of Copenhagen 2200 Copenhagen, Denmark kimstp@di.ku.dk Jan Kremer Department of Computer Science University of Copenhagen 2200 Copenhagen, Denmark jan.kremer@di.ku.dk Christian Igel Department of Computer Science University of Copenhagen 2200 Copenhagen, Denmark igel@di.ku.dk Abstract In machine learning, active learning refers to algorithms that autonomously select the data points from which they will learn. There are many data mining applications in which large amounts of unlabeled data are readily available, but labels (e.g., human annotations or results from complex experiments) are costly to obtain. In such scenarios, an active learning algorithm aims at identifying data points that, if labeled and used for training, would most improve the learned model. Labels are then obtained only for the most promising data points. This speeds up learning and reduces labeling costs. Support vector machine (SVM) classifiers are particularly well-suited for active learning due to their convenient mathematical properties. They perform linear classification, typically in a kernel-induced feature space, which makes measuring the distance of a data point from the decision boundary straightforward. Furthermore, heuristics can efficiently estimate how strongly learning from a data point influences the current model. This information can be used to actively...

Words: 9180 - Pages: 37

Premium Essay

Predicting Student Academic Performance in an Engineering Dynamics Course: a Comparison of Four Types of Predictive Mathematical Models

...include the multiple linear regression model, the multilayer perception network model, the radial basis function network model, and the support vector machine model. The inputs (i.e., predictor variables) of the models include student's cumulative GPA, grades earned in four pre-requisite courses (statics, calculus I, calculus II, and physics), and scores on three dynamics mid-term exams (i.e., the exams given to students during the semester and before the final exam). The output of the models is students' scores on the dynamics final comprehensive exam. A total of 2907 data points were collected from 323 undergraduates in four semesters. Based on the four types of mathematical models and six different combinations of predictor variables, a total of 24 predictive mathematical models were developed from the present study. The analysis reveals that the type of mathematical model has only a slight effect on the average prediction accuracy (APA, which indicates on average how well a model predicts the final exam scores of all students in the dynamics course) and on the percentage of accurate predictions (PAP, which is calculated as the number of accurate predictions divided by the total number of predictions). The combination of predictor variables has only a slight effect on the APA, but a profound effect on the PAP. In general, the support vector machine models have the highest PAP as compared to the other three types of mathematical models. The research...

Words: 432 - Pages: 2

Free Essay

Rotordynamics

...the seals. The shaft or rotor is the rotating component of the system. Many industrial applications have flexible rotors, where the shaft is designed in a relatively long and thin geometry to maximize the space available for components such as impellers and seals. Additionally, machines are operated at high rotor speeds in order to maximize the power output. The first recorded supercritical machine (operating above first critical speed or resonance mode) was a steam turbine manufactured by Gustav Delaval in 1883. Modern high performance machines normally operates above the first critical speed, generally considered to be the most important mode in the system, although they still avoid continuous operating at or near the critical speeds. Maintaining a critical speed margin of 15 % between the operating speed and the nearest critical speed is a common practice in industrial applications. The other two of the main components of rotor-dynamic systems are the bearings and the seals. The bearings support the rotating components of the system and provide the additional damping needed to stabilize the system and contain the rotor vibration. Seals, on the other hand, prevent undesired leakage flows inside the machines of the processing or lubricating fluids, however they have rotor-dynamic properties that can cause large rotor vibrations when interacting with the rotor. Generally, the vibration in rotor-dynamic systems can be categorized into synchronous or subsynchronous vibrations depending...

Words: 12839 - Pages: 52

Premium Essay

Data Mining In Computer Science

...appropriate decisions at appropriate time, to increase the profit of business. Data mining is highly related with another important area of research in Computer Science, namely, Machine Learning. Machine Learning is the field of research where machine learns from the past data and takes informed and efficient decisions for future. In number of applications, for example, optical character recognition, one needs to build the past data in the form of training patterns. These training patterns are usually taken in such an efficient way that machine can take an appropriate decision in a situation when a previously unknown pattern presents itself. The training patterns are generally taken in the form of features extracted from data. In case of data mining, creation of these patterns is not generally required as we already have the data from where knowledge is to be discovered. We however, have to be able to extract efficient features from this data, so a decision can be...

Words: 2594 - Pages: 11

Free Essay

Comp Sci as

...What is the format of an IP address? 1. 4 bytes 2. First 2 – network address 3. Second 2 – host address How is an IP address associated with a device on a network? 1. IP address specifically or automatically assigned 2. Each device has own private IP address What is the difference between a public IP address and a private IP address? 1. Public – dynamic (changes each time device connects to internet) or static (doesn’t change because used for hosting web pages or services) 2. Private – assigned on LANs (automatically or chosen by LAN administrator) and are static; able to change, but rarely. What are a URL, IP address, and a DNS? Why are they important? 1. URL – web address typed into a browser 2. IP address – series of numbers that tells computer where to find information 3. DNS (Domain Name System) – collection of domain names; translates a URL into an IP address 4. Every URL has an IP address; IP addresses were too complicated and were shortened by URLs What happens when a user types in the IP address of a website rather than its URL? 1. Web page at the IP address shows up on browser Explain how a URL is used to locate a resource on the WWW and the role of the Domain Name Service 1. A URL (Uniform Resource Locator) is the web address a user types into a browser to reach a website 2. DNS translates URL to an IP address to take user to desired site What is streaming? Explain real-time and on-demand? 1. Streaming...

Words: 1764 - Pages: 8

Free Essay

Control Architecture and Algorithms of Biped Robot

...Control Architecture and Algorithms of the Anthropomorphic Biped Robot Bip2000 Christine Azevedo and the BIP team INRIA - 655 Avenue de l’Europe 38330 Montbonnot, France ABSTRACT INRIA [1] and LMS [2] have designed and realized an anthropomorphic legged robot, BIP2000 (fig.1). A planar version achieves walking, and the whole robot is able to keep its balance on one foot while moving. The purpose of this paper is to present the principles and the architecture of the robot control we have used. After having presented the robotic system, and the software architecture, we will detail the principles of the robot control. We will finally present implementation issues and experimental results. Keywords: Robot Control, Biped Robots, Walking Machines. 1. DESCRIPTION OF THE SYSTEM The design of the robot was inspired from the human anthropometric data and his dynamic capabilities. We recall here only the main characteristics of BIP2000, the reader being referred to [5] and [9] for more details. Fig1. The Biped Robot BIP2000 Fig2. BIP without Pelvis 1.1 Mechanical Structure of BIP2000 Designed by the Laboratoire de Mécanique des Solides of Poitiers [2], the robot has 15 active joints (fig.3). It is able to walk forward thanks to the rotation of the ankles, knees and hips allowing the flexion/extension of the biped in the sagittal plane (fig.4). The ability of changing direction is given by the trunk, the pelvis and the 2 hips internal/external rotations. For the lateral equilibrium...

Words: 2364 - Pages: 10

Free Essay

Machine Learning

...Machine learning According to Alp Aydin (2010), Machine learning is an area of artificial intelligence that developed from design acknowledgment and computational learning hypothesis. It investigates the study and development of calculations that can gain from and make expectations on information. Such calculations work by building a model from sample inputs keeping in mind the end goal to settle on information driven forecasts or choices, as opposed to taking after entirely static project guidelines. Machine learning is firmly identified with computational statistics; a specialty that goes for the configuration of calculation for executing factual techniques on computers. It has solid ties to numerical enhancement, which conveys techniques, hypothesis and application areas to the field. Machine learning is utilized in a scope of registering errands where outlining and programming unequivocal calculations is in feasible (Marshland, 2009). Concepts of machine learning 1. Bayesian networks A Bayesian network is a probabilistic graphical model that speaks to an arrangement of irregular variables and their restrictive independencies through a coordinated non-cyclic chart. For instance, a Bayesian system could speak to the probabilistic connections between forex market and political unrests. Given the instances of political unrests, the system can be utilized to figure the probabilities of the forex market dropping. Proficient calculations exist that perform surmising and learning...

Words: 987 - Pages: 4