Data Mining by Mehmed Kantardzic (good book recommendations TXT) π
Read free book Β«Data Mining by Mehmed Kantardzic (good book recommendations TXT) πΒ» - read online or download for free at americanlibrarybooks.com
- Author: Mehmed Kantardzic
Read book online Β«Data Mining by Mehmed Kantardzic (good book recommendations TXT) πΒ». Author - Mehmed Kantardzic
The main advantages of SOM technology are as follows: presented results are very easy to understand and interpret; technology is very simple for implementation; and most important, it works well in many practical problems. Of course, there are also some disadvantages. SOMs are computationally expensive; they are very sensitive to measure of similarity; and finally, they are not applicable for real-world data sets with missing values. There are several possible improvements in implementations of SOMs. To reduce the number of iterations in a learning process, good initialization of weight factors is essential. Principal components of input data can make computation of the SOM orders of magnitude faster. Also, practical experience shows that hexagonal grids give output results with a better quality. Finally, selection of distance measure is important as in any clustering algorithm. Euclidean distance is almost standard, but that does not mean that it is always the best. For an improved quality (isotropy) of the display, it is advisable to select the grid of the SOM units as hexagonal.
The SOMs have been used in large spectrum of applications such as automatic speech recognition, clinical data analysis, monitoring of the condition of industrial plants and processes, classification from satellite images, analysis of genetic information, analysis of electrical signals from the brain, and retrieval from large document collections. Illustrative examples are given in Figure 7.18.
Figure 7.18. SOM applications. (a) Drugs binding to human cytochrome; (b) interest rate classification; (c) analysis of book-buying behavior.
7.8 REVIEW QUESTIONS AND PROBLEMS
1. Explain the fundamental differences between the design of an ANN and βclassicalβ information-processing systems.
2. Why is fault-tolerance property one of the most important characteristics and capabilities of ANNs?
3. What are the basic components of the neuronβs model?
4. Why are continuous functions such as log-sigmoid or hyperbolic tangent considered common activation functions in real-world applications of ANNs?
5. Discuss the differences between feedforward and recurrent neural networks.
6. Given a two-input neuron with the following parameters: bias b = 1.2, weight factors W = [w1, w2] = [3, 2], and input vector X = [β5, 6]T; calculate the neuronβs output for the following activation functions:
(a) a symmetrical hard limit
(b) a log-sigmoid
(c) a hyperbolic tangent
7. Consider a two-input neuron with the following weight factors W and input vector X:
We would like to have an output of 0.5.
(a) Is there a transfer function from Table 9.1 that will do the job if the bias is zero?
(b) Is there a bias that will do the job if the linear-transfer function is used?
(c) What is the bias that will do the job with a log-sigmoidβactivation function?
8. Consider a classification problem defined with the set of 3-D samples X, where two dimensions are inputs and the third one is the output.
(a) Draw a graph of the data points X labeled according to their classes. Is the problem of classification solvable with a single-neuron perceptron? Explain the answer.
(b) Draw a diagram of the perceptron you would use to solve the problem. Define the initial values for all network parameters.
(c) Apply single iteration of the delta-learning algorithm. What is the final vector of weight factors?
9. The one-neuron network is trained to classify inputβoutput samples:10111β1011
Show that this problem cannot be solved unless the network uses a bias.
10. Consider the classification problem based on the set of samples X:
(a) Draw a graph of the data points labeled according to their classification. Is the problem solvable with one artificial neuron? If yes, graph the decision boundaries.
(b) Design a single-neuron perceptron to solve this problem. Determine the final weight factors as a weight vector orthogonal to the decision boundary.
(c) Test your solution with all four samples.
(d) Using your network classify the following samples: (β2, 0), (1, 1), (0, 1), and (β1, β2).
(e) Which of the samples in (d) will always be classified the same way, and for which samples classification may vary depending on the solution?
11. Implement the program that performs the computation (and learning) of a single-layer perceptron.
12. For the given competitive network:
(a) find the output vector [Y1, Y2, Y3] if the input sample is [X1, X2, X3] = [1, β1, β1];
(b) what are the new weight factors in the network?
13. Search the Web to find the basic characteristics of publicly available or commercial software tools that are based on ANNs. Document the results of your search. Which of them are for learning with a teacher, and which are support learning without a teacher?
14. For a neural network, which one of these structural assumptions is the one that most affects the trade-off between underfitting (i.e., a high-bias model) and overfitting (i.e., a high-variance model):
(a) the number of hidden nodes,
(b) the learning rate,
(c) the initial choice of weights, or
(d) the use of a constant-term unit input.
15. Is it true that the Vapnik-Chervonenkis (VC) dimension of a perceptron is smaller than the VC dimension of a simple linear support vector machine (SVM)? Discuss your answer.
7.9 REFERENCES FOR FURTHER STUDY
Engel, A., C. Van den Broeck, Statistical Mechanics of Learning, Cambridge University Press, Cambridge, UK, 2001.
The subject of this book is the contribution made to machine learning over the last decade by researchers applying the techniques of statistical mechanics. The authors provide a coherent account of various important concepts and techniques that are currently only found scattered in papers. They include many examples and exercises to make a book that can be used with courses, or for self-teaching, or as a handy reference.
Haykin, S., Neural Networks and Learning Machines, 3rd edition, Prentice Hall, Upper Saddle River, NJ, 2009.
Fluid and authoritative, this well-organized book represents the first comprehensive treatment of neural networks from an engineering perspective, providing extensive, state-of-the-art coverage that will expose readers to the myriad facets of neural networks and help them appreciate the technologyβs origin, capabilities, and potential applications. The book examines all the important aspects of this emerging technology, covering the learning process, backpropagation,
Comments (0)