Database System Concepts
7th Edition
ISBN: 9780078022159
Author: Abraham Silberschatz Professor, Henry F. Korth, S. Sudarshan
Publisher: McGraw-Hill Education
expand_more
expand_more
format_list_bulleted
Question
3. Consider a training set that contains 100 positive examples and 400 negative examples. For
each of the following candidate rules,
R1: A −→ + (covers 4 positive and 1 negative examples),
R2: B −→ + (covers 30 positive and 10 negative examples),
R3: C −→ + (covers 100 positive and 90 negative examples),
determine which is the best and worst candidate rule according to:
(a) The likelihood ratio statistic.
Expert Solution
This question has been solved!
Explore an expertly crafted, step-by-step solution for a thorough understanding of key concepts.
This is a popular solution
Trending nowThis is a popular solution!
Step by stepSolved in 2 steps
Knowledge Booster
Learn more about
Need a deep-dive on the concept behind this application? Look no further. Learn more about this topic, computer-science and related others by exploring similar questions and additional content below.Similar questions
- Suppose we are trying to model a classification problem with two labels: 'sick' and 'healthy.' For the purpose of this test, we consider a positive result to be testing sick and a negative to test as healthy. After fitting the model with LDA in R, we compare predicted values from the actual values, as shown below: > lda. fit = 1da (test.result x1 + x2, data-mydata) > lda. pred > lda.class=lda.pred$class > table (lda.class, mydata$test.result) = predict (1da.fit) mydata lda.pred sick healthy sick 40 healthy 25 32 121arrow_forwardThe accuracy in the output depends on the correct inputs. The model designed is supposedto be evaluated to determine the accuracy and it is supposed to function in a supervisedway. Hence the data sets can affect the model and vice versa.Given P(D|M) ,P(M) ,and P(D'|M), solve the followinga)Draw the probability tree for the main situationB)Draw the reverse tree for the main situation.C)Using the main tree, prove all probabilities on the opposite tree.D) If P(DIM) = 0.3,P(M) = 0.1 ,and P(D'IM) = 0.65 find P(D)arrow_forwardQuestion 48. Let us return to the Titanic data set. We now have learned several models and want to choose the best one. We used three different methods to validate these models: The training error rate (apparent error rate), the error rate on an external test set and the error rate estimated by a 10-fold cross validation. Training Error | Error on the test set | Cross Validation Error 0.18 Learner Decision Tree 0.22 0.21 Random Forest 0.01 0.10 0.12 1-Nearest-Neighbour 0.18 0.19 Which of the following statements are correct? a) 1-Nearest-Neighbour has a perfect training error and hence it should be used here. b) Random Forests outperforms both 1-Nearest-Neighbour and the Decision Tree in terms of prediction error. c) Not just in this case, but in general, Cross Validation is the better validation strategy and should always be preferred over the error on a single test set. d) Not just in this case, but in general, Decision Trees always perform worse than Random Forests.arrow_forward
- Nonearrow_forwardGiven a two-category classification problem under the univariate case, where there are two training sets (one for each category) as follows: D₁ = (-3,-1,0,4} D₂ = {-2,1,2,3,6,8} Given the test example x = 5, please answer the following questions: have and a) Assume that the likelihood function of each category has certain paramétric form. Specifically, we p(x | w₁) N, 07) p(x₂)~ N(μ₂, 02). Which category should we decide on when maximum-likelihood estimation is employed to make the prediction?arrow_forwardM3arrow_forward
- 2. Can you design a binary classification experiment with 100 total population (TP+TN+FP+ FN), with precision (TP/(TP+FP)) of 1/2, with sensitivity (TP/(TP+FN)) of 2/3, and specificity (TN/(FP+TN)) of 3/5? (Please consider the population to consist of 100 individuals.)arrow_forwardGenerate test cases for the triangle problem by using the cause-effect graph and decision table testingarrow_forwardQuestion 1: Suppose that a Naive Bayesian classifier is created using a given training data set. Then, predictions are made on several testing points using this classifier. The predicted positive class probabilities and the actual class labels for these testing points are shown as follows. Please draw the ROC curve for the above prediction. Please clearly denote your splitting values and show the TP, FP, TN, FN, TPR, FPR values at each splitting value. Sample ID Actual Class Label Predicted Positive Class Probability 1 - 0.1 2 + 0.3 3 - 0.4 4 + 0.4 5 + 0.6 6 + 0.8arrow_forward
arrow_back_ios
arrow_forward_ios
Recommended textbooks for you
- Database System ConceptsComputer ScienceISBN:9780078022159Author:Abraham Silberschatz Professor, Henry F. Korth, S. SudarshanPublisher:McGraw-Hill EducationStarting Out with Python (4th Edition)Computer ScienceISBN:9780134444321Author:Tony GaddisPublisher:PEARSONDigital Fundamentals (11th Edition)Computer ScienceISBN:9780132737968Author:Thomas L. FloydPublisher:PEARSON
- C How to Program (8th Edition)Computer ScienceISBN:9780133976892Author:Paul J. Deitel, Harvey DeitelPublisher:PEARSONDatabase Systems: Design, Implementation, & Manag...Computer ScienceISBN:9781337627900Author:Carlos Coronel, Steven MorrisPublisher:Cengage LearningProgrammable Logic ControllersComputer ScienceISBN:9780073373843Author:Frank D. PetruzellaPublisher:McGraw-Hill Education
Database System Concepts
Computer Science
ISBN:9780078022159
Author:Abraham Silberschatz Professor, Henry F. Korth, S. Sudarshan
Publisher:McGraw-Hill Education
Starting Out with Python (4th Edition)
Computer Science
ISBN:9780134444321
Author:Tony Gaddis
Publisher:PEARSON
Digital Fundamentals (11th Edition)
Computer Science
ISBN:9780132737968
Author:Thomas L. Floyd
Publisher:PEARSON
C How to Program (8th Edition)
Computer Science
ISBN:9780133976892
Author:Paul J. Deitel, Harvey Deitel
Publisher:PEARSON
Database Systems: Design, Implementation, & Manag...
Computer Science
ISBN:9781337627900
Author:Carlos Coronel, Steven Morris
Publisher:Cengage Learning
Programmable Logic Controllers
Computer Science
ISBN:9780073373843
Author:Frank D. Petruzella
Publisher:McGraw-Hill Education