Data Mining - ISZC415 - Quiz 2
BITS- WILP - MTEC
1. The entropy of a fair coin toss is:
Select one:
a. 0.25
b. 0.5
c. 1
d. 0
Ans: c. 1
2. A quiz question had names of 10 algorithms of which the student
had to select only the classification algorithms. A student
identified 7 of them as classification algorithms. During evaluation
it was found that 5 of the algorithms identified by the student were
indeed classification algorithms. The student was unable to identify
2 other classification algorithms in the list.
The F-score is:
Select one:
a. 0.59
b. 0.41
c. 0.69
d. 0.71
Ans: d. 0.71
3. Hash tree is created from:
Select one:
a. transactions
b. frequent itemsets
c. strong rules
d. candidate itemsets
Ans: d. candidate itemsets
4. The following data is about a poll that occurred in 3 states. In
state1, 50% of voters support Party1, in state2, 60% of the voters
support Party1, and in state3, 35% of the voters support Party1. Of
the total population of the three states, 40% live in state1, 25%
live in state2, and 35% live in state3. Given that a voter supports
Party1, what is the probability that he lives in state2?
Select one:
a. 0.52
b. 0.42
c. 0.32
d. 0.22
Ans: c. 0.32
5. A quiz question had names of 6 algorithms of which the student had
to select only the classification algorithms. A student identified 3
of them as classification algorithms. During evaluation it was found
that 2 of the algorithms identified by the student were indeed
classification algorithms. The student was unable to identify 2 other
classification algorithms in the list.
The recall is:
Select one:
a. 0.33
b. 0.5
c. 0.73
d. 0.6
Ans: b. 0.5
6. The table below shows marks in math (x) and marks in statistics
(y).
What is the value of the slope (m) of simple regression line ?
Select one:
a. 0.744
b. 0.644
c. 0.444
d. 0.544
Ans: b. 0.644
7. In case an item occurs N times in a single transaction, the
support count of that item:
Select one:
a. zero times
b. is counted only once
c. is counted N times
d. is counted threshold times
Ans: b. is counted only once
8. In association analysis, confidence measures certainty of the
rule.
Select one:
a. True
b. False
Ans: a. True
9. Decision tree pruning is done to prevent under-fitting the data.
Select one:
a. True
b. False
Ans: b. False
10. A decision tree is split on the attribute with highest Gini
Index.
Select one:
a. True
b. False
Ans: b. False
11. Gini index cannot be used to make a ternary split on an attribute
in decision tree classification.
Select one:
a. True
b. False
Ans: b. False
12. Decision tree splitting decision can be made based upon
information gain of the attributes, but not based upon entropy of the
attributes.
Select one:
a. True
b. False
Ans: b. False
13. Laplace smoothing is applied in Naive Bayes spam classifier
because it prevents the conditional probability from becoming zero if
some words are not present in the sample.
Select one:
a. True
b. False
Ans: a. True
14. In association analysis, support is a symmetric measure of
associations.
Select one:
a. True
b. False
Ans: a. True
15. In classification, we evaluate the performance of a classifier on
training data
Select one:
a. True
b. False
Ans: b. False
No comments:
Post a Comment