
Data mining and warehousing mcq
1. What does Apriori algorithm do?
- It mines all frequent patterns through pruning rules with lesser support
- It mines all frequent patterns through pruning rules with higher support
- Both a and b
- None of these
It mines all frequent patterns through pruning rules with lesser support
2. What does FP growth algorithm do?
- It mines all frequent patterns through pruning rules with lesser support
- It mines all frequent patterns through pruning rules with higher support
- It mines all frequent patterns by constructing a FP tree
- All of these
It mines all frequent patterns by constructing a FP tree
3. What techniques can be used to improve the efficiency of apriori algorithm?
- hash based techniques
- transaction reduction
- Partitioning
- All of these
All of these
4. What do you mean by support(A)?
- Total number of transactions containing A
- Total Number of transactions not containing A
- Number of transactions containing A / Total number of transactions
- Number of transactions not containing A / Total number of transactions
Number of transactions containing A / Total number of transactions
5. Which of the following is direct application of frequent itemset mining?
- Social Network Analysis
- Market Basket Analysis
- outlier detection
- intrusion detection
Market Basket Analysis
6. What is not true about FP growth algorithms?
- It mines frequent itemsets without candidate generation
- There are chances that FP trees may not fit in the memory
- FP trees are very expensive to build
- It expands the original database to build FP trees
It expands the original database to build FP trees
7. When do you consider an association rule interesting?
- If it only satisfies min_support
- If it only satisfies min_confidence
- If it satisfies both min_support and min_confidence
- There are other measures to check so
If it satisfies both min_support and min_confidence
8. What is the difference between absolute and relative support?
- Absolute -Minimum support count threshold and Relative-Minimum support threshold
- Absolute-Minimum support threshold and Relative-Minimum support count threshold
- Both a and b
- None of these
Absolute -Minimum support count threshold and Relative-Minimum support threshold
9. What is the relation between candidate and frequent itemsets?
- A candidate itemset is always a frequent itemset
- A frequent itemset must be a candidate itemset
- No relation between the two
- None of these
A frequent itemset must be a candidate itemset
10. Which technique finds the frequent itemsets in just two database scans?
- Patitioning
- sampling
- hashing
- None of these
Patitioning
Data mining and warehousing mcq sppu
11. Which of the following is true?
- Both apriori and FP-Growth uses horizontal data format
- Both apriori and FP-Growth uses vertical data format
- Both a and b
- None of these
Both apriori and FP-Growth uses horizontal data format
12. What is the principle on which Apriori algorithm work?
- If a rule is infrequent, its specialized rules are also infrequent
- If a rule is infrequent, its generalized rules are also infrequent
- Both a and b
- None of these
If a rule is infrequent, its specialized rules are also infrequent
13. Which of these is not a frequent pattern mining algorithm
- Apriori
- FP growth
- Decision trees
- Eclat
Decision trees
14. Which algorithm requires fewer scans of data?
- Apriori
- FP growth
- Both a and b
- None of these
FP growth
15. What are Max_confidence, Cosine similarity, All_confidence?
- Frequent pattern mining algorithms
- Measures to improve efficiency of apriori
- Pattern evaluation measure
- None of these
Pattern evaluation measure
16. Linear regression – involves finding the________ line to fit two attributes (or variables)
- An itemset for which at least one proper super itemset has same support
- An item set whose no proper super- itemset has same support
- Both a and b
- None of these
An item set whose no proper super- itemset has same support
17. What are closed frequent itemsets?
- A closed itemset
- A frequent itemset
- An itemset which is both closed and frequent
- None of these
An itemset which is both closed and frequent
18. What are maximal frequent itemsets?
- A frequent item set whose no super-itemset is frequent
- A frequent itemset whose super-itemset is also frequent
- Both a and b
- None of these
A frequent item set whose no super-itemset is frequent
19. Why is correlation analysis important?
- To make apriori memory efficient
- To weed out uninteresting frequent itemsets
- To find large number of interesting itemsets
- To restrict the number of database iterations
To weed out uninteresting frequent itemsets
data mining and warehousing mcq sppu
20. What will happen if support is reduced?
- Number of frequent itemsets remains same
- Some itemsets will add to the current set of frequent itemsets
- Some itemsets will become infrequent while others will become frequent
- Can not say
Some itemsets will add to the current set of frequent itemsets
21. Can FP growth algorithm be used if FP tree cannot be fit in memory?
- Yes
- No
- Both a and b
- None of these
No
22. What is association rule mining?
- Same as frequent itemset mining
- Finding of strong association rules using frequent itemsets
- Both a and b
- None of these
Finding of strong association rules using frequent itemsets
23. What is frequent pattern growth?
- Same as frequent itemset mining
- Use of hashing to make discovery of frequent itemsets more efficient
- Mining of frequent itemsets without candidate generation
- None of these
Mining of frequent itemsets without candidate generation
24. When is sub-itemset pruning done?
- A frequent itemset ‘P’ is a proper subset of another frequent itemset ‘Q’
- Support (P) = Support(Q)
- When both a and b is true
- When a is true and b is not
When both a and b is true
25. Which of the following is not null invariant measure(that does not considers null transactions)?
- all_confidence
- max_confidence
- cosine measure
- lift
lift
26. The apriori algorithm works in a ..and ..fashion?
- top-down and depth-first
- top-down and breath-first
- bottom-up and depth-first
- bottom-up and breath-first
bottom-up and breath-first
27. Our use of association analysis will yield the same frequent itemsets and strong association rules whether a specific item occurs once or three times in an individual transaction
- TRUE
- FALSE
- Both a and b
- None of these
TRUE
28. In association rule mining the generation of the frequent itermsets is the computational intensive step.
- TRUE
- FALSE
- Both a and b
- None of these
TRUE
29. The number of iterations in apriori __
- increases with the size of the data
- decreases with the increase in size of the data
- increases with the size of the maximum frequent set
- decreases with increase in size of the maximum frequent set
increases with the size of the maximum frequent set
30. Which of the following are interestingness measures for association rules?
- recall
- lift
- accuracy
- compactness
lift
data mining and warehousing mcq questions
31. Frequent item sets is
- Superset of only closed frequent item sets
- Superset of only maximal frequent item sets
- Subset of maximal frequent item sets
- Superset of both closed frequent item sets and maximal frequent item sets
Superset of both closed frequent item sets and maximal frequent item sets
32. Assume that we have a dataset containing information about 200 individuals. A supervised data mining session has discovered the following rule: IF age < 30 & credit card insurance = yes THEN life insurance = yes Rule Accuracy: 70% and Rule Coverage: 63% How many individuals in the class life insurance= no have credit card insurance and are less than 30 years old?
- 63
- 30
- 38
- 70
38
33. In Apriori algorithm, if 1 item-sets are 100, then the number of candidate 2 item-sets are
- 100
- 4950
- 200
- 5000
4950
34. Significant Bottleneck in the Apriori algorithm is
- Finding frequent itemsets
- pruning
- Candidate generation
- Number of iterations
Candidate generation
35. Which Association Rule would you prefer
- High support and medium confidence
- High support and low confidence
- Low support and high confidence
- Low support and low confidence
Low support and high confidence
36. The apriori property means
- If a set cannot pass a test, its supersets will also fail the same test
- To decrease the efficiency, do level-wise generation of frequent item sets
- To improve the efficiency, do level-wise generation of frequent item sets
- If a set can pass a test, its supersets will fail the same test
If a set cannot pass a test, its supersets will also fail the same test
37. If an item set ‘XYZ’ is a frequent item set, then all subsets of that frequent item set are
- undefined
- not frequent
- frequent
- cant say
frequent
38. To determine association rules from frequent item sets
- Only minimum confidence needed
- Neither support not confidence needed
- Both minimum support and confidence are needed
- Minimum support is needed
Both minimum support and confidence are needed
39. If {A,B,C,D} is a frequent itemset, candidate rules which is not possible is
- C –> A
- D –> ABCD
- A –> BC
- B –> ADC
D –> ABCD
Machine Learning mcq questions and answers | ml mcq SPPU
Machine Learning mcq on unit 6 | Hierarchical Clustering mcq
Data Mining and Warehousing mcq pdf
data mining and warehousing mcq, data mining and warehousing mcq sppu, data mining and warehousing mcq questions, data mining and data warehousing mcq, data mining and warehouse mcq, high performance computing quiz, dmw mcq, data mining and warehousing multiple choice questions, data mining and warehousing mcq pdf, data warehouse mcq questions and answers, data mining quiz