MachineLearningAlgorithms / data /Decision_Trees_Classification.json
deedrop1140's picture
Upload 41 files
0d00d62 verified
{
"title": "Decision Trees Mastery: 100 MCQs",
"description": "A comprehensive 100-question collection to master Decision Trees — covering fundamentals, splitting criteria, pruning, overfitting control, ensemble integration, and real-world scenarios.",
"questions": [
{
"id": 1,
"questionText": "What is the main purpose of Decision Tree in classification task 1?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 2,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 2,
"questionText": "What is the main purpose of Decision Tree in classification task 2?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 3,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 3,
"questionText": "What is the main purpose of Decision Tree in classification task 3?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 2,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 4,
"questionText": "What is the main purpose of Decision Tree in classification task 4?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 0,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 5,
"questionText": "What is the main purpose of Decision Tree in classification task 5?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 0,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 6,
"questionText": "What is the main purpose of Decision Tree in classification task 6?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 1,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 7,
"questionText": "What is the main purpose of Decision Tree in classification task 7?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 2,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 8,
"questionText": "What is the main purpose of Decision Tree in classification task 8?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 0,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 9,
"questionText": "What is the main purpose of Decision Tree in classification task 9?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 1,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 10,
"questionText": "What is the main purpose of Decision Tree in classification task 10?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 1,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 11,
"questionText": "What is the main purpose of Decision Tree in classification task 11?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 1,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 12,
"questionText": "What is the main purpose of Decision Tree in classification task 12?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 0,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 13,
"questionText": "What is the main purpose of Decision Tree in classification task 13?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 3,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 14,
"questionText": "What is the main purpose of Decision Tree in classification task 14?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 1,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 15,
"questionText": "What is the main purpose of Decision Tree in classification task 15?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 0,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 16,
"questionText": "What is the main purpose of Decision Tree in classification task 16?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 3,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 17,
"questionText": "What is the main purpose of Decision Tree in classification task 17?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 1,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 18,
"questionText": "What is the main purpose of Decision Tree in classification task 18?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 2,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 19,
"questionText": "What is the main purpose of Decision Tree in classification task 19?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 1,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 20,
"questionText": "What is the main purpose of Decision Tree in classification task 20?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 3,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 21,
"questionText": "What is the main purpose of Decision Tree in classification task 21?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 1,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 22,
"questionText": "What is the main purpose of Decision Tree in classification task 22?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 3,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 23,
"questionText": "What is the main purpose of Decision Tree in classification task 23?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 0,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 24,
"questionText": "What is the main purpose of Decision Tree in classification task 24?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 3,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 25,
"questionText": "What is the main purpose of Decision Tree in classification task 25?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 3,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 26,
"questionText": "What is the main purpose of Decision Tree in classification task 26?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 0,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 27,
"questionText": "What is the main purpose of Decision Tree in classification task 27?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 2,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 28,
"questionText": "What is the main purpose of Decision Tree in classification task 28?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 3,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 29,
"questionText": "What is the main purpose of Decision Tree in classification task 29?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 1,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 30,
"questionText": "What is the main purpose of Decision Tree in classification task 30?",
"options": [
"To predict class labels",
"To cluster data",
"To reduce dimensions",
"To normalize data"
],
"correctAnswerIndex": 0,
"explanation": "Decision Trees are used to predict class labels based on input features."
},
{
"id": 31,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 1,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 32,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 33,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 1,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 34,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 2,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 35,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 3,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 36,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 1,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 37,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 38,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 2,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 39,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 40,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 1,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 41,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 1,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 42,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 3,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 43,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 3,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 44,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 3,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 45,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 46,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 47,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 1,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 48,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 49,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 2,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 50,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 51,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 2,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 52,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 1,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 53,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 2,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 54,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 1,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 55,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 2,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 56,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 57,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 1,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 58,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 59,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 60,
"questionText": "Scenario: A Decision Tree is overfitting the training data. What should you do?",
"options": [
"Increase tree depth",
"Prune the tree",
"Add more features",
"Decrease learning rate"
],
"correctAnswerIndex": 0,
"explanation": "Pruning helps reduce overfitting by removing unnecessary branches."
},
{
"id": 61,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 1,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 62,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 0,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 63,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 3,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 64,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 65,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 66,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 1,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 67,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 0,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 68,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 1,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 69,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 0,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 70,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 3,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 71,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 0,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 72,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 0,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 73,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 0,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 74,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 3,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 75,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 0,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 76,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 0,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 77,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 3,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 78,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 79,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 80,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 81,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 82,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 1,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 83,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 3,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 84,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 3,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 85,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 1,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 86,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 87,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 3,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 88,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 89,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 1,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 90,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 91,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 1,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 92,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 93,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 1,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 94,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 0,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 95,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 0,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 96,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 1,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 97,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 1,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 98,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 3,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 99,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
},
{
"id": 100,
"questionText": "Scenario: You are building a Decision Tree on a dataset with continuous features and high variance. What splitting criterion might perform best?",
"options": [
"Entropy",
"Information Gain",
"Gini Index",
"Chi-Square"
],
"correctAnswerIndex": 2,
"explanation": "Gini Index or Information Gain are common criteria; the best depends on data distribution, but both handle continuous attributes effectively."
}
]
}