Question: 1. Why would I choose to use an Extra Trees model vs. a Random Forest model? Select the best answer. Because it is easier to

1. Why would I choose to use an Extra Trees model vs. a Random Forest model? Select the best answer.

Because it is easier to explain.

Because it doesn't add noise by resampling with replacement.

Because it improves accuracy.

Because it reduces variance and works faster.

2. I am a real estate agent building a support vector regressor to determine the price of a house I'm listing. Why should I use this model vs. an OLS model? Select the best answer.

SVR models only consider errors falling outside of a given range, creating some error tolerance.

SVR models work better for problems with high variability, and offer a more precise prediction.

SVR models work well with non-linear data.

SVR models can produce multiple outputs.

3. When would the choice between Gradient Boosting and XGBoost not matter?

When scalability is not a concern

When explainability is not a concern

When working with only continuous data points

When the accuracy score from both models is the same

4. In which of the following scenarios would you expect K-means models to struggle?

When data consists of anisotropic clusters

When data consists of circular clusters

When data is noisy

When data consists of cylindrical clusters

5. Which of the following is not a strength of hierarchical clustering?

Hierarchical clustering takes a bottom-up approach to establish clusters and can work well with non-linearly separable data

Hierarchical clustering works well with clearly defined and separable clusters

Hierarchical clustering works best with clusters of varied spreads

6. Which of the following business metrics are useful to assess the effectiveness of recommender systems?

Accuracy Score

Customer Lifetime Value

Mean Squared Error

Product Uptake

7. How does a Random Forest model compute feature importance?

By calculating the average value of gini within each split.

By calculating a weighted impurity for each node, for each feature.

By calculating the weighted impurity for each tree, for each node.

By calculating the average number of splits until nodes are pure.

8. Which of the following is a strength for stacked ensembles?

Since models are diverse, the strengths from each approach can lead to a better prediction

They process data in parallel through many models for a faster output

They give us the ability to combine classifiers and regressors in a single process

Since models are usually simplistic, the outputs are explainable and easy to compute

9. In which scenario below is soft voting better than hard voting? Select the best answer.

Predicting whether to mine in a particular geographical region

Predicting whether a credit card transaction is fraudulent

Predicting which product to recommend to a customer

Predicting whether someone has a disease

10. Collaborative filtering can use cosine similarity to find similar customers.

True

False

11. Ensemble learning is used because ________. Select the best answer.

It reduces computational inefficiencies.

It provides a better accuracy score.

It is better than other models for highly complex problems.

It reduces overfitting by layering many models together.

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Mathematics Questions!