# Introduction to Machine Coursera Week 2 Quiz

## In this blog you will find the correct answer of the Coursera quiz Introduction to Machine Coursera Week 2 Quiz mixsaver always try to brings best blogs and best coupon codes

Week 2 Comprehensive

1.
Question 1
What does the equation for the loss function do conceptually?

1 point

• Mathematically define network outputs
• Penalize overconfidence
• Ignore historical statistical developments
• Reward indecision

2.
Question 2
What is overfitting?

1 point

• Overfitting refers to the fact that more complexity is always better, which is why deep learning works.
• Model complexity fits too well to training data and will not generalize in the real-world.
• Model complexity is perfectly matched to the data.
• Model complexity is not enough to capture the nuance of the data and will under-perform in the real-world.

3.
Question 3
Why should the test set only be used once?

1 point

• More than one use can lead to bias.
• More than one use can lead to overfitting.
• The model cannot learn anything new from subsequent uses.
• It is expensive to use more than once.

4.
Question 4
Which two of the following describe the purpose of a validation set?

1 point

• To estimate the performance of a model.
• To pick the best performing model.
• To test the performance in lieu of real-world data.
• To learn the model parameters.

5.
Question 5
How do we learn our network?

1 point

• Downhill skiing
• Monte Carlo simulation
• Analytically determine global minimum

6.
Question 6
What technique is used to minimize loss for a large data set?

1 point

• Newton’s method
• Taylor series expansion

7.
Question 7
Which of the following are benefits of stochastic gradient descent?

1 point

• With stochastic gradient descent, the update time does not scale with data size.
• Stochastic gradient descent finds the solution more accurately.
• Stochastic gradient descent can update many more times than gradient descent.
• Stochastic gradient descent gets near the solution quickly.

8.
Question 8
Why is gradient descent computationally expensive for large data sets?

1 point

• Large data sets do not permit computing the loss function, so a more expensive measure is used.
• Calculating the gradient requires looking at every single data point.
• Large data sets require deeper models, which have more parameters.
• There are too many local minima for an algorithm to find.

9.
Question 9
What are the two main benefits of early stopping?

1 point

• It helps save computation cost.
• It performs better in the real world.
• It improves the training loss.
• There is rigorous statistical theory on it.

10.
Question 10
Why are optimization and validation at odds?

1 point

• Optimization seeks to do as well as possible on a training set, while validation seeks to generalize to the real world.
• Optimization seeks to generalize to the real world, while validation seeks to do as well as possible on a validation set.
• Optimization seeks to do as well as possible on a training set, while validation seeks to do as well as possible on a validation set.
• They are not at odds—they have the same goal.