当前位置:网站首页>Deep learning Course2 first week Practical aspects of Deep Learning exercises
Deep learning Course2 first week Practical aspects of Deep Learning exercises
2022-08-01 22:07:00 【l8947943】
Practical aspects of Deep Learning
- If you have 10,000,000 examples, how would you split the train/dev/test set?
- 33% train. 33% dev. 33% test
- 60% train. 20% dev. 20% test
- 98% train. 1% dev. 1% test
- When designing a neural network to detect if a house cat is present in the picture, 500,000 pictures of cats were taken by their owners. These are used to make the training, dev and test sets. It is decided that to increase the size of the test set, 10,000 new images of cats taken from security cameras are going to be used in the test set. Which of the following is true?
- This will increase the bias of the model so the new images shouldn’t be used.
- This will be harmful to the project since now dev and test sets have different distributions.
- This will reduce the bias of the model and help improve it.
- If your Neural Network model seems to have high variance, what of the following would be promising things to try?
- Make the Neural Network deeper
- Get more training data
- Add regularization
- Get more test data
- Increase the number of units in each hidden layer
- You are working on an automated check-out kiosk for a supermarket, and are building a classifier for apples, bananas and oranges. Suppose your classifier obtains a training set error of 0.5%, and a dev set error of 7%. Which of the following are promising things to try to improve your classifier? (Check all that apply.)
- Increase the regularization parameter lambda
- Decrease the regularization parameter lambda
- Get more training data
- Use a bigger neural network
- In every case it is a good practice to use dropout when training a deep neural network because it can help to prevent overfitting. True/False?
- True
- False
- The regularization hyperparameter must be set to zero during testing to avoid getting random results. True/False?
- True
- False
- With the inverted dropout technique, at test time:
- You apply dropout (randomly eliminating units) but keep the 1/keep_prob factor in the calculations used in training.
- You do not apply dropout (do not randomly eliminate units), but keep the 1/keep_prob factor in the calculations used in training.
- You apply dropout (randomly eliminating units) and do not keep the 1/keep_prob factor in the calculations used in training
- You do not apply dropout (do not randomly eliminate units) and do not keep the 1/keep_prob factor in the calculations used in training
- Increasing the parameter keep_prob from (say) 0.5 to 0.6 will likely cause the following: (Check the two that apply)
- Increasing the regularization effect
- Reducing the regularization effect
- Causing the neural network to end up with a higher training set error
- Causing the neural network to end up with a lower training set error
- Which of the following actions increase the regularization of a model? (Check all that apply)
- Decrease the value of the hyperparameter lambda.
- Decrease the value of keep_prob in dropout.
Correct. When decreasing the keep_prob value, the probability that a node gets discarded during training is higher, thus reducing the regularization effect. - Increase the value of the hyperparameter lambda.
Correct. When increasing the hyperparameter lambda, we increase the effect of the L_2 penalization. - Increase the value of keep_prob in dropout.
- Use Xavier initialization.
- Which of the following is the correct expression to normalize the input x ? \mathbf{x}? x?
- x = x − μ σ x = \frac{x-\mu }{\sigma } x=σx−μ
边栏推荐
- AI应用第一课:支付宝刷脸登录
- VGUgarbage collector(垃圾回收器)的实现原理
- 19 Lectures on Disassembly of Multi-merchant Mall System Functions - Invoice Management on the Platform
- 将vim与系统剪贴板的交互使用
- Kubernetes Scheduler全解析
- 论文解读(GSAT)《Interpretable and Generalizable Graph Learning via Stochastic Attention Mechanism》
- Based on php hotel online reservation management system acquisition (php graduation project)
- 你居然不懂Bitmap和Drawable? 相关知识大扫盲
- 基于 OData 模型和 JSON 模型的 SAP UI5 表格控件行项目的添加和删除实现
- No more rolls!After joining ByteDance for a week, he ran decisively.
猜你喜欢

高等代数_证明_矩阵的行列式为特征值之积, 矩阵的迹为特征值之和

深度学习Course2第二周Optimization Algorithms习题整理

_ _ determinant of a matrix is higher algebra eigenvalue of the product, the characteristic value of matrix trace is combined

感觉自己好傻

【开源】Sentinel高性能高可用集群限流解决方案

Based on php online music website management system acquisition (php graduation design)

SOM Network 2: Implementation of the Code

模拟数据之mockjs

Based on php hotel online reservation management system acquisition (php graduation project)

LeetCode952三部曲之二:小幅度优化(137ms -> 122ms,超39% -> 超51%)
随机推荐
毕业十年,财富自由:那些比拼命努力更重要的事,从来没人会教你
数据分析面试手册《指标篇》
shell specification and variables
Small program -- subcontracting
小程序毕设作品之微信美食菜谱小程序毕业设计成品(5)任务书
第一讲 测试知多少
自建 Prometheus 采集腾讯云容器服务监控数据最佳实践
Uses of Anacoda
KMP 字符串匹配问题
游戏元宇宙发展趋势展望分析
FusionGAN:A generative adversarial network for infrared and visible image fusion article study notes
seaborn笔记:可视化统计关系(散点图、折线图)
19 Lectures on Disassembly of Multi-merchant Mall System Functions - Invoice Management on the Platform
【移动Web】移动端适配
No more rolls!After joining ByteDance for a week, he ran decisively.
scikit-learn no moudule named six
入门数据库Days4
Shell programming conditional statement
ModuleNotFoundError: No module named 'yaml'
【C语言实现】求两个整数的较大值