版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领
文档简介
1、Machine Learning,Survey,Some Suggestions,Teaching Top 30% English Slides Elements Requirements Basic: Exam Engineering: practice Research: theory+practice,Relations with Other Fields,Courses, Books, and Conferences,What is Machine Learning,Machine learningis a subfield ofcomputer sciencethat evolved
2、 from the study of pattern recognition andcomputational learning theoryin artificial intelligence.Machine learning explores the study and construction of algorithmsthat can learnfrom and make predictions ondata.Such algorithms operate by building amodelfrom example inputs in order to make data-drive
3、n predictions or decisions,rather than following strictly static program instructions. 1959, Arthur Samuel: Field of study that gives computers the ability to learn without being explicitly programmed”. Tom M. Mitchell: A computer program is said to learn from experience E with respect to some class
4、 of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E”. (PAC theory,7,Defining the Learning Task,Improve on task, T, with respect to performance metric, P, based on experience, E,T: Categorize email messages as spam or legitimate. P: Pe
5、rcentage of email messages correctly classified. E: Database of emails, some with human-given labels T: Recognizing hand-written words P: Percentage of words correctly classified E: Database of human-labeled images of handwritten words T: Driving on four-lane highways using vision sensors P: Average
6、 distance traveled before a human-judged error E: A sequence of images and steering commands recorded while observing a human driver. T: Playing checkers P: Percentage of games won against an arbitrary opponent E: Playing practice games against itself,A Learning Comic,Experience: Training Data What
7、is to be learned: Target Function Learning Algorithm: how to infer the target function from the experience Evaluation: Test Data,Categories of Machine Learning,Supervised Learning: direct labeled training data Unsupervised Learning: unlabeled training data Semi-Supervised Learning: unlabeled + label
8、ed training data Reinforcement Learning: indirect labeled training data Transfer Learning: training and test data are Non IID Multi-Task Learning: multiple task share representation Active Learning: actively choose training data,History of Machine Learning (From Eren Golges Blog,First step toward pr
9、evalent ML was proposed byHebb, in 1949, based on a neuropsychological learning formulation. It is calledHebbian Learningtheory. With a simple explanation, it pursues correlations between nodes of a Recurrent Neural Network (RNN). It memorizes any commonalities on the network and serves like a memor
10、y later. Formally, the argument states that,Let us assume that the persistence or repetition of a reverberatory activity (or trace) tends to induce lasting cellular changes that add to its stability. When anaxonof cellAis near enough to excite a cellBand repeatedly or persistently takes part in firi
11、ng it, some growth process or metabolic change takes place in one or both cells such thatAs efficiency, as one of the cells firingB, is increased,History of Machine Learning (From Eren Golges Blog,Arthur Samuel,In 1952,Arthur Samuelat IBM, developed a program playingCheckers. The program was able to
12、 observe positions and learn a implicit model that gives better moves for the latter cases. Samuel played so many games with the program and observed that the program was able to play better in the course of time,F. Rosenblatt,In 1957,RosenblattsPerceptronwas the second model proposed again with neu
13、roscientific background and it is more similar to todays ML models. It was a very exciting discovery at the time and it was practically more applicable than Hebbians idea. The perceptron is designed to illustrate some of the fundamental properties of intelligent systems in general, without becoming
14、too deeply enmeshed in the special, and frequently unknown, conditions which hold for particular biological organisms,WidrowengravedDelta Learning rulethat is then used as practical procedure for Perceptron training. It is also known asLeast Squareproblem. Combination of those two ideas creates a go
15、od linear classifier,History of Machine Learning (From Eren Golges Blog,However, Perceptrons excitement was hinged byMinskyin 1969 . He proposed the famousXORproblem and the inability of Perceptrons in such linearly inseparable data distributions. It was the Minskys tackle to NN community. Thereafte
16、r, NN researches would be dormant up until 1980s,History of Machine Learning (From Eren Golges Blog,There had been not to much effort until the intuition ofMulti-Layer Perceptron (MLP)was suggested byWerbosin 1981 with NN specificBackpropagation(BP)algorithm, albeit BP idea had been proposed before
17、byLinnainmaain 1970 in the name reverse mode of automatic differentiation. Still BP is the key ingredient of todays NN architectures. With those new ideas, NN researches accelerated again. In 1985 - 1986 NN researchers successively presented the idea ofMLPwith practicalBPtraining,Hecht-Nielsen, Robe
18、rt. Theory of the backpropagation neural network.Neural Networks, 1989. IJCNN., International Joint Conference on. IEEE, 1989,History of Machine Learning (From Eren Golges Blog,At the another spectrum, a very-well known ML algorithm was proposed byJ. R. Quinlanin 1986 that we callDecision Trees, mor
19、e specificallyID3algorithm. This was the spark point of the another mainstream ML.Moreover, ID3 was also released as a software able to find more real-life use case with its simplistic rules and its clear inference, contrary to still black-box NN models. After ID3, many different alternatives or imp
20、rovements have been explored by the community (e.g. ID4, Regression Trees, CART .) and still it is one of the active topic in ML,Quinlan, J. Ross. Induction of decision trees.Machine learning1.1 (1986): 81-106,History of Machine Learning (From Eren Golges Blog,One of the most important ML breakthrou
21、gh wasSupport Vector Machines(Networks) (SVM), proposed byVapnik and Cortesin1995with very strong theoretical standing and empirical results. That was the time separating the ML community into two crowds as NN or SVM advocates. However the competition between two community was not very easy for the
22、NN sideafterKernelized version of SVM bynear 2000s.(I was not able to find the first paper about the topic), SVM got the best of many tasks that were occupied by NN models before. In addition, SVM was able to exploit all the profound knowledge of convex optimization, generalization margin theory and
23、 kernels against NN models. Therefore, it could find large push from different disciplines causing very rapid theoretical and practical improvements,Cortes, Corinna, and Vladimir Vapnik. Support-vector networks.Machine learning20.3 (1995): 273-297,History of Machine Learning (From Eren Golges Blog,N
24、N took another damage by the work of Hochreiters thesis in 1991 and Hochreiter et. al. in 2001, showing the gradient loss after the saturation of NN units as we apply BP learning. Simply means, it is redundant to train NN units after a certain number of epochs owing to saturated units hence NNs are
25、very inclined to over-fit in a short number of epochs,Little before, another solid ML model was proposed byFreund and Schapirein1997prescribed with boosted ensemble of weak classifiers calledAdaboost.This work also gave the Godel Prize to the authors at the time. Adaboost trains weak set of classifi
26、ers that are easy to train, by giving more importance to hard instances. This model still the basis of many different tasks like face recognition and detection,History of Machine Learning (From Eren Golges Blog,Another ensemble model explored byBreimanin2001that ensembles multiple decision trees whe
27、re each of them is curated by a random subset of instances and each node is selected from a random subset of features. Owing to its nature,it is calledRandom Forests(RF). RF has also theoretical and empirical proofs of endurance against over-fitting. Even AdaBoost shows weakness to over-fitting and
28、outlier instances in the data, RFis more robust model against these caveats.(For more detail about RF, refer tomy old post.). RF shows its success in many different tasks like Kaggle competitions as well,Random forests are a combination of tree predictors such that each tree depends on the values of
29、 arandom vector sampled independently and with the same distribution for all trees in the forest. The generalization error for forests converges a.s. to a limit as the number of trees in the forest becomes large,History of Machine Learning (From Eren Golges Blog,As we come closer today, a new era of
30、 NN calledDeep Learninghas been commerced. This phrase simply refers NN models with many wide successive layers. The 3rd rise of NN has begun roughly in 2005with the conjunction of many different discoveries from past and present byrecent mavens Hinton, LeCun, Bengio, Andrew Ng and other valuable ol
31、der researchers,Hinton, G. E. and Salakhutdinov, R. RReducing the dimensionality of data with neural networks.Science, Vol. 313. no. 5786, pp. 504 - 507, 28 July 2006,History of Machine Learning (From Eren Golges Blog,History of Machine Learning,With the combination of all those ideas and non-listed ones, NN models are able to beat off state of art at very different tasks such as Object Recognition, Speech Recognition
温馨提示
- 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
- 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
- 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
- 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
- 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
- 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
- 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。
最新文档
- 新能源开发与利用技术发展趋势
- 陆川县(2026年)公安辅警招聘知识考试题(含答案)
- 少数民族传统建筑艺术的保护与发展
- 家庭用电:太阳能与风能的综合利用
- 2025年湖南工程学院应用技术学院辅导员招聘考试笔试模拟试题及答案解析
- (2025年)证券交易试题及答案
- 2026湖南省中南林业科技大学涉外学院人才招聘备考题库附完整答案详解【考点梳理】
- 2025年人教版六年级数学下册第三单元综合检测试卷及答案
- 2025-2026闽教院翔安一附小招聘非在编合同教师1人备考题库(二)及答案详解参考
- 2026浙江台州市温岭市滨海镇招聘编外工作人员1人备考题库含答案详解【培优b卷】
- 2024年皖西卫生职业学院单招职业适应性测试题库及答案解析
- 《爱鸟惜花守家园·考察身边的生物资源》课件 2023-2024学年辽海版《综合实践活动》七年级下册
- 全国行业职业技能竞赛(电力交易员)考试题库及答案
- 《家用电冰箱与空调器维修》课件
- GB/T 14048.11-2024低压开关设备和控制设备第6-1部分:多功能电器转换开关电器
- (完整版)初中化学新课程标准(人教版)
- 国家公园入口社区建设标准指南专项研究-国家公园研究院+自然资源保护协会-2024
- 《先进制造技术》教案
- 房建工程监理大纲范本(内容全面)
- 品管圈之降低呼吸机管路积水发生率护理课件
- 《去眼袋手术》课件
评论
0/150
提交评论