机器学习介绍(英文版:备注里有中文翻译).ppt

上传人:小飞机 文档编号:6438823 上传时间:2023-10-31 格式:PPT 页数:26 大小:313.49KB
返回 下载 相关 举报
机器学习介绍(英文版:备注里有中文翻译).ppt_第1页
第1页 / 共26页
机器学习介绍(英文版:备注里有中文翻译).ppt_第2页
第2页 / 共26页
机器学习介绍(英文版:备注里有中文翻译).ppt_第3页
第3页 / 共26页
机器学习介绍(英文版:备注里有中文翻译).ppt_第4页
第4页 / 共26页
机器学习介绍(英文版:备注里有中文翻译).ppt_第5页
第5页 / 共26页
点击查看更多>>
资源描述

《机器学习介绍(英文版:备注里有中文翻译).ppt》由会员分享,可在线阅读,更多相关《机器学习介绍(英文版:备注里有中文翻译).ppt(26页珍藏版)》请在三一办公上搜索。

1、2023/10/31,Machine Learning,2023/10/31,Machine learning,as a branch of artificial intelligence,is general terms of a kind of analytical method.It mainly utilizes computer simulate or realize the learned behavior of human.,2023/10/31,1)Machine learning just like a true champion which go haughtily;2)P

2、attern recognition in process of decline and die out;3)Deep learning is a brand-new and rapidly rising field.,the Google search index of three concept since 2004,2023/10/31,The constructed machine learning system based on computer mainly contains two core parts:representation and generalization.The

3、first step for data learning is to represent the data,i.e.detect the pattern of data.Establish a generalized model of data space according to a group of known data to predict the new data.The core target of machine learning is to generalize from known experience.Generalization means a power of which

4、 the machine learning system to be learned for known data that could predict the new data.,2023/10/31,Supervised learningInput data has labels.The common kind of learning algorithm is classification.The model has been trained via the correspondence between feature and label of input data.Therefore,w

5、hen some unknown data which has features but no label input,we can predict the label of unknown data according to the existing model.,2023/10/31,Unsupervised learningInput data has no labels.It relates to another learning algorithm,i.e.clustering.The basic definition is a course that divide the gath

6、er of physical or abstract object into multiple class which consist of similar objects.,2023/10/31,If the output eigenvector marks come from a limited set that consist of class or name variable,then the kind of machine learning belongs to classification problem.If output mark is a continuous variabl

7、e,then the kind of machine learning belongs to regression problem.,2023/10/31,Classification step,Feature extraction,Feature selection,Model training,Classification and prediction,Raw data,New data,2023/10/31,Feature selection(feature reduction),Curse of Dimensionality:Usually refer to the problem t

8、hat concerned about computation of vector.With the increase of dimension,calculated amount will jump exponentially.Cortical features of different brain regions exhibit variant effect during the classification process and may exist some redundant feature.In particular after the multimodal fusion,the

9、increase of feature dimension will cause“curse of Dimensionality”.,2023/10/31,Principal Component Analysis,PCAPCA is the most common linear dimension reduction method.Its target is mapping the data of high dimension to low-dimension space via certain linear projection,and expect the variance of data

10、 that project the corresponding dimension is maximum.It can use fewer data dimension meanwhile retain the major characteristic of raw data.,2023/10/31,Linear discriminant analysis,LDAThe basic idea of LDA is projection,mapping the N dimension data to low-dimension space and separate the between-grou

11、ps as soon as possible.i.e.the optimal separability in the space.The benchmark is the new subspace has maximum between class distance and minimal inter-object distance.,2023/10/31,Independent component analysis,ICAThe basic idea of ICA is to extract the independence signal from a group of mixed obse

12、rved signal or use independence signal to represent other signal.,2023/10/31,Recursive feature elimination algorithm,RFERFE is a greedy algorithm that wipe off insignificance feature step by step to select the feature.Firstly,cyclic ordering the feature according to the weight of sub-feature in clas

13、sification and remove the feature which rank at terminal one by one.Then,according to the final feature ordering list,select different dimension of several feature subset front to back.Assess the classification effect of different feature subset and then get the optimal feature subset.,2023/10/31,Cl

14、assification algorithm,Decision treeDecision tree is a tree structure.Each nonleaf node expresses the test of a feature property and each branch expresses the output of feature property in certain range and each leaf node stores a class.The decision-making course of decision tree is starting from ro

15、ot node,testing the corresponding feature property of waiting objects,selecting the output branch according to their values,until reaching the leaf node and take the class that leaf node store as the decision result.,2023/10/31,Naive Bayes,NBNB classification algorithm is a classification method in

16、statistics.It use probability statistics knowledge for classification.This algorithm could apply to large database and it has high classification accuracy and high speed.,2023/10/31,Artificial neural network,ANNANN is a mathematical model that apply a kind of structure which similar with synapse con

17、nection for information processing.In this model,a mass of node form a network,i.e.neural network,to reach the goal of information processing.Neural network usually need to train.The course of training is network learning.The training change the link weight of network node and make it possess the fu

18、nction of classification.The network after training apply to recognize object.,2023/10/31,k-Nearest Neighbors,kNNkNN algorithm is a kind of classification method base on living example.This method is to find the nearest k training samples with unknown sample x and examine the most of k samples belon

19、g to which class,then x belongs to that class.kNN is a lazy learning method.It stores samples but proceed classification until need to classify.If sample set are relatively complex,it may be lead to large computation overhead.So it cannot apply to strongly real-time occasion.,2023/10/31,support vect

20、or machine,SVMMapping the linearly inseparable data in low-dimension space to high-dimension space and make it linearly separable,2023/10/31,Cross validation,CV,The basic idea of CV is grouping the raw data in a sense.One part is taken as train set,the other part is taken as validation set.Primarily

21、,the classifier is trained with train set,and then use validation set to test the received model by training.,2023/10/31,K-fold cross-validationInk-fold cross-validation,the original sample is randomly partitioned intokequal sized subsamples.Of theksubsamples,a single subsample is retained as the va

22、lidation data for testing the model,and the remainingk1 subsamples are used as training data.The cross-validation process is then repeatedktimes(thefolds),with each of theksubsamples used exactly once as the validation data.Thekresults from the folds can then be averaged to produce a single estimati

23、on.The advantage of this method over repeated random sub-sampling is that all observations are used for both training and validation,and each observation is used for validation exactly once.10-fold cross-validation is commonly used.,2023/10/31,Leave-one-out cross-validation,LOOCVWhenk=n(the number o

24、f observations),thek-fold cross-validation is exactly the leave-one-out cross-validation.,2023/10/31,confusion matrix,TPgold standard and test affirm suffer from certain illness;TNgold standard and test affirm not suffer from certain illness;FPgold standard affirm not suffer from certain illness but

25、 test opposite;FNgold standard affirm suffer from certain illness but test opposite;,2023/10/31,Accuracy:(TP+TN)/(TP+TN+FP+FN),Sensitivity:TP/(TP+FN),Specificity:TN/(FP+TN),2023/10/31,ROC与AUC,TPR=TP/(TP+FN)FPR=FP/(FP+TN),The ROC curve more approach upper left,the performance of classifier exhibits w

26、ell.,2023/10/31,Area Under Curve,AUC,AUC is defined as the area under ROC curve.Obviously,its value cannot greater than 1.Take AUC as evaluation criterion because ROC cannot clearly illustrate which classifier has a better effect.But as a value,the classifier corresponding to the bigger AUC has a better effect.,2023/10/31,THANKS!,

展开阅读全文
相关资源
猜你喜欢
相关搜索

当前位置:首页 > 生活休闲 > 在线阅读


备案号:宁ICP备20000045号-2

经营许可证:宁B2-20210002

宁公网安备 64010402000987号