Browse Wiki & Semantic Web

Jump to: navigation, search
Http://dbpedia.org/resource/Sample complexity
  This page has no properties.
hide properties that link here 
  No properties link to this page.
 
http://dbpedia.org/resource/Sample_complexity
http://dbpedia.org/ontology/abstract The sample complexity of a machine learninThe sample complexity of a machine learning algorithm represents the number of training-samples that it needs in order to successfully learn a target function. More precisely, the sample complexity is the number of training-samples that we need to supply to the algorithm, so that the function returned by the algorithm is within an arbitrarily small error of the best possible function, with probability arbitrarily close to 1. There are two variants of sample complexity: * The weak variant fixes a particular input-output distribution; * The strong variant takes the worst-case sample complexity over all input-output distributions. The No free lunch theorem, discussed below, proves that, in general, the strong sample complexity is infinite, i.e. that there is no algorithm that can learn the globally-optimal target function using a finite number of training samples. However, if we are only interested in a particular class of target functions (e.g, only linear functions) then the sample complexity is finite, and it depends linearly on the VC dimension on the class of target functions.imension on the class of target functions.
http://dbpedia.org/ontology/wikiPageID 43269516
http://dbpedia.org/ontology/wikiPageLength 14205
http://dbpedia.org/ontology/wikiPageRevisionID 1068917677
http://dbpedia.org/ontology/wikiPageWikiLink http://dbpedia.org/resource/Tikhonov_regularization + , http://dbpedia.org/resource/No_free_lunch_theorem + , http://dbpedia.org/resource/Vapnik%E2%80%93Chervonenkis_theory + , http://dbpedia.org/resource/Online_machine_learning + , http://dbpedia.org/resource/Regularization_%28mathematics%29 + , http://dbpedia.org/resource/Dictionary_learning + , http://dbpedia.org/resource/Reinforcement_learning + , http://dbpedia.org/resource/Probably_approximately_correct_learning + , http://dbpedia.org/resource/Active_learning_%28machine_learning%29 + , http://dbpedia.org/resource/Machine_learning + , http://dbpedia.org/resource/VC_dimension + , http://dbpedia.org/resource/Monte_Carlo_tree_search + , http://dbpedia.org/resource/Glivenko-Cantelli_class + , http://dbpedia.org/resource/Rademacher_complexity + , http://dbpedia.org/resource/Random_variable + , http://dbpedia.org/resource/No_free_lunch_in_search_and_optimization + , http://dbpedia.org/resource/Model-free_%28reinforcement_learning%29 + , http://dbpedia.org/resource/Category:Machine_learning + , http://dbpedia.org/resource/Empirical_risk_minimization + , http://dbpedia.org/resource/Semi-supervised_learning + , http://dbpedia.org/resource/Metric_learning +
http://dbpedia.org/property/wikiPageUsesTemplate http://dbpedia.org/resource/Template:Machine_learning_bar + , http://dbpedia.org/resource/Template:Reflist +
http://purl.org/dc/terms/subject http://dbpedia.org/resource/Category:Machine_learning +
http://www.w3.org/ns/prov#wasDerivedFrom http://en.wikipedia.org/wiki/Sample_complexity?oldid=1068917677&ns=0 +
http://xmlns.com/foaf/0.1/isPrimaryTopicOf http://en.wikipedia.org/wiki/Sample_complexity +
owl:sameAs http://fa.dbpedia.org/resource/%D9%BE%DB%8C%DA%86%DB%8C%D8%AF%DA%AF%DB%8C_%D9%86%D9%85%D9%88%D9%86%D9%87 + , http://www.wikidata.org/entity/Q18354077 + , http://dbpedia.org/resource/Sample_complexity + , https://global.dbpedia.org/id/mVCu + , http://rdf.freebase.com/ns/m.0114dpwp +
rdfs:comment The sample complexity of a machine learninThe sample complexity of a machine learning algorithm represents the number of training-samples that it needs in order to successfully learn a target function. More precisely, the sample complexity is the number of training-samples that we need to supply to the algorithm, so that the function returned by the algorithm is within an arbitrarily small error of the best possible function, with probability arbitrarily close to 1. There are two variants of sample complexity:ere are two variants of sample complexity:
rdfs:label Sample complexity
hide properties that link here 
http://dbpedia.org/resource/Outline_of_machine_learning + , http://dbpedia.org/resource/Probably_approximately_correct_learning + , http://dbpedia.org/resource/Artificial_intelligence + , http://dbpedia.org/resource/Random-sampling_mechanism + , http://dbpedia.org/resource/Prior-independent_mechanism + , http://dbpedia.org/resource/M-theory_%28learning_framework%29 + , http://dbpedia.org/resource/Sample-complexity_bounds + http://dbpedia.org/ontology/wikiPageWikiLink
http://en.wikipedia.org/wiki/Sample_complexity + http://xmlns.com/foaf/0.1/primaryTopic
http://dbpedia.org/resource/Sample_complexity + owl:sameAs
 

 

Enter the name of the page to start semantic browsing from.