Lavoisier S.A.S.
14 rue de Provigny
94236 Cachan cedex
FRANCE

Heures d'ouverture 08h30-12h30/13h30-17h30
Tél.: +33 (0)1 47 40 67 00
Fax: +33 (0)1 47 40 67 02


Url canonique : www.lavoisier.fr/livre/informatique/learning-theory-lecture-notes-in-artifi-cial-intelligence-vol-4005/simon/descriptif_2298910
Url courte ou permalien : www.lavoisier.fr/livre/notice.asp?ouvrage=2298910

Learning Theory, 2006 19th Annual Conference on Learning Theory, COLT 2006, Pittsburgh, PA, USA, June 22-25, 2006, Proceedings Coll. Lecture Notes in Artificial Intelligence

Langue : Français

Coordonnateurs : Simon Hans Ulrich, Lugosi Gábor

Couverture de l’ouvrage Learning Theory

This book constitutes the refereed proceedings of the 19th Annual Conference on Learning Theory, COLT 2006, held in Pittsburgh, Pennsylvania, USA, June 2006.

The book presents 43 revised full papers together with 2 articles on open problems and 3 invited lectures. The papers cover a wide range of topics including clustering, un- and semi-supervised learning, statistical learning theory, regularized learning and kernel methods, query learning and teaching, inductive inference, and more.

Invited Presentations.- Random Multivariate Search Trees.- On Learning and Logic.- Predictions as Statements and Decisions.- Clustering, Un-, and Semisupervised Learning.- A Sober Look at Clustering Stability.- PAC Learning Axis-Aligned Mixtures of Gaussians with No Separation Assumption.- Stable Transductive Learning.- Uniform Convergence of Adaptive Graph-Based Regularization.- Statistical Learning Theory.- The Rademacher Complexity of Linear Transformation Classes.- Function Classes That Approximate the Bayes Risk.- Functional Classification with Margin Conditions.- Significance and Recovery of Block Structures in Binary Matrices with Noise.- Regularized Learning and Kernel Methods.- Maximum Entropy Distribution Estimation with Generalized Regularization.- Unifying Divergence Minimization and Statistical Inference Via Convex Duality.- Mercer’s Theorem, Feature Maps, and Smoothing.- Learning Bounds for Support Vector Machines with Learned Kernels.- Query Learning and Teaching.- On Optimal Learning Algorithms for Multiplicity Automata.- Exact Learning Composed Classes with a Small Number of Mistakes.- DNF Are Teachable in the Average Case.- Teaching Randomized Learners.- Inductive Inference.- Memory-Limited U-Shaped Learning.- On Learning Languages from Positive Data and a Limited Number of Short Counterexamples.- Learning Rational Stochastic Languages.- Parent Assignment Is Hard for the MDL, AIC, and NML Costs.- Learning Algorithms and Limitations on Learning.- Uniform-Distribution Learnability of Noisy Linear Threshold Functions with Restricted Focus of Attention.- Discriminative Learning Can Succeed Where Generative Learning Fails.- Improved Lower Bounds for Learning Intersections of Halfspaces.- Efficient Learning Algorithms Yield Circuit Lower Bounds.- OnlineAggregation.- Optimal Oracle Inequality for Aggregation of Classifiers Under Low Noise Condition.- Aggregation and Sparsity Via ?1 Penalized Least Squares.- A Randomized Online Learning Algorithm for Better Variance Control.- Online Prediction and Reinforcement Learning I.- Online Learning with Variable Stage Duration.- Online Learning Meets Optimization in the Dual.- Online Tracking of Linear Subspaces.- Online Multitask Learning.- Online Prediction and Reinforcement Learning II.- The Shortest Path Problem Under Partial Monitoring.- Tracking the Best Hyperplane with a Simple Budget Perceptron.- Logarithmic Regret Algorithms for Online Convex Optimization.- Online Variance Minimization.- Online Prediction and Reinforcement Learning III.- Online Learning with Constraints.- Continuous Experts and the Binning Algorithm.- Competing with Wild Prediction Rules.- Learning Near-Optimal Policies with Bellman-Residual Minimization Based Fitted Policy Iteration and a Single Sample Path.- Other Approaches.- Ranking with a P-Norm Push.- Subset Ranking Using Regression.- Active Sampling for Multiple Output Identification.- Improving Random Projections Using Marginal Information.- Open Problems.- Efficient Algorithms for General Active Learning.- Can Entropic Regularization Be Replaced by Squared Euclidean Distance Plus Additional Linear Constraints.

Proceedings of the 19th Annual Conference on Learning Theory, COLT 2006

Presents 43 revised full papers together with 2 articles on open problems and 3 invited lectures

Coverage includes clustering, un- and semi-supervised learning, statistical learning theory, regularized learning and kernel methods, query learning and teaching, inductive inference, and more