![[personal profile]](https://www.dreamwidth.org/img/silk/identity/user.png)
A bit of information related to machine learning and philosophy talks at MIT.
This Friday, Sep 28, at noon Vladimir Vapnik will give a philosophical talk, "Inductive principles in machine learning and philosophy of science" during a Course 9.S912: "What is Intelligence?" class by Shimon Ullman and Tomaso Poggio (Location: 46-5193 (will move to 46-3310 if a larger room is required, outsiders are welcome to come and listen)). If I can summarize it, I'll update this post later.
On Wednesday, Vapnik gave a talk "From Rosenblatt's learning model to the model of learning with nontrivial teacher" at the new Cambridge Machine Learning Colloquium and Seminar Series. The main mathematical content was that A) it is well known that the error in a support vector machine is inversely proportional to the number of training samples if the classes are well separated by the kernel in question, but is only inversely proportional to the square root of that (i.e. much more training data is needed) if the classes overlap; B) Vapnik claimed that by introducing a second kernel to be used on the training data only (e.g. some creative and not necessarily well formalizable annotations by human annotators, ranging from mundane things to assigning poetic qualities to training samples) one can make the error inversely proportional to the number of training samples even when classes overlap with respect to the main, "production" kernel. (And he was making some far-reaching philosophical conclusions from that, about importance of culture in human learning and things like that. I don't know whether his conclusions can be transferred from support vector machines to other schemas of machine learning. But it certainly looked quite interesting.)
Update: There was a videorecording during the second talk, so there is some chance that there will be a public video. Some material of the first talk was repeated during the last part of the second one (which was 2 hours long). I would not retell the philosophical part. Among the machine learning part, he said that instead of Occam Razor, there is a principle of Large Margin, more precisely, the principle of admitting as many "contradictions" as possible (but contradictions situated on the manifold, and not just anywhere in the embedded space, so to generate artificial contradictions people generate "morphs" (e.g. linear combinations, or mixtures of pixels) of objects of different classes, and this also reduces the resulting error while training on a fixed data set).
This Friday, Sep 28, at noon Vladimir Vapnik will give a philosophical talk, "Inductive principles in machine learning and philosophy of science" during a Course 9.S912: "What is Intelligence?" class by Shimon Ullman and Tomaso Poggio (Location: 46-5193 (will move to 46-3310 if a larger room is required, outsiders are welcome to come and listen)). If I can summarize it, I'll update this post later.
On Wednesday, Vapnik gave a talk "From Rosenblatt's learning model to the model of learning with nontrivial teacher" at the new Cambridge Machine Learning Colloquium and Seminar Series. The main mathematical content was that A) it is well known that the error in a support vector machine is inversely proportional to the number of training samples if the classes are well separated by the kernel in question, but is only inversely proportional to the square root of that (i.e. much more training data is needed) if the classes overlap; B) Vapnik claimed that by introducing a second kernel to be used on the training data only (e.g. some creative and not necessarily well formalizable annotations by human annotators, ranging from mundane things to assigning poetic qualities to training samples) one can make the error inversely proportional to the number of training samples even when classes overlap with respect to the main, "production" kernel. (And he was making some far-reaching philosophical conclusions from that, about importance of culture in human learning and things like that. I don't know whether his conclusions can be transferred from support vector machines to other schemas of machine learning. But it certainly looked quite interesting.)
Update: There was a videorecording during the second talk, so there is some chance that there will be a public video. Some material of the first talk was repeated during the last part of the second one (which was 2 hours long). I would not retell the philosophical part. Among the machine learning part, he said that instead of Occam Razor, there is a principle of Large Margin, more precisely, the principle of admitting as many "contradictions" as possible (but contradictions situated on the manifold, and not just anywhere in the embedded space, so to generate artificial contradictions people generate "morphs" (e.g. linear combinations, or mixtures of pixels) of objects of different classes, and this also reduces the resulting error while training on a fixed data set).
no subject
Date: 2012-09-28 09:32 am (UTC)no subject
Date: 2012-09-28 01:27 pm (UTC)no subject
Date: 2012-09-28 01:10 pm (UTC)no subject
Date: 2012-09-28 01:28 pm (UTC)no subject
Date: 2012-09-28 01:48 pm (UTC)Надеюсь, это скоро появится где-то в письменном виде.
Просто берём для обучающего множества сумму двух ядер, одно обычное, другое — как если бы мы использовали в качестве features вот эти дополнительные аннотации? А потом для классификации отбрасываем (неизвестные нам) слагаемые с дополнительными features?
no subject
Date: 2012-09-28 03:07 pm (UTC)Но у него есть электронная почта в разных местах, можно попробовать попросить слайды, задать вопросы...
no subject
Date: 2012-09-29 12:26 am (UTC)no subject
Date: 2012-09-29 09:15 am (UTC)algol@mccme.ru
no subject
Date: 2012-09-29 02:10 pm (UTC)no subject
Date: 2012-09-29 06:21 pm (UTC)no subject
Date: 2012-09-29 04:42 am (UTC)Не в этом ли и состоит регуляционная теория (см. Плохо поставленные задачи, Тихонов А.Н., Арсенин В.Я, 1974)?
no subject
Date: 2012-09-29 02:11 pm (UTC)no subject
Date: 2012-09-30 04:05 am (UTC)no subject
Date: 2012-09-30 06:21 am (UTC)no subject
Date: 2012-10-03 05:04 am (UTC)no subject
Date: 2012-10-03 10:15 am (UTC)no subject
Date: 2012-10-03 08:04 am (UTC)no subject
Date: 2012-10-03 10:15 am (UTC)no subject
Date: 2012-10-03 08:30 am (UTC)no subject
Date: 2012-10-03 10:16 am (UTC)no subject
Date: 2012-10-03 09:13 am (UTC)Happy Birthday !
no subject
Date: 2012-10-03 10:19 am (UTC)no subject
Date: 2012-10-08 09:59 pm (UTC)Интересные вещи он, видимо, рассказывал.
А он не давал ссылок на препринты или technical reports?
no subject
Date: 2012-10-08 11:15 pm (UTC)> А он не давал ссылок на препринты или technical reports?
Насколько я помню, нет, но у меня есть его слайды, и я исхожу из того, что помещать их на сеть без его разрешения не следует, но вполне можно посылать в частном порядке по электронной почте.
no subject
Date: 2012-10-08 11:30 pm (UTC)no subject
Date: 2012-10-08 11:37 pm (UTC)Ссылки на работы 2008-2010 там есть.
no subject
Date: 2012-10-08 11:42 pm (UTC)no subject
Date: 2012-10-09 12:35 am (UTC)в своих лекциях http://ium.mccme.ru/f12/nesterov.html - там разница между
следованием оракулу (black box) и методами с параллельным моделированием
оракула. Может быть, это как-то связано с новыми разработками Вапника,
а может и нет, надо будет смотреть его статьи.
no subject
Date: 2012-10-09 12:59 am (UTC)no subject
Date: 2012-10-09 08:11 pm (UTC)Если можно, попробуйте сделать трансляцию для этого rss-feeda:
http://club.pdmi.ras.ru/moodle/rss/file.php/1/2/forum/1/rss.xml
(по имени, например, clubpdmi, pdmiboard, ...)
Хотя сам сайт и feed заработают, наверно, только завтра:
http://meshulash.livejournal.com/113923.html?thread=1459459t1459459
no subject
Date: 2012-10-09 08:32 pm (UTC)"There was an error retrieving this URL. The server may be down or the content unavailable at this time. Please verify the URL you have provided and try again."
no subject
Date: 2012-10-09 08:47 pm (UTC)no subject
Date: 2012-10-10 07:03 pm (UTC)no subject
Date: 2012-10-10 08:16 pm (UTC)