<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://replica.wiki.extremist.software/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Dkritz</id>
	<title>Noisebridge - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://replica.wiki.extremist.software/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Dkritz"/>
	<link rel="alternate" type="text/html" href="https://replica.wiki.extremist.software/wiki/Special:Contributions/Dkritz"/>
	<updated>2026-04-05T13:22:51Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.39.13</generator>
	<entry>
		<id>https://replica.wiki.extremist.software/index.php?title=Machine_Learning_Meetup_Notes:2011-4-13&amp;diff=17746</id>
		<title>Machine Learning Meetup Notes:2011-4-13</title>
		<link rel="alternate" type="text/html" href="https://replica.wiki.extremist.software/index.php?title=Machine_Learning_Meetup_Notes:2011-4-13&amp;diff=17746"/>
		<updated>2011-04-19T18:38:42Z</updated>

		<summary type="html">&lt;p&gt;Dkritz: link to uploaded ppt&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Anthony Goldbloom from Kaggle Visits&lt;br /&gt;
&lt;br /&gt;
*Link to his talk: [https://www.noisebridge.net/images/e/ed/Goldbloom_-_Predictive_modeling_competitions_-_April_2011.ppt PPT presentation]&lt;br /&gt;
*Guy used random forests to win HIV competition. Word &amp;quot;random forests&amp;quot; is trademarked. Dude taught himself machine learning from watching youtube videos. Random forests are pretty robust to new data.&lt;br /&gt;
**Used [http://cran.r-project.org/web/packages/caret/ caret] package in R to deal with random forests.&lt;br /&gt;
*Kaggle splits test dataset into two, uses half for leaderboard.&lt;br /&gt;
*Often score difference between winning model and second place is not statistically significant. So they award prizes to top few. Might impose restrictions on execution time of model.&lt;br /&gt;
*Performance bottoms out in competitions within a few weeks in general. This seems to be due to all the information being &amp;quot;squeezed&amp;quot; out of the dataset at that point.&lt;br /&gt;
*Chess rating competition: build a new rating system that more accurately produces the results. The performance still plateaued, but took longer.&lt;br /&gt;
*Most users of kaggle are from computer science and statistics, followed by economics, math, biostats.&lt;br /&gt;
*Tools people use:&lt;br /&gt;
**R: lots of american users&lt;br /&gt;
**Matlab&lt;br /&gt;
**SAS&lt;br /&gt;
**Weka&lt;br /&gt;
**SPSS&lt;br /&gt;
**Python: although it&#039;s lower on the list, people are successful with it&lt;br /&gt;
*R packages used: Caret, RFE, GLM, NNET, Forecast&lt;br /&gt;
*Heritage Prize&lt;br /&gt;
**Real shit is going down may 4th, with release of all datasets.&lt;br /&gt;
**Ends in 2 years. No rush.&lt;br /&gt;
**Four prizes in total, given out throughout the next two years.&lt;/div&gt;</summary>
		<author><name>Dkritz</name></author>
	</entry>
	<entry>
		<id>https://replica.wiki.extremist.software/index.php?title=File:Goldbloom_-_Predictive_modeling_competitions_-_April_2011.ppt&amp;diff=17745</id>
		<title>File:Goldbloom - Predictive modeling competitions - April 2011.ppt</title>
		<link rel="alternate" type="text/html" href="https://replica.wiki.extremist.software/index.php?title=File:Goldbloom_-_Predictive_modeling_competitions_-_April_2011.ppt&amp;diff=17745"/>
		<updated>2011-04-19T18:36:42Z</updated>

		<summary type="html">&lt;p&gt;Dkritz: Anthony Goldbloom from Kaggle Health Prize PPT talk to the ML group Wednesday April 13th 2011&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Anthony Goldbloom from Kaggle Health Prize PPT talk to the ML group Wednesday April 13th 2011&lt;/div&gt;</summary>
		<author><name>Dkritz</name></author>
	</entry>
	<entry>
		<id>https://replica.wiki.extremist.software/index.php?title=Machine_Learning/Datasets&amp;diff=17164</id>
		<title>Machine Learning/Datasets</title>
		<link rel="alternate" type="text/html" href="https://replica.wiki.extremist.software/index.php?title=Machine_Learning/Datasets&amp;diff=17164"/>
		<updated>2011-03-16T06:07:10Z</updated>

		<summary type="html">&lt;p&gt;Dkritz: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Machine learning is a vast field and there are many different types of problems to be solved. If you find a dataset interesting, try to categorize it (or add a new category) and add it to the links below.&lt;br /&gt;
&lt;br /&gt;
===Classification===&lt;br /&gt;
*[http://yann.lecun.com/exdb/mnist/ MNIST Handwritten Digits]&lt;br /&gt;
**Classify handwritten digits using this dataset, a very popular one with lots of training examples.&lt;br /&gt;
*[http://archive.ics.uci.edu/ml/datasets/Heart+Disease Heart Disease]&lt;br /&gt;
**Predict whether a person will have heart disease based on a subset of 76 factors.&lt;br /&gt;
*[http://archive.ics.uci.edu/ml/datasets/Census-Income+%28KDD%29 Census Income]&lt;br /&gt;
**Try to predict whether a person has an income greater than or less than 50k&lt;br /&gt;
&lt;br /&gt;
===Regression===&lt;br /&gt;
*[http://www.sci.usq.edu.au/staff/dunn/Datasets/Books/Hand/Hand-R/alps-R.html Boiling point in the Alps]&lt;br /&gt;
**The boiling point of water at different barometric pressures. &lt;br /&gt;
*[http://www.sci.usq.edu.au/staff/dunn/Datasets/Books/Hand/Hand-R/shocking-R.html Shocking Rats]&lt;br /&gt;
**How does shocking a rat affect it&#039;s ability to complete a maze?&lt;br /&gt;
*[http://www.sci.usq.edu.au/staff/dunn/Datasets/Books/Hand/Hand-R/icecream-R.html Ice Cream Sales]&lt;br /&gt;
**Predict the quantity of ice cream consumed based on some other variables.&lt;br /&gt;
*[http://www.sci.usq.edu.au/staff/dunn/Datasets/applications/health/fev.html Smoking and Respiratory Function]&lt;br /&gt;
**How does smoking affect lung capacity?&lt;br /&gt;
&lt;br /&gt;
===Time Series===&lt;br /&gt;
*[http://robjhyndman.com/tsdldata/data/ausgundeaths.dat Gun-related Deaths in Australia]&lt;br /&gt;
**&amp;quot;Deaths from gun-related homicides and suicides and non-gun-related homicides and suicides. Australia: 1915-2004. Source: Neill and Leigh (2007).&amp;quot;&lt;br /&gt;
*[http://robjhyndman.com/tsdldata/data/immig.dat Immigration Rates]&lt;br /&gt;
**&amp;quot;Annual immigration into the United States: thousands. 1820 – 1962. From Kendall &amp;amp; Ord (1990), p.13.&amp;quot;&lt;br /&gt;
*[http://robjhyndman.com/tsdldata/roberts/beards.dat Percent of Men with Beards 1866-1911]&lt;br /&gt;
**&amp;quot;Percent of Men with full beards, 1866 – 1911. Source: Hipel and Mcleod (1994).&amp;quot;&lt;br /&gt;
*[http://robjhyndman.com/tsdldata/roberts/velmon.dat Velocity of Money in America 1869-1960]&lt;br /&gt;
**The [http://en.wikipedia.org/wiki/Velocity_of_money velocity of money] is basically the number of times a single unit of money changes hands over a period of time.  Theory goes, MV=PY, or Velocity = Prices * Economic Output / Quantity of Money.&lt;br /&gt;
*[http://robjhyndman.com/tsdldata/annual/globtp.dat Changes in Global Air Temperature 1880-1985]&lt;br /&gt;
**&amp;quot;Surface air temperature change for the globe, 1880-1985, Temperature change actually means temperature against an arbitrary zero point. From James Hansen and Sergej Lebedeff, &amp;quot;Global Trends of Measured Surface Air Temperature&amp;quot;, `Journal of Geophysical Research`, Vol. 92, No. D11, pages 13,345-13,372, November 20, 1987.&amp;quot;&lt;br /&gt;
*[http://robjhyndman.com/tsdldata/data/earthq.dat Number of Earthquakes per Year 1900-1988 (&amp;gt;= 7.0)]&lt;br /&gt;
**&amp;quot;Source: National Earthquake Information Center. Different lists will give different numbers depending on the formula used for calculating the magnitude.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
===Clustering===&lt;br /&gt;
*[http://archive.ics.uci.edu/ml/datasets/Plants USDA Plants Data]&lt;br /&gt;
**Automatically cluster plants based on 70 attributes.&lt;br /&gt;
*[http://www.uni-koeln.de/themen/statistik/data/cluster/ Nutriens in Meat, Fish and Fowl]&lt;br /&gt;
**Can you cluster into animal type given the data?&lt;br /&gt;
&lt;br /&gt;
===Text Data===&lt;br /&gt;
*[http://www.cs.cmu.edu/~enron/ Enron Emails]&lt;br /&gt;
**Search through Enron&#039;s publicly accessible emails.&lt;br /&gt;
*[http://archive.ics.uci.edu/ml/datasets/Bag+of+Words Bag of Words]&lt;br /&gt;
**Collection of word counts for various types of documents, including Enron emails, scientific papers, and New York Times articles.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Reinforcement Learning===&lt;/div&gt;</summary>
		<author><name>Dkritz</name></author>
	</entry>
	<entry>
		<id>https://replica.wiki.extremist.software/index.php?title=CS229&amp;diff=12569</id>
		<title>CS229</title>
		<link rel="alternate" type="text/html" href="https://replica.wiki.extremist.software/index.php?title=CS229&amp;diff=12569"/>
		<updated>2010-09-07T20:50:37Z</updated>

		<summary type="html">&lt;p&gt;Dkritz: /* Progress: Watching Lectures */  adding Dave (myself) to the lecture track list&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Overview ==&lt;br /&gt;
CS229 is the undergraduate machine learning course at Stanford. You can see the lectures from [http://itunes.apple.com/WebObjects/MZStore.woa/wa/viewiTunesUCollection?id=384233048#ls=1 iTunesU] and [http://www.youtube.com/results?search_query=stanford%20cs%20229&amp;amp;search=Search&amp;amp;sa=X&amp;amp;oi=spell&amp;amp;resnum=0&amp;amp;spell=1 Youtube]. We are going to be working through the course at one lecture a week starting 1 September 2010 and finishing in January 2011. There are four problem sets which we&#039;ll be doing one every 4 weeks.&lt;br /&gt;
&lt;br /&gt;
[http://www.stanford.edu/class/cs229/ http://www.stanford.edu/class/cs229/] &lt;br /&gt;
&lt;br /&gt;
=== Course Description ===&lt;br /&gt;
&lt;br /&gt;
This course provides a broad introduction to machine learning and&lt;br /&gt;
statistical pattern recognition. Topics include: supervised learning&lt;br /&gt;
(generative/discriminative learning, parametric/non-parametric&lt;br /&gt;
learning, neural networks, support vector machines); unsupervised&lt;br /&gt;
learning (clustering, dimensionality reduction, kernel methods);&lt;br /&gt;
learning theory (bias/variance tradeoffs; VC theory; large margins);&lt;br /&gt;
reinforcement learning and adaptive control. The course will also&lt;br /&gt;
discuss recent applications of machine learning, such as to robotic&lt;br /&gt;
control, data mining, autonomous navigation, bioinformatics, speech&lt;br /&gt;
recognition, and text and web data processing.&lt;br /&gt;
&lt;br /&gt;
== Schedule ==&lt;br /&gt;
* one lecture a week&lt;br /&gt;
* one problem set every five weeks&lt;br /&gt;
&lt;br /&gt;
[http://www.google.com/calendar/embed?src=cWE3bGFpNnZxazdpamNjbmc4bXJsY2hyNGdAZ3JvdXAuY2FsZW5kYXIuZ29vZ2xlLmNvbQ  Google Calendar of schedule]&lt;br /&gt;
&lt;br /&gt;
==Progress: Watching Lectures ==&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; cellpadding=&amp;quot;5&amp;quot; align=&amp;quot;center&amp;quot;&lt;br /&gt;
| Name&lt;br /&gt;
| Lecture 1&lt;br /&gt;
| Lecture 2&lt;br /&gt;
| Lecture 3&lt;br /&gt;
| Lecture 4&lt;br /&gt;
| Lecture 5&amp;lt;br /&amp;gt; 9/29&lt;br /&gt;
| Lecture 6&lt;br /&gt;
| Lecture 7&lt;br /&gt;
| Lecture 8&lt;br /&gt;
| Lecture 9&lt;br /&gt;
| Lecture 10&amp;lt;br /&amp;gt; 11/3&lt;br /&gt;
| Lecture 11&lt;br /&gt;
| Lecture 12&lt;br /&gt;
| Lecture 13&lt;br /&gt;
| Lecture 14&lt;br /&gt;
| Lecture 15&amp;lt;br /&amp;gt; 12/8&lt;br /&gt;
| Lecture 16&lt;br /&gt;
| Lecture 17&lt;br /&gt;
| Lecture 18&lt;br /&gt;
| Lecture 19&lt;br /&gt;
| Lecture 20&amp;lt;br /&amp;gt; 1/12&lt;br /&gt;
|-&lt;br /&gt;
| Thomas&lt;br /&gt;
| First! ;)&amp;lt;br/&amp;gt;[[Image:Gold-star.jpg|center|30px|Gold-Star]]&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
| Joe&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
| Glen&lt;br /&gt;
|[[Image:Gold-star.jpg|center|30px|Gold-Star]]&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
| Jared&lt;br /&gt;
|[[Image:Gold-star.jpg|center|30px|Gold-Star]]&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
| Dave&lt;br /&gt;
|[[Image:Gold-star.jpg|center|30px|Gold-Star]]&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
| You!&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
==Progress: Assignments ==&lt;br /&gt;
{| border=&amp;quot;1&amp;quot; cellspacing=&amp;quot;0&amp;quot; cellpadding=&amp;quot;5&amp;quot; align=&amp;quot;center&amp;quot;&lt;br /&gt;
| Name&lt;br /&gt;
| Problem set 1&amp;lt;br /&amp;gt; due 9/29&lt;br /&gt;
| Problem set 2&amp;lt;br /&amp;gt; due 11/3&lt;br /&gt;
| Problem set 3&amp;lt;br /&amp;gt; due 12/8&lt;br /&gt;
| Problem set 4&amp;lt;br /&amp;gt; due 1/20&lt;br /&gt;
|-&lt;br /&gt;
| Thomas&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|-&lt;br /&gt;
| Joe&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
| Glen&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
| Jared&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
| You!&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|-&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Dkritz</name></author>
	</entry>
</feed>