tag:blogger.com,1999:blog-13740540239174351982017-06-17T13:57:38.263-07:00Musings about Machine Learning, Technology and TeachingCsaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.comBlogger24125tag:blogger.com,1999:blog-1374054023917435198.post-63320296209492901172013-09-16T00:13:00.000-07:002013-09-16T00:13:02.160-07:00Numerical Errors, Perturbation Analysis and Machine LearningEveryone hates numerical errors. We love to think that computers are machines with infinite precision. When I was a student, I really hated error analysis. It sounded like a subject that is set out to study an annoying side-effect of our imperfect computers, a boring detail that is miles away from anything that anyone would ever consider a nice part of mathematics. I will not try to convince you today that the opposite is true. However, even in error analysis there are some nice ideas and lessons to be learned. This post asks the question whether, if you are doing machine learning, you should care about numerical errors.<br /><br />This issue should be well understood. However, I don't think that it is as well appreciated as it should be, or that it received the attention it should. In fact, I doubt that the issue is discussed in any of the recent machine learning textbooks beyond the usual caveat "beware the numerical errors" (scary!).<br /><br />In this blog, I will illustrate the question of numerical errors and stability with the problem of learning a real-valued function on the $[0,1]$ interval. Let $f$ be this unknown function. We will assume that $f$ is square integrable as our objective will be to produce an approximation to $f$ that is close to $f$ in the (unweighted) $2$-norm. For simplicity, let as assume that we have decided to approximate $f$ using polynomials of degree $d-1$. Thus, defining $G = \{ g \,:\, g(x) = \sum_{j=0}^{d-1} \alpha_j x^j, \alpha_0,\ldots,\alpha_{d-1}\in \mathbb{R} \}$, our goal is to find the minimizer<br /><br /><div style="text-align: center;">$P(f) = \arg\min_{g\in G} \|f-g\|_2$, <span style="color: red;">(Proj)</span></div><br />where $\|f\|_2^2 = \int_0^1 f^2(x) dx$. Clearly, $G$ is a $d$-dimensional subspace of $L^2([0,1])$. To represent the functions in $G$, let us choose a basis of $G$; call the $j$th basis element $p_j$ ($j=1,\ldots,d$). Thus, $p_j:[0,1] \rightarrow \mathbb{R}$ is some polynomial. One possibility, in the lack of better idea, is to choose $p_j(x) = x^j$; if you want, I use $(p_j)_{j=1}^d$ just for the sake of increasing generality (and to play with you a bit later).<br /><br />Since $(p_j)_j$ is a basis, any function $g$ of $G$ can be uniquely written as $g = \sum_{j=1}^d \theta_j p_j$. Since $g^* = P(f)$ is also an element of $G$, it can also be written in this form: $g^* = \sum_{j=1}^d \theta_j^* p_j$. To figure out $g^*$, it is thus sufficient to figure out the vector $\theta^* = (\theta_j^*)_j$. To find this vector, we can find the minimum of<br /><br /><div style="text-align: center;">$J(\theta) = \| f - \sum_{j=1}^d \theta_j p_j \|_2^2$ </div><br />(since $u \mapsto u^2$ is monotone on $[0,\infty)$). For this, it is useful to remember that for $f\in L^2([0,1])$, $\|f\|_2^2 =\langle f,f \rangle$, where the inner produce $\langle \cdot,\cdot\rangle$ is defined for all $f,g\in L^2([0,1])$ using $\langle f,g \rangle = \int_0^1 f(x)g(x) dx$. Taking the derivative of $J(\theta)$ with respect to $\theta_i$ and equating the result with zero, we get,<br /><br /><div style="text-align: center;">$0=\frac{d}{d\theta_i} J(\theta) = 2 \langle \sum_{j=1}^d \theta_j p_j-f, p_i \rangle$.</div><br />Using the symmetry and linearity of the inner product $\langle \cdot,\cdot\rangle$, introducing the matrix $A$ whose $(i,j)$th element is $a_{ij} = \langle p_i, p_j \rangle$ and the vector $b$ whose $i$th element is $\langle f,p_i \rangle$, collecting the above equations for $i=1,\ldots,d$ and reordering the terms, we see that these equations can be written in the short form<br /><br /><div style="text-align: center;">$A\theta = b$.</div><br />Note that here $A$ is a positive definite matrix. Solving this linear system thus will give us $\theta^*$ and hence $g^*$.<br /><br />So far so good. However, in practice, we rarely have access to the true function $f$. One commonly studied set-up in machine learning assumes that we have a noise "sample" of the form $((X_t,Y_t);t=1,\ldots,n)$ only, where $(X_t,Y_t)$ is i.i.d. and $\mathbb{E}[Y_t|X_t] = f(X_t)$ and $\mathrm{Var}(Y_t|X_t)\le c<\infty$. However, I do not want to go this far, but consider a conceptually simpler case when instead of $f$, we can access $\hat{f}: [0,1] \rightarrow \mathbb{R}$. Think of $\hat{f}$ as the "noise corrupted" version of $f$. Later it should become clear how the reasoning that follows extends to the case when only a finite dataset of the previously mentioned form is available.<br /><br />Realizing that we do not have access to $f$, but only to $\hat{f}=f+\Delta f$, it is clear that we won't be able to calculate $g^* = P(f)$, but only (say) $P(\hat{f})$. Let's call the result $\hat{g}^*$: $\hat{g}^* = P(\hat{f})$. How far is $\hat{g}^*$ from $f$?<br /><br />By the Pythagorean theorem, $\|f-\hat{g}^*\|_2^2 = \| f- g^* \|_2^2 + \|g^* - \hat{g}^* \|_2^2$.<br />Here, $ \| f- g^* \|_2^2$ is the error that can only be reduced by increasing the degree $d$ (or increasing $G$), while $\|g^* - \hat{g}^* \|_2^2$ is the error due to starting with the noisy function $\hat{f}$ and not with $f$. The problem of studying the size of $\|f- g^* \|_2^2$ is the subject of approximation theory (where various nice results show how, e.g., the degree of smoothness and $d$ jointly influence the rate at which this error decreases to zero as $d$ goes to infinite) and this error is called the <b>approximation error</b>. We won't get into this nice subject, perhaps in a future post.<br /><br />So it remains to calculate how big the other term is (which is normally called the estimation error, as it measures the effect of the stochastic error $\Delta f$). One way to understand this term is by using the error analysis (a.k.a. stability/perturbation analysis) of linear systems. Perhaps this is not the simplest way, but by pursuing this direction, the article will be more entertaining perhaps and hopefully we will also make some useful observations. Readers suspecting mischief are asked to stay calm..<br /><br />Let us start this error analysis by observing that if we write $\hat{g}^* =\sum_{j=1}\hat{\theta}_j^* p_j$ then the vector $\hat{\theta}^*$ will be the solution to the linear system $A \theta = b+\Delta b$, where $\hat{\Delta b}_j = \langle\Delta f,p_j\rangle$, $j=1,\ldots,d$. Now, $\| \sum_{j=1}^d \theta_j p_j \|_2^2 = \theta^\top A \theta$, hence<br /><br /><div style="text-align: center;">$\|g^* - \hat{g}^* \|_2^2 = ({\theta}^*-\hat{\theta}^*)^\top A ({\theta}^*-\hat{\theta}^*)$.</div><br />Since $A \hat{\theta}^* = b+\Delta b$ and $A {\theta}^* = {b}$, subtracting the two, we get $A ( \hat{\theta}^*- {\theta}^*) = \Delta b$, from where we get that<br /><br /><div style="text-align: center;">$({\theta}^*-\hat{\theta}^*)^\top A ({\theta}^*-\hat{\theta}^*) = (\Delta b)^\top A^{-1} \Delta b$.</div><br />Hence,<br /><br /><div style="text-align: center;"> $\|g^* - \hat{g}^* \|_2^2= (\Delta b)^\top A^{-1} \Delta b\le \lambda_{\min}^{-1}(A) \|\Delta b \|_2^2$, <span style="color: red;">(*)</span></div><br />where $\lambda_{\min}(A)$ denotes the minimum eigenvalue of $A$. Note that by appropriately choosing the error term $\Delta b$, the inequality can be made sharp. What inequality (*) thus means is that the error caused by $\hat{f}-f$ can be badly enlarged if $\Delta b$ has an unfortunate relation to $A$.<br /><br />Readers familiar with the error-analysis of linear systems of equation will recognize that (*) is similar but still different to the inequality that shows how errors propagate in the solution of linear systems due to coefficient errors (including errors in $b$). The difference is easy to understand: It is caused by the fact that in numerical analysis the error of the parameter-vector is measured in the (unweighted) 2-norm, whereas here the error is shown for the norm defined using $\|x\|_{A}^2 = x^\top A x$; in this application this error is more natural than the usual $2$-norm. Thus, this is one of the lessons: Use the norm that is the most appropriate for the application (because of this change in the norms, instead of the conditioning number of $A$, we see the inverse of the minimum eigenvalue of $A$).<br /><br />To see how bad (*) can be, consider the case when $p_j(x) = x^{j-1}$, i.e., the standard polynomial basis. In this case, $a_{ij} = \int_0^1 x^{i+j-2} dx = [x^{i+j-1}/(i+j-1)]_0^1 = 1/(i+j-1)$. The resulting matrix is an old friend, the so-called Hilbert matrix. In fact, this matrix is most infamous for its horrible conditioning, but most of the blame should go to how small the minimum eigenvalue of $A$ is. In fact, a result of Szegő (1975) shows that as $d\rightarrow \infty$, $\lambda_{\min}(A) \sim 2^{15/4} \pi^{3/2} d^{1/2} (\sqrt{2}-1)^{4d+4}$ (an amusing asymptotic relationship on its own). <b>This the reciprocal value of the minimum eigenvalue of $A$ is exponentially large in $d$, thus, potentially totally defeating the purpose of increasing $d$ to reduce the approximation error.</b><br /><br />Thus, perhaps we should choose a different system of polynomials. Indeed, once we know that the goal is to keep $\lambda_{\min}(A)$ large, we can choose the polynomials so as to make this happen. One way, which also helps considerably with speeding up the calculations, is to choose $(p_j)$ such that $A$ is diagonal and if diagonal, why not choose $(p_j)$ such that $A$ becomes the identity matrix. We can simply start by $p_1 = 1$ (because $\int_0^1 1^2 dx=1$) and then choose the coefficients of $p_2(x) = a_{20} + a_{21} x$ such that $\langle p_2,p_2 \rangle=1$ and $\langle p_1,p_2 \rangle = 0$ (this gives $[a_{20} x+ 2 a_{20}a_{21} x^2/2+a_{21}^2 x^3/3]_0^1=1$ and $[a_{20}x + a_{21} x^2/2]_0^1=0$, which can be solved for the coefficients $a_{20}$, $a_{21}$). Continuing the same way, we can find $p_1,\ldots,p_d$, which are known as the <a href="http://en.wikipedia.org/wiki/Legendre_polynomials#Shifted_Legendre_polynomials">shifted Legendre polynomials</a> (the unshifted versions are defined with the same way for the $[-1,1]$ interval).<br /><br />At this stage, however, the astute reader should really wonder whether I have lost my sense of reality! How on the earth should the choice of the basis of $G$ influence the error caused by the "noisy measurement"??? In fact, a very simple reasoning shows that the basis (and thus $A$) should not change how big $\|\hat{g}^*-g^* \|_2^2$ is. To see why note that $\hat{g}^*-g^* = P(\hat{f})-P(f)$. Up to know, I have carefully avoided giving a name to $P$, but in fact, $P$ does not have to be introduced: It is the well-known orthogonal projection operator. As such, it is a linear operator. Hence, $ P(\hat{f})-P(f) = P( \hat{f}-f ) = P(\Delta f)$. It is also known, that projections cannot make vectors larger (they are non-expansions in the 2-norm that defines them). Hence, $\|P(\Delta f)\|_2 \le \|\Delta f\|_2$, which in summary means that <br /><br /><div style="text-align: center;">$\|\hat{g}^*-g^*\|_2 \le \|\Delta f \|_2$.</div><br />The "measurement" error of $f$ cannot be enlarged by projecting the measured function to the function space $G$. How do we reconcile this with (*)? One explanation is that (*) is too conservative; it is a gross over-estimate. However, we also said that the only inequality that was used could be sharp if $\Delta b$ was selected in an "unfortunate" fashion. So that inequality could be tight. Is this a contradiction? [Readers are of course encouraged to stop here and think this through before continuing with reading the rest.]<br /><br />No, there is no contradiction. All the derivations were sounds. The answer to the puzzle is that $\Delta b$ cannot be selected arbitrarily. By writing $\Delta f = \Delta f^{\parallel}+ \Delta f^{\perp}$, where $\Delta f^{\parallel}\in G$ and $\Delta f^{\perp}$ is perpendicular to $G$, we see that $\Delta b_j = \langle \Delta f^{\parallel}, p_j\rangle$. Since $\Delta f^{\parallel}\in G$, it can be written as $\sum_j \gamma_j p_j$ with some coefficients $(\gamma_j)$. Hence, $\Delta b_j = \langle \Delta f, p_j \rangle = \sum_k \gamma_k \langle p_k,p_j\rangle$, or $\Delta b = A \gamma$! Thus $\Delta b$ is restricted to lie in the range of $A$. The rest is easy:<br /><br /><div style="text-align: center;">$\|g^* - \hat{g}^* \|_2^2= (\Delta b)^\top A^{-1} \Delta b = \gamma^\top A \gamma = \|P(\Delta f) \|_2^2$.</div><br />The conclusion? Was this a pointless exercise? Is the stability analysis of linear systems irrelevant to machine learning? No, not at all! What we can conclude is that although the basis chosen does not influence the error in the final approximator that can be attributed to errors in measuring the unknown function, this conclusion holds only if we assume infinite precision arithmetic. With finite precision arithmetic, our analysis shows that rounding errors can be blown up exponentially with the dimension of the subspace $G$ growing to infinity. Thus, with an improperly chosen basis, the rounding errors can totally offset or even reverse the benefit that is expected from increasing the dimensionality of the approximation space $G$.<br /><br /><div style="text-align: center;"> <b>If you cannot calculate it with a computer, you cannot estimate it either.</b></div><br /> (The careful reader may wonder about whether the rounding errors $\tilde{\Delta b}$ can lie in the dreaded subspace of $A$, but in this case they can.) Hence, when consider a finite dimensional approximation, care must be taken to choose the basis functions so that the resulting numerical problem is stable.<br /><br />Finally, one commenter on the blog once noted that I should post more about reinforcement learning. So, let me finish by remarking that the same issues exist in reinforcement learning, as well. A basic problem in reinforcement learning is to estimate the value function underlying a stationary Markov policy. Long story short, such a policy gives rise to a Markov chain, with a transition probability kernel $\mathcal{P}$. We can view $\mathcal{P}$ as an operator that maps bounded measurable $f:X \rightarrow \mathbb{R}$ functions to akin functions: if $g=\mathcal{P} (f)$, $g(x) = \int_{y} f(y) \mathcal{P}(dy|x)$, where the integration is over the state space $X$. Evaluating the said policy then amounts to solving<br /><br /><div style="text-align: center;">$(I-\gamma\mathcal{P}) f = r$ (PEval)</div><br />in $f$, where $0\le \gamma < 1$ is the so-called discount factor and $r:X \rightarrow \mathbb{R}$ is the so-called reward function (again, $r$ is assumed to be bounded, measurable). A common method to solve linear inverse problems like the above is Galerkin's method. Long story short, this amounts to selecting an orthogonal basis $(\psi_k)$ in $L^2(X;\mu)$ with $\mu$ being a probability measure and then solving the said system in $G= \{ f\,:\, f =\sum_{k=1}^d \alpha_k \psi_k, \alpha_1,\ldots,\alpha_d\in \mathbb{R} \}$ in the sense that the solution $g\in G$ is required to satisfy (PEval) in the 'weak sense': $\langle ((I-\gamma\mathcal{P})g,\psi_k\rangle = \langle r,\psi_k\rangle$, $k=1,\ldots,d$. As it turns out, there are numerically less and more stable ways of selecting the basis $(\psi_k)$, though figuring out how to do this in the practical settings when $I-\gamma\mathcal{P}$ is unknown but can be sampled remains for future work.<br /><br /><b>Bibliography:</b><br />Szegő G. 1982 <i>On some Hermitian forms associated with two given curves of the complex plane</i>, Collected Papers, vol 2. (Basle: Birkhauser) p 666.<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com2tag:blogger.com,1999:blog-1374054023917435198.post-1392920455950610242013-09-15T21:19:00.004-07:002013-09-15T21:19:58.635-07:00Student Response Systems -- A Year LaterI have promised to come back with my experience after the Fall semester of 2013. Though the end of that semester passed a long ago, here are some thoughts:<br /><br />Overall my experience of socrative was very positive. During the first week I polled the class to see how large a percentage of the class has some kind of wifi enabled device that they could use. 90 out of the 95 students had some kind of device that they were willing to bring to the class, so I decided to give socrative a try.<br /><br />Socrative helped me tremendously to stay on the top of what everyone in the class knows. The way I used socrative was as follows: After every block of new material (usually 5-10 slides), I inserted a few questions to verify whether the students "got" the new material. This was all the "extra work" that I had to put in designing the class material due to socrative. And I would have done something similar without socrative anyways, so in fact I did not feel much of a difference here. Once the question was projected, I started the question on socrative by pushing a button and then the students could start entering their answers. The answers are shown on the teacher's screen. Based on the feedback received, I could then decide whether I needed to re-explain something.<br /><br />I used a tablet to connect to socrative, while I used a second computer to project my slides: This was handy as I did not have to exit the presentation to start the question in socrative. In the classroom there was a projector that could project the tablet's screen. Sometimes I have shown the feedback screen even before I got all the answers in. The students seemed to like this as it created an interesting dynamic in the class.<br /><br />I used multiple choice questions mostly, but I also used yes/no questions occasionally. In addition to these, I also experimented with "short answer" questions. This latter question-type would be very useful and I generally prefer it. It is not only less work to prepare "short answer" questions, but the they are usually better at testing what the students know. To put in another way, it is really hard (and time-consuming) to design good multiple choice questions with valid-looking alternatives (if you have some ideas of how to automate this, I am all ears). Examples of the "short answer" questions I asked are: "What is the keyword that should go in this program in this place?" or "What will this program print?". <br /><br />The feedback screen after a quiz-type question shows a histogram which is updated on the fly, so if you want, you can steer things in the class by showing how the histogram evolves. The same works for yes/no questions. Unfortunately, the feedback screens for the short answer questions are less than ideal as it shows the list of answers as they are coming in. On this list the same answers in fact can be repeated many times.. Needless to say, this is not the best way of presenting this information. It would be much nicer, for example, if there was a histogram representing the top answers. <br /><br />I also replaced the end-of-class questionnaire of previous years with the "exit quiz" that socrative provided. Unfortunately, the questions on the exit quiz are not configurable and some of socrative's standard questions caused permanent confusion ("solve the problem on the board" -- we did not even have a board!). Also, to my surprise, the exit-quiz appeared to be less effective than the paper-based questionnaire in facilitating comments. Later, I figured that I can leave the exit-quiz open for a few days after class to collect more feedback; this helped, but unfortunately only just a little bit. Reading through the exit-quiz responses is also a bit of extra work if you have not done this before, but this was not the case for me. And I actually like to read through these extra comments; they are very useful most of the time.<br /><br />Once, I also experimented with socrative's "quiz" component. This allows teachers to compile a longer list of questions (a quiz) that the students can solve in the class (either timed or not timed). This quiz was not marked. Unfortunately, there were too many technical difficulties: socrative was overloaded (the class size was around 90). Also, socrative eliminated leading whitespace characters from the answers, which was quite unfortunate as whitespaces are crucially important in Python, the programming language that we used in this class. Thus, I decided not to use the quiz again. <br /><br />In conclusion, although the system is not perfect, it was usable and overall it helped both the students and me. I have received many comments from the student praising how socrative was useful to engage everyone, including the shy students. As this was my main motivation to try this system, I conclude that the experiment was successful. <br /><br />PS: Next semester, I tried again. Unfortunately, in this semester we were in a different classroom, where first the wifi service became unreliable and then it completely stopped working. I have no idea why. No one had. I asked the technical support people to fix it but all I got was promises. This was sad. Luckily, this was a smaller class (around 50), so we could still have some true interaction with the students.<br /><br />PPS: I have also used in the first semester a website where students could solve hundreds of little programming problems. The site provided the questions and it also have feedback, even including things like "did you forget to initialize your variable". Again, the tool had glitches (the question bank was not very well developed), but overall I think this was also a great tool.<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com2tag:blogger.com,1999:blog-1374054023917435198.post-36089190542534257002012-09-10T11:19:00.000-07:002013-09-15T20:04:41.839-07:00Student Response Systems<div dir="ltr" id="internal-source-marker_0.5499149360918648"><span style="background-color: transparent; color: black; font-family: Arial; font-size: 48px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;"></span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">There are plenty Student Response Systems (SRSs) out there. Which one to choose? This brief document summarizes what I have found on the web before the Fall of 2012.</span></div><h1 dir="ltr"><span style="background-color: transparent; color: black; font-family: Arial; font-size: 24px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">Radio communication-based systems</span></h1><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">The systems differ in a few characteristics. The first is what type of devices the students can use. Classical, </span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">iClicker</span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"> like systems require the student buy a device which communicates with a receiver that the teacher should possess. Since our school uses iClickers, let me focus on them. The overall cost of first generation iClickers is $10/device, assuming the student sells back the device to the bookstore (which they can). This is not a major cost but who likes to pay when you don’t have to? The first limitation of iClicker-like systems is that they are bound to the smart classrooms and the computers there. Thus, if you are like me and use your own computer for projection, you will need to switch between screens to show the results of a poll. This makes the use of iClicker quite cumbersome. A second major limitation of iClickers is that they are limited to multiple choice questions. In particular, they don’t allow free text entry, numerical responses, or other type of test questions (like matching). Free text and numerical responses are very useful for assessing the knowledge of students and designing meaningful multiple choice questions is hard and time-consuming. </span><br /><h1 dir="ltr"><span style="background-color: transparent; color: black; font-family: Arial; font-size: 24px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">Systems accepting text input from non-smart phones</span></h1><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">An alternative to iClickers is to use systems that uses </span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">Wifi</span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"> networks or the </span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">texting</span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"> capabilities of conventional phones with texting capabilities. Wifi capable devices include smartphones, tablets and laptops.</span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">The systems that I have found that support </span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">texting</span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"> (i.e., receiving input from non-smart phones) are </span><a href="http://www.polleverywhere.com/"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">Poll Everywhere</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">, </span><a href="http://www.lecturetools.com/"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">LectureTools</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"> and </span><a href="http://www.tophatmonocle.com/"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">Top Hat Monocle</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">. These also support input from Web-capable devices. Their pricing differs quite a bit. Poll Everywhere allows instructors to pay for the semester. The price currently is USD350. LectureTools requires you to pay for two semesters at the price of USD800. Top Hat Monocle does not allow the instructors to pay. The per student price is USD20 for a semester, or USD38 for five years. </span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">As said before, all these systems support Wifi and texting. I had a chance to test Poll Everywhere and LectureTools. I had some problems with LectureTools (importing my slides did not work). The concept of LectureTools is that you keep your slides on the web, tightly integrated with the questions. They support during the presentation annotation of slides, which is nice. However, overall LectureTools was not as smooth and easy to use as Poll Everywhere. For example, I could not figure out given the limited amount of time how the students will connect on the web to my questions. Poll Everywhere was really easy to use, on the other hand. It supports Powerpoint and Keynote. Compared to Top Hat Monocle, Poll Everywhere is not as feature rich (Top Hat Monocle has games, for example), but I was happy with the functionality Poll Everywhere provided.</span><br /><h1 dir="ltr"><span style="background-color: transparent; color: black; font-family: Arial; font-size: 24px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">Systems that use the Web-capable devices</span></h1><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">If one is contempt with </span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">Web-enabled devices</span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">, the number of available SRSs soars high. In fact, one can use any Web-based solution, many of them being free. Starting with the free options, of course, one can create a quiz in </span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">Moodle</span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">. Controlling what the students see and what they don’t is pretty cumbersome. Moodle is not very well suited for the purpose of live polling. </span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">Another method is to use </span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">Twitter</span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">. Students are pretty excited about Twitter (from the feedback I got), though I would be careful projecting everything that comes in to the screen -- some moderation might be essential to keep the class under control. Another problem with Twitter is that a tool for analyzing responses to questions would be needed (and I know of no such tool). </span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">The next option is to use </span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">Google Forms</span><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">. The very idea of Google Forms is that information submitted on the web is sent to a Google spreadsheet. Since you need the spreadsheet to get the URL of the form, start by creating a spreadsheet, then insert a form there. Once the form is created (give it a cool skin!), get back to the spreadsheet to get the URL to it. You will send this to the students (maybe compressing it using tinyurl.com, or a similar service). You can control the timing of when the form is accepting input from the spreadsheet. To support ad hoc questions, you can just create an empty form and recycle it through your presentation. If you want to use a fixed set of questions, you will need one form (and thus spreadsheet) per question that you can store on your google drive. The downside of Google Forms is that students can submit as many responses as they wish. With Google Apps, presumably there is a way around this, but this would need to be investigated. </span><br /><a href="http://www.flisti.com/"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">Flisti</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"> is an extremely simple web-based polling systems (I guess there are many other similar systems). You go to their webpage, create the poll there and give the poll’s URL to the students. You can view the results of the poll online. I think that users are tracked based on their IP addresses, so no multiple submissions are possible from the same IP address to the same poll. Only multiple choice questions (with multiple answers, possibly) are supported.</span><br /><a href="http://www.socrative.com/"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">Socrative</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"> is a fuller Web-based SRS currently under beta-testing. During the beta-testing phase, the system is free. The web-based interface is nice and sleek and it was extremely user-friendly. The teacher can control in real-time which questions are “live” in his/her “classroom”. The classroom is identified by a numbe, that the students go to. The only issue with Socrative is that every activity is limited to 50 students.</span><br /><a href="http://www.questionpress.com/"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">QuestionPress</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"> is another commercial system. The price for my class is $66. This seems to be a mature systems that I was truly impressed by. All interactions are Web-based. </span><a href="http://www.clickerschool.com/Pages/SignedOutHome.aspx"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">ClickerSchool</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"> is similar to QuestionPress, the price is $95 for my class for one semester. ClickerSchool are provided specialized smart phone apps (both for iOS and Android). </span><a href="http://eclicker.com/"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">eClicker</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">, on the other hand, requires the teacher to buy a software. All their software supports Apple products only (Mac OSX and iOS). </span><a href="http://studentresponsenetwork.com/"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">SRN</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"> (studentresponsenetwork.com) offers a campuswide license for $195. This is a client-server system that requires installing software on the teachers’, as well as the students’ computers. </span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">I could not find pricing information on the web for </span><a href="http://www.einstruction.com/products/student-response-systems/vclicker-mobile-edition"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">TurningPoint/ResponseWare</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">, which seems to be a mature product (but it cannot be “tried”). The same goes for </span><a href="http://www.einstruction.com/products/student-response-systems/vclicker-mobile-edition"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: underline; vertical-align: baseline;">vClicker</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">.</span><br /><h1 dir="ltr"><span style="background-color: transparent; color: black; font-family: Arial; font-size: 24px; font-style: normal; font-variant: normal; font-weight: bold; text-decoration: none; vertical-align: baseline;">Which type of system to choose?</span></h1><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">The question remains: Which type of system to choose? One factor to consider is how important it is to have other than multiple choice questions in class. Personally, I think that multiple choice questions exist only for historical reasons -- their pedagogical value is rather questionable. Good multiple choice questions are extremely hard and time consuming to create. If this is not convincing enough and you don’t mind switching the projector back and forth between your computer and the one in the classroom, you can stay with iClickers. </span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">In the opposite case, the next thing to consider whether you want to support phones with text input. I have just polled my students and out of the 52 responses so far, 43 can and are willing to bring their laptops to the classroom, 37 carry a smartphone, 6 a tablet and 8 carry a non-Web enabled phone (these are overlapping groups of students). Thus, the vast majority of students are able to use a system that is built around the Web. Since it will always be hard to achieve 100% coverage, one idea is to let the students pair up or form groups of three. Based on the statistics I gathered, overall, I am leaning towards that support of texting input should not be viewed as a major advantage. </span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">However, it should also be mentioned here that a potential danger of using laptops or other fancy, Web-enabled devices is that they represent potential sources of distraction. Thus, with the use of these devices, the teacher will need to face the challenge of competing for the students’ attention with the social networks, email, and ultimately the whole Web.</span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">The next factor consider is the ease of use of the SRS. The teacher may need to create a significant number of questions for every class. Integration with Powerpoint and Keynote may be a plus, but switching between a browser and a presentation software looks easy enough.</span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">Since I will use the chosen SRS only for formative assessment and not for grading (the present common sense is that this would be a bad idea, not talking about that this would indeed require full coverage of the whole class), I don’t care about whether the SRS supports automated grading and keeps the identities. </span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">Based on these considerations, I will probably go with Socrative.</span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;"></span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">References</span><br /><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">I have created a google spreadsheet for comparing the systems listed here, in addition to a few more. The spreadsheet can be found </span><a href="https://docs.google.com/spreadsheet/ccc?key=0AtY9i8UP2o_JdFlyZ3M3bERpMzI3c2xmdTFEaHJGeEE"><span style="background-color: transparent; color: #1155cc; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: underline; vertical-align: baseline;">here</span></a><span style="background-color: transparent; color: black; font-family: Arial; font-size: 15px; font-style: normal; font-variant: normal; font-weight: normal; text-decoration: none; vertical-align: baseline;">.</span> The spreadsheet links a few other sources that I have used during my research.<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com3tag:blogger.com,1999:blog-1374054023917435198.post-12996089465896149372012-04-14T19:53:00.000-07:002012-04-14T20:52:02.534-07:00Approximating inverse covariance matricesPhew, the last time I have posted an entry to my blog was a loong time ago.. Not that there was nothing interesting to blog about, just I always delayed things. (Btw, google changed the template which eliminated the rendering of the latex formulae, not happy.. Luckily, I could change back the template..) Now, as the actual contents:<br /><br />I have just read the PAMI paper "<a href="http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=5601741">Accuracy of Pseudo-Inverse Covariance Learning-A Random Matrix Theory Analysis</a>" by D Hoyle (IEEE T. PAMI, 2011 vol. 33 (7) pp. 1470--1481). <br /><br />The paper is about pseudo-inverse covariance matrices and their analysis based on random matrix theory analysis and I can say I enjoyed this paper quite a lot.<br /><br />In short, the author's point is this:<br />Let $d,n>0$ be integers. Let $\hat{C}$ be the sample covariance matrix of some iid data $X_1,\ldots,X_n\in \mathbb{R}^d$ based on $n$ datapoints and let $C$ be the population covariance matrix (i.e., $\hat{C}=\mathbb{E}[X_1 X_1^\top]$). Assume that $d,n\rightarrow \infty$ while $\alpha = n/d$ is kept fixed. Assume some form of "consistency" so that the eigenvalues of $C$ tend to some distribution over $[0,\infty)$. Denote by $L_{n,d} = d^{-1} \mathbb{E}[ \| C^{-1} - \hat{C}^{\dagger} \|_F^2]$ the reconstruction error of the inverse covariance matrix when one uses the pseudo-inverse.<br /><br />Then, $f(\alpha) := \lim_{d,n\rightarrow\infty} L_{n,d}$ will be well-defined (often). The main things is that $f$ becomes unbounded as $\alpha\rightarrow 1$ (also $\alpha\rightarrow 0$, but this is not that interesting). In fact, for $C=I$, there is an exact formula for $\alpha \rightarrow 1$:<br /><br />$f(\alpha) = \Theta(\alpha^3/(1-\alpha)^3)$. <br />Here is a figure that shows $f$ (on the figure $p$ denotes the dimension $d$).<br /><br /><div class="separator" style="clear: both; text-align: center;"><a href="http://4.bp.blogspot.com/-SY6ucmFwgbo/T4pFpulXU1I/AAAAAAAAE-k/gRL_eDCi8sg/s1600/inverse-covariance.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="222" src="http://4.bp.blogspot.com/-SY6ucmFwgbo/T4pFpulXU1I/AAAAAAAAE-k/gRL_eDCi8sg/s320/inverse-covariance.png" width="320" /></a></div>Nice.<br /><br />The author calls this the "peaking phenomenon": The worst case, from the point of view estimating $C^{-1}$, is when we have as many data points as is the number of dimensions (assuming full rank $C$, or this won't make sense because otherwise you could just add dummy dimensions to improve $L_{n,d}$). The explanation is that $L_{n,d}$ is just sensitive to how small the smallest positive eigenvalue of $\hat{C}$ is (this can be shown), and this smallest positive eigenvalue will become extremely small as $\alpha\rightarrow 1$ (and it does not matter whether $\alpha \rightarrow 1-$ or $\alpha\rightarrow 1+$).<br /><br />Now notice that having $\alpha=0.5$ is much better than having $\alpha=1$ (for large $n,d$). Thus, there are obvious ways of improving the sample covariance estimate!<br /><br />In fact, the author then suggests that in practice, assuming $n \ge d$, one should use bagging, while for $n\le d$ one should use random projections. Then he shows experimentally that this improves things though unfortunately this is shown for Fisher discriminants that use such inverses and the demonstration for $L_{n,d}$ is lacking. He also notes that the "peaking phenomenon" can also be dealt with by other means, such as regularization where we are referred to Bickel and Levina (AoS, 2008).<br /><br />Anyhow, one clear conclusion of this paper is that the pseudo-inverse must be abandoned when it comes to approximate the inverse covariance matrix. Why is this interesting? The issue of how well $C^{-1}$ can be estimated comes up naturally in regression, or even in reinforcement learning when using the projected Bellman error. And asymptotics is just a cool tool to study how things behave..<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com4tag:blogger.com,1999:blog-1374054023917435198.post-66006611002504650292011-05-04T08:43:00.007-07:002011-05-04T11:57:31.040-07:00Brains, Minds and MachinesFormer UofA student,<a href="http://people.csail.mit.edu/agf/Homepage/Welcome.html"> Alborz</a>, shared a <a href="http://amps-web.amps.ms.mit.edu/public/150th/may3/">link</a> to a video recording of a recent MIT150 symposium on Brains, Minds and Machines on facebook. I watched the video yesterday (guess what, I need to mark 40 something finals, hehe:)).<br />I wrote a comment back to Alborz on facebook and then I thought, why not make this a blog post? So, here it goes, edited, expanded. Warning: Spoilers ahead and the summary will be biased. Anyhow..<br /><br />The title of the panel was: "<span style="font-style: italic;">The Golden Age — A Look at the Original Roots of Artificial Intelligence, Cognitive Science, and Neuroscience</span> " and the panelist were <strong>Emilio Bizzi, </strong><strong>Sydney Brenner, </strong><strong>Noam Chomsky, </strong><strong>Marvin Minsky, </strong><strong>Barbara H. Partee and </strong><strong>Patrick H. Winston. </strong>The panel was moderated by <strong>Steven Pinker </strong>who started with a 20-30 minute introduction. Once done with this each of the panelist delivered a little speech and at the end there were like two questions asked by Pinker.<br /><br />My heroes in the panel were <span style="font-weight: bold;">Minsky</span> and <span style="font-weight: bold;">Winston</span>. They rocked! Minsky almost fell asleep during his talk, but he was well aware of this and I loved him. He told a story about Asimov not wanting to come to his lab to see the real robots (he did not want to get disappointed) and about von Neumann who said that he does not know if Minsky's thesis could qualify as a thesis on mathematics (they were both at Princeton in the math department), but soon it will be. I really enjoyed this part. Winston acted a bit like a comedian. I did not mind this either. One thing that Minsky and Winston both said is that the mistakes happened when AI became successful and everyone from that on seemed to forget the science part of AI. But they did not say much about how we can get back on the track (except that we should try). Winston blamed the short-sightedness of funding agencies and who would disagree.<br /><br /><span style="font-weight: bold;">Chomsky</span> made some interesting claims. He claimed that language is designed for thoughts and not for communication. This was a pretty interesting claim. He claimed other things supporting his idea about innate, universal grammars, but I did not know how much credit to give to those as I have many linguist friends who strongly disagree. Things became interesting when answering a question at the end, he dumped the whole of machine learning. He talked about "a novel scientific criterion" that was never heard of before referring to being able to predict "on unanalyzed data". He said that "of course" with enough data you will do better, but it seemed that he thinks that the evaluation criterion is already ridiculous. He also said that a little statistics does not hurt, but he still seems that the big deal is the engineering part. (He did not say with these words, but this is what I got from what he said).<br /><br /><strong>Sydney Brenner</strong> (pioneer in genetics and molecular biology) was puzzled about why he was invited, though he had some good stories. I liked when he said that in 50 years people will not understand why everyone talked about consciousness 50 years ago.<br /><br /><strong>Emilio Bizzi </strong>(a big shot in neuroscience, in particular in motor control) talked about modularity, "dimension reduction" and generalization, and he looked like a fine Italian gentleman, but I have to confess that I don't remember anything else, though this could have been because it was late.<br /><br /><span style="font-weight: bold;">Barbara H. Partee</span> read a script. In the first 5 minutes or so she was mostly praising Chomsky. Then she started to talk about her own work, which was foundational in semantics. She talked about how semantics is the real thing. By semantics she means formal semantics (like in logics). While in general I am fond of this work, I am not sure if anything like this is going on in our brain and if sentences really do have a meaning in a formal sense. It seems to me that the fact that sentences can have a formal meaning is more likely an illusion, a post-hoc thought than the real thing. Ad it is unclear if bringing in formal logics is going to bring us anywhere. Unfortunately, there was no discussion of this at all. At one point she made the remark that "search engines do not use semantics", but then left us in vain about how we could do any better. Oh, well..<br /><br />In summary, an impressive set of people, some nice stories, but little cutting edge science. Loads of romanticism about the 50s and 60s and no advice for the young generation. The title tells it all. It was still nice to see these people.<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com0tag:blogger.com,1999:blog-1374054023917435198.post-69522372984423856612011-04-13T20:09:00.006-07:002011-05-17T09:54:09.261-07:00Useful latex/svn tools (merge, clean, svn, diff)This blog is about some tools that I have developed (and yet another one that I have downloaded) which help me to streamline my latex work cycle. I make the tools available, hoping that other people will find them useful. However, they are admittedly limited (more about this) and as usual for free stuff they come with zero guarantee. Use them at your own risk.<br /><br />The first little tool is for creating a cleaned up file before submitting it to a publisher who asks for source files. I call it <span style="font-style: italic;">ltxclean.pl</span>, it is developed in Perl. It can be downloaded from <a href="http://www.ualberta.ca/%7Eszepesva/sourcecode/ltxclean.pl">here</a>.<br />The functionality is<br />(1) to remove latex comments<br />(2) to remove \todo{} commands<br />(3) to merge files included from a main file into the main file<br />(4) to merge the bbl file into the same main file<br /><br />If you make the tool executable (chmod a+x ltxclean.pl), you can use it like this:<br /><br />$ ltxclean.pl main.tex > cleaned.tex<br /><br /><span style="font-weight: bold;">How does this work?</span><br /><br />The tool reads in the source tex file, processes it line by line and produces some output to the standard output stream, which you can redirect (as shown above) to a file.<br />Thus, whatever the tool does is limited to the individual lines. This is a limitation, but this made it possible for me to write this tool in probably less time than I spend on writing about it now.<br />There are other limitations, see below. Now, how do we know that this worked? The advice is to run <span style="font-family:courier new;">latex+dvips</span> and then <span style="font-family:courier new;">diff original.ps new.ps</span> to see if there is any significant change. On the files I have tried, the only difference was the filename and the date.<br /><br /><span style="font-weight: bold;">Why this functionality and the glory details</span><br /><br />As it happens, <span style="font-style: italic;">removing the comments</span> before you submit a source file is <span style="font-style: italic;">crucial</span>. Not long ago, it happened to me that I have submitted a source to a publisher and I did not care about removing the comments. At the publisher, they loaded the file into a program, which wrapped the long lines, including the ones with comments! This created a lot of garbage in the middle of the text. We were pressed against time, though I could not check the text in details. The result: The text was printed with a lot of garbage! Too bad!! A painful experience for me.. I will never again submit source files with the comments kept in the file! Now, the above utility is meant to handle comments correctly. It pays attention to not to create empty lines (and thus new lines) inadvertently, not to remove end-of-line comments etc.<br /><br />The <span style="font-style: italic;">\todo{}</span> commands belong to the same category: They are better removed before submitting the file. For my todos, I use the <a href="http://www.ctan.org/tex-archive/macros/latex/contrib/todonotes/">todonotes</a> package, which puts todo notes on the margin (or within the text). This package supplies the \todo[XX]{ZZZ} command, where [XX] is optional. The above little script removes such todo commands, but only if they span a single line only. For now, you would need to remove multi-line todos by hand.<br /><br />Another service of this little tool is to <span style="font-style: italic;">merge multiple files</span> into a single one. Oftentimes, we use the latex command <span style="font-style: italic;">\input</span> to break a large source file into multiple files. However, publishers typically want just one file. So this tool reads in the main file and the recursively, whenever it sees \input{FILE} in the source, it reads in the corresponding file and processes it before it continues with the current file (just like latex would work).<br /><br />Finally, if the tool finds a \bibliography{...} command, it will take that out and open the .bbl file sharing the same base name as the input to the tool. Thus, if the tool was called on the file main.tex, when seeing a bibliography command, the tool will attempt to open main.bbl and include it in place of the \bibliography command. (If you use <a href="http://www.tug.org/applications/hyperref/manual.html">hyperref</a>, turn off <span class="ec-lmtt-10">pagebackref</span>, otherwise this functionality will not work.)<br /><br /><span style="font-weight: bold;">Managing revisions with svn</span><br /><br />Two other small utilities that I make available are <a href="http://www.ualberta.ca/%7Eszepesva/sourcecode/svnprevdiff">svnprevdiff</a> and <a href="http://www.ualberta.ca/%7Eszepesva/sourcecode/svnreviewchanges">svnreviewchanges</a>.<br />The purpose of these scripts is to help one review changes to files which are under <a href="http://en.wikipedia.org/wiki/Apache_Subversion">svn</a> control.<br />There is a third script, <a href="http://www.ualberta.ca/%7Eszepesva/sourcecode/diffmerge">diffmerge</a>, called by the above two scripts. This script takes two file arguments and loads these into the program <a href="http://www.sourcegear.com/diffmerge/">DiffMerge</a> which allows you to visually inspect the differences between the two files and make changes to the second one loaded. On a different platform/installation, or if you want to use a different tool for comparing/merging files.<br /><br />The utility <span style="font-style: italic;">svnreviewchanges</span> takes a file as an argument, compares it to its base version stored on your disk and opens up the two versions for comparison using diffmerge. The purpose is to allow one to quickly review how a file was changed before submitting a file to the svn server (so that you can write meaningful comments in the commit message).<br /><br />The utility <span style="font-style: italic;">svnprevdiff</span> takes a filename as an argument, compares it to its <span style="font-style: italic;">previous</span> version <span style="font-style: italic;">stored on the svn server</span> and then opens up the two versions using diffmerge. The purpose of this is to check the changes implemented by your pals <span style="font-style: italic;">after</span> an update. A future version will take an optional argument which when present will be interpreted as a revision number. Maybe.<br /><br /><span style="font-weight: bold;">Advice on using latex when working in a team: Break long lines<br /></span><br />A small, but useful thing is to put every sentence on its own line and generally avoiding long lines (even when writing equations). The reason is that this will make the job of diff much easier. And believe me, diffing is something people will end up doing for good or bad (mostly good) when they are on a team.<br /><br />Some of my friends, like <a href="http://www.szit.bme.hu/%7Eantos/">Antoska</a> would recommend breaking up the individual sentences into multiple lines. You can do this, but if you overdo it, you will find yourself fiddling way too much with what goes into which line.<br /><br />Finally, a tool which does this, written by <a href="http://www.math.ntnu.no/%7Estacey/">Andrew Stacey</a>, is <a href="http://www.math.ntnu.no/%7Estacey/HowDidIDoThat/LaTeX/fmtlatex">fmtlatex.pl</a>.<br />This is also in Perl and its documentation will be written on the screen if you use <code>perldoc fmtlatex. </code>I still have to try this.<br /><br /><code><br /><br /></code><div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com11tag:blogger.com,1999:blog-1374054023917435198.post-60060394754701165612011-04-13T19:22:00.005-07:002011-04-13T19:53:17.621-07:00I can run Matlab on my Mac again!After much struggling today I managed to make Matlab run again my Mac.<br />The major problem was that Matlab complained about that I have the wrong version of X11 installed on my system and it won't start. As I have finished teaching today for the semester, I thought that I am going to celebrate this by resolving this issue which I was struggling with for a year or so by now. On the internet you will see a lot of advice on what to do, and as they say, the truth is indeed out there, however, it is not so easy to find. In a nutshell what seems to happen is this:<br /><br />Why Matlab does not start when other applications do start (say, Gimp, Gnuplot using X11, etc.).<br />Matlab seems to make the assumption that the X11 libraries are located at /usr/X11/lib and it <span style="font-style: italic;">sticks to this assumption no matter how your system is configured</span>. I use XQuartz and macports' X11 and they put stuff elsewhere. I had some legacy code sitting in /usr/X11/, which I did not use. It was a remainder of some version of X11 that I used probably 2 or 3 laptops ago. Matlab reported that the lib was found, but the "architecture was wrong". The error message had something like:<br /><br />..<span style="font-size:85%;"><span style="font-family: courier new;"> </span><span style="font-family: courier new;"> Did find: </span></span><span style="font-size:85%;"><span style="font-family: courier new;">/usr/X11R6/lib/libXext.6.dylib: mach-o, but wrong architecture</span></span>..<br /><br />Anyhow, here is one solution.<br />You have to arrange that /usr/X11 points to a directory that has a working X11 copy.<br />It is probably a good idea to first clean up the old X11 installation. You can do this by following the advice on the <a href="http://xquartz.macosforge.org/trac/wiki/X11-UsersFAQ">XQuartz FAQ</a> page by issuing the following commands in the terminal:<br /><pre class="wiki"><span style="font-family: courier new;">sudo rm -rf /usr/X11* /System/Library/Launch*/org.x.* /Applications/Utilities/X11.app /etc/*paths.d/X11 sudo pkgutil --forget com.apple.pkg.X11DocumentationLeo sudo pkgutil --forget com.apple.pkg.X11User sudo pkgutil --forget com.apple.pkg.X11SDKLeo sudo pkgutil --forget org.x.X11.pkg</span><br /></pre>Then I have reinstalled the latest XQuartz copy (not all these steps might be necessary, but in order to stay on the safe side, I will describe everything I did).<br />I also have <a href="http://www.macports.org/">macports</a> and xorg-libX11, xorg-libXp, xorg-server seems necessary for the following steps to succeed (but possibly other xorg-* ports are also needed). I am guessing that XQuartz does not install all the libraries, but after installing enough xorg-* ports through macports, all the libraries will be installed which are used by Matlab.<br /><br />Now, my X11 is located at /opt/X11 and some additional libs are found at /opt/local/lib.<br /><br />So I created a bunch of symbolic links:<br /><br /><span style="font-size:85%;"><span style="font-family: courier new;">sudo ln -s /opt/X11 /usr/X11</span><br /><span style="font-family: courier new;">for i in /opt/local/lib/libX* ; do sudo ln -s $i /usr/X11/lib; done</span><br /></span><br />The first line creates a symbolic link to /opt/X11, while the second is necessary because of the additional libX* libraries which, for some reason, macports puts into /opt/local/lib instead of puttting it into /opt/X11/lib. Initially I did not know that I need these libs, and then Matlab complained that it did not find the image for some lib (it was /usr/X11/lib/libXp.6.dylib).<br /><br />Anyhow, I am really happy that this worked!<br />I hope people who will have the same trouble will find my post useful.<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com6tag:blogger.com,1999:blog-1374054023917435198.post-71279226860465498792009-11-17T08:55:00.008-07:002009-11-18T09:12:11.397-07:00Djvu vs. PdfLong blog again, so here is the <span style="font-weight: bold;">executive summary</span>: Djvu files are typically smaller than Pdf files. Why? Can we further compress pdf files? Yes, we can, but the current best solution has limitations. And you can forget all "advanced" commercial solutions. They are not as good as a free solution.<br /><br /><span style="font-weight: bold;">Introduction</span><br /><br /><a href="http://en.wikipedia.org/wiki/DjVu">DJVU</a> is a proprietary file format by LizardTech. Incidentally, it was invented by some machine learning researchers, <a href="http://yann.lecun.com/" title="Yann LeCun">Yann LeCun</a>, <a href="http://leon.bottou.org/" title="Léon Bottou">Léon Bottou</a>, <a href="http://www2.research.att.com/%7Ehaffner/" class="new" title="Patrick Haffner (page does not exist)">Patrick Haffner</a> and the image compression researcher <a href="http://www.informatik.uni-trier.de/%7Eley/db/indices/a-tree/h/Howard:Paul_G=.html" class="new" title="Paul G. Howard (page does not exist)">Paul G. Howard</a> at AT&T back in 1996. The <a href="http://djvu.sourceforge.net/">DJVULibre</a> library provides a free implementation, but is GPLd and hence is not suitable for certain commercial softwares, like <a href="http://mekentosj.com/papers/">Papers</a>, which I am using to organize my electronic paper collection. Hence, Papers, might not support djvu in the near future (the authors of Papers do not want to make it free, and, well, this is their software, their call).<br />Djvu files can converted to Pdf files using <a href="http://djvu.sourceforge.net/doc/man/ddjvu.html">ddjvu</a>, a command line tool which is part of DJVULibre (<a href="http://freshmeat.net/projects/djvu2pdf/">djvu2pdf</a> is a script that calls this tool). Djvu can also be converted into PS files using <a href="http://djvu.sourceforge.net/doc/man/djvups.html">djvups</a> (then use ps2pdf). However, all these leave us with pretty big files compared to the originals and, on the top of it, if there was an OCR layer in the Djvu file, it gets lost, but this is another story. How much bigger? Here is an illustration:<br /><br />Original djvu file: 9.9MB<br />djvu2pdf file: 427.6MB(!)<br />djvu2ps file: 1.0GB<br />djvu2ps, ps2pdf file: 162.6MB<br /><br />Note that I have turned on compression in the conversion process (-quality=50). (The quality degradation was not really noticeable at this level.) So, at best, I got more than 16 times the original file size. Going mad about it, I started to search the internet for better solutions. I have spent almost a day on this (don't do this, especially if you are a student!)..<br /><br /><span style="font-weight: bold;">JBig2 and the tale of commercial solutions</span><br /><br />First, I figured, the difference is that these use general image compression techniques (like jpeg), while djvu is specialized to text and black&white images. Thus, for example, it can recognize if the same character appears multiple times on the page, store a template and a reference to the template. This is clever. I then figured that PDF files support the so-called <a href="http://en.wikipedia.org/wiki/JBIG2">jbig2</a> encoding standard, which is built around this idea. Hence, the quest for software that would support encoding a document using a jbig2 encoder and put the result into a pdf format. The easiest would be, if such a software just existed out there. A few commercial packages indeed mention jbig2. I felt lucky (especially, seeing that there are a few cheap ones). So, I started to download trial versions. Here are the results:<br /><br /><a href="http://www.imagepdf.com/">PDFJB2</a>: 34.1MB<br /><a href="http://www.cvisiontech.com/products/general/pdfcompressor.html">CVision PdfCompressor</a>: 48MB<br />CVision PdfCompressor with OCR: 49MB<br /><a href="http://www.a-pdf.com/">A-PDF</a>: 106.8MB<br />A-PDF + PDFCompress: 106.8MB<br />djvu2pdf + <a href="http://www.bureausoft.com/products.html#PDF%20Compress">PDFCompress</a>: conversion failed<br /><br />Hmm, interesting. 34MB is much better than 160MB, but it is still a long way from 9.9MB. (After a superficial look at the resulting files I concluded that only the A-PDF compressed file lost quality. What happened with this file is that on some page in some line containing a mathematical formula, the top of the letters got chopped.)<br /><br /><span style="font-weight: bold;">Free, open source solutions</span><br /><br />Becoming desperate, I continued hunting for better solutions. Searching around, I have found <a href="http://www.lowagie.com/iText/">iText</a>, which is an open source, free Java library supporting all kinds of manipulations of Pdf files. I have figured that it "uses" Jbig2, but it was not clear if it uses it for compression or just knows how to handle the encoding. So, here I go, I wrote a java program opening a pdf file and then writing it out in "compressed" mode. Hmm, this few lines of coding allowed me to create a file of size 26MB, smaller than what I could ever get previously. Exciting! Unfortunately, opening the file revealed the `secret': Quality was gone. The file looked to be seriously downsampled (i.e., the resolution was decreased). Not good.<br /><br />Then I have found <a href="http://code.google.com/p/pdfsizeopt" style="text-decoration: none; color: rgb(0, 0, 0);">pdfsizeopt</a> on google code, which aims exactly at compressing the size of pdf files! The Holy Grail? Well, installing pdfsizeopt on my mac was far from easy (I use a Mac, which also runs Windows; quite handy as some of the above software runs only under Windows..). However, finally, I was able to run pdfsizeopt. Unfortunately, it seems to crash, without even looking at my pdf file (I hope the bug will be corrected soon and then I can report results using it). Along the way, I had to install <a href="http://github.com/agl/jbig2enc">jbig2enc</a>. For this, I just had to install <a href="http://www.leptonica.com/">leptonica</a> (version 1.62, not the latest one), which is really the part that is doing the image processing part of the process. JBig2Enc expects a tif file and produces "pdf" ready output (every page is put in a separate file), which can be concatenated into a single pdf file by a python script provided. Having jbig2enc on my system, I gave it a shot. I first used ddjvu to transform the input to a tif file (using the command line option, "-quality=75", resulting in a file of size 1GB). Then I used the jbig2 encoded with the command line arguments "-p -s". The result is this:<br /><br />jbig2enc: 3.8MB<br /><br />Wow!! Opening the file revealed a dirty little secret: Color images are gone, as well as the quality of some halftoned gray-scale images got degraded. However, line drawings were kept nicely and, in general, the quality was good (comparable to the original djvu file). Conversion to tif took 5 minutes, conversion from tif to jbig2 took ca. 4 minutes, altogether making the whole process take close to 10 minutes. (Other solutions were not faster at all either. And the tests were run on a resourceful MacBook Pro.)<br /><br /><span style="font-weight: bold;">Conclusions</span><br /><br />jbig2enc seems to work, but you will lose colors. If you are happy with this, jbig2enc is the solution, though the process should be streamlined a bit (a small script good do this). Oh yes, I did not mention that these processes are not fast. I did not attempt to measure the speed, but conversion takes a lot of time. Jbig2Enc is maybe on the faster end of the spectrum.<br /><br /><span style="font-weight: bold;">Future work</span><br /><ol><li>pdfsizeopt is a good idea. It should be made work.<br /></li><li>It would be nice to create a jbig2enc wrapper </li><li>ddjvu is open source: Maybe it can be rewritten to support jbig2 directly. The added benefit could be that one could also keep the OCR layer in the original djvu file if one existed</li><li>Along the way, I have found a cool google code project, <a href="http://code.google.com/p/tesseract-ocr/">Tesseract</a>, which is an open source OCR engine. How cool would it be if we had an OCR engine that helps the compression algorithm and eventually also puts an OCR layer on the top of documents which lack text information (think of scanned documents, or documents converted from an old postscript file). Currently, I am using Nuance's <a href="http://www.nuance.com/imaging/products/pdfconverter.asp">Pdf Converter Professional</a> (yes, I paid for it..), which I am generally very satisfied with apart from its speed. However, this could be the subject of another post.</li></ol>PS: I have tested the capabilities of Nuance's Pdf Converter Professional and Abbyy's in terms of their compression capabilities:<br />Nuance: 132MB<br />Abbyy: 129MB<br />Yes, I tried their advance "MRC" compression, in Nuance I have explicitly selected jbig2. No luck.<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com5tag:blogger.com,1999:blog-1374054023917435198.post-90341176169337219452009-11-14T12:47:00.006-07:002009-11-14T16:46:49.769-07:00Keynote vs. Powerpoint vs. BeamerA few days ago I decided to give <a href="http://www.apple.com/iwork/keynote/">Keynote</a>, Apple's presentation software, a try (part of iWork '09). Beforehand I used MS Powerpoint 2003, Impress from <a href="http://www.neooffice.org/neojava/en/index.php">NeoOffice 3.0</a> (OpenOffice's native Mac version) and LaTeX with <a href="http://latex-beamer.sourceforge.net/">beamer</a>. Here is a comparison of the ups and downs of these software, mainly to remind myself when I will reconsider my choice in half a year and also to help people decide what to use for their presentation. Comments, suggestions, critics are absolutely welcome, as usual. Btw, while preparing this note I have learned that <a href="http://go-oo.org/">go-oo.org</a> has a native Mac Aqua version of OpenOffice. Maybe I will try it some day and update the post. It would also be good to include a recent version of Powerpoint in the comparison.<br /><h3>Stability</h3><ul><li><span style="font-weight: bold;">Keynote: Excellent </span><br />After a few days of usage, so take this statement with a grain of salt..</li><li><span style="font-weight: bold;">MS Powerpoint 2003: Excellent</span> </li><li><span style="font-weight: bold;">Impress: Poor</span><br />Save your work very often</li><li style="font-weight: bold;">Beamer: Excellent</li></ul><h3>Creating visually appealing slides, graphics on slides<br /></h3><ul><li><span style="font-weight: bold;">Keynote: Excellent</span><br />Positioning rulers help a lot. The process is really smooth. Keynote forces you to use less text. Built in templates are professional looking. Adding presentation graphics (tables, basic charts) is very easy. Cooler (technical drawing) better done with <a href="http://www.omnigroup.com/applications/OmniGraffle/">OmniGraffle</a>. You can also easily animate the graphics, tables. Overall, very impressive.<br /></li><li><span style="font-weight: bold;">MS Powerpoint 2003: Good</span><br />Aligning to other objects is more cumbersome than in Keynote. The quality of fonts, color palettes, templates is not as good in Keynote.<br /></li><li><span style="font-weight: bold;">Impress: Good</span><br />Same as MS Powerpoint, maybe somewhat below (but the difference is not big).<br /></li><li><span style="font-weight: bold;">Beamer: Poor</span><br />The fonts and styles (templates) are great. However, creating slides with lively graphic is a nightmare (due to the lack of a GUI): You will end up with a few standard layouts, you will in general not use graphics, let alone animated graphics (or you will spend days on creating your slides). Also, departing from the styles is difficult and I am just bored of some of these styles that everyone seems to use.<br /></li></ul><h3>LaTeX (math) support</h3><ul><li><span style="font-weight: bold;">Keynote: Poor</span><br />Supported through <a href="http://pierre.chachatelier.fr/programmation/latexit_en.php">LatexIt</a> (free), but overall a cumbersome process. Details below.</li><li><span style="font-weight: bold;">MS Powerpoint 2003: </span><span style="font-weight: bold;">Medium</span><br />Supported through <a href="http://texpoint.necula.org/buy.html">TexPoint</a> (commercial, USD30) process is roughly same as with LatexIt and Keynote, slightly better integration.<br /></li><li><span style="font-weight: bold;">Impress: Medium</span><br />Supported through <a href="http://ooolatex.sourceforge.net/">OOoLatex</a> (free), same as MSPowerPoint + TexPoint, the integration is slightly better.<br /></li><li style="font-weight: bold;">Beamer: Excellent<br /><span style="font-weight: normal;">Beamer is built for this!</span><br /></li></ul><h3>Animations</h3><ul><li><span style="font-weight: bold;">Keynote: Near perfect</span><br />Magic slide transition helps a lot with continuity across slides. What does this do? If you have the same object on two consecutive slides, Keynote will create an animation, keeping the object on screen and flying it to its new position. Works with multiple objects, too. I have found this very helpful for presenting a multi-slide argument. In general, Keynote animations are slick, polished, the flexibility is great. I lack some features of Beamer, such as animated highlighting, in-place replacement of some text (these can all be simulated with the existing tools, but with difficulty only).<br /></li><li><span style="font-weight: bold;">MS Powerpoint 2003: Basic</span><br />I miss Keynote's magic transitions. In general, Keynote is richer in animations. Again, some features of Beamer would be nice to have.</li><li><span style="font-weight: bold;">Impress: Weak</span><br />Impress is inferior in terms of its animation caps to MS Powerpoint </li><li><span style="font-weight: bold;">Beamer: Good</span><br />If only someone added support for magic transitions between slides. Some other cool effects would also come handy.</li></ul><h3>Dual screen presentation support</h3>The idea is to show notes, time left in addition to the current and next slide on your screen, while showing the current slide on the big screen.<br /><ul><li><span style="font-weight: bold;">Keynote: Excellent</span><br />Keynote supports double screen presentations natively. If you need to swap displays, go on the notes screen in the options menu. This will be on the big screen, obviously, if you need to swap the the screens.<br /></li><li><span style="font-weight: bold;">MS Powerpoint 2003: Not available</span><br />I have no experience with this feature of MS Powerpoint. Maybe you can use and add-on or something, but the basic software does not support it. I am pretty sure newer versions of Powerpoint must support this.<br /></li><li><span style="font-weight: bold;">Impress: Excellent(?)</span><br />The "<a href="http://extensions.services.openoffice.org/project/presenter-screen">Sun Presenter Console</a>" extension supposedly supports dual screen presentations just like Keynote, but I have never had the chance to test it. Hence, the question mark. Some posts on the internet indicate that the extension might leak memory.<br /></li><li><span style="font-weight: bold;">Beamer: Basic support</span><br />Use <a href="http://code.google.com/p/splitshow/">Splitshow</a> for this purpose. However, as far as I know, you cannot show the current time or the time remaining on the notes screen.<br /></li></ul><h3>Interoparability</h3>I want to put my presentations on the web so that people can look at them no matter what (major) operating system they use, without loosing animations or any other features. Another desired feature is the ability to create a compact, printable version of the slides: That is, if you have animations spanning multiple slides, somehow they should get handled intelligently. There is a tradeoff here: The more animation rich your slides are, the more bloated/complicated your printout will be.<br /><ul><li><span style="font-weight: bold;">Keynote: OK</span><br />Proprietary file format. This is my biggest complaint. A keynote presentation is a keynote presentation. Apple likes to lock you in. Export to PDF and PPT works relatively well, but will lose some features of the presentation, like the cool animations. Exporting to PDF without animations to create printable versions seems to work well.<br /></li><li><span style="font-weight: bold;">MS Powerpoint 2003: Good</span><br />Free powerpoint viewers exist that can play any PPT file. Export to PDF will again lose some features.</li><li><span style="font-weight: bold;">Impress: Good</span><br />Same as powerpoint.<br /> </li><li><span style="font-weight: bold;">Beamer: Excellent</span><br />Produces PDF outputs: The presentations can be viewed on any computer! Also, the source is later, beamer is available on all systems. Add [handout] to the style and beamer will create an animation free version of your slides that works almost all the cases.<br /></li></ul> <h3>More about using formulae in Keynote (and why it sucks)</h3>I used LatexIt which produces a PDF that can be embedded into the presentation. Style is not matched automatically. The PDF contains the latex source for the formulae, copy paste it back to LatexIt to edit it. When done with the edit, you need to drag and drop the formula back into Keynote. This sucks, since you need to delete the original that you have edited, reposition the new formula and reapply animations if you had any. Horrible.<br /><br />Another issue is that the source saved with the formula by default does not have the preamble, thus using a command set specific to a presentation is difficult to achieve (you have to set this up manually). Another major headache is that you will not be able to use inline formula (a text is either in LaTeX, or in Keynote, the fonts in general do not match and mix well, alignment is a nightmare), nor will you be able to animate easily formulae (e.g., displaying a multiline formula line by line requires you to split the formula into multiple PDFs and use Keynote animations to show them one by one; this is problematic because formula alignment by hand is time consuming).<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com6tag:blogger.com,1999:blog-1374054023917435198.post-474168759213646352009-10-31T11:01:00.002-07:002009-10-31T11:03:05.244-07:00OptogeneticsThis is a little deviation from the usual topic.<br />Scientist are able to genetically modify neurons that respond to light. They are in fact able to do this in a targeted manner. A patient would then have some LEDs inside his skull, emitting some light. In response the selected neurons start to fire. They demonstrated the technology by making mice run counterclockwise when they turn on the light. This is input to the brain. Earlier, it was demonstrated that neurons can be genetically modified to emit light when they are firing. Are we heading towards rewiring the brain and turning it into a light computer?<br />The motivation for the research is to cure diseases like Parkinson's disease, when the patient has all the circuity and muscles but is just unable to make the movements. In fact, the researchers are already testing this technology on primates. Source: Wired Nov. 2009, "Powered by Photons" pp. 109--113. The wikipedia entry for optogenetics is <a href="http://en.wikipedia.org/wiki/Optogenetics">here</a>.<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com1tag:blogger.com,1999:blog-1374054023917435198.post-78569182794165764132009-10-25T19:28:00.006-07:002009-10-25T21:01:41.627-07:00Pitfalls of optimality in statisticsI was reading a little bit about robust statistics, as we are in the process of putting together a paper about entropy estimation where robustness comes up as an issue. While searching on the net for the best material to understand this topic (I am thinking about posting another article about what I have found), I have bumped into a nice paper (downloadable from <a href="http://projecteuclid.org/DPubS?service=UI&version=1.0&verb=Display&handle=euclid.lnms/1249305323">here</a>) by Peter J. Huber, one of the main figures in robust statistics, where he talks about a bunch of <span style="font-style: italic;">pitfalls</span> around pursuing optimality in statistics. Huber writes eloquently -- he gives plenty of examples, motivates definitions. He is just great. I can only recommend this paper or his <a href="http://www.amazon.ca/gp/product/0470129905/">book</a>. Now, what are the pitfalls he writes about? He distinguishes 4 types with the following syndromes:<br /><ul><li><span style="font-style: italic;">The fuzzy concepts syndrome: sloppy translation of concepts into mathematics</span>. Think about uniform vs. non-uniform convergence (sloppy asymptotics). In statistics a concrete example is the concept of efficiency which is defined in a non-uniform manner with respect to the estimable parameters, which allows for (weird) "super-efficient" estimators that pay special attention to some distinguished element of the parameter-space.</li><li><span style="font-style: italic;">The straitjacket syndrome</span>:<span style="font-style: italic;"> the use of overly restrictive side conditions</span>, such as requiring that an estimator is unbiased or equivariant (equivariant estimates in high dimensions are inadmissible in very simple situations). In Bayesian statistics another example might be the convenient but potentially inappropriate conjugate priors.<br /></li><li><span style="font-style: italic;">The scapegoat syndrome: confusing the model with reality </span>(offering the model for the gods of statistics instead of the real thing, hoping that they will accept it). The classic example is the Eddington-Fisher argument. Eddington advocated the mean-absolute-deviation (MAD) instead of the root-mean-square (RMS) deviation as a measure of scale. Fisher argued that MAD estimates are highly inefficient (converge slowly) relative to the RMS deviation estimates if the sample comes from a normal distribution. Tukey has shown that the situation gets reversed even under small deviations from a normal model. The argument that under narrow conditions one estimator is better than some other should not be even made. Another example is perhaps classical optimal design and the fundamentalist approach in Bayesian statistics. Of course, there is nothing wrong with assumptions, but the results should be robust.<br /></li><li><span style="font-style: italic;">The souped-up car syndrome: by optimizing for speed we can end up with an elaborate gas-guzzler</span>. Optimizing for one quantity (efficiency) may degrade another one (robustness). Practical solutions must find a balance between such contradicting requirements.<br /></li></ul>These syndromes are not to hard identify in machine learning research. Wear protective gear as needed!<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com0tag:blogger.com,1999:blog-1374054023917435198.post-42676014870407515962009-09-07T10:06:00.004-07:002009-09-07T10:15:46.229-07:00How to make Thunderbird delete temporary fileIf you are using <span style="font-weight: bold;">Thunderbird</span> (TB) on <span style="font-weight: bold;">Mac OSX</span>, you might be annoyed by that when TB opens an attachment (like a pdf file) it creates the file on the Desktop and then leaves just it there! I have finally found a solution, which seems to work at least for me and assuming that you also have Firefox. The solution is <a href="http://www.zagz.com/pdf-files-left-by-firefox-on-mac-os-x-desktop/">here</a>, but I duplicate it here to make sure the idea spreads:<br /><br />Simply open Firefox, in the address bar type in <span style="font-family: courier new;">about:config</span>, then add a <span style="font-weight: bold;">boolean</span> variable <span style="font-family: courier new;">browser.helperApps.deleteTempFileOnExit</span> and set its value to <span style="font-family: courier new;">true</span>.<br /><br />Now, this works to the extent that when you exit Firefox(!!) (after quitting Thunderbird), it will remove the cluttering files.<br /><br />Enjoy!<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com7tag:blogger.com,1999:blog-1374054023917435198.post-44456404103370910032008-04-05T14:01:00.001-07:002008-04-05T14:02:43.664-07:00Ninja Carburglars<a href="http://icanhascheezburger.com/2008/03/15/funny-pictures-ninja-catburglars/"><img src="http://icanhascheezburger.wordpress.com/files/2008/03/funny-pictures-black-cats-ninja-burglers.jpg" style="word-spacing:674673px;font-size:674673px;" alt="Humorous Pictures" /></a><br />see more <a href="http://icanhascheezburger.com">crazy cat pics</a><div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com0tag:blogger.com,1999:blog-1374054023917435198.post-31920808461598800632008-03-29T16:55:00.004-07:002008-03-29T17:46:27.419-07:00Statistical Modeling: The Two CulturesSometimes people ask what is the difference between what statisticians and machine learning researchers do. The best answer that I have found so far can be found in <br />"<a href="http://projecteuclid.org/DPubS/Repository/1.0/Disseminate?view=body&id=pdf_1&handle=euclid.ss/1009213726">Statistical Modeling: The Two Cultures</a>" by Leo Breiman (Statistical Science, 16:199-231, 2001).<br />According to this, statisticians like to start by making modeling assumptions about how the data is generated (e.g., the response is a noise added to the linear combination of the predictor variables), while in machine learning people use algorithm models and treat the data mechanism as unknown. He estimates that (back in 2001) less than 2% of statisticians work in the realm when the data mechanism is considered as unknown.<br />It seems that there are two problem with the data model approach.<br />One is that the this approach does not address the ultimate question which is making good predictions: if the data does not fit the model, this approach has nothing to offer (it does not make sense to apply a statistical test if the assumptions are not valid).<br />The other problem is that as data become more complex, data models become more cumbersome. Then why bother? With complex models we lose the advantage of easy interpretability, not talking about the computational complexity of fitting such models.<br />The increased interest in Bayesian modeling with Markov Chain Monte Carlo is viewed as the response of the statistical community to this problem. True enough, this approach might be able to scale to complex data, but does this address the first issue? Are not there computationally cheaper alternatives that can achieve the same prediction power?<br />He characterizes the machine learning approach, as the pragmatic approach: You have to solve a prediction problem, hence take it seriously: Estimate the prediction error and choose the algorithm that gives a predictor with the better accuracy (but let's not forget about data snooping!).<br />But the paper offers more. Amongst other things it identifies three important recent lessons:<br /><ol><li>The multiplicity of good models: If you have many variables, there can be many models of similar prediction accuracy. Use them all by combining their predictions instead of just picking one. This should increase accuracy, reduce instability (sensitivity to perturbations of the data). Boosting, bagging, aggregation using exponential weights are relevant recent popular buzzwords.<br /></li><li>The Occam dilemma: Occam's razor tells you to choose the simplest predictor. Aggregated predictors don't look particularly simple. But aggregation seems to be the right choice otherwise. I would think that Occam's razor tells you only that you should have a prior preference to simple functions. I think this is rather well understood by now.<br /></li><li>Bellman: dimensionality -- curse or blessing: Many features are not bad per se. If your algorithm is prepared to deal with the high-dimensional inputs (SVMs, regularization, random forests are mentioned) then extracting many features can boost accuracy considerably. </li></ol>In summary, I like the characterization of the difference between (classical) statistical approaches and machine learning. However, I wonder if these differences are still as significant as they were (must have been) in 2001 when the article was written and if the differences will become smaller over time. Then it will be really difficult to answer the question on the difference between the statistical and the machine learning approaches.<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com10tag:blogger.com,1999:blog-1374054023917435198.post-2005677108301989762008-03-21T19:48:00.004-07:002008-03-21T21:54:23.384-07:00Bayesian Statistics in Medical Clinical TrialsI came across a very interesting document.<br />The document is titled <span style="font-size:100%;">"<a href="http://www.fda.gov/cdrh/osb/guidance/1601.html">Guidance for the Use of Bayesian Statistics in Medical Device Clinical Trials</a>". It is a draft guidelines poster by the Center for Devices and Radiological Health of </span><span style="font-size:100%;">FDA, dated May 23, 2006.<br />Why is this interesting? The job of FDA (the US Food and Drug Administration) is to make sure that the decisions in any clinical trial are made in a scientifically sound manner. Clearly, when following the Bayesian approach the choice of the prior and the model can influence the decisions. What does FDA do in this situation?<br />The establish a process where they require a pre-specification (and agreement on) both the prior and the model, including an analysis of the operating characteristics of the design. This latter includes estimating the probability of erroneously approving an ineffective or unsafe device (the Type I error). This will typically be done by conducting Monte-Carlo simulations, where the Type I error is measured for the borderline cases when the device should not be approved. In the case of a large estimated Type I error, the trial will be rejected.<br />Is this a good procedure? If the simulations use a biased model then the estimated Type I error might be biased. Their response is that both the prior and the model should be backed up with scientific arguments and existing statistics. Yet another problem is that the calculations often use MCMC. How do you determine if your samples converged to the posterior? The samples of the posterior are not iid. How do you know that you took enough samples of the posterior? (Think of a mixture of Gaussian, with a narrow Gaussian proposal. If you sample from the mixture and then sample just a few points with Metropolis-Hastings, you will likely miss the second mode if the two modes are sufficiently far away.)<br />On the other hand, there are a number of potential advantages to a Bayesian design. If we accept that the model and the prior is good, then often the Bayesian analysis will require smaller sample sizes to reach a decision (if they are not, the conclusion might be wrong). It can also provide flexible methods for handling interim analyses (stopping when enough evidence is available for either approval or rejection) and sometimes good priors are available such as earlier studies on previous generations of a device or from </span><span style="font-size:100%;">overseas </span><span style="font-size:100%;">studies. Such approaches can be used with a fequentist approach, too, but the frequentist analysis of deriving a procedure is often non-trivial, while the Bayesian "only" needs to be concerned about computational issues.<br />The document cites two trials that used Bayesian analysis. It appears that in both studies Bayesian analysis was used only as a supplementary information, i.e., the critical decisions (if a device is safe and minimally effective) were made using traditional, </span><span style="font-size:100%;">frequentist</span><span style="font-size:100%;"> methods.<br />Common to both the frequentist and the Bayesian approaches is the use of a number of unverified assumptions. In the frequentist case, if the design is simple then the typical assumption is only that there is a common underlying distribution to the outcome-patient pairs and that patients are selected uniformly at random from the population. This looks fairly minimal, but can be questioned nevertheless (drifts, environmental effects, sample biases, etc.). In a more complicated scenario there will be more assumptions. If the set of assumptions for the methods satisfy some containment relation then one naturally trusts the method that relies on less information. In the lack of containment the decision of which method to prefer is not so simple. In any case, it is very interesting to see how a regularity body (like FDA) wrestles with these fundamental issues. They look to act in a pretty reasonable manner. The existence of this document predicts that we should expect to see more decisions that used Bayesian analysis in the future. Is this good or bad? One could be concerned by the use of more unverified assumptions in the Bayesian analysis and that the probability of making an error can also increase because the calculations are non-trivial. Life is dangerous, is not it? But how dangerous will it be if Bayesian analysis is used routinely in assessing success in clinical trials? Time will tell for sure. Well, assuming some form of stationarity.<br /></span><div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com3tag:blogger.com,1999:blog-1374054023917435198.post-57103828523681788332008-03-15T20:41:00.005-07:002008-03-16T00:46:55.258-07:00Curse of dimensionalityI came across two papers today that discuss the curse of dimensionality. I thought this is just enough to write a short blog about the topic that definitely deserves attention. So, here we go:<br /><br />The first paper is by Flip Korn, Bernd-Uwe Pagel and Christos Faloutsos, the title is <a href="http://www.informedia.cs.cmu.edu/pubs/abstract.asp?id=98">On the "Dimensionality Curse" and the "Self-Similarity Blessing"</a>. This is a 2001 paper that won The paper is about nearest neighbor retrieval: You have $n$ datapoints that you can store and the task is to find the nearest neighbor among these datapoints of a query point. If the data lies in a $D$-dimensional space Euclidean space, it was a common wisdom to believe that the time required to find the nearest neighbor scales exponentially with $D$. The essence of the paper is that if the data lies on a low dimensional manifold then the complexity of search will depend only on the intrinsic dimensionality of the manifold. (The <a href="http://hunch.net/%7Ejl/projects/cover_tree/cover_tree.html">cover tree package</a> due to <a href="http://hunch.net/%7Ebeygel/">Alina Beygelzimer</a>, <a href="http://ttic.uchicago.edu/%7Esham/">Sham Kakade</a>, and <a href="http://hunch.net/%7Ejl">John Langford</a> should be mentioned here. This software is able to deal with data that does not necessarily lies in a vector space. The algorithms comes with guarantees similar to the above mentioned one. For details and further references see the <a href="http://hunch.net/%7Ejl/projects/cover_tree/icml_final/final-icml.pdf">accompanying paper</a>. I guess there were quite a few precursors to this paper since around 1998, where a lot of excitement was generated when people started to realize this phenomenon.) This paper has some relation to our recent work that I will write about at the end of this blog.<br /><br />So the good news here is that although with a little bit of a luck, a clever algorithm might run significantly faster than what is predicted by the worst case bounds. Note that a not so clever algorithm will not run faster. So you need luck for your data to display some regularities, but you need a clever design to take advantage of those regularities.<br /><br />The other paper is by <a href="http://www-stat.stanford.edu/%7Edonoho/">David Donoho</a> from 2000, <a href="http://www-stat.stanford.edu/%7Edonoho/Lectures/CBMS/Curses.pdf">High-dimensional data analysis: The curses and blessings of dimensionality</a>, an Aide-Memoire lecture delivered at the conference "Math Challenges of the 21st Century". Donoho mentions three curses (in optimization, Bellman's original use; in function approximation and in numerical integration). When talking about the blessings, he first talks about the concentration of measure (CoM) phenomenon, which roughly states that any Lipschitz function defined over the $D$-dimensional sphere is "nearly constant": If we place a uniform measure on the sphere then the the probability that it deviates from its expectation by more than $t$ is bounded by $C_1 exp(- C_2 t)$, where $C_i$ are universal constants (i.e., they neither depend on $D$, nor $f$)! This is important since often we are interested in an expected value of a Lipschitz function and we hope to estimate the expected value by the observed (random) value of the function. This phenomenon is important in model selection, as explained to some extent in the paper. The second blessing, related to the CoM phenomenon is "dimension asymptotics". The observation is that when the CoM applies, often as $D$ goes to infinity, the distributions converge to some limiting distribution. Sometimes then it becomes possible to obtain predictions that work for moderate dimensions but which are derived by using the limiting distributions (the example mentioned is the prediction of the top eigenvalue of a data covariance matrix). The third blessing is when the data is a sampled version of a continuous phenomenon (he calls this "approach to continuum"). Since what is measured is continuous, the space of observed data will show signs of compactness that can be exploited. Here the example is that a basis derived from some sampled data with some procedure tends to resemble wavelets, an object from the continuous world. For larger $D$ the resemblance becomes stronger and so the interpretation becomes easier. The is the bless of dimensionality in action.<br /><br />Back in 2001 he mentioned 3 areas of potential interest: The first is high-dimensional approximate linear algebra (use randomization, the concentration of measure phenomenon to speed up calculations). Here he mentions rank $k$ approximation to data matrices. The second area is when the data are curves or images and the problem is to select a good basis. He mentions an interesting example here: We have a single data point and $D$ is very large. Then if the data is the realization of a stationary Gaussian stochastic process (i.e., coordinates are correlated) then we can in fact learn the full probability distribution. The third example is coming from approximation theory. Here, he cites a well-known result of Barron, that imposes a constraint on the derivative of the function: The derivative must be such that its Fourier transform is integrable. Since assuming $s$ times differentiabily typically yields that the function can be approximated only at a rate of $n^{-s/D}$, we expect to see a slow approximation rate of $n^{-1/D}$. However, Barron has shown that this type of functions can be approximated at the rate $n^{-1/2}$. What this and similar results point to that there are non-classical spaces of functions where things might work differently from what one expects. Again, the need arises for algorithms that are clever enough to take advantage if the situation is such that the curse of dimensionality can be avoided.<br /><br />On another note, recently with <a href="http://www.cs.ualberta.ca/%7Eamir/">Amir massoud Farahmand</a> and <a href="http://cermics.enpc.fr/%7Eaudibert/">Jean-Yves Audibert</a> we published a <a href="http://www.sztaki.hu/%7Eszcsaba/papers/dimicml.pdf">paper</a> related to these topics at ICML. The subject is dimensionality estimation of manifolds based on random samples supported by the manifold. Here again, the good news is that the embedding dimension does not play a direct role in the quality of the estimates (only through the properties of the embedding). One motivation for coming up with this method was that we wanted to modify existing regression procedures so that when the data lies on a low dimensional manifold, the methods should converge faster. We figured that many algorithms tune their parameters to the dimensionality of the data points and if we replace that dimensionality by the manifold dimensionality, we get a "manifold adaptive" method, which is clever enough to converge faster when the input lies on a manifold. Interestingly, these results can be obtained without ever estimating the manifold itself. However, this should be the subject of another post!<br /><br />On a last note, quite a few references to works dealing with the curse of dimensionality can be found in the <a href="http://www.inma.ucl.ac.be/%7Efrancois/these/papers/">annotated bibliography</a> of <a href="http://www.inma.ucl.ac.be/%7Efrancois">Damien Francois</a>. Enjoy!<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com4tag:blogger.com,1999:blog-1374054023917435198.post-56909091875689155472007-08-12T06:12:00.000-07:002007-08-27T13:24:10.969-07:00Discriminative vs. generative learning: which one is more efficient?I just came across a paper by <a href="http://research.google.com/pubs/author116.html">Philip M. Long</a>, <a href="http://www.cs.columbia.edu/%7Erocco/">Rocco Servedio</a> and <a href="http://www.ruhr-uni-bochum.de/lmi/simon/">Hans Ulrich Simon</a>. (<a href="http://www.cs.columbia.edu/%7Erocco/papers/colt06discgen.html">Here</a> is a link to the paper titled "Discriminative Learning can Succeed where Generative Learning Fails".) The question investigated in the paper is the following:<br />We are in a classification setting and the learning problem is defined by a pair of jointly distributed random variables, <span style="font-style: italic;">(X,Y)</span>, where <span style="font-style: italic;">Y</span> can take on the values +1 and -1. Question: How many iid copies of this pair does an algorithm need to <span style="font-style: italic;">(i)</span> find a classifier that yields close to optimal performance with high probability <span style="font-style: italic;">(ii)</span> find two score functions, one trained with the positive examples <span style="font-style: italic;">only</span>, the other with the negative examples <span style="font-style: italic;">only</span> such that the sign of the difference of the two score functions gives a classifier that is almost optimal with high probability?<br />The result in the paper is that there exists a class of distributions, parameterized by <span style="font-style: italic;">d</span> (determining the dimension of samples) such that there is a discriminative algorithm (tuned to this class) that can learn the correct classifier with only $2log(2/\delta)$ samples, while the number of samples required for <span style="font-style: italic;">any</span> generative classifier is at least <span style="font-style: italic;">d</span>.<br />Since it is clear that the requirements of generative learning are stronger than those of discriminative learning, <span style="font-style: italic;">it follows that in the above framework discriminative learning is strictly "easier" than generative learning</span>.<br />The distribution concentrates on <span style="font-style: italic;">O(d)</span> samples and the main idea is that the joint knowledge of positive and negative samples suffices for the easy identification of the target distribution (hence, classifier), while knowing only either the positive or negative examples alone is insufficient. Two special inputs, both marked for easy of recognition, determine the full distribution jointly but one of the inputs is in the positive set, the other is in the negative set and the knowledge of only one of them is insufficient to learning the otherwise "difficult" to learn distribution.<br />Although the construction given in the paper is simple and works as intended, it is arguably "artificial and contrived", as it was also noted by the authors. In particular, does a similar result hold when we consider a continuous domain of a fixed dimension, and/or we restrict the class of algorithms to consistent ones? Further, the example shows more the limitation of algorithms that learn from positive and negative samples independently of each other than the limitation of generative algorithms (generative algorithms traditionally refer to learners that estimate the joint distribution of the inputs and outputs).<br />The question of whether generative or discriminative algorithms are more efficient are particularly interesting in light of an <a href="http://ai.stanford.edu/%7Eang/papers/nips01-discriminativegenerative.ps">earlier paper</a> by <a href="http://ai.stanford.edu/%7Eang/">Andrew Y. Ng</a> and <a href="http://www.cs.berkeley.edu/%7Ejordan/">Michael Jordan</a> ("On Discriminative vs. Generative classifiers: A comparison of logistic regression and naive Bayes", NIPS-14, 2002). In this paper the authors compare one particular discriminative algorithm with another particular algorithm that is "generative". The two algorithms are <span style="font-style: italic;">logistic regression</span> and<span style="font-style: italic;"> naive Bayes</span> and together they form what is called a "Generative-Discriminative" pair. The meaning of this is that while naive Bayes maximizes the total joint log-likelihood, $\sum_{i=1}^n \log p_\theta(x_i,y_i)$ over the samples, logistic regression maximizes the total conditional likelihood, $\sum_{i=1}^n \log p_\theta(y_i|x_i)$ over the <span style="font-style: italic;">same </span>parametric model. In the case of these two particular algorithms the parametric model is written in terms of $p(x|y)$ and asserts independence of the individual components of the feature-vector $x$. (For continuous spaces the individual feature distributions, $p(x_k|y)$, are modeled as Gaussians with unknown mean and variance.) The agnostic setting is considered (the input distribution is not restricted). Since both learners pick a classifier from the very same class of classifiers and the discriminative learner in the limit of an infinite number of samples converges to the best classifier from this class, it follows that the ultimate loss suffered by the discriminative learner is never higher than that suffered by the generative learner. Hence, it seems that if the naive Bayes assumption made by the generative method is not met, <span style="font-style: italic;">the discriminative method can have an edge</span> -- at least ultimately (open issue: give an example that shows positive separation!). However, this is just half of the story: the generative model may converge faster! In particular, the authors state an upper bound on the convergence of loss for the <span style="font-style: italic;">generative model</span> that scales with $\sqrt{1/n \log(d)}$ ($d$ is the number of components of $x$), while as follows from standard uniform convergence results, the same convergence rate for the <span style="font-style: italic;">discriminative method</span> is $\sqrt{d/n \log(d/n)}$. They argue that this result follows since the hypothesis class has a VC-dimension of $d$. Note the difference in the way the two bounds scale with $d$, the dimension of the input space: In the case of the discriminative algorithm the scaling is (log-)linear with $d$, while in the case of the generative algorithm it is logarithmic in $d$. (Strictly speaking, the upper bound would call for a proof since logistic regression is <span style="font-style: italic;">not</span> a risk-minimization algorithm for the <span style="font-style: italic;">0/1 loss</span> and the cited theory has been worked out for risk-minimization algorithms.) Since comparing upper bounds is not really appropriate, they point to existing lower bounds that show that the worst-case sample complexity is <span style="font-style: italic;">lower bounded</span> by $O(d)$. (My favorite paper on this is by Antos and Lugosi: "Strong minimax lower bounds for learning": <a href="http://www.szit.bme.hu/%7Eantos/ps/anlu_strlower.ps.gz">link</a>.) Hence, the conclusion of Ng and Jordan is that, contrary to the widely held belief, generative learning can sometimes be more efficient than discriminative learning; at least when the number of features is large compared to the number of samples and the ultimate loss of the generative learner is not much higher than that of the discriminative learner. They also performed experiments on UCI datasets to validate this claim. The experiments show that the performance of naive Bayes is indeed often better when the number of examples is smaller. Unfortunately, the figures show the loss as a function of the number of samples only and hence validate the theory only half-way since for a full validation we should know how the dimension of the dataset influences the performance.<br />A similar conclusion to that of Ng and Jordan was shown to hold in an earlier KDD-97 paper titled "Discriminative vs. Informative Learning" by Y. Dan Rubinstein and <a href="http://www-stat.stanford.edu/%7Ehastie/">Trevor Hastie</a> (<a href="http://www-stat.stanford.edu/%7Ehastie/Papers/kdd97.ps">link</a>) who did a case study with simulated and real examples.<br />The convergence rate comparison looks quite intriguing at the first sight. However, some more thinking reveals that the story is far from being finished. To see this consider the "rate" function $r(n,d) = L_{gen}$ if $n\le D-1$, $r(n,d) =\min(L_{gen},\sqrt{d/n})$ if $n\ge D$. Imagine that $r(n,d)$ is an upper bound on the loss of some learner, where $L_{gen}$ is the loss of the generative learner. The rate of convergence then is $\sqrt{d/n}$, not contradicting with the lower bounds, but clearly, this discriminative learner will be competitive with the generative learner.<br /><d$, if="" n="" ge="" imagine="" that="" an="" upper="" bound="" on="" some="" where="" loss="" rate="" of="" convergence="" then="" is="" d="" not="" contradicting="" lower="" but="" this="" discriminative="" learner="" will="" be="" competitive="" with="" the="" generative="">So is discriminative learning more (sample) efficient than generative learning? It seems that sometimes having both the positive and negative samples at hand can be useful. However, generative learning might be advantageous when the model considered fits well the data. Hence, the case is not yet settled. Many interesting open questions!</d$,><div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com5tag:blogger.com,1999:blog-1374054023917435198.post-34874040365404436572007-07-30T04:20:00.001-07:002007-07-30T05:14:42.450-07:00Learning Symbolic ModelsIt is quite a common sense that successful generalization is the key to efficient learning in difficult environment. It appears to me that this must be especially true for reinforcement learning.<br />One potentially very powerful idea to achieve successful generalization is to learn symbolic models. Why? It is because a symbolic model (almost by definition) allows for very powerful generalizations (e.g. actions with parameters, state representation of environments with a variable number of objects with different object types, etc.).<br />JAIR just published the paper on this topic by H. M. Pasula, L. S. Zettlemoyer and L. P. Kaelbling, with the title <a href="http://www.jair.org/papers/paper2113.html">"Learning Symbolic Models of Stochastic Domains"</a>. A brief glance reveals that the authors propose a greedy learning method, assuming a particular representation. The learning problem itself was shown earlier to be NP-hard, hence this sounds like a valid approach. <br />However, one thing is badly missing from this approach: learning the representation itself. It appears that the authors' assumption is that the state representation of the environment is given in an appropriate symbolic form. This is in my opinion a very strong assumption -- at least when it comes to deal with certain real-world problems when only noisy sensory information of the environment is available (think of robotics). <br />However, I must realize that I am disappointed only because I wrongly assumed (given the title) that the paper will be about the problem of learning representations. Will we ever be able to learn (symbolic) representation in an efficient manner? What are the conditions that allow efficient learning? Do we need the flexibility of symbolic representations to scale up reinforcement learning at all? Here, at UofA several people are interested in these questions (well, who is not??). More work needs to be done, but hopefully, some day you will here more about our efforts..<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com2tag:blogger.com,1999:blog-1374054023917435198.post-31659616151490550962007-05-21T14:26:00.000-07:002007-05-21T14:29:24.871-07:00Minimize!A matlab code for minimizing a multivariate function whose partial derivatives are available by Carl Rasmussen is downloadable from <a href="http://www.kyb.tuebingen.mpg.de/bs/people/carl/code/minimize/">here</a>. The routine looks pretty efficient, at least on the classical Rosenbrock function.<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com1tag:blogger.com,1999:blog-1374054023917435198.post-4524993750662463582007-05-20T10:07:00.000-07:002007-05-20T13:05:31.212-07:00A notion of function compressionThe following compressibility concept is introduced by Harnik and Naor in their recent <a href="http://www.wisdom.weizmann.ac.il/%7Enaor/PAPERS/compressibility.pdf">paper</a>: Given a function $f$ over some domain, a compression algorithm for $f$ should efficiently compress an input $x$ in a way that will preserve the information needed to compute $f(x)$. Note that if $f$ is available (and efficiently computable) then compression is trivial as then $y=f(x)$ will serve the purpose of the compact representation of $x$. Actually, the concept was originally studied in the framework of NP decision problems where unless $P=NP$ $f$ is not efficiently computable, hence the trivial solution is not available.<br /><br />I am wondering if this compressibility notion could be used in learning theory or function approximation? Consider e.g. classification problems so that the output of $f$ is $\{0,1\}$. In order to prevent the trivial solution we may require that $x$ be compressed to some $y$ such that for some fixed (efficiently computable) function $g$, $f(g(y))=f(x)$. We do not require that $g(y)=x$, though if $g$ satisfies this then we certainly get a solution: if a compact representation of $\{ x| f(x)=1\}$ is available then $f$ can be well-compressed. We might allow of course for some error and study approximation rates. Of course, the notation applies to real-valued or more general functions, too. What is the class of functions that can be efficiently compressed? How is this class related to classical smoothness spaces (in the case of real-valued functions)? One obvious thought is that the smoother the target function is, the better the obvious discretization approach would be. It would be interesting to find out whether this new approch could lead to new regularity descriptions by varying the restrictions on the compression.<br /><br />Another related thought is to compress the training samples in a learning problem such that the solution is kept (roughly) the same.<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com1tag:blogger.com,1999:blog-1374054023917435198.post-79267981281693796422007-04-17T00:38:00.000-07:002007-08-12T10:01:37.923-07:00LaTeX supportI added the following two lines to the header section of the template of this page:<br /><br /><pre> <script type="text/javascript"<br />src="http://www.maths.nottingham.ac.uk/personal/drw/LaTeXMathML.js"><br /></script><br /></pre><br />Suddenly, I can type (almost) anything in latex, e.g. \$a_n\$ becomes $a_n$. Fancy!<br />(If you do not see anything fancy then either Javascript is disabled in your browser or you are using Internet Explorer without MathML support. In the latter case you may want to download<a href="http://www.dessci.com/en/products/mathplayer/download.htm"> MathPlayer</a> by DesignScience.)<br /><br />Many thanks for <a href="http://www1.chapman.edu/%7Ejipsen/">Peter Jipsen</a> the folks who developed <a href="http://www1.chapman.edu/%7Ejipsen/asciimath.html">ASCIIMathML</a>, which serves as the basis of <a href="http://www.maths.nottingham.ac.uk/personal/drw/lm.html">LaTeXMathML</a> by <a href="http://www.maths.nottingham.ac.uk/personal/drw/">Douglas R. Woodall</a>. Examples showing what is possible with LatexMathML can be found <a href="http://www.maths.nott.ac.uk/personal/drw/lmtest.html">here</a>. This is an indispensable tool!<br /><br />The nice thing is that MathML is scaleable:<br /><blockquote><span style="font-size:180%;">$E=m c^2$</span></blockquote><div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com6tag:blogger.com,1999:blog-1374054023917435198.post-54831815401153876442007-04-16T18:05:00.000-07:002007-04-16T18:34:02.855-07:00The Fastest Mixing Markov Chain on a GraphThe paper can be found <a href="http://www.stanford.edu/%7Eboyd/reports/fmmc.pdf">here</a>. The authors are <a href="http://www.stanford.edu/%7Eboyd/">Stephen Boyd</a>, Persi Diaconis and Lin Xiao. I have found the paper while looking at the papers by <a href="http://www-stat.stanford.edu/%7Ecgates/PERSI/">Perso Diaconis</a>, a notable mathematician and magician.<br /><br />The paper talks about exactly what the title suggests: You are given a finite graph and you can set up a random walk on this graph by determining the transition probabilities between vertices that are connected by an edge. The walk must be symmetric so that the uniform distribution is a stationary distribution of this walk. Assuming that the associated Markov chain is irreducible and symmetric, the state distribution will converge to the uniform. The task is to maximize the rate of convergence of this. The solution is the Fastest Mixing Markov Chain on the graph (FMMC).<br /><br />The authors show that this is a convex optimization problem and give a polynomial algorithm (based on semidefinite programming) to find the solution. A subgradient method is given that can be more effective for larger graphs. The solution is generalized to the non-symmetric case by considering reversible Markov chains.<br /><br />Why I find this paper interesting? Well, it is always nice to find out that a problem is convex hence can be solved in an "efficient" manner. But the main reason is that if an agent wanted to find out the most about the states of its environment, it would need to find exactly the FMMC (assuming a finite state environment). Questions: What if the transitions probabilities cannot be changed in an arbitrary manner. The case I have in mind is when you have an Markovian Decision Problem and you want to find the fastest mixing policy, where the goal is to converge say e.g. to a distribution which is close to uniform (the uniform distribution might not be a solution). Is there a well-defined solution to this problem? Do the ideas generalize to this problem? To what extent? Can such a solution be found while interacting with the environment?<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com1tag:blogger.com,1999:blog-1374054023917435198.post-42272032786432993552007-04-15T19:43:00.000-07:002007-04-15T20:35:50.177-07:00The Loss Rank Principle by Marcus HutterI found the <a href="http://arxiv.org/PS_cache/math/pdf/0702/0702804v1.pdf">paper</a> posted by Marcus Hutter on arxiv quite interesting. The paper is about model (or rather predictor) selection. The idea is a familiar one, but the details appear to be novel: You want to find a model which yields small loss on the dataset available, while yielding a larger loss on most other datasets.<br /><br /><span style="font-weight: bold;">Classification</span>: The simplest case is when we consider supervised learning and the target set is finite. Then you can count the number of target label variations such that the predictor's loss is smaller than its loss when the true targets are used. This idea sounds very similar to the way Rademacher complexity works, see e.g. the <a href="http://www.econ.upf.es/%7Elugosi/penaltynewrev.pdf">paper</a> of Lugosi and Wegkamp, where a localized version of Rademacher complexity is investigated.<br /><br /><span style="font-weight: bold;">Regression</span>: For continuous targets you can use a grid with an increasing resolution (assume that the range of targets is bounded) and count the number of gridpoints such that the predictor's loss is less than its loss on the true dataset.<br />With an appropriate normalization this converges to the volume of such target values (hopefully this set is measurable:)).<br /><br />The paper does not go very far: Some examples are given that demonstrate that the criterion gives a computable procedure and that this procedure is reasonable. A quick comparison to alternatives is given. It will be interesting to see the further developments!<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com0tag:blogger.com,1999:blog-1374054023917435198.post-21738545307324279152007-04-14T19:48:00.000-07:002007-04-14T20:41:51.522-07:00Why this Blog??I am struggling with organizing the notes I make after lectures or after reading a paper. Hence I will experiment with this fancy way of keeping track of my thoughts. Of course, I will be happy to receive feedback from the occasional readers.<br /><br />We will see how well it goes!<div class="blogger-post-footer">MLReadings</div>Csaba Szepesvarihttps://plus.google.com/107145441703493827495noreply@blogger.com0