Skip to main content

Basic Research under Attack

Since the current world-wide crisis began scientists have had very difficult times. Cuts in budgets and grants are a common difficulty that researchers have to handle in order to progress with their projects. Nevertheless, curiosity and perseverance  are major forces that can drive a scientist towards their goal even in such a harsh condition. However, there is a critical boundary that can kill the bravest researcher and this happens when the manager in charge of science in your own country; the one who has to protect you as a researcher, tells in a prestigious media that "we must now stop talking about the importance of science, and instead commit ourselves to the need for excellence in science" and "we need to change the number of researchers by maintaining and improving the quality of the contracts while reducing the quantity.You can find the full article in this link.

Of course the response of the scientific community has been quick, and with data that disqualifies these statements. For instance,, a well-known Spanish web site for the popularization of science, has shown the following chart (Fig. 1) in which depicts the number of scientists per million of inhabitants. As it shows, there seems to be a strong relation between the ratio of scientists and the progress of a country, and in the particular case of Spain it is far bellow the most industrialized (thus less affected by the crisis) countries in the EU.

Fig. 1 Number of R&D personnel per million of inhabitants. Credit:

One can argue that science has to have private funds and not public ones during crisis times; that we have to spend money in other much more important things like saving banks or preserving cultural traditions like bull fight*. This is a common reasoning of illiterate people about science. As it is pointed out by many scientists over the years, public R&D programmes are focused mainly on basic, fundamental, pure research, that is, they do not seek an immediate monetary benefit (as it happens with private companies), but to discover new things; to acquire new knowledge. This knowledge can be later exploited by companies and that is the case, just to mention one straightforward example, of the discovery of quantum physics and modern day electronics. Who could predict back in the 1930's that these crazy ideas could lead to a such an industrial revolution that move billions of dollars each year world-wide?

Investing in R&D is ensuring the future of the country and not only in the wealth of its inhabitants but also in the health and happiness as well. 

*notice the irony of the author.


Popular posts from this blog

Toward ensemble methods: A primer with Random Forest

The Kaggle-Higgs competition has attracted my attention very much lately. In this particular challenge, the goal is to generate a predictive model out of a bunch of data taken from distinct LHC’s detectors. The numbers are the following: 250000 properly labeled instances, 30 real-valued features (containing missing values), and two classes, namely background {b} and signal {s}. Also, the site provides a test set consisting of 550000 unlabeled instances. There are nearly 1300 participants while I am writing this post, and a lot of distinct methods are put to the test to win the challenge. Very interestingly, ensemble methods are all in the head of the leaderboard, and the surprise is XGBoost, a gradient boosting method that makes use of binary trees.
After checking the computational horsepower of the XGBoost algorithm by myself, I decided to take a closer look at ensemble methods. To start with, I implemented a Random Forest, an algorithm that consists of many independent binary trees…

High Performance Computing, yet another brief installation tutorial

Today’s mid- and high-end computers come with a tremendous hardware, mostly used in video games and other media software, that can be exploited for advanced computation, that is: High Performance Computing (HPC). This is a hot topic in Deep Learning as modern graphic cards come with huge streaming process power and large and quick memory. The most successful example is in Nvidia’s CUDA platform. In summary, CUDA significantly speeds up the fitting of large neural nets (for instance: from several hours to just a few minutes!).
However, the drawbacks come when setting up the scenario: it is non-trivial to install the requirements and set it running, and personally I had a little trouble the first time as many packages need to be manually compiled and installed in a specific order. The purpose of this entry is to reflect what I did for setting up Theano and Keras with HPC using an Nvidia’s graphic card (in my case a GT730) using GNU/Linux. To do so, I will start assuming a clean Debian …

A conceptual machine for cloning a human driver

If anything else, a Machine Learning practitioner has to get a global view and think on how far our conceptual machines can go. This is a little tale of my recent experience with Deep Learning. To start with, say that I am enrolled in the Udacity’s nanodegree inSelf Driving Car Engineer. Here we are learning the techniques needed for building a car controller capable of driving at the human level. The interesting part of the course (to be read as where one truly learns) is in the so called projects: these are nothing but assignments in which the student has to solve a challenging task, mainly by using Convolutional Neural Networks (convnets), the most well-known Deep Learning models. So far, the most mind-blowing task and the focus of this entry is project 3, were we have to record our driving behavior in a car simulator and let a convnet to clone it and successfully generalize the act of driving. It is remarkable that a convnet learns to control the car from raw color images jointly…