Archive

Archive for the ‘Data Mining’ Category

KDD 13, another day

August 13, 2013 Leave a comment

Stephen J. Write gave a brief, high-level, but very nice, clear keynote today at KDD’2013. Here’s the link to his talk slide, Several links he posted are also very interesting:

The panel discussion is about why and how Data Scientist should and can contribute to start up industry. People asked various questions like:

– What are the tips of choosing a startup, is a real ‘start’- type start-up better, or a start-up but kind of established better?

– As an investigator, what are the key things to evaluate to make the judgment that whether the start-up is worthy to investigate?

CEO, team, find a business co-founder, idea, execution,  hiring, patent,

– In this (big)data analysis, minding area,  does people (data scientist) still need or  better to get a PHD?

– When choose to start a ‘start-up’, when choose to leave a big company, leave a stable job & life, how do you evaluate, perceive and think about the risk behind this?  Is it rewarding  a lot (financially) ? Is financial gain your objective function?

You can find a more complete note here.

Advertisements
Categories: Data Mining

Memory on X

August 11, 2013 Leave a comment

Interesting work on “language memorability” by  Jon Kleinberg

You had me at hello: How phrasing affects memorability,  Cristian Danescu-Niculescu-Mizil, Justin Cheng, Jon Kleinberg, Lillian Lee

It’s counterpart “memory on visual data“:

* Making Personas Memorable ,  CHI, 2007, extended abstracts on Human Factors in Computing Systems.

* Konkle, T., Brady, T. F., Alvarez, G. A., & Oliva, A. (2010). Conceptual distinctiveness supports detailed visual long-term memory for real-world objects. Journal of Experimental Psychology: General, 139(3), 558-78.  <Data http://cvcl.mit.edu/MM/objectCategories.html >
Another dataset: <http://visualrecall.org/datasets.html >
More publications:  <http://visualrecall.org/publications.html >

* Understanding the intrinsic memorability of images, NIPS 2011, Phillip Isola, Devi Parikh, Antonio Torrala, Aude Oliva,

 

Let data speak of itself

July 5, 2013 Leave a comment

Very data-driven, really data-centric, talk of “big visual data” by  Alexei A. Efros.

 

Coin of two sides – big data

February 22, 2013 1 comment

What would you think about Big Data? Everyone seems excited about the 4Vs: Volume, Velocity, Variety, and Veracity. But everything comes with two sides (The curse of big data). When you flip the cois, just to be careful of which side you are really getting.

 

Categories: Data Mining, Large Scale

Notes

December 27, 2012 1 comment

Some thing may only happen when you have a large amount of data. I came across this LinkedIn discussion and found one  of the linked post are very interesting as it mention the advantages brought by large data set for ‘scene completion’ in computer vision.  One of the conclusion is ” for nearest neighbor type minimization problems with a non-negative distance function (meaning that the cost function has a lower bound of zero), that distance function will, on average, decrease monotonically with data or sample size.” 

It’s also interesting that as we modify/improve approaches, models and making them complicated, it may not always perform well as compare to simple models based on the “law of large numbers.”.

 

Correlation & Causation

December 26, 2012 1 comment

Correlation does not necessary imply Causation. There could be multiple reason for a ‘correlation’:

– Causality: X causes Y

– Reverse Causality: Y causes X

– Simultaneity: X causes Y, Y causes X;

– Endogeneity: W causes Y, and X is correlated with W.

– Spuriousness: No causation, just a fluke.

 

Determining causation( from wiki)

David Hume argued that causality is based on experience, and experience similarly based on the assumption that the future models the past, which in turn can only be based on experience – leading tocircular logic. In conclusion, he asserted that causality is not based on actual reasoning: only correlation can actually be perceived.[14]

In order for a correlation to be established as causal, the cause and the effect must be connected through an impact mechanism in accordance with known laws of nature.

Intuitively, causation seems to require not just a correlation, but a counterfactual dependence. Suppose that a student performed poorly on a test and guesses that the cause was his not studying. To prove this, one thinks of the counterfactual – the same student writing the same test under the same circumstances but having studied the night before. If one could rewind history, and change only one small thing (making the student study for the exam), then causation could be observed (by comparing version 1 to version 2). Because one cannot rewind history and replay events after making small controlled changes, causation can only be inferred, never exactly known. This is referred to as the Fundamental Problem of Causal Inference – it is impossible to directly observe causal effects.[15]

A major goal of scientific experiments and statistical methods is to approximate as best as possible the counterfactual state of the world.[16] For example, one could run an experiment on identical twins who were known to consistently get the same grades on their tests. One twin is sent to study for six hours while the other is sent to the amusement park. If their test scores suddenly diverged by a large degree, this would be strong evidence that studying (or going to the amusement park) had a causal effect on test scores. In this case, correlation between studying and test scores would almost certainly imply causation.

Well-designed experimental studies replace equality of individuals as in the previous example by equality of groups. This is achieved by randomization of the subjects to two or more groups. Although not a perfect system, the likeliness of being equal in all aspects rises with the number of subjects placed randomly in the treatment/placebo groups. From the significance of the difference of the effect of the treatment vs. the placebo, one can conclude the likeliness of the treatment having a causal effect on the disease. This likeliness can be quantified in statistical terms by the P-value[dubious – discuss].

When experimental studies are impossible and only pre-existing data are available, as is usually the case for example in economics, regression analysis can be used. Factors other than the potential causative variable of interest are controlled for by including them as regressors in addition to the regressor representing the variable of interest. False inferences of causation due to reverse causation (or wrong estimates of the magnitude of causation due the presence of bidirectional causation) can be avoided by using explanators (regressors) that are necessarily exogenous, such as physical explanators like rainfall amount (as a determinant of, say, futures prices), lagged variables whose values were determined before the dependent variable’s value was determined, instrumental variables for the explanators (chosen based on their known exogeneity), etc. See Causality#Statistics and Economics. Spurious correlation due to mutual influence from a third, common, causative variable, is harder to avoid: the model must be specified such that there is a theoretical reason to believe that no such underlying causative variable has been omitted from the model; in particular, underlying time trends of both the dependent variable and the independent (potentially causative) variable must be controlled for by including time as another independent variable.

 

 

Categories: Data Mining, e-commece

December 18, 2012 Leave a comment

Very nice blog on raising the attention for ‘Similarity Measurement’

The "Putnam Program"

Similarity appears to be a notoriously inflationary concept.

Already in 1979 a presumably even incomplete catalog of similarity measures in information retrieval listed almost 70 ways to determine similarity [1]. In contemporary philosophy, however, it is almost absent as a concept, probably because it is considered merely as a minor technical aspect of empiric activities. Often it is also related to naive realism,which claimed a similarity between a physical reality and concepts. Similarity is also a central topic in cognitive psychology, yet not often discussed, probably for the same reasons as in philosophy.

In both disciplines, understanding is usually equated with drawing conclusions. Since the business of drawing conclusions and describing the kinds and surrounds of that is considered to be the subject of logic (as a discipline), it is comprehensible that logic has been rated by many practitioners and theoreticians alike as the master discipline. While there is a…

View original post 7,860 more words

Categories: Data Mining