Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Here at Qualia (qualia.ai) we process mostly textual data from online sources (news, blogs, social media, internal data). Our background is in NLP when back in the days AI meant deep parsing, HPSG, tree-adjoining grammars, synsets, frames and speech acts, discourse, and different flavors of knowledge representations. It also meant LISP and Prolog. The domain quickly evolved from knowledge and rule-based to data-driven and statistical, mostly thanks to Brown and the IBM MT team in the 90s (that are now part of the Renaissance Fund).

We use hierarchical clustering for topic detection. We also work on topic models (Blei and his legacy). We use word embeddings for information retrieval and various ML algorithms for different applications of mood and emotional learning: Bayes, SVM, Winnow (linear models) and sometimes decision trees and lists. We also learn from past events and crises in order to create models, mostly statistical, and try to estimate how an event might evolve in the future. We have also tried graph-based community detection algorithms on Twitter (min-cut). Finally we have experimented with non-linear statistical analysis on micro-blogging data, by applying methods such as correlation functions, escape times, and multi-step Markov chains (but with limited success).

I 'd like to add here that I feel ML is well defined (supervised, semi-supervised, unsupervised and using unlabeled data), statistical learning is more fuzzy (a good starting point is Vapnik's work) and regarding AI, I am not sure I know any more what it means! I am always open to discussion and ideas. Let me know.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: