The talk introduces the language identification problem within AI, teaches about decision trees, and then asks students to write decision trees in small groups to distinguish between Hmong, Balinese, Zulu, and other languages. After a debrief on why computers are might be more effective than human-written rules, it briefly ties in themes of feature extraction and gradient descent via GBMs.
As the progenitor and leader of SAIL ON, the Stanford AI Lab’s year-round effort to attract and keep underrepresented minorities in the field of Artificial Intelligence, I engage high schoolers about artificial intelligence, machine learning, and the positive social impacts of our field. SAIL ON meets once a month in the Computer Science building at Stanford. Its trifold focus allows past participants in the SAILORS two-week summer camp to continue to learn about AI, to nurture strong relationships with each other, and to lead outreach projects that bring the technical, humanistic, and diversity missions of the AI outreach program to the wider community.
As the educational component of the October meeting of SAIL ON, we discussed and applied Naive Bayes modeling. Like other machine learning methods, Naive Bayes is a generic approach to learning from specific data such that we can make predictions in the future. Whether the application is predicting cancer/whether you’ll care about an email/who will win an election, we can use the mathematics of Naive Bayes to connect a set of input variables to a target output variable. (Of course, some problems are harder than others!)
By providing this basis, I hope to increase appreciation for applications of what students are seeing in their math classes, and to facilitate students moving further on their own with applied machine learning before November’s meeting.
Integrative complexity is a construct from political psychology that measures semantic complexity in discourse. Although this metric has been shown useful in predicting violence and understanding elections, it is very time-consuming for analysts to assess. We describe a theory-driven automated system that improves the state-of-the-art for this task from Pearson’s r = 0.57 to r = 0.73 through framing the task as ordinal regression, leveraging dense vector representations of words, and developing syntactic and semantic features that go beyond lexical phrase matching. Our approach is less labor-intensive and more transferable than the previous state-of-the-art for this task.
Convolutional neural networks transformed computer vision from “extremely hard” to “trivially achievable after a few weeks of coursework” between 2012 and 2016.
I prepared a talk for technical professional audiences that describes how neural networks extend linear classification, intuitions behind why convolutional neural networks work well for vision, and the circumstances in which they’re worth consideration. I used the “Intro to CNNs for Computer Vision” materials at two different employers in 2016, and also with the high schoolers who participated in SAIL ON in 2017. (SAIL ON extended a Stanford summer program in AI that captured underrepresented minorities; I led the summer program and extended it to two years of follow-up monthly outreach.)
Automatic processing of sign languages can only recently potentially advance beyond the toy problem of fingerspelling recognition. In just the last few years, we have leaped forward in our understanding of sign language theory, effective computer vision practices, and large-scale availability of data. This project achieves better-than-human performance on sign language identification, and it releases a dataset and benchmark for future work on the topic. It is intended as a precursor to sign language machine translation.
Whatever is on my mind. The content has varied over the past more-than-decade, but it's always been technical. In the early years I focused on improving the fabric of the internet for some niche tools. But the internet no longer needs that kind of improving, and search doesn't really work like that anymore either. This blog is currently mostly about documenting notes for my future self, and sharing those notes with anyone who is interested.