Thanks to the magic of Mathematica 9, you can now make pretty pictures of your social networks with a single command. I was impressed by how well it automatically captured the true structure of my network, which I labeled and included below. I’m not sure what you can conclude about me as a person, except that I’m happy to route all socializing through my more extroverted wife. It’s probably also pretty clear which parts of my life are geographically separated.
Greg Ver Steeg's Facebook network

Black holes


In grad school I collaborated with Chris Adami on some papers about black holes. We suggested that some of the mysteries surrounding black holes may evaporate if you correctly include the effects of stimulated emission. Chris is presenting some of these results now at the APS meeting, and on his new blog. (paper)

Can we measure the effect of one person’s words on another person? I want to describe the main idea behind my recent paper, Information-Theoretic Measures of Influence Based on Content Dynamics, which I’m presenting at WSDM 2013.

Ostensibly, information theory is a dry, academic subject about how to send 0’s and 1’s through a noisy channel (like browsing the web using your cellular signal). The goal is to encode the information so that all the bits can be recovered after they are sent through a noisy channel. Shannon’s theory tells us that the amount of information we can send through the noisy channel is related to a quantity called “mutual information”.

noisy channel

What does this have to do with influence, human speech, or social media? This abstract framework is remarkably flexible. What if the input is some statement made by Alice. Then the “noisy channel” consists of (e.g.) sound waves, the ear drum, and the brain of Bob. Now Bob “outputs” some other statement. In the example below, Bob has said something very relevant to Alice’s statement: WSDM is in Rome, so Alice should definitely have some coffee while she’s there. Bob’s statement gives us some information about what Alice’s original statement was.


If, on the other hand, Bob had proclaimed his love of borscht, it’s not obvious that this has anything to do with Alice’s statement. Clearly, Bob lives in his own universe, day-dreaming of borscht. His statements carry no information about Alice’s statements.


How do we measure this notion of whether Bob’s statements carry information about Alice’s? We just use the standard information-theoretic notion of mutual information.


Unfortunately, this quantity depends on us being able to determine the probability of every pair of statements we might see Alice and Bob utter. Unfortunately, at best we might have observed a few hundred statements from Alice and Bob (on Twitter, for example). We have to do two things:

1. Simplify our representation of these statements (or “reduce the dimensionality of the signal”). That means Alice’s statement is reduced to some keywords that it might be about. (Some details: we use a popular technique called topic modeling to achieve this.)

2. Estimate mutual information using these simplified signals. (Some details: we actually use a more nuanced measure called conditional mutual information, or transfer entropy. We use non-parametric entropy estimators that avoid the step of estimating high dimensional probability distributions.)

Surprisingly, we were able to carry this procedure out for some users on Twitter and detect signals in human speech! One way to understand this result is to say that we could find user pairs where Bob’s statements could be predicted better if we knew Alice’s recent statements. It will come as no surprise that human speech is nuanced and complex, so we were only able to detect very predictable relationships. For instance, one triad of strongly connected users were three tri-lingual sisters. If one of the sisters tweeted to the others in a specific language, the others would always respond in the same language. Other strong relationships included news dissemination and web site promotion. Can we do better and detect more complex forms of influence? Sure, we need either more data or better representation of content or better entropy estimators. We’re working on all three!

New preprint: The title is changing to: “Information-Theoretic Measures of Influence Based on Content Dynamics”. I’ll give a detailed, readable summary in a week or two. I’ll be presenting about this at WIN workshop, so please come!

For now, imagine the following problem. There are hundreds of people in a large room talking to each other. You can hear what everybody is saying, but you don’t know who’s talking to whom. How could you figure it out, based on the what they are saying? If one person says something, and another responds, we can usually tell that it fits with the original statement somehow. We quantify this intuition using information theory. If we interpret both people’s utterances as arbitrary signals, then we can use very general information theoretic tools to tell us if one signal is predictable from the other. I.e., one person’s statements are more predictable if we know what the other person is saying. 

Sadly, I was not born with a sensical one-word last name, leading to various problems throughout life. For googling purposes, the all-too-common misspellings:
GV Steeg
Greg Steeg
Greg VerSteeg

A busy month


Next month on April 9 I’ll give a talk for UC Irvine’s AI/ML seminar. The next week I’ll fly to Lyon, France for WWW (the World Wide Web conference). I’ll be giving a talk at a workshop before WWW called “Making sense of micro posts”. Then I’ll present my and Aram Galstyan‘s paper “Information Transfer in Social Media” at WWW.

In the meantime, I hope to finish a paper for UAI while continuing to teach “Physics and Computation”. After all that, perhaps a visit this summer to a research group in Singapore? More details on that later.

I’m teaching a graduate course this term at USC with Aram Galstyan called “Physics and Computation”. Basically, it shows how we can use statistical physics to understand problems in computer science. Hopefully I’ll post some details about the class later.  For now, here’s the main text we’re using: Information, computation, and physics.

In a classic two-birds-one-stone maneuver, we’re thinking about some projects that the students could do that relate to the big question: what can we do with a D-Wave chip? Both 2-SAT instances and spin glass (or LDPC) codes are exactly representable as Ising models that can be solved on the chip. Conveniently, these are both problems we’ve been studying in detail in the course: their phase transitions, average complexity, best classical algorithms, etc.  Here’s a toast to a hopefully synergistic term.


Get every new post delivered to your Inbox.

Join 90 other followers