Ad verba per numeros
Artículos
Wednesday, June 27, 2012, 08:18 AM
In my previous post I was deliberately caustic on the matter of electoral prediction from Twitter data.After writing that I decided to be a bit more constructive, try to see the glass half full instead of half empty, and improve the advice I had provided there (especially regarding baselines).Hence, I've prepared a new paper where I conduct a meta-analysis on Twitter electoral predictions (note: only those made in scholar papers) to reach some conclusions:Wednesday, June 27, 2012, 08:18 AM
- With regards to predictions based on raw counts:
- It is too dependent on arbitrary decisions such as the parties or candidates to be considered, or the selection of a period for collecting the data.
- Its performance is too unstable and strongly dependent on such parameterizations, and
- Considering the reported results as a whole it seems plausible that positive results could have been due to chance or, even, to unintentional data dredging due to post hoc analysis.
- With regards to predictions based on sentiment analysis:
- It is unclear the impact that sentiment analysis has in Twitter-based predictions. The studies applying this technique are fewer than those counting tweets and the picture they convey is confusing to say the least.
- However, taking into consideration that even naïve sentiment analysis seems to outperform a reasonable baseline it is clear that further research is needed in that line
- Both approaches share a number of weaknesses:
- All of them are post hoc analysis.
- Proposed baselines are too simplistic.
- Sentiment analysis is applied with naïveté since commonly used methods are slightly better than random classifiers and fail to catch the subtleties of political discourse.
- All of the tweets are assumed to be trustworthy when it is not the case.
- Demographics bias is neglected even when it is well known that social media is not a random sample of the population.
- Self-selection bias is also ignored although it is well known that supporters are much more vocal and responsible of most of the content.
- Period and method of collection: i.e., the dates when tweets were collected, and the parameterization used to collect them.
- Data cleansing measures:
- Purity: i.e., to guarantee that only tweets from prospective voters are used to make the prediction.
- Debiasing: i.e., to guarantee that any demographic bias in the Twitter user base is removed.
- Denoising: i.e., to remove tweets not dealing with voter opinions (e.g. spam or disinformation) or even users not corresponding to actual prospective voters (e.g. spammers, robots, or propagandists).
- Prediction method and its nature:
- The method to infer voting intentions from tweets.
- The nature of the inference: i.e., whether the method predicts individual votes or aggregated vote rates.
- The nature of the prediction: i.e., whether the method predicts just a winner or vote rates for each candidate.
- Granularity: i.e., the level at which the prediction is made (e.g. district, state, or national).
- Performance evaluation: i.e., the way in which the prediction is compared with the actual outcome of the election.
Back Next