Texifter was the first company to join as a paying customer in the alpha “Snapshot” offering from Gnip. You can still take part in that alpha by submitting a request for a free estimate of a snapshot from Twitter’s complete history. This is, however, a very fast-moving landscape for for social # bigdata. We are quickly transitioning from the alpha “Snaphot” tests to the beta of a cradle-to-grave system for building estimates for the cost of text analytic projects that feature either the real-time day-forward, Gnip-enabled PowerTrack (the Twitter fire hose), or the new historical PowerTrack. So if you have ever wished you could go back in time and collect all the tweets from an epic moment in history, your wish just came true. Contact us if you have any questions and submit a request for a free estimate today.
[Originally posted May 10, 2011]
We did it! The free, open source, Web-based, university-hosted, FISMA-compliant “Coding Analysis Toolkit” CAT recorded its one millionth coding choice. Pretty much all the credit goes to Texifter CTO and chief CAT architect Mark Hoy who has put in many paid (and unpaid) hours making sure CAT is reliable, usable, & scalable. Texifter Chief Security Officer Jim Lefcakis also played a key role ensuring the hardware and server room were maintained at the highest level of reliability and security. In honor of this milestone, I have been digging through my unpublished papers looking for material that explains in more detail where CAT, PCAT, DiscoverText, QDAP & Texifter come from. This post is the first in a series about the particular approach to coding text we have come to call the “QDAP method.”
Large political text data collections are coded and used for basic and applied research in social and computational sciences. Yet the manual annotation of the text—the coding of corpora—is often conducted in an ad hoc, inconsistent, non-replicable, invalid and unreliable manner. Even the best intentions to create the possibility for replication can, in practice, confound the most ardent followers of the creed “ Replicate, Replicate.” While mechanical, process, documentary, and other challenges exist for all approaches, practitioners of qualitative or text data analysis routinely profess to greater, even insurmountable, barriers to re-using coded data or repeating significant analyses. There are diverse approaches to coding text. They tend to be hidden away in small niche sub-fields where knowledge of them is limited to a small research community, a project team, or even a single person. While researchers classify text for a variety of reasons, it remains very difficult, for many counter-intuitive, to share these annotations with other researchers, or to work on them with partners from other disciplines for whom the coding may serve an alternate purpose. A change in the way the researchers think about, conduct, and share coded political corpora is overdue.
Coding is expensive, challenging, and too often idiosyncratic. Training and retaining student coders or producing algorithms capable of tens of thousands of reliable and valid observations requires patience, funding, and a framework for measuring and reporting effort and error. Given these factors, it is not surprising that a proprietary model of data acquisition and coding still dominates the social sciences. Despite the important role for the social in social science, researchers guard “their” privately coded text, even the raw data, fearing others will beat them to the publication finish line or challenge the validity of their inferences. The competitive approach to producing and failing to share annotations disables intriguing and highly scalable collaborative social research possibilities enabled by the Internet.
Researchers should seek to enhance and modernize their architecture for large-scale collaborative research using advanced qualitative methods of data analysis. This will require working out and attaining widespread acceptance of Internet-enabled data sharing protocols, as well as the establishment of free, open source platforms for coding text and for sharing and replicating results. We believe that when utilized in combination, “The Dataverse Network Project” and the “Coding Analysis Toolkit” (CAT) represent two important steps advancing that effort. Large-scale annotation projects conducted on CAT can be archived in the Dataverse and as a result will be more easily available for replication, review, or re-adjudication of their original coding. In Part Two of the Series “Coding Text the QDAP Way,” we’ll say more about the role of scholarly journals advancing this practice of re-using datasets.
[Originally posted May 14, 2011]
In Part One of the series “Coding Text the QDAP Way,” I wrote about the problem of idiosyncratic annotation and the lack of diverse, interesting and re-usable annotated data sets. Providing data for replication (when possible) is a requisite for step scientific approach. An important aspect of this is effort is a follow up on the agreements that were made starting in the 1990s among editors of major research journals to require replication datasets and sharing of the specifics of data coding and computer syntax. This work is now well advanced on the quantitative data sharing frontiers. Developing such an agreement for qualitative data research applications and implementing it consistently among a wide-reaching community of researchers is no simple task.
Sharing raw and coded political corpora will lead to better manual and automated text mining and analysis in political science. This is an epoch of highly accessible digital text collections. Blog posts, wikis, YouTube comments, and the like, as well as the full range of digitized traditional media, are vast sources of potentially important political data in text format. A new approach to coding and sharing annotations might help to eviscerate the prevailing perception of a zero-sum game in research, resulting in many new basic and applied research opportunities for political scientists. The manual annotation of text is a nexus for collaboration by political scientists with computer scientists, and with researchers in allied social sciences as well as in fields such as journalism, literary analysis, library science, and education where the rigorous interrogation of text is a well-established tradition.
In particular, researchers in computer science possess the tools, repositories, and methods necessary for managing studies of millions of documents over time. Just like search engines, in a very short time we can expect these emergent human language tools to become irreplaceable elements of the researcher’s electronic desktop. The next generation of language tools will be built with the “ground truth” support of high-quality coding and evaluation studies. Many researchers from a variety of disciplines stand to benefit from reliably recorded, publicly available, transparent, large-scale annotations. These collections can be produced by properly equipped and trained coders, as well as by active and machine-learning algorithms developed by computer scientists. Yet very few researchers in any discipline can say with confidence that they know where to acquire or how to produce reusable annotated corpora with widespread, multi-disciplinary appeal. Even fewer could imagine freely sharing those hard-earned text annotations with other members of a research community or publishing them on the Web to attract more diverse and sustained scholarly attention.
There is some evidence that making data available increases citation. Although a strong tradition is emerging among leading social science journals whereby scholars post their statistical data and models in repositories for those who would replicate their experiments and calculations, the same cannot currently be said about text annotations, other forms of qualitative work, and even raw text datasets. As a result, there is a dearth of well-coded contemporary and historical text datasets. This is only partly due to fact that the manual annotation of text can be conceptually very difficult, if not a bit controversial, expensive, and too often unsuccessful. It is often dreary work, a characteristic that further encourages the use of unsupervised machine annotation when possible. More fundamentally, however, only limited guidance exists in the scholarly literature about how best to recruit, train, equip, and supervise coders to get them to produce useful annotations that serve multiple research agendas in divergent disciplines.
As Eduard Hovy (Computer Science, University of Southern California-Information Sciences Institute) regularly points out, researchers need a formal science of annotation focused on cross-disciplinary text mining activities. Carefully and transparently coded corpora are a viable bridge to collaboration with computer science and computational linguistics and can open up new possibilities for large-scale text analysis. In the third and final part of this series, we look at the quest for the elusive “gold standard” in human annotation.
[Originally posted May 17, 2011]
Researchers interested in large text collections and their itinerant coders tend to muddle through with limited collaborative, cross-disciplinary resources upon which to draw. The generic criteria for high-quality codebook construction and effective coding are underdeveloped, even as the tools and techniques for measuring the limits of manual or machine coding grow ever more sophisticated. In that paradox there may be the seed of a partial solution to some of these issues.
The ability to quickly and easily pre-test coding schemes and produce on-the-fly displays of coding inconsistencies is one way to more uniformly train coders to perform reliably (hence usefully) while ensuring a satisfactory level of valid observations. By the same token, the ability to permit an unlimited number of users to review or replicate all the coding and adjudication steps using a free, web-based platform would be a large and bold step onto our methodological and metaphorical bridge. What are needed are more universal annotation metrics, a standard lexicon, and widely shared, semi-automated coding tools that make the work of humans more useful, fungible, and durable. Ideally, these tools would be interoperable, or combined in a single system. The new system would allow human coders to create annotations and allow other experts to efficiently examine, influence, and validate their work.
At a deeper level, this calls for much better and more transparently codified approaches to training and deploying coders—an annotation science subfield—so that a more coherent and collaborative research community can form around this promising methodological domain. Investigators in the social sciences use reliably coded texts to reach inferences about diverse phenomena. Many forms of public-sphere discourse and governmental records are readily amenable to coding; these include press content, policy documents, speeches, international treaties, and public comments submitted to government decision-makers, among many others. Systematic analysis of large quantities of these sorts of texts represents an appealing new avenue for both theory building and hypothesis testing. It also represents a bridge across the divide between qualitative and quantitative methodologies in the social sciences. These large text datasets are ripe for mixed-methods work that can provide a rich, data-driven approach both to the macro and micro view of large-scale political phenomena.
Traditionally, social scientists working with text use a variety of qualitative research methods for in-depth case studies. For many legitimate and pragmatic reasons, these studies generally consist of a small number of cases or even just a single case. As Steven Rothman and Ron Mitchell note, the reliability of data drawn from qualitative research comes under greater scrutiny, as increased dataset complexity requires increased interpretation and, subsequently, leads to increased opportunity for error. The case study method is plagued by concerns about limitations on its external validity and the ability to reach generalized inferences. With the proliferation of easily available, large-scale digitized text datasets, an array of new opportunities exist for large-n studies of text-based political phenomena that can yield both qualitative and quantitative findings. More to the point, high-quality manual annotation opens up the possibility for cross-disciplinary studies featuring collaboration between social and computational scientists. This second opportunity exists because researchers in the computational sciences, particularly those working in text classification, IR, opinion detection, and NLP, hunger for the elusive “gold standard” in manual annotation.
Accurate coding with high levels of inter-rater reliability and validity is possible. For example, work by the eRulemaking Research Group on near-duplicate detection in mass e-mail campaigns demonstrated that focusing on a small number of codes, each with a clear-cut rule set, has been able to produce just such a gold standard. Reliably coded corpora of sufficient size and containing consistently valid observations are essential to the process of designing and training NLP algorithms. We are likely to see more political scientists using methodologies that combine manual annotation and machine learning. In short, there are exciting possibilities for applied and basic research as techniques and tools emerge for reliably coding across the disciplines. To unleash the potential for this interdisciplinary approach, a research community must now form around the nuts and bolts questions of what and how to annotate, as well as how to train and equip the coders that make this possible.
Texifter’s most recent historical Twitter prize winners include three from the United States, one from Great Britain, and one from France. Winners receive Enterprise access to DiscoverText for six months, and Sifter credit for up to three historical Twitter days and 200,000 tweets. The following is a snapshot of the most recent winners and their proposed research projects. Diana Ascher PhD student in the Department of Information Studies at UCLA @dianaascher “Helping Companies Streamline Information” Ascher proposes exploring cultural time orientation by analyzing the Twitter feeds from three news organizations to better understand how “information agents’ cultural backgrounds affect corporate information practice,” and specifically how organizations decide what information to share and when. Ascher hopes the research will help businesses streamline their information activity and routines, and help managers understand “how employees decide what’s important and what’s not.” Stephen Barnard Assistant Professor in the Sociology Department at St. Lawrence University @socsavvy “Better Understanding Journalism via Boston Marathon Bombing Twitter Data” Barnard plans to use Sifter to collect and analyze Twitter data about the 2013 Boston Marathon bombings. He will use Twitter’s PowerTrack filters to conduct a detailed search of Tweets that reported on the bombing, and compare the results to the responses from professional and citizen journalists. “I hope to gain a better understanding of the reporting processes and outcomes emerging from both groups,” Barnard writes, adding that he will use the findings to “highlight the structural relations of the emerging journalistic field.” Oliver Haimson PhD Student in the Informatics Department at University of California, Irvine @oliverhaimson “Analyzing Hashtags” Haimson’s plans to use the prize to analyze the hashtags #nymwars and #mynameis, which were used in 2011 and 2014 to critique Google’s and Facebook’s “real name” policies. He plans to evaluate the Twitter data from these two hashtags “using computational linguistics, qualitative coding, and social network analysis.” Omar Jaafor PhD Student in the Department of Operational Research, Applied Statistics and Simulation at University of Technology of Troyes @lmhasher “Developing Algorithms for Social Networks” Jaafor and fellow researchers will use the prize to continue to develop “clustering and anomaly detection algorithms for social networks in a big data environment.” Wasim Ahmed PhD Student in the Health Informatics Research Group at the University of Sheffield’s Information Department @was3210 ” Responding to Infectious Disease Outbreaks” Ahmed will use his prize to “study how users respond to outbreaks on infectious diseases on social media platforms, such as Twitter.” He plans to use his data towards his PhD “Pandemics and epidemics: User reactions on social media and Web 2.0 platforms.” For more information on the Texifter’s social data offer and text analytics tools, please send us an email firstname.lastname@example.org. Better yet, sign up for a free 30-day trial and start collecting your own social data today.