I’ve been collecting readings for a new project focused on HCI as it pertains to organizing and activism. So far a quick skim as yielded some interesting insights and questions. Here is a link to a prelim bibliography of relevant readings [added 12/26/2017 @817PM EST – some entries are incomplete]. I will update the file on the semi-regular basis in case there is any interest in resource sharing. I’ve been working my way through these texts with an eye towards the following:
- activists unique needs/constraints (there are many)
- nature of the collaborative relationship (I like the idea that contestational design requires that designers/researchers “take sides on contentious social issues” as a necessary part of their work)
- key artifacts like TXTmob, Dialup Radio, resist.org, Protest.net, Indymedia, and many others (case studies coming if they don’t already exist)
- key methodologies (content analysis of social media seems common / embedding in volatile places and critical making workshops are less common)
- key theoretical influences, such as: Mouffe, Laclau & Mouffe, and Habermas, and
- generally interesting bits of text (always part of my approach… identify things that are interesting for reasons that aren’t immediately obvious)
I’m eager to continue this project and interested in potential collaborations with others working in this area.
Looking forward to CHI2018 in Montreal, QC, Canada! This year I am humbled to be first author on one full paper and a contributing author on a second full paper. Two full papers! Huzzah! The first author paper examines the concept of the theory-practice gap as a generative metaphor. Here is the abstract:
The theory-practice gap is a well-known concept in HCI research. It provides a way of describing a space that allegedly exists between the theory and practice of the field, and it has inspired many researchers to propose ways to “bridge the gap.” In this paper, we propose a novel interpretation of the gap as a generative metaphor that frames problems and guides researchers towards possible solutions. We examine how the metaphor has emerged in HCI discourse, and what its limitations might be. We raise concerns about treating the gap as given or obvious, which could reflect researchers’ tendencies to adopt a problem solving perspective. We discuss the value of considering problem setting in relation to the theory-practice gap, and we explore Derrida’s strategy of “reversal” as a possible way to develop new metaphors to capture the relationship between theory and practice. (https://doi.org/10.1145/3173574.3174194)
I’m excited to talk about this ongoing project and discussing its potential with other members in the HCI research community. Onward!
Theory in HCI research appears to be of interest to a number of researchers working in the field. Theory use, which refers to the different roles or functions theory may play in scholarly research or publishing, is one way of exploring the topic, but, in my view, neither topic has been framed as an HCI problem.
Each has been framed as a problem of maturity (or, more accurately, one of immaturity) and, perhaps more recently, as a problem of identity. But these framings transcend the field of HCI research. They are (and have been) relevant to many other academic disciplines.
To the extent that HCI is grappling with its maturity (or immaturity) and/or its identity as an intellectual community, theory and theory use are relevant topics of study. But they have not been formulated or engaged with in terms of human-computer interaction. Such a formulation will be a necessary, good step forward in the discourse.
Research has the potential to move in many different directions. There are constraints, sure. But regardless of where one starts, multiple paths reveal themselves at each step. Choosing a path is crucial for making progress. Moreover, revisiting and refining the intentions motivating one’s travels down a particular path is important. There is always value in asking why we’re doing the work we’re doing. Asking and answering this question is like steering the rudder on a boat.
Great news! We sent the final author proof of our article Schön’s Intellectual Legacy: A Citation Analysis of DRS Publications (2010-2016) back to the copyeditors for publication at Design Studies, which is one of the premier journals in the design field.
The article should be online within a week or so of receiving the final proof (yesterday). So, hopefully you’ll soon be able to read the outcome of a project conceived at the 2015 EAD conference, initially published at the DRS2016, and extensively revised and submitted for Design Studies this year.
I’m pleased to add that Refseer, whose origin story begins with Citeseer, was recently rekindled, and we’re going to be exploring ways to apply and expand it within the scope of our citation analysis project.. Onward!
Most of my blog entries are announcements of publications or generic thought pieces about topics of interest. One thing I’d like to do differently going forward is keeping track of current research projects here at Penn State University’s Center for Human-Computer Interaction (C4HCI) as way of providing some insight and value to folks outside of my network; especially members of community organizations, local government, and industry.
Community Data & Water Quality. A current project underway here at C4HCI has to do with water quality data as a kind of community data around which different stakeholders could organize and act. There are already several groups in State College collecting and analyzing water quality data, and our working assumption is that this data could be made accessible and interpretable to a wider group of residents.
In my view, and in an ideal case, the outcome of such would be a more data literate citizenry confident and capable to engage with local government around water quality (and other) policy and decision-making.
Tonight we held the first meeting of what we hope will be a series of conversations and workshops with folks who work for and with different water quality collection groups in Centre County, including: ClearWater Conservancy, PaSEC, WRMP, and Trout Unlimited. The goal of the meeting was to share our vision for a possible project built around water quality data and to engage in a meaningful conversation about what a collaboration between our groups could look like. The meeting was great. Our group learned a lot about the kinds of data these different organizations collect and about some of the barriers to sharing/using the data that we had not yet considered. We even identified a possible opportunity for supporting (one of) their efforts to build out an online resource for community members to learn more about water quality issues.
Looking forward to more!
This year was a good one for CHI rebuttal writing. I say that not knowing whether our rebuttal swayed any of the reviewers one way or another. But we took a different approach for this year’s CHI reviews than we have in year’s past. This year, we made changes to our paper as we wrote the rebuttal. Changing the paper became a way to think through the viability and possibility of each critique, and the rebuttal became (primarily) a record of changes already made to the submission. It may not be an approach for everyone, but I totally recommend trying it to see whether and how it works. And, I’d be curious to hear from others who take this approach when writing rebuttals (with short turnaround times) about how it has worked!
i just got an email from Amazon letting me know what’s new with Alexa! I read the very first sentences,
Life is unpredictable. Let me help.
, and i wondered whether Alexa (and other things like it) might be slowly chipping away at my capacity to deal with unpredictability. Do I cope with it less effectively? Do I get more frustrated when things don’t go according to plan?
If I had to identify one thing I do well in the classroom, it’s this: I create an environment where it’s the norm to take risks, fail, and explore half-baked ideas.
My guess is that I model all three of these things. I fail up pretty much all the time, and I’m in the habit of sharing ideas before they’re well formed. And I think that’s great.
How else does an idea become fully formed or baked unless it’s subjected to questioning and critique from a smart group of students? Yes, it feels awkward, but I think it’s a mistake to wait until you have something to say. How will you know when to stop waiting if not by sharing what you have so far and learning from the reactions? The sooner ideas make it into a conversation the sooner they become stronger.
Lately, I’ve developed a strong interest in children’s books (cf. Martin Salisbury’s research) and the important role they play in shaping crucial perceptions and actions later in life (e.g. seeing artificial constructs as natural or treating certain subjects as taboo). There are two related to the concepts of idea sharing and failing up that are worth a read. I hope you find them as interesting and important as I do:
Some additional good news to report. I submitted an abstract for a short paper to the upcoming IASDR conference in Cincinnati about some early-stage research that I’m working on with Erik Stolterman, and the abstract has been accepted! So now we’re writing the short paper and creating a poster to present at the conference.
Here is the abstract we submitted:
Scholars in a variety of academic disciplines have studied the peer review process. There are examinations of the biases that pervade peer review (Lee, Sugimoto, Zhang, & Cronin, 2013). Other studies propose tools or methods that might be useful for improving or standardizing the peer review process (Hames, 2008; Onitilo, Engel, Salzman-Scott, Stankowski, & Doi, 2013). Still others examine the kinds of criteria that ought to be relied upon in peer review processes, and in some cases these criteria are widely known and agreed upon. In the natural sciences, for example, we might say that there is a relatively stable set of criteria that can be used to assess the rigor, relevance, and validity of a scientific knowledge contribution. In this paper, our aim is to examine the process of peer review as it pertains to research through design. We aspire to build an understanding of the criteria scholars use when a design or prototype is the main contribution. How do reviewers evaluate designs as knowledge contributions? Is there any uniformity or stability to the review criteria? Are criteria from other fields (e.g. scientific criteria) used to evaluate designs? Toward this end, we report the outcome of a survey conducted with a group of meta-reviewers (n=15) from the design subcommittee for the 2017 Computer-Human Interaction (CHI) Conference, which is the flagship conference in our field of expertise. The design subcommittee reviews papers that “make a significant designerly contribution to HCI [including but not limited to] novel designs of interactive products, services, or systems that advance the state of the art.” Our findings suggest that there is little agreement on a common set of criteria for evaluating research through design.
I look forward to sharing more as this important project moves forward!