Understanding Mediated Practices: Combining Ethnographic Methods with Blog Data to Develop Insights

Share Share Share Share Share[s2If !is_user_logged_in()] [/s2If][s2If is_user_logged_in()] [s2If is_user_logged_in()]

If you cannot find a way to use software that will make a useful archive of blog content, the simplest way to automate the download of a large number of web links is to create a web page using a free online service such as Google Sites or WordPress.com. First, assemble the list of URLs that you wish to download, such as links to each blog index page or individual blog post. Once you have this list, use Excel’s text concatenation function to transform the links into valid HTML code by using a formula similar to the one below. If the links are in column A, the formula in cell B1 would read:

=CONCATENATE(“ <a href=”, , CHAR(34), A1, CHAR(34), “>”, A1, “ </a>”)

Once you fill row B with this formula, you will have a list of web links in HTML format that can be copied and pasted into the web page you made on Google Sites or WordPress.com. From there, use a program such as SiteSucker for Mac to download the contents of each web link. Be careful when setting the options so that you get the correct amount of data. In particular pay attention to settings that allow you to specify the number of levels or layers of link hierarchy that you want to retrieve. If you are studying a site with relatively few links to outside content, it might make sense to download one extra level of content. For example, this could allow you to archive other blog posts and websites referenced in links in posts in the focal blog. On the other hand, if your site is large or complex, downloading links that lead away from the original domain may result in an unmanageable mountain of data; for your first round of analysis we suggest that you download only data (text and images) that appear in posts on your target domain. After some time, the automated download software will produce a set of separate HTML files. Make a backup copy of these HTML files before proceeding with the next steps.

Because nearly all web pages have some repeated content, such as a text or graphic header, navigation, search functions, and advertising, in most cases the downloaded files will benefit from some further processing to remove repetitive content. (If not, you may find yourself wondering why “search,” “next,” or “author” are the most commonly used words in your dataset.) Using a computer tool called grep, which can be likened to a much more powerful version of the familiar search and replace function found in a word processing program, is the fastest way to do this across many text files. If you do not possess skills using grep or HTML, or wish to gain them, we recommend you bring on outside help. A savvy computer science student can make quick work of the next steps.

Using grep, compose and test a search expression that will strip out the non-desired repetitive content (headers, advertisements, navigation, and so on) from your web page. Most blogging platforms generate well-organized and commented code, often divided with the “<div>” html tag, so it is relatively easy to identify which sections you need to strip out and devise a grep expression accordingly. For example, the header section of most HTML files is used for things such as advertising tracking and will appear on every HTML page you have downloaded. It is likely this section is not needed for your analysis. The precise grep string you would use will vary with the site, but essentially you want to tell the computer to look for a string of text that starts and ends the section. Often HTML code will be commented to denote the sections of the page. These comments, which start with “<!–”, can be useful as anchors for a grep expression.

Once you have tested the grep expression on several files, you can apply it to all of the HTML files you have collected using the command line or with a batch search-and-replace function in an application such as TextWrangler for Mac. If you find that an error in your grep code or an irregularity in the HTML files you have collected has caused a problem, you can revise as necessary and run the script again on your backup. Once you have a clean set of files you can proceed with manual analysis or put the files in a database, which can help ease the process of retrieval and comparison.

For database software, we recommend DevonThink Pro Office, which runs only on Mac OS, for several reasons.

  • Built-in RSS feed support. RSS stands for really simple syndication… which can help automate the addition of new blog posts over time
  • An artificial intelligence engine. Select one database item, and DevonThink can find database items that are linguistically similar to it. A central finding in our research—and the centerpiece of a diagram explaining our theory—is, in emic terms, the “landing strip.” This term refers to an area near the entry to your home where you stash keys, mail, shoes, and outerwear to keep them from “contaminating” the rest of your living space. We identified blog posts on the topic of a landing strip and used this function to find other posts that used similar language, which helped us to extend and test our theory and also allowed us to find the most illustrative examples for our readers without reading the thousands of posts containing the term “landing strip.”
  • Sophisticated searches. In addition to Boolean searches (searching for one word AND another word OR some other word but NOT this word), DevonThink can find records containing words that are near each other. For example, to find evidence to back up a preliminary finding that Apartment Therapy linked the state of cleanliness to the feeling of calm, we looked for instances of the word “calm” near the word “clean.” These searches can be saved and as new information comes into the database they will automatically be included in the search results.
  • Redundancy. DevonThink has a built-in backup function.
  • Export tools. Records in DevonThink can be easily exported to a variety of formats, including text, for further processing in specialized text analysis software.
  • AppleScript support. AppleScript is a macro instruction language that allows repetitive tasks to be automated. We used an AppleScript and grep to change the creation date of each record in our database to match the date and time it was originally published on the AT blog. This allowed us to easily sort chronologically and divide the posts into groups according to date.

A primary disadvantage of DevonThink Pro Office is that its power and sophistication taxes a computer’s resources. If you plan to work with a database as large or larger than ours, we recommend using the fastest computer you can and making sure it has as much random access memory installed as possible. In particular this will speed complex searches. A second disadvantage is that DevonThink Pro Office is not designed for collaboration. While it is possible to save the database file using a web sharing service such as Dropbox, only one user can open and work with the database file at a time. If is opened by two or more users at once, it is highly likely that the file will become corrupted and you will lose data. To solve this problem, we developed a system of “checking out” the file for access via a quick email notification, and as a secondary measure we used the built-in label function of Mac OS X to turn the file red to show that it was not available for use. When one user was done working with the file, we changed its Finder label to green to show it was available for the use of others. While other database or qualitative analysis software can be used to the same effect, but we found that the combination of DevonThink’s ease of use and power makes it particularly well suited to researchers who are beginning to work with large datasets.

SYNTHESIZE. GO BACK AND BACKWARDS. REPEAT.

Social science knowledge is created through an iterative process that requires the investigator to go back and forth between data and provisional understandings of the phenomenon. Multi-method inquiries necessitate another layering of analysis, to merge and synthesize various types of data at different levels of specificity. This back-and-forth shifting of focus is essential because it broadens the analysis from emic to etic and synthesizes multiple interpretations, including the researcher’s own observations, the subjects’ probed narratives, and the subjects’ performed narratives (such as blog content). It also broadens the temporal and spatial spectrum of inquiry from the resource restricted boundaries of ethnographic inquiry.

While there is no universal rule to where to start, our subjective opinion is to begin by analyzing the ethnographic data first. This, we feel, holds closer to the spirit of ethnographic inquiry by allowing the researcher to begin to frame analysis based on the words and observed actions of the research participants. This can help prevent the researcher from relying too much on the somewhat decontextualized analysis a machine can make. While powerful in extracting patterns, machine generated analysis has a significant blind spot in terms of losing some of the nuances and context (Maxwell 2013). We advocate an approach where insights from interviews, content analysis, coding, or other ethnographic methods comes first, before the use of automated tools.

Once some provisional understanding is achieved, the researcher can perform analysis on the entire body of data to test and develop emergent theories. No single software is fit for all uses; in our case we chose the Stanford Part-of-Speech Tagger (Toutanova et al. 2003) because its capabilities mapped our provisional findings. We observed that the categories of objects, meanings, and doings could be roughly mapped onto parts of speech: objects were typically represented by nouns, meanings by adjectives, and doings by verbs. Using the Stanford Part-of-Speech Tagger allowed us to efficiently find the most commonly used nouns, adjectives, and verbs in our corpus. Note that the software does not do the analysis for you; as a check measure, we independently coded the results as objects, meanings, and doings, and discussed inconsistencies until they were mutually resolved.

After you have invested the time in getting your data into usable shape, it is worth investigating other analytic tools. For example, if your analysis aims to show the change in time of personal relationships, network analysis may be useful. To gain a sense of difference in how people are using words over time, software tools such as WORDij can quickly reveal shifts. If you are interested in the emotional content of a corpus of text, the Lingusitic Inquiry and Word Count software can be of use. Riopelle (2013) discusses how ethnographers can use these specific tools in an applied context. Humphreys (2010) utilizes a similar multi method approach, supplementing qualitative coding with computer-assisted content analysis to trace the changes in discourses on gambling in American newspapers.

CONCLUSION

As Warde (2005) has famously argued, consumption is only one moment in practice. Taken together, ethnography and social practice theory have the potential to illuminate both applied and academic studies of consumption. For the consultant, applying the social practice framework can show how integral practice is to the purchase and use of goods and services (Korkmann 2006), therefore offering a theoretical framework that is both substantial enough to contain the complexity of the world, but also comprehensible enough to be used by members of corporate teams regardless of previous exposure to theory. For academic audiences, applying social practice theory is one way to address the challenge of connecting potentially related research in different contexts—in effect, to show how practices intersect and draw from one another (Shove et al 2012). For example, we showed how a key resource for 21st century consumers recruited into the Apartment Therapy taste regime was the practice of soft modernism, a practice that had roots in postwar cultural changes. For those visiting the Apartment Therapy site, soft modernism provided not only a source of objects—teak coffee tables, sleek lamps, Danish chairs—but also a source of meaning. Together, objects and meanings are critical resources that become inseparably integrated with the doings (the third element of practice) of soft modernism in a way that, would become closely associated with both Apartment Therapy and our participants’ own putatively personal sense of style (Arsel and Bean 2012). Breaking down practice into these three constituent parts and showing how practice performances (Shove et al 2012) arise from the patterned combination of objects, meanings, and doings could lend some interchangeability to otherwise disparate analyses.

Mediated representations of practice can be an ideal site to confirm and understand the workings of practices observed by the ethnographer in the field. While a netnographic (Kozinets 2010) or critical visual (Schroeder 2006) approach can be used to understand mediated representations, our approach differs in applying the three-part practice heuristic of objects, meanings, and doings. Mapping representations of practice in these three categories can provide insights into the material workings of a practice and may provide hints on how a particular practice may change in the future.

ACKNOWLEDGEMENTS

The authors thank Maxwell Gillingham-Ryan for allowing them to use the Apartment Therapy site content for data analysis. Funding for this research project was provided by the Fonds Québécois de la Recherche sur la Société et la Culture (FQRSC).

Jonathan Bean is Assistant Professor of Markets, Innovation and Design at Bucknell University. He has done research for Microsoft and Intel, and worked in sales and marketing at a pioneering retailer of green building supplies. His areas of expertise are the home, consumer culture, and ethnography.

Zeynep Arsel is Associate Professor of Marketing at Concordia University. She is the recipient of the Petro Canada Young Innovator Award, and the Sidney J. Levy Award. Her research focuses on consumption using sociological, anthropological and historical methods and with particular focus on space, taste, and exchanges.


REFERENCES CITED

Arsel, Zeynep and Jonathan Bean
2013 “Taste Regimes and Market-Mediated Practice.” Journal of Consumer Research 39(5):899-917.

Arsel, Zeynep and Xin Zhao
2013 “Blogs.” In The Routledge Companion to Digital Consumption. Russell Belk and Rosa Llamas, eds. Pp 53-61. New York: Routledge.

Gebhard, David
1995 “William Wurster and His California Contemporaries: The Idea of Regionalism and Soft Modernism.” In An Everyday Modernism: The Houses of William Wurster. Marc Treib, ed. Pp. 164-183. Berkeley, CA: University of California Press.

Golec, Michael J.
2005 “Martha Stewart Living and the Marketing of Emersonian Perfectionism.” Home Cultures 3(1):5-20.

Humphreys, Ashlee
2010 “Semiotic Structure and the Legitimation of Consumption Practices: The Case of Casino Gambling.” Journal of Consumer Research 37 (3):490-510.

Korkmann, Oskar
2006 “Customer Value Formation in Practice: A Practice-Theoretical Approach.” Ph.D. diss., Swedish School of Economics and Business Administration.

[/s2If]

Pages: 1 2 3

Leave a Reply