Category Archives: Placeholders

Quest for Expertise

Will at Work Learning: People remember 10%, 20%…Oh Really?.

This post was mentioned on the mailing-list for the Society for Teaching and Learning in Higher Education (STLHE-L).

In that post, Will Thalheimer traces back a well-known claim about learning to shoddy citations. While it doesn’t invalidate the base claim (that people tend to retain more information through certain cognitive processes), Thalheimer does a good job of showing how a graph which has frequently been seen in educational fields was based on faulty interpretation of work by prominent scholars, mixed with some results from other sources.

Quite interesting. IMHO, demystification and critical thinking are among the most important things we can do in academia. In fact, through training in folkloristics, I have become quite accustomed to this specific type of debunking.

I have in mind a somewhat similar claim that I’m currently trying to trace. Preliminary searches seem to imply that citations of original statements have a similar hyperbolic effect on the status of this claim.

The claim is what a type of “rule of thumb” in cognitive science. A generic version could be stated in the following way:

It takes ten years or 10,000 hours to become an expert in any field.

The claim is a rather famous one from cognitive science. I’ve heard it uttered by colleagues with a background in cognitive science. In 2006, I first heard about such a claim from Philip E. Ross, on an episode of Scientific American‘s Science Talk podcast to discuss his article on expertise. I later read a similar claim in Daniel Levitin’s 2006 This Is Your Brain On Music. The clearest statement I could find back in Levitin’s book is the following (p. 193):

The emerging picture from such studies is that ten thousand hours of practice is required to achieve the level of mastery associated with being a world-class expert – in anything.

More recently, during a keynote speech he was giving as part of his latest book tour, I heard a similar claim from presenter extraordinaire Malcolm Gladwell. AFAICT, this claim runs at the centre of Gladwell’s recent book: Outliers: The Story of Success. In fact, it seems that Gladwell uses the same quote from Levitin, on page 40 of Outliers (I just found that out).

I would like to pinpoint the origin for the claim. Contrary to Thalheimer’s debunking, I don’t expect that my search will show that the claim is inaccurate. But I do suspect that the “rule of thumb” versions may be a bit misled. I already notice that most people who set up such claims are doing so without direct reference to the primary literature. This latter comment isn’t damning: in informal contexts, constant referal to primary sources can be extremely cumbersome. But it could still be useful to clear up the issue. Who made this original claim?

I’ve tried a few things already but it’s not working so well. I’m collecting a lot of references, to both online and printed material. Apart from Levitin’s book and a few online comments, I haven’t yet read the material. Eventually, I’d probably like to find a good reference on the cognitive basis for expertise which puts this “rule of thumb” in context and provides more elaborate data on different things which can be done during that extensive “time on task” (including possible skill transfer).

But I should proceed somewhat methodically. This blogpost is but a preliminary step in this process.

Since Philip E. Ross is the first person on record I heard talk about this claim, a logical first step for me is to look through this SciAm article. Doing some text searches on the printable version of his piece, I find a few interesting things including the following (on page 4 of the standard version):

Simon coined a psychological law of his own, the 10-year rule, which states that it takes approximately a decade of heavy labor to master any field.

Apart from the ten thousand (10,000) hours part of the claim, this is about as clear a statement as I’m looking for. The “Simon” in question is Herbert A. Simon, who did research on chess at the Department of Psychology at Carnegie-Mellon University with colleague William G. Chase.  So I dig for diverse combinations of “Herbert Simon,” “ten(10)-year rule,” “William Chase,” “expert(ise),” and/or “chess.” I eventually find two primary texts by those two authors, both from 1973: (Chase and Simon, 1973a) and (Chase and Simon, 1973b).

The first (1973a) is an article from Cognitive Psychology 4(1): 55-81, available for download on ScienceDirect (toll access). Through text searches for obvious words like “hour*,” “year*,” “time,” or even “ten,” it seems that this article doesn’t include any specific statement about the amount of time required to become an expert. The quote which appears to be the most relevant is the following:

Behind this perceptual analysis, as with all skills (cf., Fitts & Posner, 1967), lies an extensive cognitive apparatus amassed through years of constant practice.

While it does relate to the notion that there’s a cognitive basis to practise, the statement is generic enough to be far from the “rule of thumb.”

The second Chase and Simon reference (1973b) is a chapter entitled “The Mind’s Eye in Chess” (pp. 215-281) in the proceedings of the Eighth Carnegie Symposium on Cognition as edited by William Chase and published by Academic Press under the title Visual Information Processing. I borrowed a copy of those proceedings from Concordia and have been scanning that chapter visually for some statements about the “time on task.” Though that symposium occurred in 1972 (before the first Chase and Simon reference was published), the proceedings were apparently published after the issue of Cognitive Psychology since the authors mention that article for background information.

I do find some interesting quotes, but nothing that specific:

By a rough estimate, the amount of time each player has spent playing chess, studying chess, and otherwise staring at chess positions is perhaps 10,000 to 50,000 hours for the Master; 1,000 to 5,000 hours for the Class A player; and less than 100 horus for the beginner. (Chase and Simon 1973b: 219)

or:

T
he organization of the Master’s elaborate repertoire of information takes thousands of hours to build up, and the same is true of any skilled task (e.g., football, music). That is why practice is the major independent variable in the acquisition of skill. (Chase and Simon 1973b: 279, emphasis in the original, last sentences in the text)

Maybe I haven’t scanned these texts properly but those quotes I find seem to imply that Simon hadn’t really devised his “10-year rule” in a clear, numeric version.

I could probably dig for more Herbert Simon wisdom. Before looking (however cursorily) at those 1973 texts, I was using Herbert Simon as a key figure in the origin of that “rule of thumb.” To back up those statements, I should probably dig deeper in the Herbert Simon archives. But that might require more work than is necessary and it might be useful to dig through other sources.

In my personal case, the other main written source for this “rule of thumb” is Dan Levitin. So, using online versions of his book, I look for comments about expertise. (I do own a copy of the book and I’m assuming the Index contains page numbers for references on expertise. But online searches are more efficient and possibly more thorough on specific keywords.) That’s how I found the statement, quoted above. I’m sure it’s the one which was sticking in my head and, as I found out tonight, it’s the one Gladwell used in his first statement on expertise in Outliers.

So, where did Levitin get this? I could possibly ask him (we’ve been in touch and he happens to be local) but looking for those references might require work on his part. A preliminary step would be to look through Levitin’s published references for Your Brain On Music.

Though Levitin is a McGill professor, Your Brain On Music doesn’t follow the typical practise in English-speaking academia of ladling copious citations onto any claim, even the most truistic statements. Nothing strange in this difference in citation practise.  After all, as Levitin explains in his Bibliographic Notes:

This book was written for the non-specialist and not for my colleagues, and so I have tried to simplify topics without oversimplifying them.

In this context, academic-style citation-fests would make the book too heavy. Levitin does, however, provide those “Bibliographic Notes” at the end of his book and on the website for the same book. In the Bibliographic Notes of that site, Levitin adds a statement I find quite interesting in my quest for “sources of claims”:

Because I wrote this book for the general reader, I want to emphasize that there are no new ideas presented in this book, no ideas that have not already been presented in scientific and scholarly journals as listed below.

So, it sounds like going through those references is a good strategy to locate at least solid references on that specific “10,000 hour” claim. Among relevant references on the cognitive basis of expertise (in Chapter 7), I notice the following texts which might include specific statements about the “time on task” to become an expert. (An advantage of the Web version of these bibliographic notes is that Levitin provides some comments on most references; I put Levitin’s comments in parentheses.)

  • Chi, Michelene T.H., Robert Glaser, and Marshall J. Farr, eds. 1988. The Nature of Expertise. Hillsdale, New Jersey: Lawrence Erlbaum Associates. (Psychological studies of expertise, including chess players)
  • Ericsson, K. A., and J. Smith, eds. 1991. Toward a General Theory of Expertise: prospects and limits. New York: Cambridge University Press. (Psychological studies of expertise, including chess players)
  • Hayes, J. R. 1985. Three problems in teaching general skills. In Thinking and Learning Skills: Research and Open Questions, edited by S. F. Chipman, J. W. Segal and R. Glaser. Hillsdale, NJ: Erlbaum. (Source for the study of Mozart’s early works not being highly regarded, and refutation that Mozart didn’t need 10,000 hours like everyone else to become an expert.)
  • Howe, M. J. A., J. W. Davidson, and J. A. Sloboda. 1998. Innate talents: Reality or myth? Behavioral & Brain Sciences 21 (3):399-442. (One of my favorite articles, although I don’t agree with everything in it; an overview of the “talent is a myth” viewpoint.)
  • Sloboda, J. A. 1991. Musical expertise. In Toward a general theory of expertise, edited by K. A. Ericcson (sic) and J. Smith. New York: Cambridge University Press. (Overview of issues and findings in musical expertise literature)

I have yet to read any of those references. I did borrow Ericsson and Smith when I first heard about Levitin’s approach to talent and expertise (probably through a radio and/or podcast appearance). But I had put the issue of expertise on the back-burner. It was always at the back of my mind and I did blog about it, back then. But it took Gladwell’s talk to wake me up. What’s funny, though, is that the “time on task” statements in (Ericsson and Smith,  1991) seem to lead back to (Chase and Simon, 1973b).

At this point, I get the impression that the “it takes a decade and/or 10,000 hours to become an expert”:

  • was originally proposed as a vague hypothesis a while ago (the year 1899 comes up);
  • became an object of some consideration by cognitive psychologists at the end of the 1960s;
  • became more widely accepted in the 1970s;
  • was tested by Benjamin Bloom and others in the 1980s;
  • was precised by Ericsson and others in the late 1980s;
  • gained general popularity in the mid-2000s;
  • is being further popularized by Malcolm Gladwell in late 2008.

Of course, I’ll have to do a fair bit of digging and reading to verify any of this, but it sounds like the broad timeline makes some sense. One thing, though, is that it doesn’t really seem that anybody had the intention of spelling it out as a “rule” or “law” in such a format as is being carried around. If I’m wrong, I’m especially surprised that a clear formulation isn’t easier to find.

As an aside, of sorts… Some people seem to associate the claim with Gladwell, at this point. Not very surprsing, given the popularity of his books, the effectiveness of his public presentations, the current context of his book tour, and the reluctance of the general public to dig any deeper than the latest source.

The problem, though, is that it doesn’t seem that Gladwell himself has done anything to “set the record straight.” He does quote Levitin in Outliers, but I heard him reply to questions and comments as if the research behind the “ten years or ten thousand hours” claim had some association with him. From a popular author like Gladwell, it’s not that awkward. But these situations are perfect opportunities for popularizers like Gladwell to get a broader public interested in academia. As Gladwell allegedly cares about “educational success” (as measured on a linear scale), I would have expected more transparency.

Ah, well…

So, I have some work to do on all of this. It will have to wait but this placeholder might be helpful. In fact, I’ll use it to collect some links.

 

Some relevant blogposts of mine on talent, expertise, effort, and Levitin.

And a whole bunch of weblinks to help me in my future searches (I have yet to really delve in any of this).

Gender and Culture

A friend sent me a link to the following video:

JC Penney: Beware of the Doghouse | Creativity Online.

In that video, a man is “sent to the doghouse” (a kind of prison for insensitive men) because he offered a vacuum cleaner to his wife. It’s part of a marketing campaign through which men are expected to buy diamonds to their wives and girlfriends.

The campaign is quite elaborate and the main website for the campaign makes interesting uses of social media.

For instance, that site makes use of Facebook Connect as a way to tap viewers’ online social network. FC is a relatively new feature (the general release was last week) and few sites have been putting it to the test. In this campaign’s case, a woman can use her Facebook account to connect to her husband or boyfriend and either send him a warning about his insensitivity to her needs (of diamonds) or “put him in the doghouse.” From a social media perspective, it can accurately be described as “neat.”

The site also uses Share This to facilitate the video‘s diffusion  through various social media services, from WordPress.com to Diigo. This tends to be an effective strategy to encourage “viral marketing.” (And, yes, I fully realize that I actively contribute to this campaign’s “viral spread.”)

The campaign could be a case study in social marketing.

But, this time, I’m mostly thinking about gender.

Simply put, I think that this campaign would fare rather badly in Quebec because of its use of culturally inappropriate gender stereotypes.

As I write this post, I receive feedback from Swedish ethnomusicologist Maria Ljungdahl who shares some insight about gender stereotypes. As Maria says, the stereotypes in this ad are “global.” But my sense is that these “global stereotypes” are not that compatible with local culture, at least among Québécois (French-speaking Quebeckers).

See, as a Québécois born and raised as a (male) feminist, I tend to be quite gender-conscious. I might even say that my gender awareness may be somewhat above the Québécois average and gender relationships are frequently used in definitions of Québécois identity.

In Québécois media, advertising campaigns portraying men as naïve and subservient have frequently been discussed. Ten or so years ago, these portrayals were a hot topic (searches for Brault & Martineau, Tim Hortons, and Un gars, une fille should eventually lead to appropriate evidence). Current advertising campaigns seem to me more subtle in terms of male figures, but careful analysis would be warranted as discussions of those portrayals are more infrequent than they have been in the past.

That video and campaign are, to me, very US-specific. Because I spent a significant amount of time in Indiana, Massachusetts, and Texas, my initial reaction while watching the video had more to do with being glad that it wasn’t the typical macrobrewery-style sexist ad. This reaction also has to do with the context for my watching that video as I was unclear as to the gender perspective of the friend who sent me the link (a male homebrewer from the MidWest currently living in Texas).

By the end of the video, however, I reverted to my Québécois sensibility. I also reacted to the obvious commercialism, partly because one of my students has been working on engagement rings in our material culture course.

But my main issue was with the presumed insensitivity of men.

Granted, part of this is personal. I define myself as a “sweet and tendre man” and I’m quite happy about my degree of sensitivity, which may in fact be slightly higher than average, even among Québécois. But my hunch is that this presumption of male insensitivity may not have very positive effects on the perception of such a campaign. Québécois watching this video may not groan but they may not find it that funny either.

There’s a generational component involved and, partly because of a discussion of writing styles in a generational perspective, I have been thinking about “generations” as a useful model for explaining cultural diversity to non-ethnographers.

See, such perceived generational groups as “Baby Boomers” and “Generation X” need not be defined as monolithic, monadic, bounded entities and they have none of the problems associated with notions of “ethnicity” in the general public. “Generations” aren’t “faraway tribes” nor do they imply complete isolation. Some people may tend to use “generational” labels in such terms that they appear clearly defined (“Baby Boomers are those individuals born between such and such years”). And there is some confusion between this use of “historical generations” and what the concept of “generation” means in, say, the study of kinship systems. But it’s still relatively easy to get people to think about generations in cultural terms: they’re not “different cultures” but they still seem to be “culturally different.”

Going back to gender… The JC Penney marketing campaign visibly lumps together people of different ages. The notion seems to be that doghouse-worthy male insensitivity isn’t age-specific or related to inexperience. The one man who was able to leave the doghouse based on his purchase of diamonds is relatively “age-neutral” as he doesn’t really seem to represent a given age. Because this attempt at crossing age divisions seems so obvious, I would assume that it came in the context of perceived differences in gender relationships. Using the logic of those who perceive the second part of the 20th Century as a period of social emancipation, one might presume that younger men are less insensitive than older men (who were “brought up” in a cultural context which was “still sexist”). If there are wide differences in the degree of sensitivity of men of different ages, a campaign aiming at a broad age range needs to diminish the importance of these differences. “The joke needs to be funny to men of all ages.”

The Quebec context is, I think, different. While we do perceive the second part of the 20th Century (and, especially, the 1970s) as a period of social emancipation (known as the “Quiet Revolution” or «Révolution Tranquille»), the degree of sensitivity to gender issues appears to be relatively level, across the population. At a certain point in time, one might have argued that older men were still insensitive (at the same time as divorcées in their forties might have been regarded as very assertive) but it seems difficult to make such a distinction in the current context.

All this to say that the JC Penney commercial is culturally inappropriate for Québécois society? Not quite. Though the example I used was this JC Penney campaign, I’m thinking about broader contexts for Québécois identity (for a variety of personal reasons, including the fact that I have been back in Québec for several months, now).

My claim is…

Ethnographic field research would go a long way to unearth culturally appropriate categories which might eventually help marketers cater to Québécois.

Of course, the agency which produced that JC Penney ad (Saatchi & Saatchi) was targeting the US market (JC Penney doesn’t have locations in Quebec) and I received the link through a friend in the US. But it was an interesting opportunity for me to think and write about a few issues related to the cultural specificity of gender stereotypes.

Blogging and Literary Standards

I wrote the following comment in response to a conversation between novelist Rick Moody and podcasting pioneer Chris Lydon:

Open Source » Blog Archive » In the Obama Moment: Rick Moody.

In keeping with the RERO principle I describe in that comment, the version on the Open Source site is quite raw. As is my habit, these days, I pushed the “submit” button without rereading what I had written. This version is edited, partly because I noticed some glaring mistakes and partly because I wanted to add some links. (Blog comments are often tagged for moderation if they contain too many links.) As I started editing that comment, I changed a few things, some of which have consequences to the meaning of my comment. There’s this process, in both writing and editing, which “generates new thoughts.” Yet another argument for the RERO principle.

I can already think of an addendum to this post, revolving on my personal position on writing styles (informed by my own blogwriting experience) along with my relative lack of sensitivity for Anglo writing. But I’m still blogging this comment on a standalone basis.

Read on, please… Continue reading Blogging and Literary Standards

My Problem With Journalism

I hate having an axe to grind. Really, I do. “It’s unlike me.” When I notice that I catch myself grinding an axe, I “get on my own case.” I can be quite harsh with my own self.

But I’ve been trained to voice my concerns. And I’ve been perceiving an important social problem for a while.

So I “can’t keep quiet about it.”

If everything goes really well, posting this blog entry might be liberating enough that I will no longer have any axe to grind. Even if it doesn’t go as well as I hope, it’ll be useful to keep this post around so that people can understand my position.

Because I don’t necessarily want people to agree with me. I mostly want them to understand “where I come from.”

So, here goes:

Journalism may have outlived its usefulness.

Like several other “-isms” (including nationalism, colonialism, imperialism, and racism) journalism is counterproductive in the current state of society.

This isn’t an ethical stance, though there are ethical positions which go with it. It’s a statement about the anachronic nature of journalism. As per functional analysis, everything in society needs a function if it is to be maintained. What has been known as journalism is now taking new functions. Eventually, “journalism as we know it” should, logically, make way for new forms.

What these new forms might be, I won’t elaborate in this post. I have multiple ideas, especially given well-publicised interests in social media. But this post isn’t about “the future of journalism.”

It’s about the end of journalism.

Or, at least, my looking forward to the end of journalism.

Now, I’m not saying that journalists are bad people and that they should just lose their jobs. I do think that those who were trained as journalists need to retool themselves, but this post isn’t not about that either.

It’s about an axe I’ve been grinding.

See, I can admit it, I’ve been making some rather negative comments about diverse behaviours and statements, by media people. It has even become a habit of mine to allow myself to comment on something a journalist has said, if I feel that there is an issue.

Yes, I know: journalists are people too, they deserve my respect.

And I do respect them, the same way I respect every human being. I just won’t give them the satisfaction of my putting them on a pedestal. In my mind, journalists are people: just like anybody else. They deserve no special treatment. And several of them have been arrogant enough that I can’t help turning their arrogance back to them.

Still, it’s not about journalist as people. It’s about journalism “as an occupation.” And as a system. An outdated system.

Speaking of dates, some context…

I was born in 1972 and, originally,I was quite taken by journalism.

By age twelve, I was pretty much a news junkie. Seriously! I was “consuming” a lot of media at that point. And I was “into” media. Mostly television and radio, with some print mixed in, as well as lots of literary work for context: this is when I first read French and Russian authors from the late 19th and early 20th centuries.

I kept thinking about what was happening in The World. Back in 1984, the Cold War was a major issue. To a French-Canadian tween, this mostly meant thinking about the fact that there were (allegedly) US and USSR “bombs pointed at us,” for reasons beyond our direct control.

“Caring about The World” also meant thinking about all sorts of problems happening across The Globe. Especially poverty, hunger, diseases, and wars. I distinctly remember caring about the famine in Ethiopia. And when We Are the World started playing everywhere, I felt like something was finally happening.

This was one of my first steps toward cynicism. And I’m happy it occured at age twelve because it allowed me to eventually “snap out of it.” Oh, sure, I can still be a cynic on occasion. But my cynicism is contextual. I’m not sure things would have been as happiness-inducing for me if it hadn’t been for that early start in cynicism.

Because, you see, The World disinterested itself quite rapidly with the plight of Ethiopians. I distinctly remember asking myself, after the media frenzy died out, what had happened to Ethiopians in the meantime. I’m sure there has been some report at the time claiming that the famine was over and that the situation was “back to normal.” But I didn’t hear anything about it, and I was looking. As a twelve-year-old French-Canadian with no access to a modem, I had no direct access to information about the situation in Ethiopia.

Ethiopia still remained as a symbol, to me, of an issue to be solved. It’s not the direct cause of my later becoming an africanist. But, come to think of it, there might be a connection, deeper down than I had been looking.

So, by the end of the Ethiopian famine of 1984-85, I was “losing my faith in” journalism.

I clearly haven’t gained a new faith in journalism. And it all makes me feel quite good, actually. I simply don’t need that kind of faith. I was already training myself to be a critical thinker. Sounds self-serving? Well, sorry. I’m just being honest. What’s a blog if the author isn’t honest and genuine?

Flash forward to 1991, when I started formal training in anthropology. The feeling was exhilarating. I finally felt like I belonged. My statement at the time was to the effect that “I wasn’t meant for anthropology: anthropology was meant for me!” And I was learning quite a bit about/from The World. At that point, it already did mean “The Whole Wide World,” even though my knowledge of that World was fairly limited. And it was a haven of critical thinking.

Ideal, I tell you. Moan all you want, it felt like the ideal place at the ideal time.

And, during the summer of 1993, it all happened: I learnt about the existence of the “Internet.” And it changed my life. Seriously, the ‘Net did have a large part to play in important changes in my life.

That event, my discovery of the ‘Net, also has a connection to journalism. The person who described the Internet to me was Kevin Tuite, one of my linguistic anthropology teachers at Université de Montréal. As far as I can remember, Kevin was mostly describing Usenet. But the potential for “relatively unmediated communication” was already a big selling point. Kevin talked about the fact that members of the Caucasian diaspora were able to use the Internet to discuss with their relatives and friends back in the Caucasus about issues pertaining to these independent republics after the fall of the USSR. All this while media coverage was sketchy at best (sounded like journalism still had a hard time coping with the new realities).

As you can imagine, I was more than intrigued and I applied for an account as soon as possible. In the meantime, I bought at 2400 baud modem, joined some local BBSes, and got to chat about the Internet with several friends, some of whom already had accounts. Got my first email account just before semester started, in August, 1993. I can still see traces of that account, but only since April, 1994 (I guess I wasn’t using my address in my signature before this). I’ve been an enthusiastic user of diverse Internet-based means of communication since then.

But coming back to journalism, specifically…

Journalism missed the switch.

During the past fifteen years, I’ve been amazed at how clueless members of mainstream media institutions have been to “the power of the Internet.” This was during Wired Magazine’s first year as a print magazine and we (some friends and I) were already commenting upon the fact that print journalists should look at what was coming. Eventually, they would need to adapt. “The Internet changes everything,” I thought.

No, I didn’t mean that the Internet would cause any of the significant changes that we have seeing around us. I tend to be against technological determinism (and other McLuhan tendencies). Not that I prefer sociological determinism yet I can’t help but think that, from ARPAnet to the current state of the Internet, most of the important changes have been primarily social: if the Internet became something, it’s because people are making it so, not because of some inexorable technological development.

My enthusiastic perspective on the Internet was largely motivated by the notion that it would allow people to go beyond the model from the journalism era. Honestly, I could see the end of “journalism as we knew it.” And I’m surprised, fifteen years later, that journalism has been among the slowest institutions to adapt.

In a sense, my main problem with journalism is that it maintains a very stratified structure which gives too much weight to the credibility of specific individuals. Editors and journalists, who are part of the “medium” in the old models of communication, have taken on a gatekeeping role despite the fact that they rarely are much more proficient thinkers than people who read them. “Gatekeepers” even constitute a “textbook case” in sociology, especially in conflict theory. Though I can easily perceive how “constructed” that gatekeeping model may be, I can easily relate to what it entails in terms of journalism.

There’s a type of arrogance embedded in journalistic self-perception: “we’re journalists/editors so we know better than you; you need us to process information for you.” Regardless of how much I may disagree with some of his words and actions, I take solace in the fact that Murdoch, a key figure in today’s mainstream media, talked directly at this arrogance. Of course, he might have been pandering. But the very fact that he can pay lip-service to journalistic arrogance is, in my mind, quite helpful.

I think the days of fully stratified gatekeeping (a “top-down approach” to information filtering) are over. Now that information is easily available and that knowledge is constructed socially, any “filtering” method can be distributed. I’m not really thinking of a “cream rises to the top” model. An analogy with water sources going through multiple layers of mountain rock would be more appropriate to a Swiss citizen such as myself. But the model I have in mind is more about what Bakhtin called “polyvocality” and what has become an ethical position on “giving voice to the other.” Journalism has taken voice away from people. I have in mind a distributed mode of knowledge construction which gives everyone enough voice to have long-distance effects.

At the risk of sounding too abstract (it’s actually very clear in my mind, but it requires a long description), it’s a blend of ideas like: the social butterfly effect, a post-encyclopedic world, and cultural awareness. All of these, in my mind, contribute to this heightened form of critical thinking away from which I feel journalism has led us.

The social butterfly effect is fairly easy to understand, especially now that social networks are so prominent. Basically, the “butterfly effect” from chaos theory applied to social networks. In this context, a “social butterfly” is a node in multiple networks of varying degrees of density and clustering. Because such a “social butterfly” can bring things (ideas, especially) from one such network to another, I argue that her or his ultimate influence (in agregate) is larger than that of someone who sits at the core of a highly clustered network. Yes, it’s related to “weak ties” and other network classics. But it’s a bit more specific, at least in my mind. In terms of journalism, the social butterfly effect implies that the way knowledge is constructed needs not come from a singular source or channel.

The “encyclopedic world” I have in mind is that of our good friends from the French Enlightenment: Diderot and the gang. At that time, there was a notion that the sum of all knowledge could be contained in the Encyclopédie. Of course, I’m simplifying. But such a notion is still discussed fairly frequently. The world in which we now live has clearly challenged this encyclopedic notion of exhaustiveness. Sure, certain people hold on to that notion. But it’s not taken for granted as “uncontroversial.” Actually, those who hold on to it tend to respond rather positively to the journalistic perspective on human events. As should be obvious, I think the days of that encyclopedic worldview are counted and that “journalism as we know it” will die at the same time. Though it seems to be built on an “encyclopedia” frame, Wikipedia clearly benefits from distributed model of knowledge management. In this sense, Wikipedia is less anachronistic than Britannica. Wikipedia also tends to be more insightful than Britannica.

The cultural awareness point may sound like an ethnographer’s pipe dream. But I perceive a clear connection between Globalization and a certain form of cultural awareness in information and knowledge management. This is probably where the Global Voices model can come in. One of the most useful representations of that model comes from a Chris Lydon’s Open Source conversation with Solana Larsen and Ethan Zuckerman. Simply put, I feel that this model challenges journalism’s ethnocentrism.

Obviously, I have many other things to say about journalism (as well as about its corrolate, nationalism).

But I do feel liberated already. So I’ll leave it at that.

Ethnographic Disciplines

Just because this might be useful in the future…
I perceive a number of academic disciplines to be “ethnographic” in the sense that they use the conceptual and epistemic apparatus of “ethnography.” (“Ethnography” taken here as an epistemological position in human research, not as “the description of a people” in either literary or methodological uses.)

I don’t mean by this that practitioners are all expected to undertake ethnographic field research or that their methods are exclusively ethnographic. I specifically wish to point out that ethnography is not an “exclusive prerogative” of anthropology. And I perceive important connections between these disciplines.

In no particular order:

  • Ethnohistory
  • Ethnolinguistics (partly associated with Linguistic Anthropology)
  • Folkloristics
  • Ethnomusicology
  • Ethnocinematography (partly associated with Visual Anthropology)
  • Ethnology (Cultural Anthropology)

The following disciplines (the “micros”), while not ethnographic per se, often have ethnographic components at the present time.

  • Microhistory
  • Microsociology
  • Microeconomics

Health research and market research also make frequent use of ethnographic methods, these days (especially through “qualitative data analysis” software). But I’m not clear on how dedicated these researchers are to the epistemological bases for ethnography.

It may all sound very idiosyncratic. But I still think it works, as a way to provide working definitions for disciplines and approaches.

Thoughts, comments, suggestions, questions?

I Am Not a Guru

“Nor do I play one online!”

The “I am not a ” phrase is often used as a disclaimer when one is giving advice. Especially in online contexts having to do with law, in which case the IANAL acronym can be used, and understood.
I’m not writing this to give advice. (Even though I could!) I’ve simply been thinking about social media a fair deal, recently, and thought I’d share a few thoughts.

I’ve been on the record as saying that I have a hard time selling my expertise. It’s not through lack of self-confidence (though I did have problems with this in the past), nor is it that my expertise is difficult to sell. It’s simply a matter of seeing myself as a friendly humanist, not as a brand to sell. To a certain extent, this post is an extension of the same line of thinking.

I’m also going back to my post about “the ‘social’ in ‘social media/marketing/web'” as I tend to position myself as an ethnographer and social scientist (I teach anthropology, sociology, and folkloristics). Simply put, I do participant-observation in social media spheres. Haven’t done formal research on the subject, nor have I taught in that field. But I did gain some insight in terms if what social media entails.

Again, I’m no guru. I’m just a social geek.

The direct prompt for this blogpost is a friend’s message in which he asked me for advice on the use of social media to market his creative work. Not that he framed his question in precisely those terms but the basic idea was there.

As he’s a friend, I answered him candidly, not trying to sell my social media expertise to him. But, after sending that message, I got to think about the fact that I’m not selling my social media expertise to anyone.

One reason is that I’m no salesman. Not only do I perceive myself as “too frank to be a salesman” (more on the assumptions later), but I simply do not have the skills to sell anything. Some people are so good at sales pitches that they could create needs where they is none (the joke about refrigerators and “Eskimos” is too much of an ethnic slur to be appropriate). I’ve been on the record saying that “I couldn’t sell bread for a penny” (to a rich yet starving person).

None of this means that I haven’t had any influence on any purchasing pattern. In fact, that long thread in which I confessed my lack of salesman skills was the impulse (direct or indirect) behind the purchase of a significant number of stovetop coffee devices and this “influence” has been addressed explicitly. It’s just that my influence tends to be more subtle, more “diffuse.” Influence based on participation in diverse groups. It’s one reason I keep talking about the “social butterfly effect.”

Coming back to social media and social marketing.

First, some working definitions. By “social media” I usually mean blogs, podcasts, social networking systems, and microblogs. My usage also involves any participatory use of the Internet and any alternative to “mainstream media” (MSM) which makes use of online contacts between human beings. “Social marketing” is, to me, the use of social media to market and sell a variety of things online, including “people as brands.” This notion connects directly to a specific meaning of “social capital” which, come to think of it, probably has more to do with Putnam than Bourdieu (PDF version of an atricle about both versions).

Other people, I admit, probably have much better ways to define those concepts. But those definitions are appropriate in the present context. I mostly wanted to talk about gurus.

Social Guru

I notice guru-like behaviour in the social media/marketing sphere.

I’m not targetting individuals, though the behaviour is adopted by specific people. Not every one is acting as a “social media guru” or “social marketing guru.” The guru-like behaviour is in fact quite specific and not as common as some would think.

Neither am I saying that guru-like behaviour is inappropriate. I’m not blaming anyone for acting like a guru. I’m mostly distancing myself from that behaviour. Trying to show that it’s one model for behaviour in the social media/marketing sphere.

It should go without saying: I’m not using the term “guru” in a literal sense it might have in South Asia. That kind of guru I might not distance myself from as quickly. Especially if we think about “teachers as personal trainers.” But I’m using “guru” in reference to an Anglo-American phenomenon having to do with expertise and prestige.

Guru-like behaviour, as noticed in the social media/marketing sphere, has to do with “portraying oneself as an expert holding a secret key which can open the doors to instant success.” Self-assurance is involved, of course. But there’s also a degree of mystification. And though this isn’t a rant against people who adopt this kind of behaviour, I must admit that I have negative reactions to any kind of mystification.

There’s a difference between mystery and mystification. Something that is mysterious is difficult to explain “by its very nature.” Mystification involves withholding information to prevent knowledge. As an academic, I have been trained to fight obscurantism of any kind. Mystification seems counterproductive. “Information Wants to be Free.”

This is not to say that I dislike ambiguity, double-entendres, or even secrets. In fact, I’m often using ambiguity in playful manner and, working with a freemasonry-like secret association, I do understand the value of the most restrictive knowledge management practises. But I find limited value in restricting information when knowledge can be beneficial to everyone. As in Eco’s The Name of the Rose, subversive ideas find their way out of attempts to hide them.

Another aspect of guru-like behaviour which tends to bother me is that I can’t help but find it empty. As some would say, “there needs to be a ‘there’ there.” With social media/marketing, the behaviour I’m alluding to seems to imply that there is, in fact, some “secret key to open all doors.” Yet, as I scratch beneath the surface, I find something hollow. (The image I have in mind is that of a chocolate Easter egg. But any kind of trompe-l’œil would work.)

Obviously, I’m not saying that there’s “nothing to” social media/marketing. Those who dismiss social media and/or social marketing sound to me like curmudgeons or naysayers. “There’s nothing new, here. It’s just the same thing as what it always was. Buy my book to read all about what nonsense this all is.” (A bit self-serving, don’t you think?)

And I’m not saying that I know what there is in social media and marketing which is worth using. That would not only be quite presumptuous but it would also represent social media and marketing in a more simplified manner than I feel it deserves.

I’m just saying that caution should be used with people who claim they know everything there is to know about social media and social marketing. In other words, “be careful when someone promises to make you succeed through the Internet.” Sounds obvious, but some people still fall prey to grandiose claims.

Having said this, I’ll keep on posting some of thoughts about social media and social marketing. I might be way off, so “don’t quote me on this.” (You can actually quote me but don’t give my ideas too much credit.)

Café à la montréalaise

Montréal est en passe de (re)devenir une destination pour le café. Mieux encore, la «Renaissance du café à Montréal» risque d’avoir des conséquences bénéfiques pour l’ensemble du milieu culinaire de la métropole québécoise.

Cette thèse peut sembler personnelle et je n’entends pas la proposer de façon dogmatique. Mais en me mêlant au milieu du café à Montréal, j’ai accumulé un certain nombre d’impressions qu’il me ferait plaisir de partager. Il y a même de la «pensée magique» dans tout ça en ce sens qu’il me semble plus facile de rebâtir la scène montréalaise du café si nous avons une idée assez juste de ce qui constitue la spécificité montréalaise.

Continue reading Café à la montréalaise

Blogging Academe

LibriVox founder and Montreal geek Hugh McGuire recently posted a blog entry in which he gave a series of nine arguments for academics to blog:

Why Academics Should Blog

Hugh’s post reminded me of one of my favourite blogposts by an academic, a pointed defence of blogging by Mark Liberman, of Language Log fame.
Raising standards –by lowering them

While I do agree with Hugh’s points, I would like to reframe and rephrase them.

Clearly, I’m enthusiastic about blogging. Not that I think every academic should, needs to, ought to blog. But I do see clear benefits of blogging in academic contexts.

Academics do a number of different things, from search committees to academic advising. Here, I focus on three main dimensions of an academic’s life: research, teaching, and community outreach. Other items in a professor’s job description may benefit from blogging but these three main components tend to be rather prominent in terms of PTR (promotion, tenure, reappointment). What’s more, blogging can help integrate these dimensions of academic life in a single set of activities.

Impact

In relation to scholarship, the term “impact” often refers to the measurable effects of a scholar’s publication through a specific field. “Citation impact,” for instance, refers to the number of times a given journal article has been cited by other scholars. This kind of measurement is directly linked to Google’s PageRank algorithm which is used to assess the relevance of their search results. The very concept of “citation impact” relates very directly to the “publish or perish” system which, I would argue, does more to increase stress levels among full-time academic than to enhance scholarship. As such, it may need some rethinking. What does “citation impact” really measure? Is the most frequently cited text on a given subject necessarily the most relevant? Isn’t there a clustering effect, with some small groups of well-known scholars citing one another without paying attention to whatever else may happen in their field, especially in other languages?

An advantage of blogging is that this type of impact is easy to monitor. Most blogging platforms have specific features for “statistics,” which let bloggers see which of their posts have been visited (“hit”) most frequently. More sophisticated analysis is available on some blogging platforms, especially on paid ones. These are meant to help bloggers monetize their blogs through advertising. But the same features can be quite useful to an academic who wants to see which blog entries seem to attract the most traffic.

Closer to “citation impact” is the fact that links to a given post are visible within that post through the ping and trackback systems. If another blogger links to this very blogpost, a link to that second blogger’s post will appear under mine as a link. In other words, a blogpost can embed future references.

In terms of teaching, thinking about impact through blogging can also have interesting effects. If students are blogging, they can cite and link to diverse items and these connections can serve as a representation of the constructive character of learning. But even if students don’t blog, a teacher blogging course-related material can increase the visibility of that course. In some cases, this visibility may lead to inter-institutional collaboration or increased enrollment.

Transparency

While secrecy may be essential in some academic projects, most academics tend to adopt a favourable attitude toward transparency. Academia is about sharing information and spreading knowledge, not about protecting information or about limiting knowledge to a select few.

Bloggers typically value transparency.

There are several ethical issues which relate to transparency. Some ethical principles prevent transparency (for instance, most research projects involving “human subjects” require anonymity). But academic ethics typically go with increased transparency on the part of the researcher. For instance, informed consent by a “human subject” requires complete disclosure of how the data will be used and protected. There are usually requirements for the primary investigator to be reachable during the research project.

Transparency is also valuable in teaching. While some things should probably remain secret (say, answers to exam questions), easy access to a number of documents makes a lot of sense in learning contexts.

Public Intellectuals

It seems that the term “intellectual” gained currency as a label for individuals engaged in public debates. While public engagement has taken a different type of significance, over the years, but the responsibility for intellectuals to communicate publicly is still a matter of interest.

Through blogging, anyone can engage in public debate, discourse, or dialogue.

Reciprocity

Scholars working with “human subjects” often think about reciprocity. While remuneration may be the primary mode of retribution for participation in a research project, a broader concept of reciprocity is often at stake. Those who participated in the project usually have a “right to know” about the results of that study. Even when it isn’t the case and the results of the study remain secret, the asymmetry of human subjects revealing something about themselves to scholars who reveal nothing seems to clash with fundamental principles in contemporary academia.

Reciprocity in teaching can lead directly to some important constructivist principles. The roles of learners and teachers, while not completely interchangeable, are reciprocal. A teacher may learn and a learner may teach.

Playing with Concepts

Blogging makes it easy to try concepts out. More than “thinking out loud,” the type of blogging activity I’m thinking about can serve as a way to “put ideas on paper” (without actual paper) and eventually get feedback on those ideas.

In my experience, microblogging (Identi.ca, Twitter…) has been more efficient than extended blogging in terms of getting conceptual feedback. In fact, social networks (Facebook, more specifically) have been even more conducive to hashing out concepts.

Many academics do hash concepts out with students, especially with graduate students. The advantage is that students are likely to understand concepts quickly as they already share some of the same references as the academic who is playing with those concepts. There’s already a context for mutual understanding. The disadvantage is that a classroom context is fairly narrow to really try out the implications of a concept.

A method I like to use is to use fairly catchy phrases and leave concepts fairly raw, at first. I then try the same concept in diverse contexts, on my blogs or off.

The main example I have in mind is the “social butterfly effect.” It may sound silly at first but I find it can be a basis for discussion, especially if it spreads a bit.

A subpoint, here, is that this method allows for “gauging interest” in new concepts and it can often lead one in completely new directions. By blogging about concepts, an academic can tell if this concept has a chance to stick in a broad frame (outside the Ivory Tower) and may be given insight from outside disciplines.

Playing with Writing

This one probably applies more to “junior academics” (including students) but it can also work with established academics who enjoy diversifying their writing styles. Simply put: blogwriting is writing practise.

A common idea, in cognitive research on expertise, is that it takes about ten thousand hours to become an expert. For better or worse, academics are experts at writing. And we gain that expertise through practise. In this context, it’s easy to see blogging as a “writing exercise.” At least, that would be a perspective to which I can relate.

My impression is that writing skills are most efficiently acquired through practise. The type of practise I have in mind is “low-stakes,” in the sense that the outcomes of a writing exercise are relatively inconsequential. The basis for this perspective is that self-consciousness, inhibition, and self-censorship tend to get in the way of fluid writing. High-stakes writing (such as graded assignments) can make a lot of sense at several stages in the learning process, but overemphasis on evaluating someone’s writing skills will likely stress out the writer more than make her/him motivated to write.

This impression is to a large extent personal. I readily notice that when I get too self-conscious about my own writing (self-unconscious, even), my writing becomes much less fluid. In fact, because writing about writing tends to make one self-conscious, my writing this post is much less efficient than my usual writing sessions.

In my mind, there’s a cognitive basis to this form of low-stakes, casual writing. As with language acquisition, learning occurs whether or not we’re corrected. According to most research in language acquisition, children acquire their native languages through exposure, not through a formal learning process. My guess is that the same apply to writing.

In some ways, this is a defence of drafts. “Draft out your ideas without overthinking what might be wrong about your writing.” Useful advice, at least in my experience. The further point is to do something with those drafts, the basis for the RERO principle: “release your text in the wild, even if it may not correspond to your standards.” Every text is a work in progress. Especially in a context where you’re likely to get feedback (i.e., blogging). Trial and error, with a feedback mechanism. In my experience, feedback on writing tends to be given in a thoughtful and subtle fashion while feedback on ideas can be quite harsh.

The notion of writing styles is relevant, here. Some of Hugh’s arguments about the need for blogging in academia revolve around the notion that “academics are bad writers.” My position is that academics are expert writers but that academic writing is a very specific beast. Hugh’s writing standards might clash with typical writing habits among academics (which often include neologisms and convoluted metaphors). Are Hugh’s standards appropriate in terms of academic writing? Possibly, but why then are academic texts rating so low on writing standards after having been reviewed by peers and heavily edited? The relativist’s answer is, to me, much more convincing: academic texts are typically judged through standards which are context-specific. Judging academic writing with outside standards is like judging French writing with English standards (or judging prose through the standards of classic poetry).

Still, there’s something to be said about readability. Especially when these texts are to be used outside academia. Much academic writing is meant to remain within the walls of the Ivory Tower yet most academic disciplines benefit from some interaction with “the general public.” Though it may not be taught in universities and colleges, the skill of writing for a broader public is quite valuable. In fact, it may easily be transferable to teaching, especially if students come from other disciplines. Furthermore, writing outside one’s discipline is required in any type of interdisciplinary context, including project proposals for funding agencies.

No specific writing style is implied in blogging. A blogger can use whatever style she/he chooses for her/his posts. At the same time, blogging tends to encourage writing which is broadly readable and makes regular use of hyperlinks to connect to further information. In my opinion, this type of writing is a quite appropriate one in which academics can extend their skills.

“Public Review”

Much of the preceding connects with peer review, which was the basis of Mark Liberman’s post.

In academia’s recent history, “peer reviewed publications” have become the hallmark of scholarly writing. Yet, as Steve McIntyre claims, the current state of academic peer review may not be as efficient at ensuring scholarly quality as its proponents claim it to be. As opposed to financial auditing, for instance, peer review implies very limited assessment based on data. And I would add that the very notion of “peer” could be assessed more carefully in such a context.

Overall, peer review seems to be relatively inefficient as a “reality check.” This might sound like a bold claim and I should provide data to support it. But I mostly want to provoke some thought as to what the peer review process really implies. This is not about reinventing the wheel but it is about making sure we question assumptions about the process.

Blogging implies public scrutiny. This directly relates to transparency, discussed above. But there is also the notion of giving the public the chance to engage with the outcomes of academic research. Sure, the general public sounds like a dangerous place to propose some ideas (especially if they have to do with health or national security). But we may give some thought to Linus’s law and think about the value of “crowdsourcing” academic falsification.

Food for Thought

There’s a lot more I want to add but I should heed my call to RERO. Otherwise, this post will remain in my draft posts for an indefinite period of time, gathering dust and not allowing any timely discussion. Perhaps more than at any other point, I would be grateful for any thoughtful comment about academic blogging.

In fact, I will post this blog entry “as is,” without careful proofreading. Hopefully, it will be the start of a discussion.

I will “send you off” with a few links related to blogging in academic contexts, followed by Hugh’s list of arguments.

Links on Academic Blogging

(With an Anthropological emphasis)

Hugh’s List

  1. You need to improve your writing
  2. Some of your ideas are dumb
  3. The point of academia is to expand knowledge
  4. Blogging expands your readership
  5. Blogging protects and promotes your ideas
  6. Blogging is Reputation
  7. Linking is better than footnotes
  8. Journals and blogs can (and should) coexist
  9. What have journals done for you lately?

L'intellectuel s'assume

Le personnage de l’intellectuel(le) mérite bien son petit billet. D’autant que son identité est venue se loger à plusieurs reprises dans ma vie, ces derniers temps.

(Pour simplifier, et par référence à un contexte universaliste, j’utiliserai le terme «intellectuel» au masculin comme s’il était neutre.)

Oui, bien entendu, je suis moi-même un intellectuel et je m’assume en tant que tel. D’ailleurs, j’ai d’abord pensé intituler ce billet «Confessions d’un intellectuel solidaire» ou quelque-chose du genre. Mais la formule «Confessions d’un <nom><adjectif>» est déjà assez fréquente, sur ce blogue. Et je ne pense pas seulement de façon introspective à ce personnage.

D’ailleurs, c’est en lisant certains trucs au sujet de la fameuse Affaire Dreyfus que m’est venue l’idée d’écrire un billet sur la notion d’«intellectuel». Il s’avère que l’adoption du terme «intellectuel» pour désigner une certaine catégorie d’individu puisse dater de la France de la fin du XIXè siècle, y compris dans son usage anglais. Cette période historique m’a fortement influencé, surtout par la lecture de divers écrivains français de l’époque. Mais c’est moins par désir de reconstituer une réalité historique que je me mets à parler d’intellectuel que par intérêt pour la construction de personnages sociaux, quels qu’ils soient. Penser au fait que l’intellectuel est construit me permet de remettre en contexte social un ensemble de réalités qui m’apparaissent intéressantes. Surtout qu’elles peuvent facilement être liées à la «culture geek» qui m’intéresse tant, en plus de me toucher directement.

Évidemment, ce n’est pas la première fois que l’intellectuel comme personnage se retrouve sur ce blogue. Mais le contexte semblait particulièrement approprié, aujourd’hui.

Faut dire que je suis allé à un petit brunch avec des amis du primaire. Ça ne surprendra personne de savoir que ces amis me considéraient déjà comme un intellectuel à l’époque. Pas qu’ils aient utilisé le terme. Mais l’étiquette était là. Sauf que, contrairement à ce que j’ai ressenti il y a près de trente ans, cette étiquette n’était pas la base d’un rejet.

D’ailleurs, je pense souvent à la théorie de l’étiquetage. Elle était même présente dans un cours de sociologie que j’enseignais l’été dernier. Pour simplifier: les étiquettes qui nous sont collées ont des implications durables dans nos agissements sociaux. Ou, pour citer Howie Becker selon un dictionnaire suisse:

Le comportement déviant est ce que les gens étiquettent comme tel ; le déviant est celui à qui on a réussi à coller cette étiquette

(Évidemment, j’étends la notion d’étiquetage hors de la déviance au sens strict.)

Dans ce contexte, le comportement d’intellectuel est celui qui est étiquetté comme tel. L’intellectuel est celui à qui on a réussi à coller cette étiquette.

Version personnelle (que j’ai même eu l’occasion d’exposer à un ami du primaire): je me comporte en fonction de l’étiquette d’intellectuel qui a été posée sur moi, dès le jeune âge. Pas que cette étiquette est abusive: elle colle parce qu’elle trouve une surface qui s’y prête. Mais le personnage de l’intellectuel n’est pas naturel, universel, atemporel ou dénué d’ambiguïté.

Parlant d’ambiguïté, faudrait penser à le définir, cet intellectuel.

Selon Wikipedia:

Un intellectuel est une personne qui, du fait de sa position sociale, dispose d’une forme d’autorité et s’engage dans la sphère publique pour défendre des valeurs.

Pas mal. C’est un peu la base de mon premier billet sur les intellectuels. L’engagement public prend diverses formes et on comprend le lien avec l’Affaire Dreyfus.

Mais les usages communs du terme (et d’«intellectualisme» et “intellectualism“) semblent aller dans diverses autres directions. D’abord, la notion d’une intelligence «supérieure» (que les cognitivistes relativisent si bien mais qui semble consensuelle, socialement). Cette perception de l’intelligence est liée à une forme d’élitisme, l’intellectuel fait partie d’une élite particulière et exclue parfois ceux qui n’en font pas partie. Puis il y a la notion de «rationalité», l’intellectuel conçu comme étant «loin de ses émotions». Ou la maladresse et le manque d’aptitudes manuelles, le terme «intellectuel» alors utilisé pour exprimer un certain mépris. Pour aller plus loin, on peut même dire que le fait de souscrire à un certain dualisme «corps/esprit» est souvent teinté d’«intellectualisme».

Ces dénotations et connotations me semblent toutes appropriées pour décrire un type précis d’intellectuel: le «geek» (j’aime bien «geekette» pour le féminin; il y a relativement peu de femmes geeks). Le personnage du geek est une part important du stéréotype contemporain lié à l’intellectuel. Contrairement au «nerd» des années 1980, le geek a désormais une place de choix au sein de la culture populaire. Et la réhabilitation du geek constitue un mouvement contraire à une vague d’anti-intellectualisme très patente aux États-Unis et dans d’autres sociétés post-industrielles.

Penser au geek en tant qu’intellectuel permet de situer le personnage dans son contexte social. D’un point de vue professionnel, le geek typique est souvent ingénieur, informatien ou scientifique. Le contexte scolaire a souvent accordé beaucoup d’importance aux notes qu’il obtenait. Il est peut-être très apte à entreprendre diverses activités manuelles, il peut même «travailler de ses mains autant que de sa tête», mais son intellect demeure valorisé. C’est «un cerveau», un “brainiac”. Pas que son «niveau d’intelligence» est nécessairement plus élevé que la moyenne, mais le type particulier d’intelligence qui le caractérise correspond largement à l’idée qu’on se fait généralement du «quotient intellectuel»: capacité d’abstraction, sens logique, rapidité à résoudre des équations ou à se remémorer une information, minutie…

Pour revenir à la construction sociale du personnage de l’intellectuel. Malgré certaines transformations au cours du dernier siècle, l’intellectuel conserve un statut social particulier. Dans un modèle d’économie politique (à la fois dans sa version capitaliste que socialiste), l’intellectuel fait partie d’une espèce de classe sociale avec ses caractéristiques particulières. C’est un type de «col blanc» qui ne fait pas un travail très routinier. C’est aussi l’individu qui bénéficie du privilège lié à l’éducation post-secondaire dans les sociétés post-industrielles. C’est celui qui a le loisir de lire et de parfaire son apprentissage. C’est le public-cible de «La Culture», au sens raffiné du terme. C’est peut-être même un snob, un personnage hautain, l’opposé du «vrai monde».

Et c’est là que le mode introspectif me fait réagir: je suis peut-être un intellectuel, mais je suis pas snob. Si je suis «anti-» quoi que ce soit, c’est anti-snob. Et je ne considère pas l’intellectuel comme plus intelligent qu’un autre. Je considère surtout l’intellectuel comme une création des sociétés post-industrielles, basées sur la division pointue du travail social. Même que, ce snobisme, c’est ce qui me dérange le plus du fait d’être intellectuel. C’est probablement pour ça que, même si je m’assume comme intellectuel, je tente souvent d’effacer cette étiquette. «Je suis un intellectuel mais je suis aussi un bon gars.»

Dans mon cas, le fait d’être considéré comme un intellectuel a beaucoup de lien avec mon éloquence perçue. On m’a toujours considéré comme éloquent. Enfant, déjà, je «parlais bien». Du moins, c’est ce qu’on a dit de moi (pas plus tard qu’aujourd’hui). Bon, d’accord, comme l’art oratoire a toujours été valorisé dans ma famille, j’ai probablement été porté à m’amuser avec le verbe. Aussi, je lisais déjà beaucoup, enfant. Et j’écrivais: à l’âge de dix ans, je tapais à la dactylo un petit texte au sujet de la perfection (qui semble logiquement impossible puisqu’elle est une absence de défaut). Et j’avais l’occasion de m’exprimer. Auprès d’adultes, surtout.

D’ailleurs, c’est probablement un point très important. Tout jeune, j’avais des rapports assez étroits avec plusieurs adultes (des amis de mes parents, surtout). J’étais souvent le seul enfant parmi de nombreux adultes. Plusieurs d’entre eux étaient profs (comme mon père). On m’écoutait avec intérêt. Dans une certaine mesure, j’étais presque pavané comme un animal de cirque qui pouvait discourir sur tout et sur rien. Mon père a souvent parlé de tout ça comme d’un problème fondamental. Peut-être par extension, mon étiquette d’intellectuel était perçue comme un problème. Fondamental.

Je considère aujourd’hui que je me suis bien développé. Je suis ce que j’ai toujours voulu être et je peux parfois faire ce que j’ai toujours voulu faire. Je devrais pas avoir honte.

D’être un intellectuel.

Taste and Judgement (Draft)

This post isn’t ready to be written. So this is just a placeholder. But, given my RERO mantra, I guess I should still publish it as a “placeholder” of sorts.

I recently served as judge in the Canadian Barista Championship (CBC), here in Montreal. That championship is the national competition pitting against one another baristas (espresso artists) from regional competitions. Rules and regulations (PDF) for this championship closely follow those set by the World Barista Championship (WBC).

Participating in this event, I got to think about taste, evaluation, (inter)subjectivity, coffee, Montreal’s culinary scene, and food culture generally.

Some videos from the event are available, through the event organizer’s uStream channel.

The event was blogged by Anthony Benda. Despite being busy preparing for his café’s grand opening, Anthony managed to give an excellent performance during the championship, especially on the first day. I wasn’t on the panel of judge for his performance but I have reason to believe that Anthony’s performance was really quite good.

I also got to think about my own involvement in such events.

Being a judge at barista championships is still somewhat new to me. I judged during the Eastern regional championship, back in June, and this was my first national championship. I still think that the “barista judge” label fits and I did mention it on occasion, with lots of disclaimers. Most judges at that event were coffee professionals of one type or the other (from equipment distributors to barista champions). My impression is that, despite my limited experience and my somewhat indirect connections to the coffee industry, I was accepted as a peer by other judges.

More importantly, I sincerely think that the judging at this competition was exceedingly fair. My strong perception is that we achieved a high degree of consistency in our judging, both at an individual level and through the group. A large part of what I perceive to be a resounding success comes from the work of WBC’s Brent Fortune, who trained and calibrated the judging team.

One thing I kept thinking about was how different barista judging is from judging homebrewed beer. I haven’t acted as a homebrew judge myself but many of my friends have and I proctored an exam for the Beer Judge Certification Program (BJCP). Put simply, homebrew competitions are stricter than barista championships. And I mean this to imply that BJCP competitions are in some ways less reasonable than barista championships, though the WBC could learn a thing or two from the BJCP.

Barista championships are based on fairness and impartiality. Though it’s mentioned on occasion, “objectivity” isn’t the core principle in judging. Not emphasizing “objectivity” allows for a sensible approach to tasting since, after all, tasting is as subjective as any other form of sensory perception. In other words, acknowledging the subjective nature of tasting brings realism to WBC-style competitions.

The reason I emphasize the “subjectivity” issue is that homebrewing competitions seem to exist in a radically different world, a world in which “objectivity” is an absolute goal. Though the BJCP style guidelines may allow for some room for variation in judging beer aroma, appearance, flavour, and mouthfeel, the main approach is object-based and a direct connection between a judge’s experience and precise measurements of the beer’s characteristics is assumed. Several homebrew judges do use the term “objective” fairly frequently and “subjectivity” is a “bad word” in many of the homebrew circles which give a lot of weight to those homebrew competitions.

One thing I find fascinating about this distinction between WBC and BJCP competitions is that, to some extent, the coffee professionals are less, well, “anal” than the homebrewers. The level of technical expertise may be as high in both domains. The drinks themselves are comparable on many levels, including in terms of chemical complexity. But the approaches taken to evaluate those drinks are radically different.

I also got to think about the connections (actual and potential) between Montreal’s strong beer scene and its renascent coffee scene. It certainly was fun to have beers at Benelux with a number of participants in the Canadian Barista Championship, including coffee writer Felipe Gonzalez. Myriade’s grand opening, on Monday, will likely serve as an opportunity for me to discuss Montreal’s coffee scene in more depth.

Of course, any of this could be the start of a long monologue on my part. But it’s probably better if I leave this post as it is, to serve as a placeholder for further discussion of taste, evaluation, subjectivity, coffee scenes