Tag Archives: linkfest

WordPress as Content Directory: Getting Somewhere

{I tend to ramble a bit. If you just want a step-by-step tutorial, you can skip to here.}

Woohoo!

I feel like I’ve reached a milestone in a project I’ve had in mind, ever since I learnt about Custom Post Types in WordPress 3.0: Using WordPress as a content directory.

The concept may not be so obvious to anyone else, but it’s very clear to me. And probably much clearer for anyone who has any level of WordPress skills (I’m still a kind of WP newbie).

Basically, I’d like to set something up through WordPress to make it easy to create, review, and publish entries in content databases. WordPress is now a Content Management System and the type of “content management” I’d like to enable has to do with something of a directory system.

Why WordPress? Almost glad you asked.

These days, several of the projects on which I work revolve around WordPress. By pure coincidence. Or because WordPress is “teh awsum.” No idea how representative my sample is. But I got to work on WordPress for (among other things): an academic association, an adult learners’ week, an institute for citizenship and social change, and some of my own learning-related projects.

There are people out there arguing about the relative value of WordPress and other Content Management Systems. Sometimes, WordPress may fall short of people’s expectations. Sometimes, the pro-WordPress rhetoric is strong enough to sound like fanboism. But the matter goes beyond marketshare, opinions, and preferences.

In my case, WordPress just happens to be a rather central part of my life, these days. To me, it’s both a question of WordPress being “the right tool for the job” and the work I end up doing being appropriate for WordPress treatment. More than a simple causality (“I use WordPress because of the projects I do” or “I do these projects because I use WordPress”), it’s a complex interaction which involves diverse tools, my skillset, my social networks, and my interests.

Of course, WordPress isn’t perfect nor is it ideal for every situation. There are cases in which it might make much more sense to use another tool (Twitter, TikiWiki, Facebook, Moodle, Tumblr, Drupal..). And there are several things I wish WordPress did more elegantly (such as integrating all dimensions in a single tool). But I frequently end up with WordPress.

Here are some things I like about WordPress:

This last one is where the choice of WordPress for content directories starts making the most sense. Not only is it easy for me to use and build on WordPress but the learning curves are such that it’s easy for me to teach WordPress to others.

A nice example is the post editing interface (same in the software and service). It’s powerful, flexible, and robust, but it’s also very easy to use. It takes a few minutes to learn and is quite sufficient to do a lot of work.

This is exactly where I’m getting to the core idea for my content directories.

I emailed the following description to the digital content editor for the academic organization for which I want to create such content directories:

You know the post editing interface? What if instead of editing posts, someone could edit other types of contents, like syllabi, calls for papers, and teaching resources? What if fields were pretty much like the form I had created for [a committee]? What if submissions could be made by people with a specific role? What if submissions could then be reviewed by other people, with another role? What if display of these items were standardised?

Not exactly sure how clear my vision was in her head, but it’s very clear for me. And it came from different things I’ve seen about custom post types in WordPress 3.0.

For instance, the following post has been quite inspiring:

I almost had a drift-off moment.

But I wasn’t able to wrap my head around all the necessary elements. I perused and read a number of things about custom post types, I tried a few things. But I always got stuck at some point.

Recently, a valuable piece of the puzzle was provided by Kyle Jones (whose blog I follow because of his work on WordPress/BuddyPress in learning, a focus I share).

Setting up a Staff Directory using WordPress Custom Post Types and Plugins | The Corkboard.

As I discussed in the comments to this post, it contained almost everything I needed to make this work. But the two problems Jones mentioned were major hurdles, for me.

After reading that post, though, I decided to investigate further. I eventually got some material which helped me a bit, but it still wasn’t sufficient. Until tonight, I kept running into obstacles which made the process quite difficult.

Then, while trying to solve a problem I was having with Jones’s code, I stumbled upon the following:

Rock-Solid WordPress 3.0 Themes using Custom Post Types | Blancer.com Tutorials and projects.

This post was useful enough that I created a shortlink for it, so I could have it on my iPad and follow along: http://bit.ly/RockSolidCustomWP

By itself, it might not have been sufficient for me to really understand the whole process. And, following that tutorial, I replaced the first bits of code with use of the neat plugins mentioned by Jones in his own tutorial: More Types, More Taxonomies, and More Fields.

I played with this a few times but I can now provide an actual tutorial. I’m now doing the whole thing “from scratch” and will write down all steps.

This is with the WordPress 3.0 blogging software installed on a Bluehost account. (The WordPress.com blogging service doesn’t support custom post types.) I use the default Twenty Ten theme as a parent theme.

Since I use WordPress Multisite, I’m creating a new test blog (in Super Admin->Sites, “Add New”). Of course, this wasn’t required, but it helps me make sure the process is reproducible.

Since I already installed the three “More Plugins” (but they’re not “network activated”) I go in the Plugins menu to activate each of them.

I can now create the new “Product” type, based on that Blancer tutorial. To do so, I go to the “More Types” Settings menu, I click on “Add New Post Type,” and I fill in the following information: post type names (singular and plural) and the thumbnail feature. Other options are set by default.

I also set the “Permalink base” in Advanced settings. Not sure it’s required but it seems to make sense.

I click on the “Save” button at the bottom of the page (forgot to do this, the last time).

I then go to the “More Fields” settings menu to create a custom box for the post editing interface.

I add the box title and change the “Use with post types” options (no use in having this in posts).

(Didn’t forget to click “save,” this time!)

I can now add the “Price” field. To do so, I need to click on the “Edit” link next to the “Product Options” box I just created and add click “Add New Field.”

I add the “Field title” and “Custom field key”:

I set the “Field type” to Number.

I also set the slug for this field.

I then go to the “More Taxonomies” settings menu to add a new product classification.

I click “Add New Taxonomy,” and fill in taxonomy names, allow permalinks, add slug, and show tag cloud.

I also specify that this taxonomy is only used for the “Product” type.

(Save!)

Now, the rest is more directly taken from the Blancer tutorial. But instead of copy-paste, I added the files directly to a Twenty Ten child theme. The files are available in this archive.

Here’s the style.css code:

/*
Theme Name: Product Directory
Theme URI: http://enkerli.com/
Description: A product directory child theme based on Kyle Jones, Blancer, and Twenty Ten
Author: Alexandre Enkerli
Version: 0.1
Template: twentyten
*/
@import url("../twentyten/style.css");

The code for functions.php:

<!--?php /**  * ProductDir functions and definitions  * @package WordPress  * @subpackage Product_Directory  * @since Product Directory 0.1  */ /*Custom Columns*/ add_filter("manage_edit-product_columns", "prod_edit_columns"); add_action("manage_posts_custom_column",  "prod_custom_columns"); function prod_edit_columns($columns){ 		$columns = array( 			"cb" =--> "<input type="\&quot;checkbox\&quot;" />",
			"title" => "Product Title",
			"description" => "Description",
			"price" => "Price",
			"catalog" => "Catalog",
		);

		return $columns;
}

function prod_custom_columns($column){
		global $post;
		switch ($column)
		{
			case "description":
				the_excerpt();
				break;
			case "price":
				$custom = get_post_custom();
				echo $custom["price"][0];
				break;
			case "catalog":
				echo get_the_term_list($post->ID, 'catalog', '', ', ','');
				break;
		}
}
?>

And the code in single-product.php:

<!--?php /**  * Template Name: Product - Single  * The Template for displaying all single products.  *  * @package WordPress  * @subpackage Product_Dir  * @since Product Directory 1.0  */ get_header(); ?-->
<div id="container">
<div id="content">
<!--?php the_post(); ?-->

<!--?php 	$custom = get_post_custom($post--->ID);
	$price = "$". $custom["price"][0];

?>
<div id="post-<?php the_ID(); ?><br />">>
<h1 class="entry-title"><!--?php the_title(); ?--> - <!--?=$price?--></h1>
<div class="entry-meta">
<div class="entry-content">
<div style="width: 30%; float: left;">
			<!--?php the_post_thumbnail( array(100,100) ); ?-->
			<!--?php the_content(); ?--></div>
<div style="width: 10%; float: right;">
			Price
<!--?=$price?--></div>
</div>
</div>
</div>
<!-- #content --></div>
<!-- #container -->

<!--?php get_footer(); ?-->

That’s it!

Well, almost..

One thing is that I have to activate my new child theme.

So, I go to the “Themes” Super Admin menu and enable the Product Directory theme (this step isn’t needed with single-site WordPress).

I then activate the theme in Appearance->Themes (in my case, on the second page).

One thing I’ve learnt the hard way is that the permalink structure may not work if I don’t go and “nudge it.” So I go to the “Permalinks” Settings menu:

And I click on “Save Changes” without changing anything. (I know, it’s counterintuitive. And it’s even possible that it could work without this step. But I spent enough time scratching my head about this one that I find it important.)

Now, I’m done. I can create new product posts by clicking on the “Add New” Products menu.

I can then fill in the product details, using the main WYSIWYG box as a description, the “price” field as a price, the “featured image” as the product image, and a taxonomy as a classification (by clicking “Add new” for any tag I want to add, and choosing a parent for some of them).

Now, in the product management interface (available in Products->Products), I can see the proper columns.

Here’s what the product page looks like:

And I’ve accomplished my mission.

The whole process can be achieved rather quickly, once you know what you’re doing. As I’ve been told (by the ever-so-helpful Justin Tadlock of Theme Hybrid fame, among other things), it’s important to get the data down first. While I agree with the statement and its implications, I needed to understand how to build these things from start to finish.

In fact, getting the data right is made relatively easy by my background as an ethnographer with a strong interest in cognitive anthropology, ethnosemantics, folk taxonomies (aka “folksonomies“), ethnography of communication, and ethnoscience. In other words, “getting the data” is part of my expertise.

The more technical aspects, however, were a bit difficult. I understood most of the principles and I could trace several puzzle pieces, but there’s a fair deal I didn’t know or hadn’t done myself. Putting together bits and pieces from diverse tutorials and posts didn’t work so well because it wasn’t always clear what went where or what had to remain unchanged in the code. I struggled with many details such as the fact that Kyle Jones’s code for custom columns wasn’t working first because it was incorrectly copied, then because I was using it on a post type which was “officially” based on pages (instead of posts). Having forgotten the part about “touching” the Permalinks settings, I was unable to get a satisfying output using Jones’s explanations (the fact that he doesn’t use titles didn’t really help me, in this specific case). So it was much harder for me to figure out how to do this than it now is for me to build content directories.

I still have some technical issues to face. Some which are near essential, such as a way to create archive templates for custom post types. Other issues have to do with features I’d like my content directories to have, such as clearly defined roles (the “More Plugins” support roles, but I still need to find out how to define them in WordPress). Yet other issues are likely to come up as I start building content directories, install them in specific contexts, teach people how to use them, observe how they’re being used and, most importantly, get feedback about their use.

But I’m past a certain point in my self-learning journey. I’ve built my confidence (an important but often dismissed component of gaining expertise and experience). I found proper resources. I understood what components were minimally necessary or required. I succeeded in implementing the system and testing it. And I’ve written enough about the whole process that things are even clearer for me.

And, who knows, I may get feedback, questions, or advice..

Minds of All Sizes Think Alike

Or «les esprits de toutes tailles se rencontrent».

This post is a response to the following post about Social Network Analysis (SNA), social change, and communication.

…My heart’s in Accra » Shortcuts in the social graph.

I have too many disparate things to say about that post to make it into a neat and tidy “quickie,” yet I feel like I should probably be working on other things. So we’ll see how this goes.

First, a bit of context..

[This “bit of context” may be a bit long so, please bear with me. Or you could get straight to the point, if you don’t think you can bear the context bit.]

I’ve never met Ethan Zuckerman (@EthanZ), who wrote the post to which I’m responding. And I don’t think we’ve had any extended conversation in the past. Further, I doubt that I’m on his radar. He’s probably seen my name, since I’ve commented on some of his posts and some of his contacts may have had references to me through social media. But I very much doubt that he’s ever mentioned me to anyone. I’m not noticeable to him.

I, on the other hand, have mentioned Zuckerman on several occasions. Latest time I remember was in class, a few weeks ago. It’s a course on Africa and I was giving students a list of online sources with relevance to our work. Zuckerman’s connection to Africa may not be his main thing, despite his blog’s name, but it’s part of the reason I got interested in his work, a few years ago.

In fact, there’s something embarrassing, here.. I so associate Zuckerman to Africa that my mind can’t help but link him to Erik Hersman, aka White African. I did meet Herman. [To be exact, I met Erik at BarCampAustin, which is quite possibly the conference-like event which has had the most influence on me, in the past few years (I go to a lot of these events).] When I did meet Hersman, I made a faux-pas in associating him with Zuckerman. Good-natured as he seemed to be, Hersman smiled as he corrected me.

EthanZ and I have other contacts in common. Jeremy Clarke, for instance, who co-organizes WordCamp Montreal and has been quite active in Montreal’s geek scene. Jeremy’s also a developer for Global Voices, a blogging community that Zuckerman co-founded. I’m assuming Clarke and Zuckerman know each other.

Another mutual contact is Christopher Lydon, host of Radio Open Source. Chris and I have exchanged a few emails, and Zuckerman has been on ROS on a few occasions.

According to Facebook, Zuckerman and I have four contacts in common. Apart from Clarke and Hersman, there’s P. Kerim Friedman and Gerd Leonhard. Kerim is a fellow linguistic anthropologist and we’ve collaborated on the official Society for Linguistic Anthropology (SLA) site. I got in touch with Leonhard through “Music 2.0” issues, as he was interviewed by Charles McEnerney on Well-Rounded Radio.

On LinkedIn, Zuckerman is part of my third degree, with McEnerney as one of my first-degree contacts who could connect me to Zuckerman, through Zuckerman’s contacts.

(Yes, I’m fully aware of the fact that I haven’t name a single woman in this list. Nor someone who doesn’t write in English with some frequency, for that matter.)

By this time, my guess is that you may be either annoyed or confused. “Surely, he can’t be that obsessed with Zuckerman as to stalk him in every network.”

No, I’m not at all obsessed with Ethan Zuckerman in any way, shape, or form. Though I mention him on occasion and I might have a good conversation with him if the occasion arises, I wouldn’t go hang out in Cambridge just in case I might meet him. Though I certainly respect his work, I wouldn’t treat him as my “idol” or anything like that. In other words, he isn’t a focus in my life.

And that’s a key point, to me.

In certain contexts, when social networks are discussed, too much is made of the importance of individuals. Yet, there’s something to be said about relative importance.

In his “shortcuts” post, Zuckerman talks about a special kind of individuals. Those who are able to bypass something of a clustering effect happening in many human networks. Malcolm Gladwell (probably “inspired” by somebody else) has used “connectors” to label a fairly similar category of people and, given Gladwell’s notoriety in some circles, the name has resonance in some contexts (mostly “business-focused people,” I would say, with a clear idea in my mind of the groupthink worldview implied).

In one of my earliest blogposts, I talked about an effect happening through a similar mechanism, calling it the “Social Butterfly Effect” (SBE). I still like it, as a concept. Now, I admit that it focuses on a certain type of individuals. But it’s more about their position in “the grand scheme of things” than about who they are, though I do associate myself with this “type.”

The basic idea is quite simple. People who participate in different (sub)networks, who make such (sub)networks sparser, are having unpredictable and unmeasurable effects on what is transmitted through the network(s).

On one hand, it’s linked to my fragmentary/naïve understanding of the Butterfly Effect in the study of climate and as a component of Chaos Theory.

On the other hand, it’s related to Granovetter‘s well-known notion of “weak ties.” And it seems like Granovetter is making something of a comeback, as we discuss different mechanisms behind social change.

Interestingly, much of what is being said about weak ties, these past few weeks, relates to Gladwell’s flamebait apparent lack of insight in describing current social processes. Sounds like Gladwell may be too caught up in the importance of individuals to truly grok the power of networks.

Case in point.. One of the most useful pieces I’ve read about weak ties, recently, was Jonah Lehrer‘s direct response to Gladwell:

Weak Ties, Twitter and Revolution | Wired Science | Wired.com.

Reading Lehrer’s piece, one gets the clear impression that Gladwell hadn’t “done his homework” on Granovetter before launching his trolling “controversial” piece on activism.

But I digress. Slightly.

Like the Gladwell-specific coverage, Zuckerman’s blogpost is also about social change and he’s already responded to Gladwell. One way to put it is that, as a figure, Gladwell has shaped the discussion in a way similar to a magnetic field orienting iron filings around it. Since it’s a localized effect having to do with polarization, the analogy is fairly useful, as analogies go.

Which brings me to groupthink, the apparent target of Zuckerman’s piece.

Still haven’t read Irving Janis but I’ve been quite interested in groupthink for a while. Awareness of the concept is something I immediately recognize, praise, and associate with critical thinking.

In fact, it’s one of several things I was pleasantly surprised to find in an introductory sociology WikiBook I ended up using in my  “Intro. to Society” course, last year. Critical thinking was the main theme of that course, and this short section was quite fitting in the overall discussion.

So, what of groupthink and networks? Zuckerman sounds worried:

This is interesting to me because I’m intrigued – and worried – by information flows through social networks. If we’re getting more (not lots yet, but more) information through social networks and less through curated media like newspapers, do we run the risk of encountering only information that our friends have access to? Are we likely to be overinformed about some conversations and underinformed about others? And could this isolation lead to ideological polarization, as Cass Sunstein and others suggest? And if those fears are true, is there anything we can do to rewire social networks so that we’re getting richer, more diverse information?

Similar questions have animated many discussions in media-focused circles, especially in those contexts where the relative value (and meaning) of “old vs. new media” may be debated. At about the same time as I started blogging, I remember discussing things with a statistician friend about the polarization effect of media, strong confirmation bias in reading news stories, and political lateralization.

In the United States, especially, there’s a narrative (heard loud and clear) that people who disagree on some basic ideas are unable to hear one another. “Shockingly,” some say, “conservatives and liberals read different things.” Or “those on (the) two sides of (the) debate understand things in completely different ways.” It even reminds me of the connotations of Tannen’s booktitle, You Just Don’t Understand. Irreconciliable differences. (And the first time I mention a woman in this decidedly imbalanced post.)

While, as a French-Canadian ethnographer, my perspective is quite different from Zuckerman, I can’t help but sympathize with the feeling. Not that I associate groupthink with a risk in social media (au contraire!). But, like Zuckerman, I wish to find ways to move beyond these boundaries we impose on ourselves.

Zuckerman specifically discusses the attempt by Onnik Krikorian (@OneWMPhoto) to connect Armenians (at least those in Hayastan) and Azeris, with Facebook “affording” Krikorian some measure of success. This case is now well-known in media-centric circles and it has almost become shorthand for the power of social media. Given a personal interest in Armenians (at least in the Diaspora), my reaction to Krikorian’s success are less related to the media aspect than to the personal one.

At a personal level, boundaries may seem difficult to surmount but they can also be fairly porous and even blurry. Identity may be negotiated. Individuals crossing boundaries may be perceived in diverse ways, some of which have little to do with other people crossing the same boundaries. Things are lived directly, from friendships to wars, from breakups to reconciliations. Significant events happen regardless of the way  they’re being perceived across boundaries.

Not that boundaries don’t matter but they don’t necessarily circumscribe what happens in “personal lives.” To use an seemingly-arbitrary example, code-switching doesn’t “feel” strange at an individual level. It’s only when people insist on separating languages using fairly artificial criteria that alternance between them sounds awkward.

In other words, people cross boundaries all the time and “there’s nothing to it.”

Boundaries have quite a different aspect, at the macrolevel implied by the journalistic worldview (with nation-based checkbox democracy at its core and business-savvy professionalization as its mission). To “macros” like journos and politicos, boundaries look like borders, appearing clearly on maps (including mind ones) and implying important disconnects. The border between Armenia and Azerbaijan is a boundary separating two groups and the conflicts between these two groups reify that boundary. Reaching out across the border is a diplomatic process and necessitates finding the right individuals for the task. Most of the important statuses are ascribed, which may sound horrible to some holding neoliberal ideas about freewill and “individual freedoms.”

Though it’s quite common for networked activities to be somewhat constrained by boundaries, a key feature of networks is that they’re typically boundless. Sure, there are networks which are artificially isolated from the rest. The main example I can find is that of a computer virology laboratory.

Because, technically, you only need one link between two networks to transform them into a single network. So, it’s quite possible to perceive Verizon’s wireless network as a distinct entity, limited by the national boundaries of the U.S. of A. But the simple fact that someone can use Verizon’s network to contact someone in Ségou shows that the network isn’t isolated. Simple, but important to point out.

Especially since we’re talking about a number of things happening on a single network: The Internet. (Yes, there is such a thing as Internet2 and there are some technical distinctions at stake. But we’re still talking about an interconnected world.)

As is well-known, there are significant clusters in this One Network. McLuhan’s once-popular “Global Village” fallacy used to hide this, but we now fully realize that language barriers, national borders, and political lateralization go with “low-bandwidth communication,” in some spots of The Network. “Gs don’t talk to Cs so even though they’re part of the same network, there’s a weak spot, there.” In a Shannon/Weaver view, it sounds quite important to identify these weak spots. “Africa is only connected to North America via a few lines so access is limited, making things difficult for Africans.” Makes sense.

But going back to weak ties, connectors, Zuckerman’s shortcuts, and my own social butterflies, the picture may be a little bit more fleshed out.

Actually, the image I have in mind has, on one side, a wire mesh serving as the floor of an anechoic chamber  and on the other some laser beams going in pseudorandom directions as in Entrapment or Mission Impossible. In the wire mesh, weaker spots might cause a person to fall through and land on those artificial stalagmites. With the laser beams, the pseudorandom structure makes it more difficult to “find a path through the maze.” Though some (engineers) may see the mesh as the ideal structure for any network, there’s something humanly fascinating about the pseudorandom structure of social networks.

Obviously, I have many other ideas in mind. For instance, I wanted to mention “Isabel Wilkerson’s Leaderless March that Remade America.” Or go back to that intro soci Wikibook to talk about some very simple and well-understood ideas about social movements, which often seem to be lacking in discussions of social change. I even wanted to recount some anecdotes of neat network effects in my own life, such as the serendipity coming from discuss disparate subjects to unlike people or the misleading impression that measuring individualized influence is a way to understand social media. Not to mention a whole part I had in my mind about Actor Network Theory, non-human actors, and material culture (the other course I currently teach).

But I feel like going back to more time-sensitive things.

Still, I should probably say a few words about this post’s title.

My mother and I were discussing parallel inventions and polygenesis with the specific theme of moving away from the focus on individualized credit. My favourite example, and one I wish Gladwell (!) had used in Outliers (I actually asked him about it) is that of Gregor Mendel and the “rediscovery” of his laws by de Vries, Correns, and Tschermak. A semi-Marxian version of the synchronous polygenesis part might hold that “ideas are in the air” or that the timing of such dicoveries and inventions has to do with zeitgeist. A neoliberal version could be the “great minds think alike” expression or its French equivalent «Les grands esprits se rencontrent» (“The great spirits meet each other”). Due to my reluctance in sizing up minds, I’d have a hard time using that as a title. In the past, I used a similar title to refer to another form of serendipity:

To me, most normally constituted minds are “great,” so I still could have used the expression as a title. But an advantage of tweaking an expression is that it brings attention to what it implies.

In this case, the “thinking alike” may be a form of groupthink.

 

Academics and Their Publics

Misunderstood by Raffi Asdourian
Misunderstood by Raffi Asdourian

Academics are misunderstood.

Almost by definition.

Pretty much any academic eventually feels that s/he is misunderstood. Misunderstandings about some core notions in about any academic field are involved in some of the most common pet peeves among academics.

In other words, there’s nothing as transdisciplinary as misunderstanding.

It can happen in the close proximity of a given department (“colleagues in my department misunderstand my work”). It can happen through disciplinary boundaries (“people in that field have always misunderstood our field”). And, it can happen generally: “Nobody gets us.”

It’s not paranoia and it’s probably not self-victimization. But there almost seems to be a form of “onedownmanship” at stake with academics from different disciplines claiming that they’re more misunderstood than others. In fact, I personally get the feeling that ethnographers are more among the most misunderstood people around, but even short discussions with friends in other fields (including mathematics) have helped me get the idea that, basically, we’re all misunderstood at the same “level” but there are variations in the ways we’re misunderstood. For instance, anthropologists in general are mistaken for what they aren’t based on partial understanding by the general population.

An example from my own experience, related to my decision to call myself an “informal ethnographer.” When you tell people you’re an anthropologist, they form an image in their minds which is very likely to be inaccurate. But they do typically have an image in their minds. On the other hand, very few people have any idea about what “ethnography” means, so they’re less likely to form an opinion of what you do from prior knowledge. They may puzzle over the term and try to take a guess as to what “ethnographer” might mean but, in my experience, calling myself an “ethnographer” has been a more efficient way to be understood than calling myself an “anthropologist.”

This may all sound like nitpicking but, from the inside, it’s quite impactful. Linguists are frequently asked about the number of languages they speak. Mathematicians are taken to be number freaks. Psychologists are perceived through the filters of “pop psych.” There are many stereotypes associated with engineers. Etc.

These misunderstandings have an impact on anyone’s work. Not only can it be demoralizing and can it impact one’s sense of self-worth, but it can influence funding decisions as well as the use of research results. These misunderstandings can underminine learning across disciplines. In survey courses, basic misunderstandings can make things very difficult for everyone. At a rather basic level, academics fight misunderstandings more than they fight ignorance.

The  main reason I’m discussing this is that I’ve been given several occasions to think about the interface between the Ivory Tower and the rest of the world. It’s been a major theme in my blogposts about intellectuals, especially the ones in French. Two years ago, for instance, I wrote a post in French about popularizers. A bit more recently, I’ve been blogging about specific instances of misunderstandings associated with popularizers, including Malcolm Gladwell’s approach to expertise. Last year, I did a podcast episode about ethnography and the Ivory Tower. And, just within the past few weeks, I’ve been reading a few things which all seem to me to connect with this same issue: common misunderstandings about academic work. The connections are my own, and may not be so obvious to anyone else. But they’re part of my motivations to blog about this important issue.

In no particular order:

But, of course, I think about many other things. Including (again, in no particular order):

One discussion I remember, which seems to fit, included comments about Germaine Dieterlen by a friend who also did research in West Africa. Can’t remember the specifics but the gist of my friend’s comment was that “you get to respect work by the likes of Germaine Dieterlen once you start doing field research in the region.” In my academic background, appreciation of Germaine Dieterlen’s may not be unconditional, but it doesn’t necessarily rely on extensive work in the field. In other words, while some parts of Dieterlen’s work may be controversial and it’s extremely likely that she “got a lot of things wrong,” her work seems to be taken seriously by several French-speaking africanists I’ve met. And not only do I respect everyone but I would likely praise someone who was able to work in the field for so long. She’s not my heroine (I don’t really have heroes) or my role-model, but it wouldn’t have occurred to me that respect for her wasn’t widespread. If it had seemed that Dieterlen’s work had been misunderstood, my reflex would possibly have been to rehabilitate her.

In fact, there’s  a strong academic tradition of rehabilitating deceased scholars. The first example which comes to mind is a series of articles (PDF, in French) and book chapters by UWO linguistic anthropologist Regna Darnell.about “Benjamin Lee Whorf as a key figure in linguistic anthropology.” Of course, saying that these texts by Darnell constitute a rehabilitation of Whorf reveals a type of evaluation of her work. But that evaluation comes from a third person, not from me. The likely reason for this case coming up to my mind is that the so-called “Sapir-Whorf Hypothesis” is among the most misunderstood notions from linguistic anthropology. Moreover, both Whorf and Sapir are frequently misunderstood, which can make matters difficulty for many linguistic anthropologists talking with people outside the discipline.

The opposite process is also common: the “slaughtering” of “sacred cows.” (First heard about sacred cows through an article by ethnomusicologist Marcia Herndon.) In some significant ways, any scholar (alive or not) can be the object of not only critiques and criticisms but a kind of off-handed dismissal. Though this often happens within an academic context, the effects are especially lasting outside of academia. In other words, any scholar’s name is likely to be “sullied,” at one point or another. Typically, there seems to be a correlation between the popularity of a scholar and the likelihood of her/his reputation being significantly tarnished at some point in time. While there may still be people who treat Darwin, Freud, Nietzsche, Socrates, Einstein, or Rousseau as near divinities, there are people who will avoid any discussion about anything they’ve done or said. One way to put it is that they’re all misunderstood. Another way to put it is that their main insights have seeped through “common knowledge” but that their individual reputations have decreased.

Perhaps the most difficult case to discuss is that of Marx (Karl, not Harpo). Textbooks in introductory sociology typically have him as a key figure in the discipline and it seems clear that his insight on social issues was fundamental in social sciences. But, outside of some key academic contexts, his name is associated with a large series of social events about which people tend to have rather negative reactions. Even more so than for Paul de Man or  Martin Heidegger, Marx’s work is entangled in public opinion about his ideas. Haven’t checked for examples but I’m quite sure that Marx’s work is banned in a number of academic contexts. However, even some of Marx’s most ardent opponents are likely to agree with several aspects of Marx’s work and it’s sometimes funny how Marxian some anti-Marxists may be.

But I digress…

Typically, the “slaughtering of sacred cows” relates to disciplinary boundaries instead of social ones. At least, there’s a significant difference between your discipline’s own “sacred cows” and what you perceive another discipline’s “sacred cows” to be. Within a discipline, the process of dismissing a prior scholar’s work is almost œdipean (speaking of Freud). But dismissal of another discipline’s key figures is tantamount to a rejection of that other discipline. It’s one thing for a physicist to show that Newton was an alchemist. It’d be another thing entirely for a social scientist to deconstruct James Watson’s comments about race or for a theologian to argue with Darwin. Though discussions may have to do with individuals, the effects of the latter can widen gaps between scholarly disciplines.

And speaking of disciplinarity, there’s a whole set of issues having to do with discussions “outside of someone’s area of expertise.” On one side, comments made by academics about issues outside of their individual areas of expertise can be very tricky and can occasionally contribute to core misunderstandings. The fear of “talking through one’s hat” is quite significant, in no small part because a scholar’s prestige and esteem may greatly decrease as a result of some blatantly inaccurate statements (although some award-winning scholars seem not to be overly impacted by such issues).

On the other side, scholars who have to impart expert knowledge to people outside of their discipline  often have to “water down” or “boil down” their ideas and, in effect, oversimplifying these issues and concepts. Partly because of status (prestige and esteem), lowering standards is also very tricky. In some ways, this second situation may be more interesting. And it seems unavoidable.

How can you prevent misunderstandings when people may not have the necessary background to understand what you’re saying?

This question may reveal a rather specific attitude: “it’s their fault if they don’t understand.” Such an attitude may even be widespread. Seems to me, it’s not rare to hear someone gloating about other people “getting it wrong,” with the suggestion that “we got it right.”  As part of negotiations surrounding expert status, such an attitude could even be a pretty rational approach. If you’re trying to position yourself as an expert and don’t suffer from an “impostor syndrome,” you can easily get the impression that non-specialists have it all wrong and that only experts like you can get to the truth. Yes, I’m being somewhat sarcastic and caricatural, here. Academics aren’t frequently that dismissive of other people’s difficulties understanding what seem like simple concepts. But, in the gap between academics and the general population a special type of intellectual snobbery can sometimes be found.

Obviously, I have a lot more to say about misunderstood academics. For instance, I wanted to address specific issues related to each of the links above. I also had pet peeves about widespread use of concepts and issues like “communities” and “Eskimo words for snow” about which I sometimes need to vent. And I originally wanted this post to be about “cultural awareness,” which ends up being a core aspect of my work. I even had what I might consider a “neat” bit about public opinion. Not to mention my whole discussion of academic obfuscation (remind me about “we-ness and distinction”).

But this is probably long enough and the timing is right for me to do something else.

I’ll end with an unverified anecdote that I like. This anecdote speaks to snobbery toward academics.

[It’s one of those anecdotes which was mentioned in a course I took a long time ago. Even if it’s completely fallacious, it’s still inspiring, like a tale, cautionary or otherwise.]

As the story goes (at least, what I remember of it), some ethnographers had been doing fieldwork  in an Australian cultural context and were focusing their research on a complex kinship system known in this context. Through collaboration with “key informants,” the ethnographers eventually succeeded in understanding some key aspects of this kinship system.

As should be expected, these kinship-focused ethnographers wrote accounts of this kinship system at the end of their field research and became known as specialists of this system.

After a while, the fieldworkers went back to the field and met with the same people who had described this kinship system during the initial field trip. Through these discussions with their “key informants,” the ethnographers end up hearing about a radically different kinship system from the one about which they had learnt, written, and taught.

The local informants then told the ethnographers: “We would have told you earlier about this but we didn’t think you were able to understand it.”

In Phase

Lissajous curve
Lissajous curve

Something which happens to me on a rather regular basis (and about which I blogged before) is that I’ll hear about something right after thinking about it. For instance, if I think about the fact that a given tool should exist, it may be announced right at that moment.

Hey, I was just thinking about this!

The effect is a bit strange but it’s quite easy to explain. It feels like a “premonition,” but it probably has more to do with “being in phase.” In some cases, it may also be that I heard about that something but hadn’t registered the information. I know it happens a lot and  it might not be too hard to trace back. But I prefer thinking about phase.

And, yes, I am thinking about phase difference in waves. Not in a very precise sense, but the image still works, for me. Especially with the Lissajous representation, as above.

See, I don’t particularly want to be “ahead of the curve” and I don’t particularly mind being “behind the curve.” But when I’m right “in the curve,” something interesting happens. I’m “in the now.”

I originally thought about being “in tune” and it could also be about “in sync” or even “matching impedances.” But I still like the waves analogy. Especially since, when two waves are in phase, they reinforce one another. As analogies go, it’s not only a beautiful one, but a powerful one. And, yes, I do think about my sweetheart.

One reason I like the concept of phase difference is that I think through sound. My first exposure to the concept comes from courses in musical acoustics, almost twenty years ago. It wasn’t the main thing I’d remember from the course and it’s not something I investigated at any point since. Like I keep telling students, some things hit you long after you’ve heard about it in a course. Lifelong learning and “landminds” are based on such elements, even tiny unimportant ones. Phase difference is one such thing.

And it’s no big deal, of course. It’s not like I spent days thinking about these concepts. But I’ve been feeling like writing, lately, and this is as good an opportunity as any.

The trigger for this particular thing is rather silly and is probably explained more accurately, come to think of it, by “unconsciously registering” something before consciously registering it.

Was having breakfast and started thinking about the importance of being environmentally responsible, the paradox of “consumption as freedom,” the consequences of some lifestyle choices including carfree living, etc. This stream of thought led me, not unexpectedly, to the perspectives on climate change, people’s perception of scientific evidence, and the so-called ClimateGate. I care a lot about critical thinking, regardless of whether or not I agree with a certain idea, so I think the email controversy shows the importance of transparency. So far, nothing unexpected. Within a couple of minutes, I had covered a few of the subjects du jour. And that’s what struck me, because right then, I (over)heard a radio host introduce a guest whose talk is titled:

What is the role of climate scientists in the climate change debate?

Obviously, Tremblay addressed ClimateGate quite directly. So my thoughts were “in phase” with Tremblay’s.

A few minutes prior to (over)hearing this introduction, I (over)heard a comment about topics of social conversations at different points in recent history. According to screenwriter Fabienne Larouche, issues covered in the first seasons of her “flagship” tv series are still at the forefront in Quebec society today, fourteen years later. So I was probably even more “in tune” with the notion of being “in phase.” Especially with my society.

I said “(over)heard” because I wasn’t really listening to that radio show. It was just playing in the background and I wasn’t paying much attention. I don’t tend to listen to live radio but I do listen to some radio recordings as podcasts. One reason I like doing so is that I can pay much closer attention to what I hear. Another is that I can listen to what I want when I feel like listen to it, which means that I can prepare myself for a heady topic or choose some tech-fluff to wind down after a course. There’s also the serendipity of listening to very disparate programmes in the same listening session, as if I were “turning the dial” after each show on a worldwide radio (I often switch between French and English and/or between European and North American sources). For a while now, I’ve been listening to podcasts at double-speed, which helps me focus on what’s most significant.

(In Jazz, we talk about “top notes,” meaning the ones which are more prominent. It’s easier to focus on them at double-speed than at normal speed so “double-times” have an interesting cognitive effect.)

So, I felt “in phase.” As mentioned, it probably has much more to do with having passively heard things without paying attention yet letting it “seep into my brain” to create connections between a few subjects which get me to the same point as what comes later. A large part of this is well-known in psychology, especially in terms of cognition. We start noticing things when they enter into a schema we have in our mind. These things we start noticing were there all along so the “discovery” is only in our mind (in the sense that it wouldn’t be a discovery for others). When we learn a new word, for instance, we start hearing it everywhere.

But there are also words which start being used by everyone because they have been diffused largely at a given point in time. An actual neologism can travel quickly and a word in our passive vocabulary can also come to prominence, especially in mainstream media. Clearly, this is an issue of interest to psychologists, folklorists, and media analysts alike. I’m enough of a folklorist and media observer to think about the social processes behind the diffusion of terms regardless of what psychologists think.

A few months back, I got the impression that the word “nimble” had suddenly increased in currency after it was used in a speech by the current PotUS. Since I’m a non-native speaker of English, I’m likely to be accused of noticing the word because it’s part my own passive vocabulary. I have examples in French, though some are with words which were new to me, at the time («peoplisation», «battante»…). I probably won’t be able to defend myself from those who say that it’s just a matter of my own exposure to those terms. Though there are ways to analyze the currency of a given term, I’m not sure I trust this type of analysis a lot more than my gut feeling, at least in terms of realtime trends.

Which makes me think of “memetics.” Not in the strict sense that Dawkins would like us to use. But in the way popular culture cares about the propagation of “units of thought.” I recently read a fascinating blogpost (in French) about  memetics from this perspective, playing Dawkins against himself. As coincidences keep happening (or, more accurately, as I’m accutely tuned to find coincidences everywhere), I’ve been having a discussion about Mahir‘s personal homepage (aka “I kiss you”), who became an “Internet celebrity” through this process which is now called memetic. The reason his page was noticed isn’t that it was so unique. But it had this je ne sais quoi which captured the imagination, at the time (the latter part of the “Dot-Com Bubble”). As some literary critics and many other humanists teach us, it’s not the item itself which counts, it’s how we receive it (yes, I tend to be on the “reception” and “eye of the beholder” side of things). Mahir was striking because he was, indeed, “out of phase” with the times.

As I think about phase, I keep hearing the other acoustic analogy: the tuning of sine waves. When a sine wave is very slightly “out of tune” with another, we hear a very slow oscillation (interference beats) until they produce resonance. There’s a direct relationship between beat tones and phase, but I think “in tune” and “in phase” remain separate analogies.

One reason I like to think about waves for these analogies is that I tend to perceive temporal change through these concepts. If we think of historical change through cycles, being “in phase” is a matter of matching two change processes until they’re aligned but the cycles may be in harmonic relationships. One can move twice as fast as society and still be “in phase” with it.

Sure, I’m overextending the analogies, and there’s something far-fetched about this. But that’s pretty much what I like about analogical thinking. As I’m under the weather, this kind of rambling is almost therapeutic.

No Office Export in Keynote/Numbers for iPad?

To be honest, I’m getting even more excited about the iPad. Not that we get that much more info about it, but:

For one thing, the Pages for iPad webpage is explicitly stating Word support:

Attach them to an email as Pages files for Mac, Microsoft Word files, or PDF documents.

Maybe this is because Steve Jobs himself promised it to Walt Mossberg?
Thing is, the equivalent pages about Keynote for iPad and about Numbers for iPad aren’t so explicit:

The presentations you create in Keynote on your iPad can be exported as Keynote files for Mac or PDF documents

and…

To share your work, export your spreadsheet as a Numbers file for Mac or PDF document

Not a huge issue, but it seems strange that Apple would have such an “export to Microsoft Office” feature on only one of the three “iWork for iPad” apps. Now, the differences in the way exports are described may not mean that Keynote won’t be able to export to Microsoft PowerPoint or that Numbers won’t be able to export to Microsoft Excel. After all, these texts may have been written at different times. But it does sound like PowerPoint and Excel will be import-only, on the iPad.

Which, again, may not be that big an issue. Maybe iWork.com will work well enough for people’s needs. And some other cloud-based tools do support Keynote. (Though Google Docs and Zoho Show don’t.)

The reason I care is simple: I do share most of my presentation files. Either to students (as resources on Moodle) or to whole wide world (through Slideshare). My desktop outliner of choice, OmniOutliner, exports to Keynote and Microsoft Word. My ideal workflow would be to send, in parallel, presentation files to Keynote for display while on stage and to PowerPoint for sharing. The Word version could also be useful for sharing.

Speaking of presenting “slides” on stage, I’m also hoping that the “iPad Dock Connector to VGA Adapter” will support “presenter mode” at some point (though it doesn’t seem to be the case, right now). I also dream of a way to control an iPad presentation with some kind of remote. In fact, it’s not too hard to imagine it as an iPod touch app (maybe made by Appiction, down in ATX).

To be clear: my “presentation files” aren’t really about presenting so much as they are a way to package and organize items. Yes, I use bullet points. No, I don’t try to make the presentation sexy. My presentation files are acting like cue cards and like whiteboard snapshots. During a class, I use the “slides” as a way to keep track of where I planned the discussion to go. I can skip around, but it’s easier for me to get at least some students focused on what’s important (the actual depth of the discussion) because they know the structure (as “slides”) will be available online. Since I also podcast my lectures, it means that they can go back to all the material.

I also use “slides” to capture things we build in class, such as lists of themes from the readings or potential exam questions.  Again, the “whiteboard” idea. I don’t typically do the same thing during a one-time talk (say, at an unconference). But I still want to share my “slides,” at some point.

So, in all of these situations, I need a file format for “slides.” I really wish there were a format which could work directly out of the browser and could be converted back and forth with other formats (especially Keynote, OpenOffice, and PowerPoint). I don’t need anything fancy. I don’t even care about transitions, animations, or even inserting pictures. But, despite some friends’ attempts at making me use open solutions, I end up having to use presentation files.

Unfortunately, at this point, PowerPoint is the de facto standard for presentation files. So I need it, somehow. Not that I really need PowerPoint itself. But it’s still the only format I can use to share “slides.”

So, if Keynote for iPad doesn’t export directly to PowerPoint, it means that I’ll have to find another way to make my workflow fit.

Ah, well…

Scriptocentrism and the Freedom to Think

As a comment on my previous blogpost on books, a friend sent me (through Facebook) a link to a blogpost about a petition to Amazon with the following statement:

The freedom to read is tantamount to the freedom to think.

As this friend and I are both anthros+africanists, I’m reacting (perhaps a bit strongly) to that statement.

Given my perspective, I would dare say that I find this statement (brought about by DbD)… ethnocentric.

There, I said it.

And I’ll try to back it up in this blogpost in order to spark even more discussion.

We won’t exhaust this topic any time soon, but I feel there’s a lot we can do about it which has rarely been done.

I won’t use the textbook case of “Language in the Inner City,” but it could help us talk about who decides, in a given social context, what is important. We both come from a literacy-focused background, so we may have to take a step back. Not sure if Bourdieu has commented on Labov, especially in terms of what all this means for “education,” but I’d even want to bring in Ivan Illich, at some point.

Hunters with whom I’ve been working, in Mali, vary greatly in terms of literacy. Some of them have a strong university background and one can even write French legalese (he’s a judge). Others (or some of the same) have gone to Koranic school long enough that can read classical Arabic. Some have the minimal knowledge of Arabic which suffices, for them, to do divination. Many of them have a very low level of functional literacy. There’s always someone around them who can read and write, so they’re usually not out of the loop and it’s not like the social hierarchy stereotypical of the Catholic Church during the Middle Ages in Europe. It’s a very different social context which can hardly be superimposed with the history of writing and the printing press in Europe.

In terms of “freedom to thinik,” I really wouldn’t say that they’re lacking. Of course, “free thinker” has a specific meaning in liberal societies with a European background. But even this meaning can be applied to many people I’ve met in Mali.

And I go back to the social context. Those with the highest degree of functional literacy aren’t necessarily those with the highest social status. And unlike Harlem described by Labov, it’s a relatively independent context from the one in which literacy is a sine qua non. Sure, it’s a neocolonial context and Euro-Americans keep insisting that literacy in Latin script is “the most important thing ever” if they are to become a true liberal democracy. Yet, internally, it’s perfectly possible for someone to think freely, get recognition, and help other people to think without going through the written medium.

Many of those I know who have almost nonexistent skills in the written medium also have enough power (in a Weberian sense) that they get others to do the reading and writing for them. And because there are many social means to ensure that communication has worked appropriately, these “scribes” aren’t very likely to use this to take anything away from those for whom they read and write.

In Switzerland, one of my recent ancestors was functionally illiterate. Because of this, she “signed away” most of her wealth. Down the line, I’m one of her very few heirs. So, in a way, I lost part of my inheritance due to illiteracy.

Unless the switch to a European model for notarial services becomes complete, a case like this is unlikely to occur among people I know in Mali. If it does happen, it’s clearly not a failure of the oral system but a problem with this kind of transition. It’s somewhat similar to the situation with women in diverse parts of the continent during the period of direct colonialism: the fact that women have lost what powers they had (say, in a matrilineal/matrilocal society) has to do with the switch to a hierarchical system which put the emphasis on new factors which excluded the type of influence women had.

In other words, I fully understand the connections between liberalism and literacy and I’ve heard enough about the importance of the printing press and journalism in these liberal societies to understand what role reading has played in those contexts. I simply dispute the notion that these connections should be universal.

Yes, I wish the “Universal Declaration of Human Rights” (including the (in)famous Article 26, which caused so many issues) were more culturally aware.

I started reading Deschooling Society a few weeks ago. In terms of “insight density,” it’s much higher than the book which prompted this discussion. While reading the first chapter, I constructed a number of ideas which I personally find useful.

I haven’t finished reading the book. Yet. I might eventually finish it. But much of what I wanted to get from that book, I was able to get from diverse sources. Including that part of the book I did read, sequentially. But, also, everything which has been written about Illich since 1971. And I’ll be interested in reading comments by the reading group at Wikiversity.

Given my background, I have as many “things to say” about the issues surrounding schooling as what I’ve read. If I had the time, I could write as much on what I’ve read from that book and it’d probably bring me a lot of benefits.

I’ve heard enough strong reactions against this attitude I’m displaying that I can hear it, already: “how can you talk about a book you haven’t read.” And I sincerely think these people miss an important point. I wouldn’t go so far as to say that their reading habits are off (that’d be mean), especially since those are well-adapted to certain contexts, including what I call scriptocentrism. Not that these people are scriptocentric. But their attitude “goes well with” scriptocentrism.

Academia, despite being to context for an enormous amount of writing and reading, isn’t displaying that kind of scriptocentrism. Sure, a lot of what we do needs to be written (although, it’s often surprising how much insight goes unwritten in the work of many an academic). And we do get evaluated through our writing. Not to mention that we need to write in a very specific mode, which almost causes a diglossia.

But we simply don’t feel forced to “read the whole text.”

A colleague has described this as the “dirty little secret” of academia. And one which changes many things for students, to the point that it almost sounds as if it remains a secret so as to separate students into categories of “those who get it” and “the mass.”

It doesn’t take a semester to read a textbook so there are students who get the impression that they can simply read the book in a weekend and take the exams. These students may succeed, depending on the course. In fact, they may get really good grades. But they run into a wall if they want to go on with a career making any use of knowledge construction skills.

Bill Reimer has interesting documents about “better reading.” It’s a PowerPoint presentation accompanied by exercises in a PDF format. (No, I won’t discuss format here.)

I keep pointing students to those documents for a simple reason: Reimer isn’t advocating reading every word in sequence. His “skim then focus” advice might be the one piece which is harder to get through to people but it’s tremendously effective in academic contexts. It’s also one which is well-adapted to the kind of online reading I’m thinking about. And not necessarily that good for physical books. Sure, you can efficiently flip pages in a book. But skimming a text on paper is more likely to be about what stands out visually than about the structure of the text. Especially with book-length texts. The same advice holds with physical books, of course. After all, this kind of advice originally comes from that historical period which I might describe as the “heyday of books”: the late 20th Century. But I’d say that the kind of “better reading” Reimer describes is enhanced in the context of online textuality. Not just the “Read/Write Web” but Instant Messaging, email, forums, ICQ, wikis, hypertext, Gopher, even PowerPoint…

Much of this has to do with different models of human communication. The Shannon/Weaver crowd have a linear/directional model, based on information processing. Codec and modem. Something which, after Irvine’s Shadow Conversations, I tend to call “the football theory of communication.” This model might be the best-known one, especially among those who study in departments of communication along with other would-be journalists. Works well for a “broadcast” medium with mostly indirect interaction (books, television, radio, cinema, press conferences, etc.). Doesn’t work so well for the backchannel-heavy “smalltalk”  stuff of most human communication actually going on in this world.

Some cognitivists (including Chomsky) have a schema-based model. Constructivists (from Piaget on) have an elaborate model based on knowledge. Several linguistic anthropologists (including yours truly but also Judith Irvine, Richard Bauman, and Dell Hymes) have a model which gives more than lipservice to the notion of performance. And there’s a functional model of any human communication in Jakobson’s classic text on verbal communication. It’s a model which can sound as if it were linear/bidirectional but it’s much broader than this. His six “functions of verbal communication” do come from six elements of the communication process (channel, code, form, context, speaker, listener). But each of these elements embeds a complex reality and Jakobson’s model seems completely compatible with a holistic approach to human communication. In fact, Jakobson has had a tremendous impact on a large variety of people, including many key figures in linguistic anthropology along with Lévi-Strauss and, yes, even Chomsky.

(Sometimes, I wish more people knew about Jakobson. Oh, wait! Since Jakobson was living in the US, I need to americanize this statement: “Jakobson is the most underrated scholar ever.”)

All these models do (or, in my mind, should) integrate written communication. Yet scriptocentrism has often led us far away from “texts as communication” and into “text as an object.” Scriptocentrism works well with modernity. Going away from scriptocentrism is a way to accept our postmodern reality.

Why I Need an iPad

I’m one of those who feel the iPad is the right tool for the job.

This is mostly meant as a reply to this blogthread. But it’s also more generally about my personal reaction to Apple’s iPad announcement.

Some background.

I’m an ethnographer and a teacher. I read a fair deal, write a lot of notes, and work in a variety of contexts. These days, I tend to spend a good amount of time in cafés and other public places where I like to work without being too isolated. I also commute using public transit, listen to lots of podcast, and create my own. I’m also very aural.

I’ve used a number of PDAs, over the years, from a Newton MessagePad 130 (1997) to a variety of PalmOS devices (until 2008). In fact, some people readily associated me with PDA use.

As soon as I learnt about the iPod touch, I needed one. As soon as I’ve heard about the SafariPad, I wanted one. I’ve been an intense ‘touch user since the iPhone OS 2.0 release and I’m a happy camper.

(A major reason I never bought an iPhone, apart from price, is that it requires a contract.)

In my experience, the ‘touch is the most appropriate device for all sorts of activities which are either part of an other activity (reading during a commute) or are simply too short in duration to constitute an actual “computer session.” You don’t “sit down to work at your ‘touch” the way you might sit in front of a laptop or desktop screen. This works great for “looking up stufff” or “checking email.” It also makes a lot of sense during commutes in crowded buses or metros.

In those cases, the iPod touch is almost ideal. Ubiquitous access to Internet would be nice, but that’s not a deal-breaker. Alternative text-input methods would help in some cases, but I do end up being about as fast on my ‘touch as I was with Graffiti on PalmOS.

For other tasks, I have a Mac mini. Sure, it’s limited. But it does the job. In fact, I have no intention of switching for another desktop and I even have an eMachines collecting dust (it’s too noisy to make a good server).

What I miss, though, is a laptop. I used an iBook G3 for several years and loved it. For a little while later, I was able to share a MacBook with somebody else and it was a wonderful experience. I even got to play with the OLPC XO for a few weeks. That one was not so pleasant an experience but it did give me a taste for netbooks. And it made me think about other types of iPhone-like devices. Especially in educational contexts. (As I mentioned, I’m a teacher)

I’ve been laptop-less for a while, now. And though my ‘touch replaces it in many contexts, there are still times when I’d really need a laptop. And these have to do with what I might call “mobile sessions.”

For instance: liveblogging a conference or meeting. I’ve used my ‘touch for this very purpose on a good number of occasions. But it gets rather uncomfortable, after a while, and it’s not very fast. A laptop is better for this, with a keyboard and a larger form factor. But the iPad will be even better because of lower risks of RSI. A related example: just imagine TweetDeck on iPad.

Possibly my favourite example of a context in which the iPad will be ideal: presentations. Even before learning about the prospect of getting iWork on a tablet, presentations were a context in which I really missed a laptop.

Sure, in most cases, these days, there’s a computer (usually a desktop running XP) hooked to a projector. You just need to download your presentation file from Slideshare, show it from Prezi, or transfer it through USB. No biggie.

But it’s not the extra steps which change everything. It’s the uncertainty. Even if it’s often unfounded, I usually get worried that something might just not work, along the way. The slides might not show the same way as you see it because something is missing on that computer or that computer is simply using a different version of the presentation software. In fact, that software is typically Microsoft PowerPoint which, while convenient, fits much less in my workflow than does Apple Keynote.

The other big thing about presentations is the “presenter mode,” allowing you to get more content than (or different content from) what the audience sees. In most contexts where I’ve used someone else’s computer to do a presentation, the projector was mirroring the computer’s screen, not using it as a different space. PowerPoint has this convenient “presenter view” but very rarely did I see it as an available option on “the computer in the room.” I wish I could use my ‘touch to drive presentations, which I could do if I installed software on that “computer in the room.” But it’s not something that is likely to happen, in most cases.

A MacBook solves all of these problems. and it’s an obvious use for laptops. But how, then, is the iPad better? Basically because of interface. Switching slides on a laptop isn’t hard, but it’s more awkward than we realize. Even before watching the demo of Keynote on the iPad, I could simply imagine the actual pleasure of flipping through slides using a touch interface. The fit is “natural.”

I sincerely think that Keynote on the iPad will change a number of things, for me. Including the way I teach.

Then, there’s reading.

Now, I’m not one of those people who just can’t read on a computer screen. In fact, I even grade assignments directly from the screen. But I must admit that online reading hasn’t been ideal, for me. I’ve read full books as PDF files or dedicated formats on PalmOS, but it wasn’t so much fun, in terms of the reading process. And I’ve used my ‘touch to read things through Stanza or ReadItLater. But it doesn’t work so well for longer reading sessions. Even in terms of holding the ‘touch, it’s not so obvious. And, what’s funny, even a laptop isn’t that ideal, for me, as a reading device. In a sense, this is when the keyboard “gets in the way.”

Sure, I could get a Kindle. I’m not a big fan of dedicated devices and, at least on paper, I find the Kindle a bit limited for my needs. Especially in terms of sources. I’d like to be able to use documents in a variety of formats and put them in a reading list, for extended reading sessions. No, not “curled up in bed.” But maybe lying down in a sofa without external lighting. Given my experience with the ‘touch, the iPad is very likely the ideal device for this.

Then, there’s the overall “multi-touch device” thing. People have already been quite creative with the small touchscreen on iPhones and ‘touches, I can just imagine what may be done with a larger screen. Lots has been said about differences in “screen real estate” in laptop or desktop screens. We all know it can make a big difference in terms of what you can display at the same time. In some cases, two screens isn’t even a luxury, for instance when you code and display a page at the same time (LaTeX, CSS…). Certainly, the same qualitative difference applies to multitouch devices. Probably even more so, since the display is also used for input. What Han found missing in the iPhone’s multitouch was the ability to use both hands. With the iPad, Han’s vision is finding its space.

Oh, sure, the iPad is very restricted. For instance, it’s easy to imagine how much more useful it’d be if it did support multitasking with third-party apps. And a front-facing camera is something I was expecting in the first iPhone. It would just make so much sense that a friend seems very disappointed by this lack of videoconferencing potential. But we’re probably talking about predetermined expectations, here. We’re comparing the iPad with something we had in mind.

Then, there’s the issue of the competition. Tablets have been released and some multitouch tablets have recently been announced. What makes the iPad better than these? Well, we could all get in the same OS wars as have been happening with laptops and desktops. In my case, the investment in applications, files, and expertise that I have made in a Mac ecosystem rendered my XP years relatively uncomfortable and me appreciate returning to the Mac. My iPod touch fits right in that context. Oh, sure, I could use it with a Windows machine, which is in fact what I did for the first several months. But the relationship between the iPhone OS and Mac OS X is such that using devices in those two systems is much more efficient, in terms of my own workflow, than I could get while using XP and iPhone OS. There are some technical dimensions to this, such as the integration between iCal and the iPhone OS Calendar, or even the filesystem. But I’m actually thinking more about the cognitive dimensions of recognizing some of the same interface elements. “Look and feel” isn’t just about shiny and “purty.” It’s about interactions between a human brain, a complex sensorimotor apparatus, and a machine. Things go more quickly when you don’t have to think too much about where some tools are, as you’re working.

So my reasons for wanting an iPad aren’t about being dazzled by a revolutionary device. They are about the right tool for the job.

Homeroasting and Coffee Geekness

I’m a coffee geek. By which I mean that I have a geeky attitude to coffee. I’m passionate about the crafts and arts of coffee making, I seek coffee-related knowledge wherever I can find it, I can talk about coffee until people’s eyes glaze over (which happens more quickly than I’d guess possible), and I even dream about coffee gadgets. I’m not a typical gadget freak, as far as geek culture goes, but coffee is one area where I may invest in some gadgetry.

Perhaps my most visible acts of coffee geekery came in the form of updates I posted through diverse platforms about my home coffee brewing experiences. Did it from February to July. These posts contained cryptic details about diverse measurements, including water temperature and index of refraction. It probably contributed to people’s awareness of my coffee geek identity, which itself has been the source of fun things like a friend bringing me back coffee from Ethiopia.

But I digress, a bit. This is both about coffee geekness in general and about homeroasting in particular.

See, I bought myself this Hearthware i-Roast 2 dedicated homeroasting device. And I’m dreaming about coffee again.

Been homeroasting since December 2002, at the time I moved to Moncton, New Brunswick and was lucky enough to get in touch with Terry Montague of Down Esst Coffee.

Though I had been wishing to homeroast for a while before that and had become an intense coffee-lover fifteen years prior to contacting him, Terry is the one who enabled me to start roasting green coffee beans at home. He procured me a popcorn popper, sourced me some quality green beans, gave me some advice. And off I was.

Homeroasting is remarkably easy. And it makes a huge difference in one’s appreciation of coffee. People in the coffee industry, especially baristas and professional roasters, tend to talk about the “channel” going from the farmer to the “consumer.” In some ways, homeroasting gets the coffee-lover a few steps closer to the farmer, both by eliminating a few intermediaries in the channel and by making coffee into much less of a commodity. Once you’ve spent some time smelling the fumes emanated by different coffee varietals and looking carefully at individual beans, you can’t help but get a deeper appreciation for the farmer’s and even the picker’s work. When you roast 150g or less at a time, every coffee bean seems much more valuable. Further, as you experiment with different beans and roast profiles, you get to experience coffee in all of its splendour.

A popcorn popper may sound like a crude way to roast coffee. And it might be. Naysayers may be right in their appraisal of poppers as a coffee roasting method. You’re restricted in different ways and it seems impossible to produce exquisite coffee. But having roasted with a popper for seven years, I can say that my poppers gave me some of my most memorable coffee experiences. Including some of the most pleasant ones, like this organic Sumatra from Theta Ridge Coffee that I roasted in my campus appartment at IUSB and brewed using my beloved Brikka.

Over the years, I’ve roasted a large variety of coffee beans. I typically buy a pound each of three or four varietals and experiment with them for a while.

Mostly because I’ve been moving around quite a bit, I’ve been buying green coffee beans from a rather large variety of places. I try to buy them locally, as much as possible (those beans have travelled far enough and I’ve had enough problems with courier companies). But I did participate in a few mail orders or got beans shipped to me for some reason or another. Sourcing green coffee beans has almost been part of my routine in those different places where I’ve been living since 2002: Moncton, Montreal, Fredericton, South Bend, Northampton, Brockton, Cambridge, and Austin. Off the top of my head, I’ve sourced beans from:

  1. Down East
  2. Toi, moi & café
  3. Brûlerie Saint-Denis
  4. Brûlerie des quatre vents
  5. Terra
  6. Theta Ridge
  7. Dean’s Beans
  8. Green Beanery
  9. Cuvée
  10. Fair Bean
  11. Sweet Maria’s
  12. Evergreen Coffee
  13. Mon café vert
  14. Café-Vrac
  15. Roastmasters
  16. Santropol

And probably a few other places, including this one place in Ethiopia where my friend Erin bought some.

So, over the years, I got beans from a rather large array of places and from a wide range of regional varietals.

I rapidly started blending freshly-roasted beans. Typically, I would start a blend by roasting three batches in a row. I would taste some as “single origin” (coffee made from a single bean varietal, usually from the same farm or estate), shortly after roasting. But, typically, I would mix my batches of freshly roasted coffee to produce a main blend. I would then add fresh batches after a few days to fine-tune the blend to satisfy my needs and enhance my “palate” (my ability to pick up different flavours and aromas).

Once the quantity of green beans in a particular bag would fall below an amount I can reasonably roast as a full batch (minimum around 100g), I would put those green beans in a pre-roast blend, typically in a specially-marked ziplock bag. Roasting this blend would usually be a way for me to add some complexity to my roasted blends.

And complexity I got. Lots of diverse flavours and aromas. Different things to “write home about.”

But I was obviously limited in what I could do with my poppers. The only real controls that I had in homeroasting, apart from blending, consisted in the bean quantity and roasting time. Ambient temperature was clearly a factor, but not one over which I was able to exercise much control. Especially since I frequently ended up roasting outside, so as to not incommodate people with fumes, noise, and chaff. The few homeroast batches which didn’t work probably failed because of low ambient temperature.

One reason I stuck with poppers for so long was that I had heard that dedicated roasters weren’t that durable. I’ve probably used three or four different hot air popcorn poppers, over the years. Eventually, they just stop working, when you use them for coffee beans. As I’d buy them at garage sales and Salvation Army stores for 3-4$, replacing them didn’t feel like such a financially difficult thing to do, though finding them could occasionally be a challenge. Money was also an issue. Though homeroasting was important for me, I wasn’t ready to pay around 200$ for an entry-level dedicated roaster. I was thinking about saving money for a Behmor 1600, which offers several advantages over other roasters. But I finally gave in and bought my i-Roast as a kind of holiday gift to myself.

One broad reason is that my financial situation has improved since I started a kind of partial professional reorientation (PPR). I have a blogpost in mind about this PPR, and I’ll probably write it soon. But this post isn’t about my PPR.

Although, the series of events which led to my purchase does relate to my PPR, somehow.

See, the beans I (indirectly) got from Roastmasters came from a friend who bought a Behmor to roast cocoa beans. The green coffee beans came with the roaster but my friend didn’t want to roast coffee in his brand new Behmor, to avoid the risk of coffee oils and flavours getting into his chocolate. My friend asked me to roast some of these beans for his housemates (he’s not that intensely into coffee, himself). When I went to drop some homeroasted coffee by the Station C co-working space where he spends some of his time, my friend was discussing a project with Duncan Moore, whom I had met a few times but with whom I had had few interactions. The three of us had what we considered a very fruitful yet very short conversation. Later on, I got to do a small but fun project with Duncan. And I decided to invest that money into coffee.

A homeroaster seemed like the most appropriate investment. The Behmor was still out of reach but the i-Roast seemed like a reasonable purchase. Especially if I could buy it used.

But I was also thinking about buying it new, as long as I could get it quickly. It took me several years to make a decision about this purchase but, once I made it, I wanted something as close to “instant gratification” as possible. In some ways, the i-Roast was my equivalent to Little Mrs Sommers‘s “pair of silk stockings.”

At the time, Mon café vert seemed like the only place where I could buy a new i-Roast. I tried several times to reach them to no avail. As I was in the Mile-End as I decided to make that purchase, I went to Caffè in Gamba, both to use the WiFi signal and to check if, by any chance, they might not have started selling roasters. They didn’t, of course, homeroasters isn’t mainstream enough. But, as I was there, I saw the Hario Ceramic Coffee Mill Skerton, a “hand-cranked” coffee grinder about which I had read some rather positive reviews.

For the past few years, I had been using a Bodum Antigua conical burr electric coffee grinder. This grinder was doing the job, but maybe because of “wear and tear,” it started taking a lot longer to grind a small amount of coffee. The grind took so long, at some points, that the grounds were warm to the touch and it seemed like the grinder’s motor was itself heating.

So I started dreaming about the Baratza Vario, a kind of prosumer electric grinder which seemed like the ideal machine for someone who uses diverse coffee making methods. The Vario is rather expensive and seemed like overkill, for my current coffee setup. But I was lusting over it and, yes, dreaming about it.

One day, maybe, I’ll be able to afford a Vario.

In the meantime, and more reasonably, I had been thinking about “Turkish-style mills.” A friend lent me a box-type manual mill at some point and I did find it produced a nice grind, but it wasn’t that convenient for me, partly because the coffee drops into a small drawer which rapidly gets full. A handmill seemed somehow more convenient and there are some generic models which are sold in different parts of the World, especially in the Arab World. So I got the impression that I might be able to find handmills locally and started looking for them all over the place, enquiring at diverse stores and asking friends who have used those mills in the past. Of course, they can be purchased online. But they end up being relatively expensive and my manual experience wasn’t so positive as to convince me to spend so much money on one.

The Skerton was another story. It was much more convenient than a box-type manual mill. And, at Gamba, it was inexpensive enough for me to purchase it on the spot. I don’t tend to do this very often so I did feel strange about such an impulse purchase. But I certainly don’t regret it.

Especially since it complements my other purchases.

So, going to the i-Roast.

Over the years, I had been looking for the i-Roast and Behmor at most of the obvious sites where one might buy used devices like these. eBay, Craig’s List, Kijiji… As a matter of fact, I had seen an i-Roast on one of these, but I was still hesitating. Not exactly sure why, but it probably had to do with the fact that these homeroasters aren’t necessarily that durable and I couldn’t see how old this particular i-Roast was.

I eventually called to find out, after taking my decision to get an i-Roast. Turns out that it’s still under warranty, is in great condition, and was being sold by a very interesting (and clearly trustworthy) alto singer who happens to sing with a friend of mine who is also a local beer homebrewer. The same day I bought the roaster, I went to the cocoa-roasting friend’s place and saw a Behmor for the first time. And I tasted some really nice homemade chocolate. And met other interesting people including a couple that I saw, again, while taking the bus after purchasing the roaster.

The series of coincidences in that whole situation impressed me in a sense of awe. Not out of some strange superstition or other folk belief. But different things are all neatly packaged in a way that most of my life isn’t. Nothing weird about this. The packaging is easy to explain and mostly comes from my own perception. The effect is still there that it all fits.

And the i-Roast 2 itself fits, too.

It’s clearly not the ultimate coffee geek’s ideal roaster. But I get the impression it could become so. In fact, one reason I hesitated to buy the i-Roast 2 is that I was wondering if Hearthware might be coming out with the i-Roast 3, in the not-so-distant future.

I’m guessing that Hearthware might be getting ready to release a new roaster. I’m using unreliable information, but it’s still an educated guess. So, apparently…

I could just imagine what the i-Roast 3 might be. As I’m likely to get, I have a number of crazy ideas.

One “killer feature” actually relates both to the differences between the i-Roast and i-Roast 2 as well as to the geek factor behind homeroasting: roast profiles as computer files. Yes, I know, it sounds crazy. And, somehow, it’s quite unlikely that Hearthware would add such a feature on an entry-level machine. But I seriously think it’d make the roaster much closer to a roasting geek’s ultimate machine.

For one thing, programming a roast profile on the i-Roast is notoriously awkward. Sure, you get used to it. But it’s clearly suboptimal. And one major improvement of the i-Roast 2 over the original i-Roast is that the original version didn’t maintain profiles if you unplugged it. The next step, in my mind, would be to have some way to transfer a profile from a computer to the roaster, say via a slot for SD cards or even a USB port.

What this would open isn’t only the convenience of saving profiles, but actually a way to share them with fellow homeroasters. Since a lot in geek culture has to do with sharing information, a neat effect could come out of shareable roast profiles. In fact, when I looked for example roast profiles, I found forum threads, guides, and incredibly elaborate experiments. Eventually, it might be possible to exchange roasting profiles relating to coffee beans from the same shipment and compare roasting. Given the well-known effects of getting a group of people using online tools to share information, this could greatly improve the state of homeroasting and even make it break out of the very small niche in which it currently sits.

Of course, there are many problems with that approach, including things as trivial as voltage differences as well as bigger issues such as noise levels:

[youtube=http://www.youtube.com/watch?v=c88nEXYCbfs]

But I’m still dreaming about such things.

In fact, I go a few steps further. A roaster which could somehow connect to a computer might also be used to track data about temperature and voltage. In my own experiments with the i-Roast 2, I’ve been logging temperatures at 15 second intervals along with information about roast profile, quantity of beans, etc. It may sound extreme but it already helped me achieve a result I wanted to achieve. And it’d be precisely the kind of information I would like to share with other homeroasters, eventually building a community of practice.

Nothing but geekness, of course. Shall the geek inherit the Earth?

Development and Quality: Reply to Agile Diary

[youtube=http://www.youtube.com/watch?v=iry_CKAlI3g]

Former WiZiQ product manager Vikrama Dhiman responded to one of my tweets with a full-blown blogpost, thereby giving support to Matt Mullenweg‘s point that microblogging goes hand-in-hand with “macroblogging.”

My tweet:

enjoys draft æsthetics yet wishes more developers would release stable products. / adopte certains produits trop rapidement.

Vikrama’s post:

Good Enough Software Does Not Mean Bad Software « Agile Diary, Agile Introduction, Agile Implementation.

My reply:

“To an engineer, good enough means perfect. With an artist, there’s no such thing as perfect.” (Alexander Calder)

Thanks a lot for your kind comments. I’m very happy that my tweet (and status update) triggered this.

A bit of context for my tweet (actually, a post from Ping.fm, meant as a status update, thereby giving support in favour of conscious duplication, «n’en déplaise aux partisans de l’action contre la duplication».)

I’ve been thinking about what I call the “draft æsthetics.” In fact, I did a podcast episode about it. My description of that episode was:

Sometimes, there is such a thing as “Good Enough.”

Though I didn’t emphasize the “sometimes” part in that podcast episode, it was an important part of what I wanted to say. In fact, my intention wasn’t to defend draft æsthetics but to note that there seems to be a tendency toward this æsthetic mode. I do situate myself within that mode in many things I do, but it really doesn’t mean that this mode should be the exclusive one used in any context.

That aforequoted tweet was thus a response to my podcast episode on draft æsthetics. “Yes, ‘good enough’ may work, sometimes. But it needs not be applied in all cases.”

As I often get into convoluted discussions with people who seem to think that I condone or defend a position because I take it for myself, the main thing I’d say there is that I’m not only a relativist but I cherish nuance. In other words, my tweet was a way to qualify the core statement I was talking about in my podcast episode (that “good enough” exists, at times). And that statement isn’t necessarily my own. I notice a pattern by which this statement seems to be held as accurate by people. I share that opinion, but it’s not a strongly held belief of mine.

Of course, I digress…

So, the tweet which motivated Vikrama had to do with my approach to “good enough.” In this case, I tend to think about writing but in view of Eric S. Raymond’s approach to “Release Early, Release Often” (RERO). So there is a connection to software development and geek culture. But I think of “good enough” in a broader sense.

Disclaimer: I am not a coder.

The Calder quote remained in my head, after it was mentioned by a colleague who had read it in a local newspaper. One reason it struck me is that I spend some time thinking about artists and engineers, especially in social terms. I spend some time hanging out with engineers but I tend to be more on the “artist” side of what I perceive to be an axis of attitudes found in some social contexts. I do get a fair deal of flack for some of my comments on this characterization and it should be clear that it isn’t meant to imply any evaluation of individuals. But, as a model, the artist and engineer distinction seems to work, for me. In a way, it seems more useful than the distinction between science and art.

An engineer friend with whom I discussed this kind of distinction was quick to point out that, to him, there’s no such thing as “good enough.” He was also quick to point out that engineers can be creative and so on. But the point isn’t to exclude engineers from artistic endeavours. It’s to describe differences in modes of thought, ways of knowing, approaches to reality. And the way these are perceived socially. We could do a simple exercise with terms like “troubleshooting” and “emotional” to be assigned to the two broad categories of “engineer” and “artist.” Chances are that clear patterns would emerge. Of course, many concepts are as important to both sides (“intelligence,” “innovation”…) and they may also be telling. But dichotomies have heuristic value.

Now, to go back to software development, the focus in Vikrama’s Agile Diary post…

What pushed me to post my status update and tweet is in fact related to software development. Contrary to what Vikrama presumes, it wasn’t about a Web application. And it wasn’t even about a single thing. But it did have to do with firmware development and with software documentation.

The first case is that of my Fonera 2.0n router. Bought it in early November and I wasn’t able to connect to its private signal using my iPod touch. I could connect to the router using the public signal, but that required frequent authentication, as annoying as with ISF. Since my iPod touch is my main WiFi device, this issue made my Fonera 2.0n experience rather frustrating.

Of course, I’ve been contacting Fon‘s tech support. As is often the case, that experience was itself quite frustrating. I was told to reset my touch’s network settings which forced me to reauthenticate my touch on a number of networks I access regularly and only solved the problem temporarily. The same tech support person (or, at least, somebody using the same name) had me repeat the same description several times in the same email message. Perhaps unsurprisingly, I was also told to use third-party software which had nothing to do with my issue. All in all, your typical tech support experience.

But my tweet wasn’t really about tech support. It was about the product. Thougb I find the overall concept behind the Fonera 2.0n router very interesting, its implementation seems to me to be lacking. In fact, it reminds me of several FLOSS development projects that I’ve been observing and, to an extent, benefitting from.

This is rapidly transforming into a rant I’ve had in my “to blog” list for a while about “thinking outside the geek box.” I’ll try to resist the temptation, for now. But I can mention a blog thread which has been on my mind, in terms of this issue.

Firefox 3 is Still a Memory Hog — The NeoSmart Files.

The blogpost refers to a situation in which, according to at least some users (including the blogpost’s author), Firefox uses up more memory than it should and becomes difficult to use. The thread has several comments providing support to statements about the relatively poor performance of Firefox on people’s systems, but it also has “contributions” from an obvious troll, who keeps assigning the problem on the users’ side.

The thing about this is that it’s representative of a tricky issue in the geek world, whereby developers and users are perceived as belonging to two sides of a type of “class struggle.” Within the geek niche, users are often dismissed as “lusers.” Tech support humour includes condescending jokes about “code 6”: “the problem is 6″ from the screen.” The aforementioned Eric S. Raymond wrote a rather popular guide to asking questions in geek circles which seems surprisingly unaware of social and cultural issues, especially from someone with an anthropological background. Following that guide, one should switch their mind to that of a very effective problem-solver (i.e., the engineer frame) to ask questions “the smart way.” Not only is the onus on users, but any failure to comply with these rules may be met with this air of intellectual superiority encoded in that guide. IOW, “Troubleshoot now, ask questions later.”

Of course, many users are “guilty” of all sorts of “crimes” having to do with not reading the documentation which comes with the product or with simply not thinking about the issue with sufficient depth before contacting tech support. And as the majority of the population is on the “user” side, the situation can be described as both a form of marginalization (geek culture comes from “nerd” labels) and a matter of elitism (geek culture as self-absorbed).

This does have something to do with my Fonera 2.0n. With it, I was caught in this dynamic whereby I had to switch to the “engineer frame” in order to solve my problem. I eventually did solve my Fonera authentication problem, using a workaround mentioned in a forum post about another issue (free registration required). Turns out, the “release candidate” version of my Fonera’s firmware does solve the issue. Of course, this new firmware may cause other forms of instability and installing it required a bit of digging. But it eventually worked.

The point is that, as released, the Fonera 2.0n router is a geek toy. It’s unpolished in many ways. It’s full of promise in terms of what it may make possible, but it failed to deliver in terms of what a router should do (route a signal). In this case, I don’t consider it to be a finished product. It’s not necessarily “unstable” in the strict sense that a software engineer might use the term. In fact, I hesitated between different terms to use instead of “stable,” in that tweet, and I’m not that happy with my final choice. The Fonera 2.0n isn’t unstable. But it’s akin to an alpha version released as a finished product. That’s something we see a lot of, these days.

The main other case which prompted me to send that tweet is “CivRev for iPhone,” a game that I’ve been playing on my iPod touch.

I’ve played with different games in the Civ franchise and I even used the FLOSS version on occasion. Not only is “Civilization” a geek classic, but it does connect with some anthropological issues (usually in a problematic view: Civ’s worldview lacks anthro’s insight). And it’s the kind of game that I can easily play while listening to podcasts (I subscribe to a number of th0se).

What’s wrong with that game? Actually, not much. I can’t even say that it’s unstable, unlike some other items in the App Store. But there’s a few things which aren’t optimal in terms of documentation. Not that it’s difficult to figure out how the game works. But the game is complex enough that some documentation is quite useful. Especially since it does change between one version of the game and another. Unfortunately, the online manual isn’t particularly helpful. Oh, sure, it probably contains all the information required. But it’s not available offline, isn’t optimized for the device it’s supposed to be used with, doesn’t contain proper links between sections, isn’t directly searchable, and isn’t particularly well-written. Not to mention that it seems to only be available in English even though the game itself is available in multiple languages (I play it in French).

Nothing tragic, of course. But coupled with my Fonera experience, it contributed to both a slight sense of frustration and this whole reflection about unfinished products.

Sure, it’s not much. But it’s “good enough” to get me started.

Actively Reading: OLPC Critique

Critical thinking has been on my mind, recently. For one thing, I oriented an  “intro. to sociology” course I teach toward critical skills and methods. To me, it’s a very important part of university education, going much beyond media literacy.
And media literacy is something about which I care a great deal. Seems to me that several journalists have been giving up on trying to help the general population increase and enhance their own media literacy skills. It’s almost as if they were claiming they’re the only ones who can reach a significant level of media literacy. Of course, many of them seem unable to have a critical approach to their own work. I’m with Bourdieu on this one. And I make my problem with journalism known.
As a simple example, I couldn’t help but notice a number of problems with this CBC coverage of a new citizenship guidebook. My approach to this coverage is partly visible in short discussions I’ve had on Aardvark about bylines.
A bit over a week ago, I heard about something interesting related to “making technology work,” on WTP (a technology podcast for PRI/BBC/Discovery The World, a bit like Search Engine from bigger media outlets). It was a special forum discussion related to issues broader than simply finding the right tool for the right task. In fact, it sounded like it could become a broad discussion of issues and challenges going way beyond the troubleshooting/problem-solving approach favoured by some technology enthusiasts. Given my ethnographic background, my interest in geek culture, and my passion for social media, I thought I’d give it a try.
The first thing I noticed was a link to a critique of the OLPC project. I’ve personally been quite critical of that project, writing several blogposts about it. So I had to take a look.
And although I find the critical stance of this piece relatively useful (there was way too much groupthink with the original coverage of the OLPC), I couldn’t help but use my critical sense as I was reading this piece.
Which motivated me to do some Diigo annotations on it. For some reason, there are things that I wanted to highlight which aren’t working and I think I may have lost some annotations in the process. But the following is the result of a relatively simple reading of this piece. True to the draft aesthetics, I made no attempt to be thorough, clean, precise, or clear.
http://www.miller-mccune.com/business_economics/computer-error-1390.print
http://www.miller-mccune.com/business_economics/computer-error-1390.print
  • appealing
  • World Economic Forum
  • 50 percent of staff were being laid off and a major restructuring was under way
    • The dramatic version which sends the message: OLPC Inc. was in big trouble. (The fact that it’s allegedly a non-profit is relatively irrelevant.)
  • the project seems nearly dead in the water
    • A strong statement. Stronger than all those “beleaguered company” ones made about Apple in the mid90s before Jobs went back.
  • And that may be great news for children in the developing world.
    • Tadaa! Here’s the twist! The OLPC is dead, long live the Child!
  • lobbied national governments and international agencies
    • Right. The target was institutional. Kind of strange for a project which was billed as a way to get tools in the hands of individual children. And possibly one of the biggest downfalls of the project.
  • Negroponte and other techno-luminati
    • Oh, snap!
      It could sound relatively harmless an appellation. But the context and the piece’s tone make it sound like a rather deep insult.
  • Innovate
    • Ah, nice! Not “create” or “build.” But “innovate.” Which is something the project has been remarkably good at. It was able to achieve a number of engineering feats. Despite Negroponte’s repeated claims to the contrary, the OLPC project can be conceived as an engineering project. In fact, it’s probably the most efficient way to shed the most positive light on it. As an engineering project, it was rather successful. As an “education project” (as Negroponte kept calling it), it wasn’t that successful. In fact, it may have delayed a number of things which matter in terms of education.
  • take control of their education
    • Self-empowerment, at the individual level. In many ways, it sounds like a very Protestant ideal. And it’s clearly part of the neoliberal agenda (or the neoconservative one, actually). Yet it doesn’t sound strange at all. It sounds naturally good and pure.
  • technology optimists
    • Could be neutral in denotation but does connote a form of idealistic technological determinism.
  • Child
  • school attendance
    • “Children who aren’t in school can’t be learning anything, right?”
  • trending dramatically upward
    • Fascinating choice of words.
  • tens of millions of dollars
  • highly respected center
    • Formulas such as these are often a way to prevent any form of source criticism. Not sure Wikipedians would consider these “peacock terms,” but they don’t clearly represent a “neutral point of view.”
  • they don’t seem to be learning much
    • Nothing which can be measured with our tools, at least. Of course, nothing else matters. But still…
  • international science exam
    • Of course, these tend to be ideally suited for most learning contexts…
  • There’s no question that improving education in the developing world is necessary.
    • Although, there could be a question or two about this. Not politically expedient, perhaps. But still…
  • powerful argument
    • Tools in a rhetorical process.
  • instinctive appeal
    • Even the denotative sense is polarized.
  • precious little evidence
    • Switching to the “studies have shown” mode. In this mode, lack of proof is proof of lack, critical thinking is somewhat discouraged, and figures are significant by themselves.
  • circumstantial evidence
    • The jury isn’t out, on this one.
  • co-founder of J-PA
    • Did Esther co-write the article? Honest question.
  • the technology didn’t work any better than a normal classroom teacher
    • A very specific point. If the goal of tool use is to improve performance over “regular teaching,” it’s a particular view of technology. One which, itself, is going by the wayside. And which has been a large part of the OLPC worldview.
  • the goal is improving education for children in the developing world, there are plenty of better, and cheaper, alternatives.
    • A core belief, orienting the piece. Cost is central. The logic is one of “bang for the buck.”
  • the teachers simply weren’t using the computers
    • We’re touching on something, here. People have to actually use the computers for the “concept” to work. Funny that there’s rarely a lot of discussion on how that works. A specific version of “throwing money at a problem” is to “throw technology at” people.
  • few experimental studies to show a positive impact from the use of computers
    • Is the number of studies going one way or another the main issue, here? Can’t diverse studies look at different things and be understood as a way to describe a more complex reality than “technology is good and/or bad?”
  • substituting computers for teachers
    • Still oriented toward the “time to task” approach. But that’s good enough for cognitive science, which tends to be favourably viewed in educational fields.
  • supplement
    • Kept thinking about the well-known Hawthorne effect. In this case, the very idea that providing students with supplementary “care” can be seen as an obvious approach which is most often discussed in the field instead of at the higher levels of decision-making.
  • The OLPC concept has been pioneered in a number of school districts in the United States over the last decade
    • From a 2005 project targeting “countries with inconsistent power grids,” we get to a relatively long series of initiatives in individual school districts in the USofA since last century. Telescoping geographical and temporal scales. And, more importantly, assigning the exact same “concept” to diverse projects.
  • Negroponte has explicitly derided
    • Not the only thing Negroponte derides. He’s been a professional derider for a while, now.
      Negroponte’s personality is part of the subtext of any OLPC-related piece. It’d be interesting to analyse him in view of the “mercurial CEO” type which fascinates a number of people.
  • It must be said
    • Acknowledging the fact that there is more to the situation than what this piece is pushing.
  • academic
    • In this context, “academic” can have a variety of connotations, many of which are relatively negative.
  • teachers limited access to the computers
    • Typically, teachers have relatively little control in terms of students’ access to computers so it sounds likely that the phrase should have read “had limited access.” But, then again, maybe teachers in Hollow’s research were in fact limiting access to computers, which would be a very interesting point to bring and discuss. In fact, part of what is missing in many of those pieces about technology and learning is what access really implies. Typically, most discussions on the subject have to do with time spent alone with such a tool, hence the “one…per child” part of the OLPC approach. But it’s hard to tell if there has been any thought about the benefits of group access to tools or limited access to such tools.
      To go even further, there’s a broad critique of the OLPC approach, left unaddressed in this piece, about the emphasis on individual ownership of tools. In the US, it’s usually not ok for neighbours to ask about using others’ lawnmowers and ladders. It’s unsurprising that pushing individual ownership would seem logical to those who design projects from the US.
  • had not been adequately trained
    • In the OLPC context, it has been made as a case for the dark side of constructionism. The OLPC project might have been a learning project, but it wasn’t a teaching one. Some explicit comments from project members were doing little to dispel the notion that constructivism isn’t about getting rid of teachers. Even documentation for the OLPC XO contained precious little which could help teachers. Teachers weren’t the target audience. Children and governments were.
  • not silver bullets
    • Acknowledging, in an oblique way, that the situation is more complex.
  • surveys of students
    • With a clear Hawthorne effect.
  • parents rolling their eyes
    • Interesting appeal to parenting experience. Even more than teachers, they’re absent from many of these projects. Not a new pattern. Literacy projects often forget parents and the implications in terms of a generation gap. But what is perhaps more striking is that parents are also invisible in coverage of many of these issues. Contrary to “our” children, children in “those poor countries over there” are “ours to care for,” through development projects, adoptions, future immigration, etc.
  • evaluation of an OLPC project in Haiti
    • Sounds more like a pilot project than like field research. But maybe it’s more insightful.
  • Repeated calls and e-mails to OLPC and Negroponte seeking comment on OLPC did not receive a response
    • Such statements are “standard procedure” for journalists. But what is striking about this one is where it’s placed in the piece. Not only is it near the end of the argumentation but it’s in a series of comments about alternative views on the OLPC project. Whether or not it was done on purpose, the effect that we get is that there are two main voices, pro and con. Those on the con side can only have arguments in the same line of thought (about the project’s cost and “efficacy,” with possible comments about management). Those on the pro side are put in a defensive position.
      In such cases, responsiveness is often key. Though Negroponte has been an effective marketer of his pet project, the fact that he explicitly refuses to respond to criticisms and critiques makes for an even more constrained offense/defense game.
  • ironic
    • Strong words, in such a context. Because it’s not the situation which is ironic. It’s a lack of action in a very specific domain.
  • the Third World
    • Interesting that the antiquated “Third World” expression comes in two contexts: the alleged target of the OLPC project (with little discussion as to what was meant by that relationship) and as the J-PAL field of expertise.
  • a leader in
    • Peacock terms or J-PAL are on the Miller-McCune lovelist?
  • There are
    • This is where the piece switches. We’re not talking about the OLPC, anymore. We reduce OLPC to a single goal, which has allegedly not been met, and propose that there are better ways to achieve this goal. Easy and efficient technique, but there still seems to be something missing.
  • etting children in developing countries into school and helping them learn more while they are there
    • A more specific goal than it might seem, at first blush.
      For a very simple example: how about homeschooling?
  • proven successful
    • “We have proof!”
  • cheap
    • One might have expected “inexpensive,” here, instead of “cheap.” But, still, the emphasis is on cost.
  • deworming
    • Sounds a little bit surprising a switch from computer tech to public health.
  • 50 cents per child per year
  • $4 per student per year
  • 30 percent increase in lifetime earnings
  • technology-based approaches to improving student learning in the developing world
    • Coming back to technology, to an extent, but almost in passing. Technology, here, can still be a saviour. The issue would be to find the key technology to solve that one problem (student learning in the developing world needs calls for improvement). Rather limited in scope, depth, insight.
  • show more promise than one laptop per child
    • Perhaps the comment most directly related to opinions. “Showing promise” is closer to “instinctive appeal” but, in this case, it’s a positive. We don’t need to apply critical thinking to something which shows promise. It’s undeniably good. Right?
  • the J-PAL co-founder
    • There we are!
  • $2.20
  • Remedial education
  • A study in Kenya
    • Reference needed.
  • it didn’t matter
    • Sounds like a bold statement, as it’s not expressly linked to the scope of the study. It probably did matter. Just not in terms of what was measured. Mattering has to do with significance in general, not just with statistical significance.
  • expensive
    • Cost/benefits are apparently the only two “factors” to consider.
  • quarter of the cost
  • cheaper
  • $2 per month
  • $3 per month