Category Archives: mass media

Twenty Years Online

This month marks the 20th anniversary of my first Internet account. I don’t remember the exact date but I know it was in late summer 1993, right before what became known as “Eternal September”. The Internet wasn’t new, but it still wasn’t on most people’s proverbial “radars”.

Had heard one of my professors, Kevin Tuite, talk about the Internet as a system through which people from all over the World were communicating. Among the examples Tuite gave of possibilities offered by the ‘Net were conversations among people from former Soviet Republics, during this period of broad transitions. As a specialist of Svaneti, in present-day Georgia, Kevin was particularly interested in these conversations.

During that fated Summer of ‘93, I was getting ready to begin the last year of my B.Sc. in anthropology, specializing in linguistic anthropology and ethnomusicology. As I had done during previous summers, I was working BOH at a French restaurant. But, in my free time, I was exploring a brand new world.

In retrospect, it might not be a complete coincidence that my then-girlfriend of four years left me during that Fall 1993 semester.

It started with a local BBS, WAJU (“We Are Joining You”). I’m not exactly sure when I got started, but I remember being on WAJU in July. Had first been lent a 300 baud modem but I quickly switched to a 2400 baud one. My current ISP plan is 15Mbps, literally 50,000 times faster than my original connection.

By August 1993, thanks to the aforementioned Kevin Tuite, I was able to get an account on UdeM’s ERE network, meant for teaching and research (it stood for «Environnement de recherche et d’enseignement»). That network was running on SGI machines which weren’t really meant to handle large numbers of external connections. But it worked for my purpose of processing email (through Pine), Usenet newsgroups, FTP downloads (sometimes through Archie), IRC sessions, individual chats (though Talk), Gopher sites, and other things via Telnet. As much as possible, I did all of these things from campus, through one of the computer rooms, which offered amazingly fast connections (especially compared to my 2.4kbps modem). I spent enough time in those computer rooms that I still remember a distinct smell from them.

However, at some point during that period, I was able to hack a PPP connection going through my ERE account. In fact, I ended up helping some other people (including a few professors) do the same. It then meant we could use native applications to access the ’Net from home and, eventually, browse the Web graphically.

But I’m getting ahead of myself.

By the time I got online, NCSA Mosaic hadn’t been released. In fact, it took a little while before I even heard of the “World Wide Web”. I seem to remember that I only started browsing the Web in 1994. At the same time, I’m pretty sure one of my most online-savvy friends (likely Alex Burton or Martin Dupras) had told me about the Web as soon as version 1.0 of Mosaic was out, or even before.

The Web was a huge improvement, to be sure. But it was neither the beginning nor the end of the ‘Net, for those of us who had been there a little while. Yes, even a few months. Keep in mind that, at the time, there weren’t that many sites, on the Web. Sure, most universities had a Web presence and many people with accounts on university networks had opportunities to create homepages. But there’s a reason there could be Web directories (strongly associated with Yahoo!, now, but quite common at the time). Pages were “static” and there wasn’t much which was “social” on the Web, at the time.

But the ’Net as a whole was very social. At least, for the budding ethnographer that I was, the rest of the ‘Net was a much more interesting context for observation than the Web. Especially newsgroups and mailinglists.

Especially since the ‘Net was going through one of its first demographic explosions. Some AOLers were flooding the ‘Net. Perhaps more importantly, newbie bashing was peaking and comments against AOL or other inexperienced “Netizens” were frequently heard. I personally heard a lot more from people complaining about AOL than from anyone accessing the ’Net through AOL.

Something about the influx which was clear, though, is that the “democratization” was being accompanied by commercialization. A culture of open sharing was being replaced by corporate culture. Free culture was being preempted by a culture of advertising. The first .com domains were almost a novelty, in a ‘Net full of country-specific domains along with lots of .edu, .net, .org, .gov, and even .mil servers.

The ‘Net wasn’t yet about “paying for content”. That would come a few years later, when media properties pushed “user-generated content” into its own category (instead of representing most of what was available online). The ‘Net of the mid-1990s was about gaining as much attention as possible. We’re still in that mode, of course. But the contrast was striking. Casual conversations were in danger of getting drowned by megaphones. The billboard overtook the café. With the shift, a strong sense of antagonism emerged. The sense of belonging to a community of early adopters increased with the sense of being attacked by old “media types”. People less interested in sharing knowledge and more interested in conveying their own corporate messages. Not that individuals had been agenda-free until that point. But there was a big difference between geeks arguing about strongly-held opinions and “brands” being pushed onto the scene.

Early on, the thing I thought the Internet would most likely disrupt was journalism. I had a problem with journalism so, when I saw how the ‘Net could provide increased access to information, I was sure it’d imply a reappropriation of news by people themselves, with everything this means in the spread of critical thinking skills. Some of this has happened, to an extent. But media consolidation had probably a more critical role to play in journalism’s current crisis than online communication. Although, I like to think of these things as complex systems of interrelated trends and tendencies instead of straightforward causal scenarios.

In such a situation, the ‘Net becoming more like a set of conventional mass media channels was bad news. More specifically, the logic of “getting your corporate message across” was quite offputting to a crowd used to more casual (though often heated and loud) conversations. What comes to mind is a large agora with thousands of people having thousands of separate conversations being taken over by a massive PA system. Regardless of the content of the message being broadcast by this PA system, the effect is beyond annoying.

Through all of this, I distinctly remember mid-April, 1994. At that time, the Internet changed.  One might say it never recovered.

At that time, two unscrupulous lawyers sent the first commercial spam on Usenet newsgroups. They apparently made a rather large sum of money from their action but, more importantly, they ended the “Netiquette” era. From this point on, a conflict has emerged between those who use and those who abuse the ‘Net. Yes, strong words. But I sincerely think they’re fitting. Spammers are like Internet’s cancer. They may “serve a function” and may inspire awe. Mostly, though, they’re “cells gone rogue”. Not that I’m saying the ‘Net was free of disease before this “Green Card lottery” moment. For one thing, it’s possible (though unlikely) that flamewars were somewhat more virulent then than they are now. It’s just that the list of known online woes expanded quickly with the addition of cancer-like diseases. From annoying Usenet spam, we went rather rapidly to all sorts of malevolent large-scale actions. Whatever we end up doing online, we carry the shadow of such actions.

Despite how it may sound, my stance isn’t primarily moral. It’s really about a shift from a “conversational” mode to a “mass media” one. Spammers exploited Usenet by using it as a “mass media” channel, at a time when most people online were using it as a large set of “many-to-many” channels.

The distinction between Usenet spam and legitimate advertising may be extremely important, to a very large number of people. But the gates spammers opened were the same ones advertisers have been using ever since.

My nostalgia of the early Internet has a lot to do with this shift. I know we gained a lot, in the meantime. I enjoy many benefits from the “democratization” of the ‘Net. I wouldn’t trade the current online services and tools for those I was using in August, 1993. But I do long for a cancer-free Internet.

Wearable Hub: Getting the Ball Rolling

Statement

After years of hype, wearable devices are happening. What wearable computing lacks is a way to integrate devices into a broader system.

Disclaimer/Disclosure/Warning

  • For the past two months or so, I’ve been taking notes about this “wearable hub” idea (started around CES’s time, as wearable devices like the Pebble and Google Glass were discussed with more intensity). At this point, I have over 3000 words in notes, which probably means that I’d have enough material for a long essay. This post is just a way to release a few ideas and to “think aloud” about what wearables may mean.
  • Some of these notes have to do with the fact that I started using a few wearable devices to monitor my activities, after a health issue pushed me to start doing some exercise.
  • I’m not a technologist nor do I play one on this blog. I’m primarily an ethnographer, with diverse interests in technology and its implications for human beings. I do research on technological appropriation and some of the course I teach relate to the social dimensions of technology. Some of the approaches to technology that I discuss in those courses relate to constructionism and Actor-Network Theory.
  • I consider myself a “geek ethnographer” in the sense that I take part in geek culture (and have come out as a geek) but I’m also an outsider to geekdom.
  • Contrary to the likes of McLuhan, Carr, and Morozov, my perspective on technology and society is non-deterministic. The way I use them, “implication” and “affordance” aren’t about causal effects or, even, about direct connections. I’m not saying that society is causing technology to appear nor am I proposing a line from tools to social impacts. Technology and society are in a complex system.
  • Further, my approach isn’t predictive. I’m not saying what will happen based on technological advances nor am I saying what technology will appear. I’m thinking about the meaning of technology in an intersubjective way.
  • My personal attitude on tools and gadgets is rather ambivalent. This becomes clear as I go back and forth between techno-enthusiastic contexts (where I can almost appear like a Luddite) and techno-skeptical contexts (where some might label me as a gadget freak). I integrate a number of tools in my life but I can be quite wary about them.
  • I’m not wedded to the ideas I’m putting forth, here. They’re just broad musings of what might be. More than anything, I hope to generate thoughtful discussion. That’s why I start this post with a broad statement (not my usual style).
  • Of course, I know that other people have had similar ideas and I know that a concept of “wearable hub” already exists. It’s obvious enough that it’s one of these things which can be invented independently.

From Wearables to Hubs

Back in the 1990s, “wearable computing” became something of a futuristic buzzword, often having to do with articles of clothing. There have been many experiments and prototypes converging on an idea that we would, one day, be able to wear something resembling a full computer. Meanwhile, “personal digital assistants” became something of a niche product and embedded systems became an important dimension of car manufacturing.

Fast-forward to 2007, when a significant shift in the use of smartphones occurred. Smartphones existed before that time, but their usages, meanings, and positions in the public discourse changed quite radically around the time of the iPhone’s release. Not that the iPhone itself “caused a smartphone revolution” or that smartphone adoption suddenly reached a “tipping point”. I conceive of this shift as a complex interplay between society and tools. Not only more Kuhn than Popper, but more Latour than Kurzweil.

Smartphones, it may be argued, “happened”.

Without being described as “wearable devices”, smartphones started playing some of the functions people might have assigned to wearable devices. The move was subtle enough that Limor Fried recently described it as a realization she’s been having. Some tech enthusiasts may be designing location-aware purses and heads-up displays in the form of glasses. Smartphones are already doing a lot of the things wearables were supposed to do. Many people “wear” smartphones at most times during their waking lives and these Internet-connected devices are full of sensors. With the proliferation of cases, one might even perceive some of them as fashion accessories, like watches and sunglasses.

Where smartphones become more interesting, in terms of wearable computing, is as de facto wearable hubs.

My Wearable Devices

Which brings me to mention the four sensors I’ve been using more extensively during the past two months:

Yes, these all have to do with fitness (and there’s quite a bit of overlap between them). And, yes, I started using them a few days after the New Year. But it’s not about holiday gifts or New Year’s resolutions. I’ve had some of these devices for a while and decided to use them after consulting with a physician about hypertension. Not only have they helped me quite a bit in solving some health issues, but these devices got me to think.

(I carry several other things with me at most times. Some of my favourites include Tenqa REMXD Bluetooth headphones and the LiveScribe echo smartpen.)

One aspect is that they’re all about the so-called “quantified self”. As a qualitative researcher, I tend to be skeptical of quants. In this case, though, the stats I’m collecting about myself fit with my qualitative approach. Along with quantitative data from these devices, I’ve started collecting qualitative data about my life. The next step is to integrate all those data points automatically.

These sensors are also connected to “gamification”, a tendency I find worrisome, preferring playfulness. Though game mechanics are applied to the use of these sensors, I choose to rely on my intrinsic motivation, not paying much attention to scores and badges.

But the part which pushed me to start taking the most notes was that all these sensors connect with my iOS ()and Android) devices. And this is where the “wearable hub” comes into play. None of these devices is autonomous. They’re all part of my personal “arsenal”, the equipment I have on my me on most occasions. Though there are many similarities between them, they still serve different purposes, which are much more limited than those “wearable computers” might have been expected to serve. Without a central device serving as a type of “hub”, these sensors wouldn’t be very useful. This “hub” needs not be a smartphone, despite the fact that, by default, smartphones are taken to be the key piece in this kind of setup.

In my personal scenario, I do use a smartphone as a hub. But I also use tablets. And I could easily use an existing device of another type (say, an iPod touch), or even a new type of device meant to serve as a wearable hub. Smartphones’ “hub” affordances aren’t exclusive.

From Digital Hub to Wearable Hub

Most of the devices which would likely serve as hubs for wearable sensors can be described as “Post-PC”. They’re clearly “personal” and they’re arguably “computers”. Yet they’re significantly different from the “Personal Computers” which have been so important at the end of last century (desktop and laptop computers not used as servers, regardless of the OS they run).

Wearability is a key point, here. But it’s not just a matter of weight or form factor. A wearable hub needs to be wireless in at least two important ways: independent from a power source and connected to other devices through radio waves. The fact that they’re worn at all times also implies a certain degree of integration with other things carried throughout the day (wallets, purses, backpacks, pockets…). These devices may also be more “personal” than PCs because they may be more apparent and more amenable to customization than PCs.

Smartphones fit the bill as wearable hubs. Their form factors and battery life make them wearable enough. Bluetooth (or ANT+, Nike+, etc.) has been used to pair them wirelessly with sensors. Their connectivity to GPS and cellular networking as well as their audio and visual i/o can have interesting uses (mapping a walk, data updates during a commute, voice feedback…). And though they’re far from ubiquitous, smartphones have become quite common in key markets.

Part of the reason I keep thinking about “hubs” has to do with comments made in 2001 by then Apple CEO Steve Jobs about the “digital lifestyle” age in “PC evolution” (video of Jobs’s presentation; as an anthropologist, I’ll refrain from commenting on the evolutionary analogies):

We believe the PC, or more… importantly, the Mac can become the “digital hub” of our emerging digital lifestyle, with the ability to add tremendous value to … other digital devices.

… like camcorders, portable media players, cellphones, digital cameras, handheld organizers, etc. (Though they weren’t mentioned, other peripherals like printers and webcams also connect to PCs.)

The PC was thus going to serve as a hub, “not only adding value to these devices but interconnecting them, as well”.

At the time, key PC affordances which distinguished them from those other digital devices:

  • Big screen affording more complex user interfaces
  • Large, inexpensive hard disk storage
  • Burning DVDs and CDs
  • Internet connectivity, especially broadband
  • Running complex applications (including media processing software like the iLife suite)

Though Jobs pinpointed iLife applications as the basis for this “digital hub” vision, it sounds like FireWire was meant to be an even more important part of this vision. Of course, USB has supplanted FireWire in most use cases. It’s interesting, then, to notice that Apple only recently started shipping Macs with USB 3. In fact, DVD burning is absent from recent Macs. In 2001, the Mac might have been at the forefront of this “digital lifestyle” age. In 2013, the Mac has moved away from its role as “digital hub”.

In the meantime, the iPhone has become one of the best known examples of what I’m calling “wearable hubs”. It has a small screen and small, expensive storage (by today’s standards). It also can’t burn DVDs. But it does have nearly-ubiquitous Internet connectivity and can run fairly complex applications, some of which are adapted from the iLife suite. And though it does have wired connectivity (through Lightning or the “dock connector”), its main hub affordances have to do with Bluetooth.

It’s interesting to note that the same Steve Jobs, who used the “digital hub” concept to explain that the PC wasn’t dead in 2001, is partly responsible for popularizing the concept of “post-PC devices” six years later. One might perceive hypocrisy in this much delayed apparent flip-flop. On the other hand, Steve Jobs’s 2007 comments (video) were somewhat nuanced, as to the role of post-PC devices. What’s more interesting, though, is to think about the implications of the shift between two views of digital devices, regardless of Apple’s position through that shift.

Some post-PC devices (including the iPhone, until quite recently) do require a connection to a PC. In this sense, a smartphone might maintain its position with regards to the PC as digital hub. Yet, some of those devices are used independently of PCs, including by some people who never owned PCs.

Post-Smartphone Hubs

It’s possible to imagine a wearable hub outside of the smartphone (and tablet) paradigm. While smartphones are a convenient way to interconnect wearables, their hub-related affordances still sound limited: they lack large displays and their storage space is quite expensive. Their battery life may also be something to consider in terms of serving as hubs. Their form factors make some sense, when functioning as phones. Yet they have little to do with their use as hubs.

Part of the realization, for me, came from the fact that I’ve been using a tablet as something of an untethered hub. Since I use Bluetooth headphones, I can listen to podcasts and music while my tablet is in my backpack without being entangled in a cable. Sounds trivial but it’s one of these affordances I find quite significant. Delegating music playing functions to my tablet relates in part to battery life and use of storage. The tablet’s display has no importance in this scenario. In fact, given some communication between devices, my smartphone could serve as a display for my tablet. So could a “smartwatch” or “smartglasses”.

The Body Hub

Which led me to think about other devices which would work as wearable hubs. I originally thought about backpackable and pocketable devices.

But a friend had a more striking idea:

Under Armour’s Recharge Energy Suit may be an extreme version of this, one which would fit nicely among things Cathi Bond likes to discuss with Nora Young on The Sniffer. Nora herself has been discussing wearables on her blog as well as on her radio show. Sure, part of this concept is quite futuristic. But a sensor mesh undershirt is a neat idea for several reasons.

  • It’s easy to think of various sensors it may contain.
  • Given its surface area, it could hold enough battery power to supplement other devices.
  • It can be quite comfortable in cold weather and might even help diffuse heat in warmer climates.
  • Though wearable, it needs not be visible.
  • Thieves would probably have a hard time stealing it.
  • Vibration and haptic feedback on the body can open interesting possibilities.

Not that it’s the perfect digital hub and I’m sure there are multiple objections to a connected undershirt (including issues with radio signals). But I find the idea rather fun to think, partly because it’s so far away from the use of phones, glasses, and watches as smart devices.

Another thing I find neat, and it may partly be a coincidence, is the very notion of a “mesh”.

The Wearable Mesh

Mesh networking is a neat concept, which generates more hype than practical uses. As an alternative to WiFi access points and cellular connectivity, it’s unclear that it may “take the world by storm”. But as a way to connect personal devices, it might have some potential. After all, as Bernard Benhamou recently pointed out on France Culture’s Place de la toile, the Internet of Things may not require always-on full-bandwith connectivity. Typically, wearable sensors use fairly little bandwidth or only use it for limited amounts of time. A wearable mesh could connect wearable devices to one another while also exchanging data through the Internet itself.

Or with local devices. Smart cities, near field communication, and digital appliances occupy interesting positions among widely-discussed tendencies in the tech world. They may all have something to do with wearable devices. For instance, data exchanged between transit systems and their users could go through wearable devices. And while mobile payment systems can work through smartphones and other cellphones, wallet functions can also be fulfilled by other wearable devices.

Alternative Futures

Which might provide an appropriate segue into the ambivalence I feel toward the “wearable hub” concept I’m describing. Though I propose these ideas as if I were enthusiastic about them, they all give me pause. As a big fan of critical thinking, I like to think about “what might be” to generate questions and discussions exposing a diversity of viewpoints about the future.

Mass media discussions about these issues tend to focus on such things as privacy, availability, norms, and usefulness. Google Glass has generated quite a bit of buzz about all four. Other wearables may mainly raise issues for one or two of these broad dimensions. But the broad domain of wearable computing raises a lot more issues.

Technology enthusiasts enjoy discussing issues through the dualism between dystopia and utopia. An obvious issue with this dualism is that humans disagree about the two categories. Simply put, one person’s dystopia can be another person’s utopia, not to mention the nuanced views of people who see complex relationships between values and social change.

In such a context, a sociologist’s reflex may be to ask about the implications of these diverse values and opinions. For instance:

  • How do people construct these values?
  • Who decides which values are more important?
  • How might social groups cope with changes in values?

Discussing these issues and more, in a broad frame, might be quite useful. Some of the trickiest issues are raised after some changes in technology have already happened. From writing to cars, any technological context has unexpected implications. An ecological view of these implications could broaden the discussion.

I tend to like the concept of the “drift-off moment”, during which listeners (or readers) start thinking about the possibilities afforded a new tool (or concept). In the context of a sales pitch, the idea is that these possibilities are positive, a potential buyer is thinking about the ways she might use a newfangled device. But I also like the deeper process of thinking about all sorts of implications, regardless of their value.

So…

What might be the implications of a wearable hub?

The Magazine and Social Media

Megaphone red
Megaphone red by Adamantios (via Wikimedia Commons, (GFDL, CC-BY-SA)

The following is my App Store review of The Magazine, a Newsstand offering by Instapaper developer Marco Arment.

Though I like Marco Arment’s work and there’s nothing specifically wrong about this implementation of the magazine model, I don’t find the magazine model particularly useful, at this point. And, make no mistake. The Magazine is indeed a magazine.

Oh, sure, this format overcomes several of the limitations set by advertising-based models and hierarchical boards. But it maintains something of the magazine logic: a tight bundle of a few articles authored by people connected through the same “editorial intent”. It’s not a conversation with the public. In this first issue, it’s not even a conversation among co-authors.

The “linked list” aspect of the “Fireball Format” (from John Gruber’s Daring Fireball media property) is described in one of the pieces in this first issue. Other distinguishing factors of the “Fireball Format” aren’t discussed in that same piece. They include a “no comment” policy which has become rather common among high-profile blogs. Unlike most blogs of the pioneer era in social media, these blogs don’t allow readers to comment directly.

A justification for this policy is that comments can be posted elsewhere. And since most of these bloggers are active on microblogging platforms like App.net and Twitter, there’s a chance that a comment might be noticed by those authors. What’s missing, though, is the sense of belonging which bloggers created among themselves before MySpace.

In other words, now that there are large social networking services online, the social aspects of blogging have been deemphasized and authorial dimensions have come to prominence. Though Arment dislikes the word, blog authors have become “brands”. It still works when these authors are in conversation with one another, when there’s a likelihood of a “followup” (FU in 5by5 parlance), when authors are responsive.

None of that interaction potential seems to be part of the core model for The Magazine. You can scream at your iOS device all you want, Jason Snell will probably not respond to you in a future edition of The Magazine. You can attempt dialogue on Twitter, but any conversation you may succeed in starting there is unlikely to have any impact on The Magazine. You’re talking with authors, now, not with members of a community.

With The Magazine, the transition from social to authorial is almost complete. Not only are posts set apart from the conversation but the editorial act of bundling posts together brings back all the problems media scholars have been pointing out for the past several decades. The issue at stake isn’t merely the move to online delivery. It’s the structure of authority and the one-to-many broadcast-style transmission. We’ve taken a step back.

So, while The Magazine has certain technical advantages over old school magazines like The Daily and Wired, it represents a step away from social media and towards mass media. Less critical thinking, more pedestals.

A new model could emerge using the infrastructure and business model that Arment built. But it’d require significant work outside of the application. The Feature might contribute something to this new model, especially if the way posts are bundled together became more flexible.

So, all in all, I consider The Magazine to be a step in the wrong direction by someone whose work I respect.

Good thing we still have podcasts.

Activism and Journalism

In yesterday’s “Introduction to Society” class, we discussed a number of things related to activism, journalism, labour issues, and even Apple and Foxconn (along with slacktivism, Kony 2012, mass media, moral entrepreneurs, and Wal-Mart).

This discussion was sparked, in part, from a student’s question:

What good are the finding the sociologists obtain if the sociologists themselves are passive to the issues observed?

Very good question, and I feel that the discussion we’ve had in class scratched the surface of the issue.

My response could have related to my current work, which I have mentioned in class on several occasions. These days, an important part of my work outside of the Ivory Tower has to do with community organizations. More specifically, I do fieldwork for Communautique, whose mission is to:

Support civic participation by promoting information literacy, appropriation of information and communications technologies and contribution to their development.

Though I’m no activist, I see a clear role for activism and my work directly supports a form of activism. The goal here is social change, toward increased participation by diverse citizens. Thankfully, this is no “us/them” campaign. There’s no demonization, here. Many of us may disagree on a course of action, but inclusion, not confrontation, is among this work’s main goals.

I sincerely think that my work, however modest, may have a positive impact. Not that I delude myself into thinking that there’s a “quick fix” to problems associated with social exclusion. But I see a fairly clear bifurcation between paths and I choose one which might lead to increased inclusiveness.

I didn’t talk about my work during out classroom discussion. Though I love to talk about it, I try to make these discussions as interactive as possible. Even when I end up talking more than anybody else, I do what I can not to lead the discussion in too specific a direction. So, instead of talking about Communautique, we talked about Foxconn. I’m pretty sure I brought it up, but it was meant as a way to discuss a situation with which students can relate.

Turns out, there was an ideal case to discuss many of these themes. Here’s a message about this case that I just sent to the class’s forum:

Some of you might have heard of this but I hadn’t, before going to class. Sounds to me like it brings together several points we’ve discussed yesterday (activism, journalism, message dissemination, labour conditions, Foxconn, Apple…). It also has a lot to do with approaches to truth, which do tend to differ.

 

So… An episode of This American Life about Foxconn factories making Apple products contained a number of inaccurate things, coming from Mike Daisey, a guy who does monologues as stage plays. These things were presented as facts (and had gone through an elaborate “factchecking” process) and Daisey defends them as theatre, meant to make people react.

 

Here’s a piece about it, from someone who was able to pinpoint some inaccuracies: “An acclaimed Apple critic made up the details”.

 

The retraction from the team at This American Life took a whole show, along with an apparently difficult blogpost.

Interesting stuff, if you ask me. Especially since people might argue that the whole event may negatively impact the cause. After all, the problems of factory workers in China may appeal to more than people’s quickest emotional responses. Though I’m a big fan of emotions, I also think there’s an opportunity to discuss these issues thoughtfully and critically. The issue goes further than Apple or even Foxconn. And it has a lot to do with Wallerstein’s “World Systems Theory”.

 

Anyhoo… Just thought some of you may be interested.

Intimacy, Network Effect, Hype

Is “intimacy” a mere correlate of the network effect?

Can we use the network effect to explain what has been happening with Quora?

Is the Quora hype related to network effect?

I really don’t feel a need to justify my dislike of Quora. Oh, sure, I can explain it. At length. Even on Quora itself. And elsewhere. But I tend to sense some defensiveness on the part of Quora fans.

[Speaking of fans, I have blogposts on fanboism laying in my head, waiting to be hatched. Maybe this will be part of it.]

But the important point, to me, isn’t about whether or not I like Quora. It’s about what makes Quora so divisive. There are people who dislike it and there are some who defend it.

Originally, I was only hearing from contacts and friends who just looooved Quora. So I was having a “Ionesco moment”: why is it that seemingly “everyone” who uses it loves Quora when, to me, it represents such a move in the wrong direction? Is there something huge I’m missing? Or has that world gone crazy?

It was a surreal experience.

And while I’m all for surrealism, I get this strange feeling when I’m so unable to understand a situation. It’s partly a motivation for delving into the issue (I’m surely not the only ethnographer to get this). But it’s also unsettling.

And, for Quora at least, this phase seems to be over. I now think I have a good idea as to what makes for such a difference in people’s experiences with Quora.

It has to do with the network effect.

I’m sure some Quora fanbois will disagree, but it’s now such a clear picture in my mind that it gets me into the next phase. Which has little to do with Quora itself.

The “network effect” is the kind of notion which is so commonplace that few people bother explaining it outside of introductory courses (same thing with “group forming” in social psychology and sociology, or preferential marriage patterns in cultural anthropology). What someone might call (perhaps dismissively): “textbook stuff.”

I’m completely convinced that there’s a huge amount of research on the network effect, but I’m also guessing few people looking it up. And I’m accusing people, here. Ever since I first heard of it (in 1993, or so), I’ve rarely looked at explanations of it and I actually don’t care about the textbook version of the concept. And I won’t “look it up.” I’m more interested in diverse usage patterns related to the concept (I’m a linguistic anthropologist).

So, the version I first heard (at a time when the Internet was off most people’s radar) was something like: “in networked technology, you need critical mass for the tools to become truly useful. For instance, the telephone has no use if you’re the only one with one and it has only very limited use if you can only call a single person.” Simple to the point of being simplistic, but a useful reminder.

Over the years, I’ve heard and read diverse versions of that same concept, usually in more sophisticated form, but usually revolving around the same basic idea that there’s a positive effect associated with broader usage of some networked technology.

I’m sure specialists have explored every single implication of this core idea, but I’m not situating myself as a specialist of technological networks. I’m into social networks, which may or may not be associated with technology (however defined). There are social equivalents of the “network effect” and I know some people are passionate about those. But I find that it’s quite limiting to focus so exclusively on quantitative aspects of social networks. What’s so special about networks, in a social science perspective, isn’t scale. Social scientists are used to working with social groups at any scale and we’re quite aware of what might happen at different scales. But networks are fascinating because of different features they may have. We may gain a lot when we think of social networks as acephalous, boundless, fluid, nameless, indexical, and impactful. [I was actually lecturing about some of this in my “Intro to soci” course, yesterday…]

So, from my perspective, “network effect” is an interesting concept when talking about networked technology, in part because it relates to the social part of those networks (innovation happens mainly through technological adoption, not through mere “invention”). But it’s not really the kind of notion I’d visit regularly.

This case is somewhat different. I’m perceiving something rather obvious (and which is probably discussed extensively in research fields which have to do with networked technology) but which strikes me as missing from some discussions of social networking systems online. In a way, it’s so obvious that it’s kind of difficult to explain.

But what’s coming up in my mind has to do with a specific notion of “intimacy.” It’s actually something which has been on my mind for a while and it might still need to “bake” a bit longer before it can be shared properly. But, like other University of the Streets participants, I perceive the importance of sharing “half-baked thoughts.”

And, right now, I’m thinking of an anecdotal context which may get the point across.

Given my attendance policy, there are class meetings during which a rather large proportion of the class is missing. I tend to call this an “intimate setting,” though I’m aware that it may have different connotations to different people. From what I can observe, people in class get the point. The classroom setting is indeed changing significantly and it has to do with being more “intimate.”

Not that we’re necessarily closer to one another physically or intellectually. It needs not be a “bonding experience” for the situation to be interesting. And it doesn’t have much to do with “absolute numbers” (a classroom with 60 people is relatively intimate when the usual attendance is close to 100; a classroom with 30 people feels almost overwhelming when only 10 people were showing up previously). But there’s some interesting phenomenon going on when there are fewer people than usual, in a classroom.

Part of this phenomenon may relate to motivation. In some ways, one might expect that those who are attending at that point are the “most dedicated students” in the class. This might be a fairly reasonable assumption in the context of a snowstorm but it might not work so well in other contexts (say, when the incentive to “come to class” relates to extrinsic motivation). So, what’s interesting about the “intimate setting” isn’t necessarily that it brings together “better people.” It’s that something special goes on.

What’s going on, with the “intimate classroom,” can vary quite a bit. But there’s still “something special” about it. Even when it’s not a bonding experience, it’s still a shared experience. While “communities of practice” are fascinating, this is where I tend to care more about “communities of experience.” And, again, it doesn’t have much to do with scale and it may have relatively little to do with proximity (physical or intellectual). But it does have to do with cognition and communication. What is special with the “intimate classroom” has to do with shared assumptions.

Going back to Quora…

While an online service with any kind of network effect is still relatively new, there’s something related to the “intimate setting” going on. In other words, it seems like the initial phase of the network effect is the “intimacy” phase: the service has a “large enough userbase” to be useful (so, it’s achieved a first type of critical mass) but it’s still not so “large” as to be overwhelming.

During that phase, the service may feel to people like a very welcoming place. Everyone can be on a “first-name basis. ” High-status users mingle with others as if there weren’t any hierarchy. In this sense, it’s a bit like the liminal phase of a rite of passage, during which communitas is achieved.

This phase is a bit like the Golden Age for an online service with a significant “social dimension.” It’s the kind of time which may make people “wax nostalgic about the good ole days,” once it’s over. It’s the time before the BYT comes around.

Sure, there’s a network effect at stake.  You don’t achieve much of a “sense of belonging” by yourself. But, yet again, it’s not really a question of scale. You can feel a strong bond in a dyad and a team of three people can perform quite well. On the other hand, the cases about which I’m thinking are orders of magnitude beyond the so-called “Dunbar number” which seems to obsess so many people (outside of anthro, at least).

Here’s where it might get somewhat controversial (though similar things have been said about Quora): I’d argue that part of this “intimacy effect” has to do with a sense of “exclusivity.” I don’t mean this as the way people talk about “elitism” (though, again, there does seem to be explicit elitism involved in Quora’s case). It’s more about being part of a “select group of people.” About “being there at the time.” It can get very elitist, snobbish, and self-serving very fast. But it’s still about shared experiences and, more specifically, about the perceived boundedness of communities of experience.

We all know about early adopters, of course. And, as part of my interest in geek culture, I keep advocating for more social awareness in any approach to the adoption part of social media tools. But what I mean here isn’t about a “personality type” or about the “attributes of individual actors.” In fact, this is exactly a point at which the study of social networks starts deviating from traditional approaches to sociology. It’s about the special type of social group the “initial userbase” of such a service may represent.

From a broad perspective (as outsiders, say, or using the comparativist’s “etic perspective”), that userbase is likely to be rather homogeneous. Depending on the enrollment procedure for the service, the structure of the group may be a skewed version of an existing network structure. In other words, it’s quite likely that, during that phase, most of the people involved were already connected through other means. In Quora’s case, given the service’s pushy overeagerness on using Twitter and Facebook for recruitment, it sounds quite likely that many of the people who joined Quora could already be tied through either Twitter or Facebook.

Anecdotally, it’s certainly been my experience that the overwhelming majority of people who “follow me on Quora” have been part of my first degree on some social media tool in the recent past. In fact, one of my main reactions as I’ve been getting those notifications of Quora followers was: “here are people with whom I’ve been connected but with whom I haven’t had significant relationships.” In some cases, I was actually surprised that these people would “follow” me while it appeared like they actually weren’t interested in having any kind of meaningful interactions. To put it bluntly, it sometimes appeared as if people who had been “snubbing” me were suddenly interested in something about me. But that was just in the case of a few people I had unsuccessfully tried to engage in meaningful interactions and had given up thinking that we might not be that compatible as interlocutors. Overall, I was mostly surprised at seeing the quick uptake in my follower list, which doesn’t tend to correlate with meaningful interaction, in my experience.

Now that I understand more about the unthinking way new Quora users are adding people to their networks, my surprise has transformed into an additional annoyance with the service. In a way, it’s a repeat of the time (what was it? 2007?) when Facebook applications got their big push and we kept receiving those “app invites” because some “social media mar-ke-tors” had thought it wise to force people to “invite five friends to use the service.” To Facebook’s credit (more on this later, I hope), these pushy and thoughtless “invitations” are a thing of the past…on those services where people learnt a few lessons about social networks.

Perhaps interestingly, I’ve had a very similar experience with Scribd, at about the same time. I was receiving what seemed like a steady flow of notifications about people from my first degree online network connecting with me on Scribd, whether or not they had ever engaged in a meaningful interaction with me. As with Quora, my initial surprise quickly morphed into annoyance. I wasn’t using any service much and these meaningless connections made it much less likely that I would ever use these services to get in touch with new and interesting people. If most of the people who are connecting with me on Quora and Scribd are already in my first degree and if they tend to be people I have limited interactions, why would I use these services to expand the range of people with whom I want to have meaningful interactions? They’re already within range and they haven’t been very communicative (for whatever reason, I don’t actually assume they were consciously snubbing me). Investing in Quora for “networking purposes” seemed like a futile effort, for me.

Perhaps because I have a specific approach to “networking.”

In my networking activities, I don’t focus on either “quantity” or “quality” of the people involved. I seriously, genuinely, honestly find something worthwhile in anyone with whom I can eventually connect, so the “quality of the individuals” argument doesn’t work with me. And I’m seriously, genuinely, honestly not trying to sell myself on a large market, so the “quantity” issue is one which has almost no effect on me. Besides, I already have what I consider to be an amazing social network online, in terms of quality of interactions. Sure, people with whom I interact are simply amazing. Sure, the size of my first degree network on some services is “well above average.” But these things wouldn’t matter at all if I weren’t able to have meaningful interactions in these contexts. And, as it turns out, I’m lucky enough to be able to have very meaningful interactions in a large range of contexts, both offline and on. Part of it has to do with the fact that I’m teaching addict. Part of it has to do with the fact that I’m a papillon social (social butterfly). It may even have to do with a stage in my life, at which I still care about meeting new people but I don’t really need new people in my circle. Part of it makes me much less selective than most other people (I like to have new acquaintances) and part of it makes me more selective (I don’t need new “friends”). If it didn’t sound condescending, I’d say it has to do with maturity. But it’s not about my own maturity as a human being. It’s about the maturity of my first-degree network.

There are other people who are in an expansionist phase. For whatever reason (marketing and job searches are the best-known ones, but they’re really not the only ones), some people need to get more contacts and/or contacts with people who have some specific characteristics. For instance, there are social activists out there who need to connect to key decision-makers because they have a strong message to carry. And there are people who were isolated from most other people around them because of stigmatization who just need to meet non-judgmental people. These, to me, are fine goals for someone to expand her or his first-degree network.

Some of it may have to do with introversion. While extraversion is a “dominant trait” of mine, I care deeply about people who consider themselves introverts, even when they start using it as a divisive label. In fact, that’s part of the reason I think it’d be neat to hold a ShyCamp. There’s a whole lot of room for human connection without having to rely on devices of outgoingness.

So, there are people who may benefit from expansion of their first-degree network. In this context, the “network effect” matters in a specific way. And if I think about “network maturity” in this case, there’s no evaluation involved, contrary to what it may seem like.

As you may have noticed, I keep insisting on the fact that we’re talking about “first-degree network.” Part of the reason is that I was lecturing about a few key network concepts just yesterday so, getting people to understand the difference between “the network as a whole” (especially on an online service) and “a given person’s first-degree network” is important to me. But another part relates back to what I’m getting to realize about Quora and Scribd: the process of connecting through an online service may have as much to do with collapsing some degrees of separation than with “being part of the same network.” To use Granovetter’s well-known terms, it’s about transforming “weak ties” into “strong” ones.

And I specifically don’t mean it as a “quality of interaction.” What is at stake, on Quora and Scribd, seems to have little to do with creating stronger bonds. But they may want to create closer links, in terms of network topography. In a way, it’s a bit like getting introduced on LinkedIn (and it corresponds to what biz-minded people mean by “networking”): you care about having “access” to that person, but you don’t necessarily care about her or him, personally.

There’s some sense in using such an approach on “utilitarian networks” like professional or Q&A ones (LinkedIn does both). But there are diverse ways to implement this approach and, to me, Quora and Scribd do it in a way which is very precisely counterproductive. The way LinkedIn does it is context-appropriate. So is the way Academia.edu does it. In both of these cases, the “transaction cost” of connecting with someone is commensurate with the degree of interaction which is possible. On Scribd and Quora, they almost force you to connect with “people you already know” and the “degree of interaction” which is imposed on users is disproportionately high (especially in Quora’s case, where a contact of yours can annoy you by asking you personally to answer a specific question). In this sense, joining Quora is a bit closer to being conscripted in a war while registering on Academia.edu is just a tiny bit more like getting into a country club. The analogies are tenuous but they probably get the point across. Especially since I get the strong impression that the “intimacy phase” has a lot to do with the “country club mentality.”

See, the social context in which these services gain much traction (relatively tech-savvy Anglophones in North America and Europe) assign very negative connotations to social exclusion but people keep being fascinating by the affordances of “select clubs” in terms of social capital. In other words, people may be very vocal as to how nasty it would be if some people had exclusive access to some influential people yet there’s what I perceive as an obsession with influence among the same people. As a caricature: “The ‘human rights’ movement leveled the playing field and we should never ever go back to those dark days of Old Boys’ Clubs and Secret Societies. As soon as I become the most influential person on the planet, I’ll make sure that people who think like me get the benefits they deserve.”

This is where the notion of elitism, as applied specifically to Quora but possibly expanding to other services, makes the most sense. “Oh, no, Quora is meant for everyone. It’s Democratic! See? I can connect with very influential people. But, isn’t it sad that these plebeians are coming to Quora without a proper knowledge of  the only right way to ask questions and without proper introduction by people I can trust? I hate these n00bz! Even worse, there are people now on the service who are trying to get social capital by promoting themselves. The nerve on these people, to invade my own dedicated private sphere where I was able to connect with the ‘movers and shakers’ of the industry.” No wonder Quora is so journalistic.

But I’d argue that there’s a part of this which is a confusion between first-degree networks and connection. Before Quora, the same people were indeed connected to these “influential people,” who allegedly make Quora such a unique system. After all, they were already online and I’m quite sure that most of them weren’t more than three or four degrees of separation from Quora’s initial userbase. But access to these people was difficult because connections were indirect. “Mr. Y Z, the CEO of Company X was already in my network, since there were employees of Company X who were connected through Twitter to people who follow me. But I couldn’t just coldcall CEO Z to ask him a question, since CEOs are out of reach, in their caves. Quora changed everything because Y responded to a question by someone ‘totally unconnected to him’ so it’s clear, now, that I have direct access to my good ol’ friend Y’s inner thoughts and doubts.”

As RMS might say, this type of connection is a “seductive mirage.” Because, I would argue, not much has changed in terms of access and whatever did change was already happening all over this social context.

At the risk of sounding dismissive, again, I’d say that part of what people find so alluring in Quora is “simply” an epiphany about the Small World phenomenon. With all sorts of fallacies caught in there. Another caricature: “What? It takes only three contacts for me to send something from rural Idaho to the head honcho at some Silicon Valley firm? This is the first time something like this happens, in the History of the Whole Wide World!”

Actually, I do feel quite bad about these caricatures. Some of those who are so passionate about Quora, among my contacts, have been very aware of many things happening online since the early 1990s. But I have to be honest in how I receive some comments about Quora and much of it sounds like a sudden realization of something which I thought was a given.

The fact that I feel so bad about these characterizations relates to the fact that, contrary to what I had planned to do, I’m not linking to specific comments about Quora. Not that I don’t want people to read about this but I don’t want anyone to feel targeted. I respect everyone and my characterizations aren’t judgmental. They’re impressionistic and, again, caricatures.

Speaking of what I had planned, beginning this post… I actually wanted to talk less about Quora specifically and more about other issues. Sounds like I’m currently getting sidetracked, and it’s kind of sad. But it’s ok. The show must go on.

So, other services…

While I had a similar experiences with Scribd and Quora about getting notifications of new connections from people with whom I haven’t had meaningful interactions, I’ve had a very different experience on many (probably most) other services.

An example I like is Foursquare. “Friendship requests” I get on Foursquare are mostly from: people with whom I’ve had relatively significant interactions in the past, people who were already significant parts of my second-degree network, or people I had never heard of. Sure, there are some people with whom I had tried to establish connections, including some who seem to reluctantly follow me on Quora. But the proportion of these is rather minimal and, for me, the stakes in accepting a friend request on Foursquare are quite low since it’s mostly about sharing data I already share publicly. Instead of being able to solicit my response to a specific question, the main thing my Foursquare “friends” can do that others can’t is give me recommendations, tips, and “notifications of their presence.” These are all things I might actually enjoy, so there’s nothing annoying about it. Sure, like any online service with a network component, these days, there are some “friend requests” which are more about self-promotion. But those are usually easy to avoid and, even if I get fooled by a “social media mar-ke-tor,” the most this person may do to me is give usrecommendation about “some random place.” Again, easy to avoid. So, the “social network” dimension of Foursquare seems appropriate, to me. Not ideal, but pretty decent.

I never really liked the “game” aspect and while I did play around with getting badges and mayorships in my first few weeks, it never felt like the point of Foursquare, to me. As Foursquare eventually became mainstream in Montreal and I was asked by a journalist about my approach to Foursquare, I was exactly in the phase when I was least interested in the game aspect and wished we could talk a whole lot more about the other dimensions of the phenomenon.

And I realize that, as I’m saying this, I may sound to some as exactly those who are bemoaning the shift out of the initial userbase of some cherished service. But there are significant differences. Note that I’m not complaining about the transition in the userbase. In the Foursquare context, “the more the merrier.” I was actually glad that Foursquare was becoming mainstream as it was easier to explain to people, it became more connected with things business owners might do, and generally had more impact. What gave me pause, at the time, is the journalistic hype surrounding Foursquare which seemed to be missing some key points about social networks online. Besides, I was never annoyed by this hype or by Foursquare itself. I simply thought that it was sad that the focus would be on a dimension of the service which was already present on not only Dodgeball and other location-based services but, pretty much, all over the place. I was critical of the seemingly unthinking way people approached Foursquare but the service itself was never that big a deal for me, either way.

And I pretty much have the same attitude toward any tool. I happen to have my favourites, which either tend to fit neatly in my “workflow” or otherwise have some neat feature I enjoy. But I’m very wary of hype and backlash. Especially now. It gets old very fast and it’s been going for quite a while.

Maybe I should just move away from the “tech world.” It’s the context for such hype and buzz machine that it almost makes me angry. [I very rarely get angry.] Why do I care so much? You can say it’s accumulation, over the years. Because I still care about social media and I really do want to know what people are saying about social media tools. I just wish discussion of these tools weren’t soooo “superlative”…

Obviously, I digress. But this is what I like to do on my blog and it has a cathartic effect. I actually do feel better now, thank you.

And I can talk about some other things I wanted to mention. I won’t spend much time on them because this is long enough (both as a blogpost and as a blogging session). But I want to set a few placeholders, for further discussion.

One such placeholder is about some pet theories I have about what worked well with certain services. Which is exactly the kind of thing “social media entrepreneurs” and journalists are so interested in, but end up talking about the same dimensions.

Let’s take Twitter, for instance. Sure, sure, there’s been a lot of talk about what made Twitter a success and probably-everybody knows that it got started as a side-project at Odeo, and blah, blah, blah. Many people also realize that there were other microblogging services around as Twitter got traction. And I’m sure some people use Twitter as a “textbook case” of “network effect” (however they define that effect). I even mention the celebrity dimensions of the “Twitter phenomenon” in class (my students aren’t easily starstruck by Bieber and Gaga) and I understand why journalists are so taken by Twitter’s “broadcast” mission. But something which has been discussed relatively rarely is the level of responsiveness by Twitter developers, over the years, to people’s actual use of the service. Again, we all know that “@-replies,” “hashtags,” and “retweets” were all emerging usage patterns that Twitter eventually integrated. And some discussion has taken place when Twitter changed it’s core prompt to reflect the fact that the way people were using it had changed. But there’s relatively little discussion as to what this process implies in terms of “developing philosophy.” As people are still talking about being “proactive” (ugh!) with users, and crude measurements of popularity keep being sold and bandied about, a large part of the tremendous potential for responsiveness (through social media or otherwise) is left untapped. People prefer to hype a new service which is “likely to have Twitter-like success because it has the features users have said they wanted in the survey we sell.” Instead of talking about the “get satisfaction” effect in responsiveness. Not that “consumers” now have “more power than ever before.” But responsive developers who refrain from imposing their views (Quora, again) tend to have a more positive impact, socially, than those which are merely trying to expand their userbase.

Which leads me to talk about Facebook. I could talk for hours on end about Facebook, but I almost feel afraid to do so. At this point, Facebook is conceived in what I perceive to be such a narrow way that it seems like anything I might say would sound exceedingly strange. Given the fact that it was part one of the first waves of Web tools with explicit social components to reach mainstream adoption, it almost sounds “historical” in timeframe. But, as so many people keep saying, it’s just not that old. IMHO, part of the implication of Facebook’s relatively young age should be that we are able to discuss it as a dynamic process, instead of assigning it to a bygone era. But, whatever…

Actually, I think part of the reason there’s such lack of depth in discussing Facebook is also part of the reason it was so special: it was originally a very select service. Since, for a significant period of time, the service was only available to people with email addresses ending in “.edu,” it’s not really surprising that many of the people who keep discussing it were actually not on the service “in its formative years.” But, I would argue, the fact that it was so exclusive at first (something which is often repeated but which seems to be understood in a very theoretical sense) contributed quite significantly to its success. Of course, similar claims have been made but, I’d say that my own claim is deeper than others.

[Bang! I really don’t tend to make claims so, much of this blogpost sounds to me as if it were coming from somebody else…]

Ok, I don’t mean it so strongly. But there’s something I think neat about the Facebook of 2005, the one I joined. So I’d like to discuss it. Hence the placeholder.

And, in this placeholder, I’d fit: the ideas about responsiveness mentioned with Twitter, the stepwise approach adopted by Facebook (which, to me, was the real key to its eventual success), the notion of intimacy which is the true core of this blogpost, the notion of hype/counterhype linked to journalistic approaches, a key distinction between privacy and intimacy, some non-ranting (but still rambling) discussion as to what Google is missing in its “social” projects, anecdotes about “sequential network effects” on Facebook as the service reached new “populations,” some personal comments about what I get out of Facebook even though I almost never spent any significant amount of time on it, some musings as to the possibility that there are online services which have reached maturity and may remain stable in the foreseeable future, a few digressions about fanboism or about the lack of sophistication in the social network models used in online services, and maybe a bit of fun at the expense of “social media expert marketors”…

But that’ll be for another time.

Cheers!

Minds of All Sizes Think Alike

Or «les esprits de toutes tailles se rencontrent».

This post is a response to the following post about Social Network Analysis (SNA), social change, and communication.

…My heart’s in Accra » Shortcuts in the social graph.

I have too many disparate things to say about that post to make it into a neat and tidy “quickie,” yet I feel like I should probably be working on other things. So we’ll see how this goes.

First, a bit of context..

[This “bit of context” may be a bit long so, please bear with me. Or you could get straight to the point, if you don’t think you can bear the context bit.]

I’ve never met Ethan Zuckerman (@EthanZ), who wrote the post to which I’m responding. And I don’t think we’ve had any extended conversation in the past. Further, I doubt that I’m on his radar. He’s probably seen my name, since I’ve commented on some of his posts and some of his contacts may have had references to me through social media. But I very much doubt that he’s ever mentioned me to anyone. I’m not noticeable to him.

I, on the other hand, have mentioned Zuckerman on several occasions. Latest time I remember was in class, a few weeks ago. It’s a course on Africa and I was giving students a list of online sources with relevance to our work. Zuckerman’s connection to Africa may not be his main thing, despite his blog’s name, but it’s part of the reason I got interested in his work, a few years ago.

In fact, there’s something embarrassing, here.. I so associate Zuckerman to Africa that my mind can’t help but link him to Erik Hersman, aka White African. I did meet Herman. [To be exact, I met Erik at BarCampAustin, which is quite possibly the conference-like event which has had the most influence on me, in the past few years (I go to a lot of these events).] When I did meet Hersman, I made a faux-pas in associating him with Zuckerman. Good-natured as he seemed to be, Hersman smiled as he corrected me.

EthanZ and I have other contacts in common. Jeremy Clarke, for instance, who co-organizes WordCamp Montreal and has been quite active in Montreal’s geek scene. Jeremy’s also a developer for Global Voices, a blogging community that Zuckerman co-founded. I’m assuming Clarke and Zuckerman know each other.

Another mutual contact is Christopher Lydon, host of Radio Open Source. Chris and I have exchanged a few emails, and Zuckerman has been on ROS on a few occasions.

According to Facebook, Zuckerman and I have four contacts in common. Apart from Clarke and Hersman, there’s P. Kerim Friedman and Gerd Leonhard. Kerim is a fellow linguistic anthropologist and we’ve collaborated on the official Society for Linguistic Anthropology (SLA) site. I got in touch with Leonhard through “Music 2.0” issues, as he was interviewed by Charles McEnerney on Well-Rounded Radio.

On LinkedIn, Zuckerman is part of my third degree, with McEnerney as one of my first-degree contacts who could connect me to Zuckerman, through Zuckerman’s contacts.

(Yes, I’m fully aware of the fact that I haven’t name a single woman in this list. Nor someone who doesn’t write in English with some frequency, for that matter.)

By this time, my guess is that you may be either annoyed or confused. “Surely, he can’t be that obsessed with Zuckerman as to stalk him in every network.”

No, I’m not at all obsessed with Ethan Zuckerman in any way, shape, or form. Though I mention him on occasion and I might have a good conversation with him if the occasion arises, I wouldn’t go hang out in Cambridge just in case I might meet him. Though I certainly respect his work, I wouldn’t treat him as my “idol” or anything like that. In other words, he isn’t a focus in my life.

And that’s a key point, to me.

In certain contexts, when social networks are discussed, too much is made of the importance of individuals. Yet, there’s something to be said about relative importance.

In his “shortcuts” post, Zuckerman talks about a special kind of individuals. Those who are able to bypass something of a clustering effect happening in many human networks. Malcolm Gladwell (probably “inspired” by somebody else) has used “connectors” to label a fairly similar category of people and, given Gladwell’s notoriety in some circles, the name has resonance in some contexts (mostly “business-focused people,” I would say, with a clear idea in my mind of the groupthink worldview implied).

In one of my earliest blogposts, I talked about an effect happening through a similar mechanism, calling it the “Social Butterfly Effect” (SBE). I still like it, as a concept. Now, I admit that it focuses on a certain type of individuals. But it’s more about their position in “the grand scheme of things” than about who they are, though I do associate myself with this “type.”

The basic idea is quite simple. People who participate in different (sub)networks, who make such (sub)networks sparser, are having unpredictable and unmeasurable effects on what is transmitted through the network(s).

On one hand, it’s linked to my fragmentary/naïve understanding of the Butterfly Effect in the study of climate and as a component of Chaos Theory.

On the other hand, it’s related to Granovetter‘s well-known notion of “weak ties.” And it seems like Granovetter is making something of a comeback, as we discuss different mechanisms behind social change.

Interestingly, much of what is being said about weak ties, these past few weeks, relates to Gladwell’s flamebait apparent lack of insight in describing current social processes. Sounds like Gladwell may be too caught up in the importance of individuals to truly grok the power of networks.

Case in point.. One of the most useful pieces I’ve read about weak ties, recently, was Jonah Lehrer‘s direct response to Gladwell:

Weak Ties, Twitter and Revolution | Wired Science | Wired.com.

Reading Lehrer’s piece, one gets the clear impression that Gladwell hadn’t “done his homework” on Granovetter before launching his trolling “controversial” piece on activism.

But I digress. Slightly.

Like the Gladwell-specific coverage, Zuckerman’s blogpost is also about social change and he’s already responded to Gladwell. One way to put it is that, as a figure, Gladwell has shaped the discussion in a way similar to a magnetic field orienting iron filings around it. Since it’s a localized effect having to do with polarization, the analogy is fairly useful, as analogies go.

Which brings me to groupthink, the apparent target of Zuckerman’s piece.

Still haven’t read Irving Janis but I’ve been quite interested in groupthink for a while. Awareness of the concept is something I immediately recognize, praise, and associate with critical thinking.

In fact, it’s one of several things I was pleasantly surprised to find in an introductory sociology WikiBook I ended up using in my  “Intro. to Society” course, last year. Critical thinking was the main theme of that course, and this short section was quite fitting in the overall discussion.

So, what of groupthink and networks? Zuckerman sounds worried:

This is interesting to me because I’m intrigued – and worried – by information flows through social networks. If we’re getting more (not lots yet, but more) information through social networks and less through curated media like newspapers, do we run the risk of encountering only information that our friends have access to? Are we likely to be overinformed about some conversations and underinformed about others? And could this isolation lead to ideological polarization, as Cass Sunstein and others suggest? And if those fears are true, is there anything we can do to rewire social networks so that we’re getting richer, more diverse information?

Similar questions have animated many discussions in media-focused circles, especially in those contexts where the relative value (and meaning) of “old vs. new media” may be debated. At about the same time as I started blogging, I remember discussing things with a statistician friend about the polarization effect of media, strong confirmation bias in reading news stories, and political lateralization.

In the United States, especially, there’s a narrative (heard loud and clear) that people who disagree on some basic ideas are unable to hear one another. “Shockingly,” some say, “conservatives and liberals read different things.” Or “those on (the) two sides of (the) debate understand things in completely different ways.” It even reminds me of the connotations of Tannen’s booktitle, You Just Don’t Understand. Irreconciliable differences. (And the first time I mention a woman in this decidedly imbalanced post.)

While, as a French-Canadian ethnographer, my perspective is quite different from Zuckerman, I can’t help but sympathize with the feeling. Not that I associate groupthink with a risk in social media (au contraire!). But, like Zuckerman, I wish to find ways to move beyond these boundaries we impose on ourselves.

Zuckerman specifically discusses the attempt by Onnik Krikorian (@OneWMPhoto) to connect Armenians (at least those in Hayastan) and Azeris, with Facebook “affording” Krikorian some measure of success. This case is now well-known in media-centric circles and it has almost become shorthand for the power of social media. Given a personal interest in Armenians (at least in the Diaspora), my reaction to Krikorian’s success are less related to the media aspect than to the personal one.

At a personal level, boundaries may seem difficult to surmount but they can also be fairly porous and even blurry. Identity may be negotiated. Individuals crossing boundaries may be perceived in diverse ways, some of which have little to do with other people crossing the same boundaries. Things are lived directly, from friendships to wars, from breakups to reconciliations. Significant events happen regardless of the way  they’re being perceived across boundaries.

Not that boundaries don’t matter but they don’t necessarily circumscribe what happens in “personal lives.” To use an seemingly-arbitrary example, code-switching doesn’t “feel” strange at an individual level. It’s only when people insist on separating languages using fairly artificial criteria that alternance between them sounds awkward.

In other words, people cross boundaries all the time and “there’s nothing to it.”

Boundaries have quite a different aspect, at the macrolevel implied by the journalistic worldview (with nation-based checkbox democracy at its core and business-savvy professionalization as its mission). To “macros” like journos and politicos, boundaries look like borders, appearing clearly on maps (including mind ones) and implying important disconnects. The border between Armenia and Azerbaijan is a boundary separating two groups and the conflicts between these two groups reify that boundary. Reaching out across the border is a diplomatic process and necessitates finding the right individuals for the task. Most of the important statuses are ascribed, which may sound horrible to some holding neoliberal ideas about freewill and “individual freedoms.”

Though it’s quite common for networked activities to be somewhat constrained by boundaries, a key feature of networks is that they’re typically boundless. Sure, there are networks which are artificially isolated from the rest. The main example I can find is that of a computer virology laboratory.

Because, technically, you only need one link between two networks to transform them into a single network. So, it’s quite possible to perceive Verizon’s wireless network as a distinct entity, limited by the national boundaries of the U.S. of A. But the simple fact that someone can use Verizon’s network to contact someone in Ségou shows that the network isn’t isolated. Simple, but important to point out.

Especially since we’re talking about a number of things happening on a single network: The Internet. (Yes, there is such a thing as Internet2 and there are some technical distinctions at stake. But we’re still talking about an interconnected world.)

As is well-known, there are significant clusters in this One Network. McLuhan’s once-popular “Global Village” fallacy used to hide this, but we now fully realize that language barriers, national borders, and political lateralization go with “low-bandwidth communication,” in some spots of The Network. “Gs don’t talk to Cs so even though they’re part of the same network, there’s a weak spot, there.” In a Shannon/Weaver view, it sounds quite important to identify these weak spots. “Africa is only connected to North America via a few lines so access is limited, making things difficult for Africans.” Makes sense.

But going back to weak ties, connectors, Zuckerman’s shortcuts, and my own social butterflies, the picture may be a little bit more fleshed out.

Actually, the image I have in mind has, on one side, a wire mesh serving as the floor of an anechoic chamber  and on the other some laser beams going in pseudorandom directions as in Entrapment or Mission Impossible. In the wire mesh, weaker spots might cause a person to fall through and land on those artificial stalagmites. With the laser beams, the pseudorandom structure makes it more difficult to “find a path through the maze.” Though some (engineers) may see the mesh as the ideal structure for any network, there’s something humanly fascinating about the pseudorandom structure of social networks.

Obviously, I have many other ideas in mind. For instance, I wanted to mention “Isabel Wilkerson’s Leaderless March that Remade America.” Or go back to that intro soci Wikibook to talk about some very simple and well-understood ideas about social movements, which often seem to be lacking in discussions of social change. I even wanted to recount some anecdotes of neat network effects in my own life, such as the serendipity coming from discuss disparate subjects to unlike people or the misleading impression that measuring individualized influence is a way to understand social media. Not to mention a whole part I had in my mind about Actor Network Theory, non-human actors, and material culture (the other course I currently teach).

But I feel like going back to more time-sensitive things.

Still, I should probably say a few words about this post’s title.

My mother and I were discussing parallel inventions and polygenesis with the specific theme of moving away from the focus on individualized credit. My favourite example, and one I wish Gladwell (!) had used in Outliers (I actually asked him about it) is that of Gregor Mendel and the “rediscovery” of his laws by de Vries, Correns, and Tschermak. A semi-Marxian version of the synchronous polygenesis part might hold that “ideas are in the air” or that the timing of such dicoveries and inventions has to do with zeitgeist. A neoliberal version could be the “great minds think alike” expression or its French equivalent «Les grands esprits se rencontrent» (“The great spirits meet each other”). Due to my reluctance in sizing up minds, I’d have a hard time using that as a title. In the past, I used a similar title to refer to another form of serendipity:

To me, most normally constituted minds are “great,” so I still could have used the expression as a title. But an advantage of tweaking an expression is that it brings attention to what it implies.

In this case, the “thinking alike” may be a form of groupthink.

 

Academics and Their Publics

Misunderstood by Raffi Asdourian
Misunderstood by Raffi Asdourian

Academics are misunderstood.

Almost by definition.

Pretty much any academic eventually feels that s/he is misunderstood. Misunderstandings about some core notions in about any academic field are involved in some of the most common pet peeves among academics.

In other words, there’s nothing as transdisciplinary as misunderstanding.

It can happen in the close proximity of a given department (“colleagues in my department misunderstand my work”). It can happen through disciplinary boundaries (“people in that field have always misunderstood our field”). And, it can happen generally: “Nobody gets us.”

It’s not paranoia and it’s probably not self-victimization. But there almost seems to be a form of “onedownmanship” at stake with academics from different disciplines claiming that they’re more misunderstood than others. In fact, I personally get the feeling that ethnographers are more among the most misunderstood people around, but even short discussions with friends in other fields (including mathematics) have helped me get the idea that, basically, we’re all misunderstood at the same “level” but there are variations in the ways we’re misunderstood. For instance, anthropologists in general are mistaken for what they aren’t based on partial understanding by the general population.

An example from my own experience, related to my decision to call myself an “informal ethnographer.” When you tell people you’re an anthropologist, they form an image in their minds which is very likely to be inaccurate. But they do typically have an image in their minds. On the other hand, very few people have any idea about what “ethnography” means, so they’re less likely to form an opinion of what you do from prior knowledge. They may puzzle over the term and try to take a guess as to what “ethnographer” might mean but, in my experience, calling myself an “ethnographer” has been a more efficient way to be understood than calling myself an “anthropologist.”

This may all sound like nitpicking but, from the inside, it’s quite impactful. Linguists are frequently asked about the number of languages they speak. Mathematicians are taken to be number freaks. Psychologists are perceived through the filters of “pop psych.” There are many stereotypes associated with engineers. Etc.

These misunderstandings have an impact on anyone’s work. Not only can it be demoralizing and can it impact one’s sense of self-worth, but it can influence funding decisions as well as the use of research results. These misunderstandings can underminine learning across disciplines. In survey courses, basic misunderstandings can make things very difficult for everyone. At a rather basic level, academics fight misunderstandings more than they fight ignorance.

The  main reason I’m discussing this is that I’ve been given several occasions to think about the interface between the Ivory Tower and the rest of the world. It’s been a major theme in my blogposts about intellectuals, especially the ones in French. Two years ago, for instance, I wrote a post in French about popularizers. A bit more recently, I’ve been blogging about specific instances of misunderstandings associated with popularizers, including Malcolm Gladwell’s approach to expertise. Last year, I did a podcast episode about ethnography and the Ivory Tower. And, just within the past few weeks, I’ve been reading a few things which all seem to me to connect with this same issue: common misunderstandings about academic work. The connections are my own, and may not be so obvious to anyone else. But they’re part of my motivations to blog about this important issue.

In no particular order:

But, of course, I think about many other things. Including (again, in no particular order):

One discussion I remember, which seems to fit, included comments about Germaine Dieterlen by a friend who also did research in West Africa. Can’t remember the specifics but the gist of my friend’s comment was that “you get to respect work by the likes of Germaine Dieterlen once you start doing field research in the region.” In my academic background, appreciation of Germaine Dieterlen’s may not be unconditional, but it doesn’t necessarily rely on extensive work in the field. In other words, while some parts of Dieterlen’s work may be controversial and it’s extremely likely that she “got a lot of things wrong,” her work seems to be taken seriously by several French-speaking africanists I’ve met. And not only do I respect everyone but I would likely praise someone who was able to work in the field for so long. She’s not my heroine (I don’t really have heroes) or my role-model, but it wouldn’t have occurred to me that respect for her wasn’t widespread. If it had seemed that Dieterlen’s work had been misunderstood, my reflex would possibly have been to rehabilitate her.

In fact, there’s  a strong academic tradition of rehabilitating deceased scholars. The first example which comes to mind is a series of articles (PDF, in French) and book chapters by UWO linguistic anthropologist Regna Darnell.about “Benjamin Lee Whorf as a key figure in linguistic anthropology.” Of course, saying that these texts by Darnell constitute a rehabilitation of Whorf reveals a type of evaluation of her work. But that evaluation comes from a third person, not from me. The likely reason for this case coming up to my mind is that the so-called “Sapir-Whorf Hypothesis” is among the most misunderstood notions from linguistic anthropology. Moreover, both Whorf and Sapir are frequently misunderstood, which can make matters difficulty for many linguistic anthropologists talking with people outside the discipline.

The opposite process is also common: the “slaughtering” of “sacred cows.” (First heard about sacred cows through an article by ethnomusicologist Marcia Herndon.) In some significant ways, any scholar (alive or not) can be the object of not only critiques and criticisms but a kind of off-handed dismissal. Though this often happens within an academic context, the effects are especially lasting outside of academia. In other words, any scholar’s name is likely to be “sullied,” at one point or another. Typically, there seems to be a correlation between the popularity of a scholar and the likelihood of her/his reputation being significantly tarnished at some point in time. While there may still be people who treat Darwin, Freud, Nietzsche, Socrates, Einstein, or Rousseau as near divinities, there are people who will avoid any discussion about anything they’ve done or said. One way to put it is that they’re all misunderstood. Another way to put it is that their main insights have seeped through “common knowledge” but that their individual reputations have decreased.

Perhaps the most difficult case to discuss is that of Marx (Karl, not Harpo). Textbooks in introductory sociology typically have him as a key figure in the discipline and it seems clear that his insight on social issues was fundamental in social sciences. But, outside of some key academic contexts, his name is associated with a large series of social events about which people tend to have rather negative reactions. Even more so than for Paul de Man or  Martin Heidegger, Marx’s work is entangled in public opinion about his ideas. Haven’t checked for examples but I’m quite sure that Marx’s work is banned in a number of academic contexts. However, even some of Marx’s most ardent opponents are likely to agree with several aspects of Marx’s work and it’s sometimes funny how Marxian some anti-Marxists may be.

But I digress…

Typically, the “slaughtering of sacred cows” relates to disciplinary boundaries instead of social ones. At least, there’s a significant difference between your discipline’s own “sacred cows” and what you perceive another discipline’s “sacred cows” to be. Within a discipline, the process of dismissing a prior scholar’s work is almost œdipean (speaking of Freud). But dismissal of another discipline’s key figures is tantamount to a rejection of that other discipline. It’s one thing for a physicist to show that Newton was an alchemist. It’d be another thing entirely for a social scientist to deconstruct James Watson’s comments about race or for a theologian to argue with Darwin. Though discussions may have to do with individuals, the effects of the latter can widen gaps between scholarly disciplines.

And speaking of disciplinarity, there’s a whole set of issues having to do with discussions “outside of someone’s area of expertise.” On one side, comments made by academics about issues outside of their individual areas of expertise can be very tricky and can occasionally contribute to core misunderstandings. The fear of “talking through one’s hat” is quite significant, in no small part because a scholar’s prestige and esteem may greatly decrease as a result of some blatantly inaccurate statements (although some award-winning scholars seem not to be overly impacted by such issues).

On the other side, scholars who have to impart expert knowledge to people outside of their discipline  often have to “water down” or “boil down” their ideas and, in effect, oversimplifying these issues and concepts. Partly because of status (prestige and esteem), lowering standards is also very tricky. In some ways, this second situation may be more interesting. And it seems unavoidable.

How can you prevent misunderstandings when people may not have the necessary background to understand what you’re saying?

This question may reveal a rather specific attitude: “it’s their fault if they don’t understand.” Such an attitude may even be widespread. Seems to me, it’s not rare to hear someone gloating about other people “getting it wrong,” with the suggestion that “we got it right.”  As part of negotiations surrounding expert status, such an attitude could even be a pretty rational approach. If you’re trying to position yourself as an expert and don’t suffer from an “impostor syndrome,” you can easily get the impression that non-specialists have it all wrong and that only experts like you can get to the truth. Yes, I’m being somewhat sarcastic and caricatural, here. Academics aren’t frequently that dismissive of other people’s difficulties understanding what seem like simple concepts. But, in the gap between academics and the general population a special type of intellectual snobbery can sometimes be found.

Obviously, I have a lot more to say about misunderstood academics. For instance, I wanted to address specific issues related to each of the links above. I also had pet peeves about widespread use of concepts and issues like “communities” and “Eskimo words for snow” about which I sometimes need to vent. And I originally wanted this post to be about “cultural awareness,” which ends up being a core aspect of my work. I even had what I might consider a “neat” bit about public opinion. Not to mention my whole discussion of academic obfuscation (remind me about “we-ness and distinction”).

But this is probably long enough and the timing is right for me to do something else.

I’ll end with an unverified anecdote that I like. This anecdote speaks to snobbery toward academics.

[It’s one of those anecdotes which was mentioned in a course I took a long time ago. Even if it’s completely fallacious, it’s still inspiring, like a tale, cautionary or otherwise.]

As the story goes (at least, what I remember of it), some ethnographers had been doing fieldwork  in an Australian cultural context and were focusing their research on a complex kinship system known in this context. Through collaboration with “key informants,” the ethnographers eventually succeeded in understanding some key aspects of this kinship system.

As should be expected, these kinship-focused ethnographers wrote accounts of this kinship system at the end of their field research and became known as specialists of this system.

After a while, the fieldworkers went back to the field and met with the same people who had described this kinship system during the initial field trip. Through these discussions with their “key informants,” the ethnographers end up hearing about a radically different kinship system from the one about which they had learnt, written, and taught.

The local informants then told the ethnographers: “We would have told you earlier about this but we didn’t think you were able to understand it.”

Scriptocentrism and the Freedom to Think

As a comment on my previous blogpost on books, a friend sent me (through Facebook) a link to a blogpost about a petition to Amazon with the following statement:

The freedom to read is tantamount to the freedom to think.

As this friend and I are both anthros+africanists, I’m reacting (perhaps a bit strongly) to that statement.

Given my perspective, I would dare say that I find this statement (brought about by DbD)… ethnocentric.

There, I said it.

And I’ll try to back it up in this blogpost in order to spark even more discussion.

We won’t exhaust this topic any time soon, but I feel there’s a lot we can do about it which has rarely been done.

I won’t use the textbook case of “Language in the Inner City,” but it could help us talk about who decides, in a given social context, what is important. We both come from a literacy-focused background, so we may have to take a step back. Not sure if Bourdieu has commented on Labov, especially in terms of what all this means for “education,” but I’d even want to bring in Ivan Illich, at some point.

Hunters with whom I’ve been working, in Mali, vary greatly in terms of literacy. Some of them have a strong university background and one can even write French legalese (he’s a judge). Others (or some of the same) have gone to Koranic school long enough that can read classical Arabic. Some have the minimal knowledge of Arabic which suffices, for them, to do divination. Many of them have a very low level of functional literacy. There’s always someone around them who can read and write, so they’re usually not out of the loop and it’s not like the social hierarchy stereotypical of the Catholic Church during the Middle Ages in Europe. It’s a very different social context which can hardly be superimposed with the history of writing and the printing press in Europe.

In terms of “freedom to thinik,” I really wouldn’t say that they’re lacking. Of course, “free thinker” has a specific meaning in liberal societies with a European background. But even this meaning can be applied to many people I’ve met in Mali.

And I go back to the social context. Those with the highest degree of functional literacy aren’t necessarily those with the highest social status. And unlike Harlem described by Labov, it’s a relatively independent context from the one in which literacy is a sine qua non. Sure, it’s a neocolonial context and Euro-Americans keep insisting that literacy in Latin script is “the most important thing ever” if they are to become a true liberal democracy. Yet, internally, it’s perfectly possible for someone to think freely, get recognition, and help other people to think without going through the written medium.

Many of those I know who have almost nonexistent skills in the written medium also have enough power (in a Weberian sense) that they get others to do the reading and writing for them. And because there are many social means to ensure that communication has worked appropriately, these “scribes” aren’t very likely to use this to take anything away from those for whom they read and write.

In Switzerland, one of my recent ancestors was functionally illiterate. Because of this, she “signed away” most of her wealth. Down the line, I’m one of her very few heirs. So, in a way, I lost part of my inheritance due to illiteracy.

Unless the switch to a European model for notarial services becomes complete, a case like this is unlikely to occur among people I know in Mali. If it does happen, it’s clearly not a failure of the oral system but a problem with this kind of transition. It’s somewhat similar to the situation with women in diverse parts of the continent during the period of direct colonialism: the fact that women have lost what powers they had (say, in a matrilineal/matrilocal society) has to do with the switch to a hierarchical system which put the emphasis on new factors which excluded the type of influence women had.

In other words, I fully understand the connections between liberalism and literacy and I’ve heard enough about the importance of the printing press and journalism in these liberal societies to understand what role reading has played in those contexts. I simply dispute the notion that these connections should be universal.

Yes, I wish the “Universal Declaration of Human Rights” (including the (in)famous Article 26, which caused so many issues) were more culturally aware.

I started reading Deschooling Society a few weeks ago. In terms of “insight density,” it’s much higher than the book which prompted this discussion. While reading the first chapter, I constructed a number of ideas which I personally find useful.

I haven’t finished reading the book. Yet. I might eventually finish it. But much of what I wanted to get from that book, I was able to get from diverse sources. Including that part of the book I did read, sequentially. But, also, everything which has been written about Illich since 1971. And I’ll be interested in reading comments by the reading group at Wikiversity.

Given my background, I have as many “things to say” about the issues surrounding schooling as what I’ve read. If I had the time, I could write as much on what I’ve read from that book and it’d probably bring me a lot of benefits.

I’ve heard enough strong reactions against this attitude I’m displaying that I can hear it, already: “how can you talk about a book you haven’t read.” And I sincerely think these people miss an important point. I wouldn’t go so far as to say that their reading habits are off (that’d be mean), especially since those are well-adapted to certain contexts, including what I call scriptocentrism. Not that these people are scriptocentric. But their attitude “goes well with” scriptocentrism.

Academia, despite being to context for an enormous amount of writing and reading, isn’t displaying that kind of scriptocentrism. Sure, a lot of what we do needs to be written (although, it’s often surprising how much insight goes unwritten in the work of many an academic). And we do get evaluated through our writing. Not to mention that we need to write in a very specific mode, which almost causes a diglossia.

But we simply don’t feel forced to “read the whole text.”

A colleague has described this as the “dirty little secret” of academia. And one which changes many things for students, to the point that it almost sounds as if it remains a secret so as to separate students into categories of “those who get it” and “the mass.”

It doesn’t take a semester to read a textbook so there are students who get the impression that they can simply read the book in a weekend and take the exams. These students may succeed, depending on the course. In fact, they may get really good grades. But they run into a wall if they want to go on with a career making any use of knowledge construction skills.

Bill Reimer has interesting documents about “better reading.” It’s a PowerPoint presentation accompanied by exercises in a PDF format. (No, I won’t discuss format here.)

I keep pointing students to those documents for a simple reason: Reimer isn’t advocating reading every word in sequence. His “skim then focus” advice might be the one piece which is harder to get through to people but it’s tremendously effective in academic contexts. It’s also one which is well-adapted to the kind of online reading I’m thinking about. And not necessarily that good for physical books. Sure, you can efficiently flip pages in a book. But skimming a text on paper is more likely to be about what stands out visually than about the structure of the text. Especially with book-length texts. The same advice holds with physical books, of course. After all, this kind of advice originally comes from that historical period which I might describe as the “heyday of books”: the late 20th Century. But I’d say that the kind of “better reading” Reimer describes is enhanced in the context of online textuality. Not just the “Read/Write Web” but Instant Messaging, email, forums, ICQ, wikis, hypertext, Gopher, even PowerPoint…

Much of this has to do with different models of human communication. The Shannon/Weaver crowd have a linear/directional model, based on information processing. Codec and modem. Something which, after Irvine’s Shadow Conversations, I tend to call “the football theory of communication.” This model might be the best-known one, especially among those who study in departments of communication along with other would-be journalists. Works well for a “broadcast” medium with mostly indirect interaction (books, television, radio, cinema, press conferences, etc.). Doesn’t work so well for the backchannel-heavy “smalltalk”  stuff of most human communication actually going on in this world.

Some cognitivists (including Chomsky) have a schema-based model. Constructivists (from Piaget on) have an elaborate model based on knowledge. Several linguistic anthropologists (including yours truly but also Judith Irvine, Richard Bauman, and Dell Hymes) have a model which gives more than lipservice to the notion of performance. And there’s a functional model of any human communication in Jakobson’s classic text on verbal communication. It’s a model which can sound as if it were linear/bidirectional but it’s much broader than this. His six “functions of verbal communication” do come from six elements of the communication process (channel, code, form, context, speaker, listener). But each of these elements embeds a complex reality and Jakobson’s model seems completely compatible with a holistic approach to human communication. In fact, Jakobson has had a tremendous impact on a large variety of people, including many key figures in linguistic anthropology along with Lévi-Strauss and, yes, even Chomsky.

(Sometimes, I wish more people knew about Jakobson. Oh, wait! Since Jakobson was living in the US, I need to americanize this statement: “Jakobson is the most underrated scholar ever.”)

All these models do (or, in my mind, should) integrate written communication. Yet scriptocentrism has often led us far away from “texts as communication” and into “text as an object.” Scriptocentrism works well with modernity. Going away from scriptocentrism is a way to accept our postmodern reality.

Actively Reading: "Teach Naked" sans PowerPoint

Some Diigo comments on a Chronicle piece on moving lectures out of the classroom. (Or, if you ask the piece’s author and some commenters, on PowerPoint as a source of boredom.)

I’d like to transform some of my own comments in a standalone blog entry, especially given the discussions Pamthropologist and I have been having through comments on her blog and mine. (And I just noticed Pamthropologist had written her own blogpost about this piece…) As I’m preparing for the Fall semester, I tend to think a lot about learning and teaching but I also get a bit less time.

Semi-disclaimer: John Bentley, instructional developer and programme coordinator at Concordia’s CTLS pointed me to this piece. John used to work for the Open University and the BBC. Together, John and I are currently developing a series of workshops on the use of online tools in learning and teaching. We’ve been discussing numerous dimensions of the connection between learning, teaching, and online tools. Our current focus is on creating communities of learners. One thing that I find especially neat about this collaboration is that our perspectives and spheres of expertise are quite different. Makes for interesting and thoughtful discussions.

‘Teach Naked’ Effort Strips Computers From Classrooms – Technology – The Chronicle of Higher Education

  • Not to be too snarky but… I can’t help but feel this is typical journalism. Take a complex issue, get a diverse array of comments on it, boil it down to an overly simplistic point about some polarizing question (PPT: is it evil?). Tadaa! You got an article and you’ve discouraged critical thinking.Sorry. I’m bad. I really shouldn’t go there.But I guess I’m disappointed in myself. When I first watched the video interview, I was reacting fairly strongly against Bowen. After reading (very actively!) the whole piece, I now realize that Jeff Young is the one who set the whole thing up.The problem with this is that I should know better. Right?Well, ok, I wasn’t that adamantly opposed to Bowen. I didn’t shout at my computer screen or anything. But watching the video interview again, after reading the piece, I notice that I interpret as much more open a discussion than the setup made it sound like. In other words, I went from thinking that Bowen was imposing a radical view on members of his faculty to hearing Bowen proposing ideas about ways to cope with social changes surrounding university education.The statement about most on-campus lectures being bad is rather bold, but it’s nothing we haven’t heard and it’s a reasonable comment to make in such a context. The stronger statement against PPT is actually weakened by Bowen himself in two ways: he explicitly talks about using PPT online and he frames his comment in comparison with podcasts. It then sounds like his problem isn’t with PPT itself. It’s with the use of PPT in the classroom by comparison to both podcasts and PPTs online. He may be wrong about the relative merits of podcasts, online “presentations,” and classroom lectures using PPT. But his opinion is much less radical than what I originally thought.Still, there’s room for much broader discussion of what classroom lectures and PPT presentations imply in teaching. Young’s piece and several Diigo comments on it focus on the value of PPT either in the abstract or through appropriate use. But there’s a lot more ground to cover, including such apparently simple issues as the effort needed to create compelling “presentation content” or students’ (and future employers’) expectations about PPT presentations.
  • Mr. Bowen wants to discourage professors from using PowerPoint, because they often lean on the slide-display program as a crutch rather using it as a creative tool.
    • damn you got there first! comment by dean groom
    • I think the more important point that’s being made by the article – is something that many of us in edtech world realised very quickly – that being able to teach well is a prerequisite to being able to effectively and creatively engage technology to help others learn…Powerpoint is probably the most obvious target because oif its ubiquity – but I suspect that there will also be a backlash when the masses start adopting other technologies… they’ll be misused just as effectively as PPT is.When we can assume that all university lecturers/tutors are effective teachers then the argument will be moot… until then we’ll continue to see death by powerpoint and powerpointlessness…I’m a drama teacher and love the idea of active rooms filled with proactive engaged learners… and if we have proactive engaged learners we can more effectively deploy technology in the mix…The world of teaching and learning is far from perfect and expectations seem to be geared towards a paradigm that says : “professors should tell me every last thing I need to know in order to get good grades and if students sat still and shut up long enough they might just learn something useful.”I even had one “lecturer” recently tell me “I’m a subject specialist, why do I need to know about pedagogy?” – sadly he was serious. comment by Kim FLINTOFF
    • On the subject specialist uninterested in pedagogy…It’s not an uncommon perspective, in university teaching. In fact, it might be more common among French-speakers, as most of those I’ve heard say something like this were French-speakers.I reacted quite negatively when I first heard some statement about university teachers not needing pedagogy. Don’t they care about learning?But… Isn’t there a point to be made about “non-pedagogy?”Not trying to be contrarian, here. Not playing devil’s advocate. Nor am I going on the kind of “anti-anti” PoMo mode which seems not to fit too well in English-speaking communities. I’m just thinking about teacher-less learning. And a relativist’s attitude to not judge before I know more. After all, can we safely assume that courses given by someone with such a reluctant attitude to learning pedagogy are inherently bad?There are even some people out there who take constructivism and constructionism to such an extreme that they’d say teachers aren’t needed. To an extent, the OLPC project has been going in that direction. “Students will teach themselves. We don’t need to train teachers or to engage with them in building this project.”There’s also a lot of discussion about learning outside of formal institutions. Including “on-the-job training” but also all sorts of learning strategies which don’t rely on the teacher/student (mentee, apprentice, pupil…) hierarchy. For instance, actual learning occurs in a large set of online activities. Enthusiastic people learn about things that passion them by reading about the subject, participating in online discussions, presenting their work for feedback, etc. Oftentimes, there is a hierarchy in terms of prestige, but it’s mostly negotiated through actions and not set in advance. More like “achieved status” than “ascribed status” (to use a convenient distinction from SOC101 courses). As this kind of training not infrequently leads to interesting careers, we’d be remiss to ignore the trend.Speaking of trends… It’s quite clear that many universities tend toward a more consumer-based approach. Students register and pay tuition to get “credentials” (good grades and impressive degrees). The notion that they might be there to do the actual learning is going by the wayside. In some professional contexts, people are quite explicit about how little they learnt in classrooms. It makes for difficult teaching contexts (especially at prestigious universities in the US), but it’s also something with which people learn to cope.My personal attitude is that “learning happens despite teachers.” I still think teachers make a difference, that we should learn about learners and learning, that pedagogy matters a whole lot. In fact, I’m passionate about pedagogy and I do what I can to improve my teaching.Yet the bottomline is: do people learn? If they do, does it matter what pedagogical training the teacher has? This isn’t a rhetorical question. comment by Alexandre Enkerli
  • A study published in the April issue of British Educational Research Journal
    • http://www.informaworld.com/smpp/content~db=all~content=a902053143 comment by Alexandre Enkerli
  • PowerPoint was one of the dullest methods they saw.
    • Can somebody post links to especially good PowerPoint files? comment by Bill Chapman
    • I don’t think this is really about PPT, but more about blind use of technology. It’s not the software to blame but the user.Also if you’re looking for great PPT examples, check out slideshare.net comment by Dean Shareski
    • Looking forward to reading what their criteria are for boredom.And the exact justification they give for lectures needing not to be boring.Or if they discuss the broad implications of lecturing, as opposed to the many other teaching methods that we use.Now, to be honest, I do use PPT in class. In fact, my PPT slides are the very example of what many people would consider boring: text outlines transformed into bullet points. Usually black on white, without images.But, overall, students seem to find me engaging. In student evaluations, I do get the occasional comment about the course being boring, but that’s also about the book and the nature of what we discuss.I upload these PPT files to Slideshare before going to class. In seminars, I use the PPT file to outline some topics, themes, and questions brought up by students and I upload the updated file after class.The PPT files on Slideshare are embedded into Moodle and serve as “course notes,” in conjunction with the audio recordings from the class meetings. These slides may include material which wasn’t covered in class.During “lecture,” I often spend extend periods of time discussing things with the class as a whole, leaving a slide up as a reminder of the general topic. Going from a bullet point to an extended discussion has the benefit of providing context for the discussion. When I started teaching, several students were saying that I’m “disorganized.” I still get a few comments like that but they’re much less frequent. And I still go on tangents, based on interactions with the group.Once in a while, I refrain from using PPT altogether. Which can lead to interesting challenges, in part because of student expectations and the fact that the screen becomes an indicator that “teaching is going on.”Perhaps a more important point: I try to lecture as little as possible. My upper-level courses are rapidly transformed into seminars. Even in large classes, the last class meetings of the semester involve just a few minutes of lecturing.This may all sound like a justification for my teaching method. But it’s also a reaction to the frequent discussions about PPT as evil. I do hate PPT, but I still use it.If only Google Wave could be released soon, we could use it to replace PPT. Wikis and microblogging tools are good and well, but they’re not as efficient in terms of real-time collaboration on complex material. comment by Alexandre Enkerli
  • seminars, practical sessions, and group discussions
  • In other words, tech-free classrooms were the most engaging.
    • Does it follow so directly? It’s quite easy to integrate technology with “seminars, practical sessions, and group discussions.” comment by Alexandre Enkerli
  • better than many older classroom technologies, like slate chalkboards or overhead transparencies
    • Which seems to support a form of technological determinism or, at least, a notion of a somewhat consistent improvement in the use of tools, if not in the tools themselves. comment by Alexandre Enkerli
  • But technology has hardly revolutionized the classroom experience for most college students, despite millions of dollars in investment and early predictions that going digital would force professors to rethink their lectures and would herald a pedagogical renaissance.
    • If so, then it’s only because profs aren’t bringing social technologies into their classrooms. Does the author of this article understand what’s current in ed tech? comment by Shelly Blake-Plock
    • the problem here is that in higher education, student satisfaction drives a service mentality – and students WANT summised PPTs and the want PODCASTS. Spoooon feeeeeed me – for I am paying. comment by dean groom
    • A rather broad statement which might be difficult to support with evidence.If we look at “classroom experience” in different contexts, we do notice large differences. Not necessarily in a positive sense. Technology is an integral part of all sorts of changes happening in, around, and away from the classroom.It would be quite different if that sentence said: “But institutional programs based on the adoption of specific tools in the classroom have hardly revolutionized…” It’s still early to assess the effectiveness of these programs, especially if we think about lifelong learning and about ongoing social changes related to technology use. But the statement would make more sense if it were more directly tied to specific programs instead of being a blanket critique of “technology” (left undefined). comment by Alexandre Enkerli
  • dream of shaking up college instruction
    • One of the most interesting parts of the interview with Bowen has to do with the notion that this isn’t, in fact, about following a dream. It’s about remaining relevant in a changing world. There’s a lot about Bowen’s perspective which sounds quite strange, to me. But the notion that universities should “wake up and smell the coffee” is something I wish were the object of more discussion in academic circles. comment by Alexandre Enkerli
  • Here’s the kicker, though: The biggest resistance to Mr. Bowen’s ideas has come from students, some of whom have groused about taking a more active role during those 50-minute class periods.
    • Great points, here. Let’s wish more students were involved in this conversation. It’s not just “about” them.One thing we should probably not forget about student populations is that they’re diverse. Chances are, some students in Meadows are delighted by the discussion focus. Others may be puzzled. It’s likely an adaptation for most of them. And it doesn’t sound like they were ever consulted about those changes. comment by Alexandre Enkerli
  • lecture model is pretty comfortable
    • And, though many of us are quick to criticize it, it’s difficult to avoid in the current systems of formal education in which we work. comment by Alexandre Enkerli
  • cool gadgets
    • The easiest way to dismiss the social role of technology is to call tools “gadgets.” But are these tools really just gadgets? In fact, some tools which are put to good use really aren’t that cool or even new. Are we discussing them enough? Are we aware of how they fit in the grand scheme of things?An obvious example would be cellphones. Some administrators and teachers perceive them as a nuisance. Rather few people talk about educational opportunities with cellphones, even though they already are used by people in different parts of the World to empower themselves and to learn. Negroponte has explicltly dimissed the educational potential of cellphones but the World isn’t waiting for approval from designers. comment by Alexandre Enkerli
  • seasoned performer,
    • There’s a larger point to be about performance in teaching. Including through a reference to Dick Bauman’s “Verbal Art as Performance” or other dimensions of Performance Theory.There’s also a more “mundane” point about a kind of conflict in universities between academic material and performance. In French-speaking universities, at least, it’s not uncommon to hear teachers talk about the necessity to be a “performer” as something of a distraction in teaching. Are teachers in front of the class to entertain students or is the classroom an environment in which to think and learn about difficult concepts? The consumer approach to universities, pushed in part by administrators who run universities like businesses, tends to emphasize the “entertainment paradigm,” hence the whole “boredom” issue.Having said all of this, Bowen’s own attitude goes beyond this simplistic “entertainment paradigm.” In fact, it sounds like he’s specifically not advocating for lectures to become a series of TEDtalks. Judging from the interview, it sounds like he might say that TEDtalk-style presentation should be put online and classroom-time should be devoted to analyzing those presentations.I do consider myself a performer, as I’ve been playing saxophone in a rather broad range of circumstances, from outdoor stages at festivals to concert halls. And my experience as a performer does influence the way I teach large classes. At the same time, it probably makes more salient the distinction between teaching and performing. comment by Alexandre Enkerli
  • The goateed administrator sported a suit jacket over a dark T-shirt
    • Though I’d be the first one to say that context is key, I fail to see what Bowen’s clothes contribute to the discussion. comment by Alexandre Enkerli
  • philosophical argument about the best way to engage students, he grounded it
  • information delivery common in today’s classroom lectures should be recorded and delivered to students as podcasts or online videos before class sessions
    • Fully agreed. Especially if we throw other things in the mix such as journal articles and collaboratively-created learning material. comment by Alexandre Enkerli
  • short online multiple-choice tests.
    • I don’t think he’s using the mc tests with an essessment focus rather an engagement focus – noit necessarily the most sophisticated but done playfully and creatively it can be a good first step to getting reluctatnt students to engage in first instance… comment by Kim FLINTOFF
    • I would also “defend” the use of MCTs in this context. Especially if the stakes are relatively low, the questions are well-crafted, and students do end up engaging.Like PPT, MCTs have some advantages, including because of student expectations.But, of course, it’s rather funny to hear Bowen talk about shaking things up and find out that he uses such tools. Still, the fact that these tests are online (and, one would think, taken outside of class time) goes well with Bowen’s main point about class time vs. tech-enabled work outside of class. comment by Alexandre Enkerli
  • Introduce issues of debate within the discipline and get the students to weigh in based on the knowledge they have from those lecture podcasts, Mr. Bowen says.
    • This wouldn’t be too difficult to do in social sciences and there are scenarios in which it would work wonderfully for lab sciences (if we think of “debate” as something similar to “discussion” sections in scientific articles).At the same time, some people do react negatively to such approaches based not on discipline but on “responsibilities of the university.” Some people even talk about responsibilities toward students’ parents! comment by Alexandre Enkerli
  • But if the student believes they can contribute, they’re a whole lot more motivated to enter the discourse, and to enter the discipline.
    • Sounds a bit like some of the “higher” positions in William Perry’s scheme. comment by Alexandre Enkerli
  • don’t be boring
    • Is boredom induced exclusively by the teacher? Can a student bored during a class meeting still be motivated and engaged in the material at another point? Should we apply the same principle to the readings we assign? Is there a way to efficiently assess the “boredom factor” of an academic article? How can we convince academic publishers that fighting boredom isn’t necessarily done through the addition of pretty pictures? comment by Alexandre Enkerli
  • you need a Ph.D. to figure it out
    • While I agree that these panels are difficult to use and could afford a redesign, the joke about needing a PhD sounds a bit strange in context. comment by Alexandre Enkerli
  • plug in their laptops
    • There’s something of a more general move toward getting people to use their own computers in the workplace. In fact, classroom computers are often so restricted as to be quite cumbersome to use in teaching. comment by Alexandre Enkerli
  • allow students to work in groups more easily
    • Not a bad idea. A good number of classrooms are structured in a way that makes it very hard to get students to do group work. Of course, it’s possible to do group work in any setting, but it’s remarkable how some of these seemingly trivial matters as the type of desk used can be enough to discourage some teachers from using certain teaching strategies. comment by Alexandre Enkerli
  • The classroom computers were old and needed an upgrade when Mr. Bowen arrived, so ditching them instead saved money.
    • Getting into the core of the issue. The reason it’s so important to think about “new ways” to do things isn’t necessarily that “old ways” weren’t conducive to learning. It’s because there are increased pressures on the system and some seem to perceive that cost-cutting and competition from online learning, making the issue so pressing. comment by Alexandre Enkerli
  • eliminate one staff position for a technician
    • Sounds sad, especially since support staff is already undervalued. But, at the same time, it does sound like relatively rational cost-cutting. One would just wish that they replaced that position with, say, teaching support. comment by Alexandre Enkerli
  • gave every professor a laptop
    • Again, this is a rather common practise outside of universities. Knowing how several colleagues think, this may also function as a way to “keep them happy.” comment by Alexandre Enkerli
  • support so they could create their own podcasts and videos.
    • This is where the tech support position which was cut could be useful. Recording and podcasting aren’t difficult to set up or do. But it’s an area where support can mean more than answering questions about which button to press. In fact, podcasting projects are an ideal context for collaboration between tech, teach, and research. comment by Alexandre Enkerli
  • lugging their laptops to class,
    • It can be an issue, especially if there wasn’t a choice in the type of laptop which could be used. comment by Alexandre Enkerli
  • She’s made podcasts for her course on “Critical Scholarship in Communication” that feature interviews she recorded with experts in the field.
    • One cool thing about these podcasting projects is that people can build upon them, one semester after the other. Interviews with practitioners do help provide a multiplicity of voices. And, yes, getting students to produce their own content is often a good way to go, especially if the topic is somehow related to the activity. Getting students in applied communication to create material does sound appropriate. comment by Alexandre Enkerli
  • they come in actually much more informed
    • Sounds effective. Especially since Bowen’s approach seems to be oriented toward pre-class preparation. comment by Alexandre Enkerli
  • if they had been assigned a reading.
    • There’s a lot to be said about this. One reason this method may be more efficient than reading assignments could have to do with the particularities of written language, especially the very formal style of those texts we often assign as readings. Not that students shouldn’t read, of course! But there’s a case to be made for some readings being replaced by oral sources, especially those which have to do with people’s experience in a field. Reading primary source material, integrating some reference texts, and using oral material can all be part of an appropriate set of learning strategies. comment by Alexandre Enkerli
  • created podcast lectures
    • An advantage of such “lecturecasts,” “profcasts,” and “slidecasts” is that they’re relatively easy to build and can be tightly structured. It’s not the end-all of learning material, but it’s a better substitute for classroom lectures than one might think.Still, there’s room for improvement in the technology itself. For instance, it’d be nice to have Revver-style comments in the timeline. comment by Alexandre Enkerli
  • shows movie clips from his laptop
    • This one is slightly surprising because one would expect that these clips could easily be shown, online, ahead of class. It might have to do with the chilling effect of copyright regulation or Heffernan’s strategy of getting “fresh” feedback. There would have been good questions to ask Heffernan in preparation for this piece. comment by Alexandre Enkerli
  • “Strangely enough, the people who are most resistant to this model are the students, who are used to being spoon-fed material that is going to be quote unquote on the test,” says Mr. Heffernan. “Students have been socialized to view the educational process as essentially passive. The only way we’re going to stop that is by radically refiguring the classroom in precisely the way José wants to do it.”
    • This interpretation sounds a tiny bit lopsided. After all, aren’t there students who were already quite active and engaged in the “old system” who have expressed some resistance to the “new system?” Sounds likely to me. But maybe my students are different.One fascinating thing is the level of agreement, among teachers, about the necessity to have students who aren’t passive. I certainly share this opinion but there are teachers in this World who actually do prefer students who are somewhat passive and… “obedient.” comment by Alexandre Enkerli
  • The same sequence of events
    • That part is quite significant: Bowen was already a reformer and already had gone through the process. In this case, he sounds more like one of those CEOs who are hired to save a company from a difficult situation. He originally sounded more like someone who wanted to impose specific views about teaching. comment by Alexandre Enkerli
  • ‘I paid for a college education and you’re not going to lecture?'”
    • A fairly common reaction, in certain contexts. A combination of the infamous “sense of entitlement,” the “customer-based approach to universities,” and student expectations about the way university teaching is supposed to go.One version I’ve had in student evaluations is that the student felt like s/he was hearing too much from other students instead of from me. It did pain me, because of the disconnect between what I was trying to do and that student’s notion of what university courses are supposed to bring her/him. comment by Alexandre Enkerli
  • PowerPoint lecture
    • As a commenter to my blog was saying about lectures in general, some of us (myself included) have been building a straw man. We may have negative attitudes toward certain teaching strategies, including some lecturing techniques. But that shouldn’t prevent us from discussing a wide array of teaching and learning strategies.In this case, it’s remarkable that despite the radical nature of Bowen’s reform, we learn that there are teachers who record PPT-based presentations. It then sounds like the issue isn’t so much about using PPT as it is about what is done in the classroom as opposed to what is done during the rest of the week.Boring or not, PPT lectures, even some which aren’t directly meant to engage students, can still find their place in the “teaching toolbox.” A dogmatic anti-PPT stance (such as the one displayed by this journalist) is unlikely to foster conversations about tools and learning. Based on the fact that teachers are in fact doing PPT lectures to be used outside the classroom, one ends up seeing Bowen’s perspective as much more open than that of the Chronicle’s editorial staff. comment by Alexandre Enkerli
  • Sandi Mann, the British researcher who led the recent study on student attitudes toward teaching, argues that boredom has serious implications in an educational setting.
    • Unsurprising perspective. Wonder if it had any impact on Mann’s research results. Makes the research sound more oriented than one might hope. comment by Alexandre Enkerli
  • according to some studies
  • low-cost online alternatives to the traditional campus experience
    • This could have been the core issue discussed in an article about Bowen. Especially if we are to have a thoughtful conversation about the state of higher education in a changing context. Justification for high tuition fees, the latent functions of “college life,” the likely outcome of “competing with free,” the value of the complete learning experience as opposed to the value of information transmission… comment by Alexandre Enkerli
  • give away videos
    • This is the “competing with free” part, to which record companies have been oblivious for so long but which makes OCW appear like quite a forward-looking proposition. comment by Alexandre Enkerli
  • colleges must make sure their in-person teaching really is superior to those alternatives
    • It’s both a free-market argument, which goes so well with the customer-based approach to learning, and a plea to consider learning in a broader way than the mere transmission of information from authoritative source to passive mass. An old saw, for sure, but one which surprisingly hasn’t been heard by everyone. comment by Alexandre Enkerli
  • add value
    • This might be appropriate language to convince trustees. At some institutions, this might be more important than getting students’ or teachers’ approval. comment by Alexandre Enkerli
  • not being online
    • Although, they do have an online presence. The arguments used have more to do with blended learning than with exclusively face-to-face methods. comment by Alexandre Enkerli
  • might need to stay a low-tech zone to survive.
    • Rubbish there is no reason to dumb down learning; and he obviously is not teaching 2500 students at one time. PPT is not the problem here, and this really is a collection of facile arguements that are not ironically substantiated. Lowering his overhead does not increase student learning – wheres the evidence? comment by dean groom
    • Come to think of it, it sounds like the argument was made more forcefully by Young than by Bowen himself. Bowen is certainly quite vocal but the “need… to survive” sounds a tad bit stronger than Bowen’s project.What’s funny is that the video made Bowen sound almost opinionated. The article makes Young sound like he has his own axe to grind comment by Alexandre Enkerli

A Glocal Network of City-States?

This one should probably be in a fictive mode, maybe even in a science-fiction genre. In fact, I’m reconnecting with literature after a long hiatus and now would be an interesting time to start writing fiction. But I’ll still start this as one of those  “ramblings” blogposts that I tend to build or which tend to come to me.

The reason this should be fiction is that it might sound exceedingly naïve, especially for a social scientist. I tend to “throw ideas out there” and see what sticks to other ideas, but this broad idea about which I’ve been thinking for a while may sound rather crazy, quaint, unsophisticated.

See, while my academic background is rather solid, I don’t have formal training in political science. In fact, I’ve frequently avoided several academic activities related to political science as a discipline. Or to journalism as a discipline. Part of my reluctance to involve myself in academic activities related political science relates to my reaction to journalism. The connection may not seem obvious to everyone but I see political science as a discipline in the same frame, and participating in the same worldview, as what I find problematic in journalism.

The simplest way to contextualize this connection is the (“modern”) notion of the “Nation-State.” That context involves me personally. As an anthropologist, as a post-modernist, as a “dual citizen” of two countries, as a folklorist, as a North American with a relatively salient European background, as a “citizen of the World,” and as a member of a community which has switched in part from a “nationalist” movement to other notions of statehood. Simply put: I sincerely think that the notion of a “Nation-State” is outdated and that it will (whether it should or not) give way to other social constructs.

A candidate to replace the conceptual apparatus of the “Nation-State” is both global and local, both post-modern and ancient: a glocal network of city-states (GNoCS).

Yes, I know, it sounds awkward. No, I’m not saying that things would necessarily be better in a post-national world. And I have no idea when this shift from the “nation-states” frame to a network of city-states may happen. But I sincerely think that it could happen. And that it could happen rather quickly.

Not that the shift would be so radical as to obliterate the notion of “nation-state” overnight. In this case, I’m closer to Foucault’s épistémè than to Kuhn’s paradigm. After all, while the “Democratic Nation-State” model is global, former social structures are still present around the Globe and the very notion of a “Nation-State” takes different values in different parts of the world. What I envision has less to do with the linear view of history than with a perspective in which different currents of social change interact with one another over time, evoking shifts in polarity for those who hold a binary perspective on social issues.

I started “working on” this post four months ago. I was just taking some notes in a blog draft, in view of a blogpost, instead of simply keeping general notes, as I tend to do. This post remained on my mind and I’ve been accumulating different threads which can connect to my basic idea. I now realize that this blogpost will be more of a placeholder for further thinking than a “milestone” in my reflection on the topic. My reluctance to publish this blog entry had as much to do with an idiosyncratic sense of prudence as with time-management or any other issue. In other words, I was wary of sticking my neck out. Which might explain why this post is so personal as compared to most of my posts in English.

As uninformed as I may seem of the minutiae of national era political science, I happen to think that there’s a lot of groupthink involved in the way several people describe political systems. For instance, there’s a strong tendency for certain people, journalists especially, to “count countries.” With relatively few exceptions (especially those which have to do with specific international institutions like the United Nations or the “G20”) the number of countries involved in an event only has superficial significance. Demographic discrepancies between these national entities, not tio mention a certain degree of diversity in their social structures or even government apparatus, makes “counting countries” appear quite misleading, especially when the issue has to do with, say, social dynamics or geography. It sounds at times like people have a vague “political map of the World” in their heads and that this image preempts other approaches to global diversity. This may sound like a defensive stance on my part, as I try to position myself as “perhaps crazy but not more than others are.” But the issue goes deeper. In fact, it seems that “countries” are so ingrained  in some people’s minds and political borders are so obvious that local and regional issues are perceived as micro-version of what happens at the “national level.” This image doesn’t seem so strange when we talk about partisan politics but it appears quite inappropriate when we talk about a broad range of other subjects, from epidemiology to climate change, from online communication to geology, from language to religion.

An initial spark in my thinking about several of these issues came during Beverly Stoeltje‘s interdisciplinary Ph.D. seminar on nationalism at Indiana University Bloomington, back in 2000. Not only was this seminar edifying on many levels, but it represented a kind of epiphany moment in my reflections on not only nationalism itself (with related issues of patriotism, colonialism, and citizenship) but on a range of social issues and changes.

My initial “realization” was on the significance of the shift from Groulx-style French-Canadian nationalism to what Lévesque called «souveraineté-association» (“sovereignty-association”) and which served as the basis for the Quebec sovereignty movement.

While this all connects to well-known issues in political science and while it may (again) sound exceedingly naïve, I mean it in a very specific way which, I think, many people who discuss Quebec’s political history may rarely visit. As with other shifts about which I think, I don’t envision the one from French-Canadian nationalism (FCN) to Quebec sovereignty movement (QSM) to be radical or complete. But it was significant and broad-reaching.

Regardless of Lévesque’s personal view on nationalism (a relatively recent television series on his life had it that he became anti-nationalist after a visit to concentration camps), the very idea that there may exist a social movement oriented toward sovereignty outside of the nationalist logic seems quite important to me personally. The fact that this movement may only be represented in partisan politics as nationalism complicates the issue and may explain a certain confusion in terms of the range of Quebec’s current social movements. In other words, the fact that anti-nationalists are consistently lumped together with nationalists in the public (and journalistic) eye makes it difficult to discuss post-nationalism in this part of the Globe.

But Quebec’s history is only central to my thinking because I was born and Montreal and grew up through the Quiet Revolution. My reflections on a post-national shift are hopefully broader than historical events in a tiny part of the Globe.

In fact, my initial attempt at drafting this blogpost came after I attended a talk by Satoshi Ikeda entitled The Global Financial Crisis and the End of Neoliberalism. (November 27, 2008, Concordia University, SGW H-1125-12; found thanks to Twistory). My main idea at this point was that part of the solution to global problems were local.

But I was also thinking about The Internet.

Contrary to what technological determinists tend to say, the ‘Net isn’t changing things as much as it is part of a broad set of changes. In other words, the global communication network we now know as the Internet is embedded in historical contexts, not the ultimate cause of History. At the risk of replacing technological determinism with social determinism, one might point out that the ‘Net existed (both technologically and institutionally) long before its use became widespread. Those of us who observed a large influx of people online during the early to mid-1990s might even think that social changes were more significant in making the ‘Net what it is today than any “immanent” feature of the network as it was in, say, 1991.

Still, my thinking about the ‘Net has to do with the post-national shift. The ‘Net won’t cause the shift to new social and political structures. But it’s likely to “play a part” in that shift, to be prominently places as we move into a post-national reality.

There’s a number of practical and legal issues with a wide range of online activities which make it clear that the ‘Net fits more in a global structure than in an “international” one. Examples I have in mind include issues of copyright, broadcast rights, “national content,” and access to information, not to mention the online setting for some grassroots movements and the notion of “Internet citizenry.” In all of these cases, “Globalization” expands much beyond trade and currency-based economy.

Then, there’s the notion of “glocalization.” Every time I use the term “glocal,” I point out how “ugly” it is. The term hasn’t gained any currency (AFAICT) but I keep thinking that the concept can generate something interesting. What I personally have in mind is a movement away from national structures into both a globally connected world and a more local significance. The whole “Think Local, Act Global” idea (which I mostly encountered as “Think Global, Drink Local” as a motto). “Despite” the ‘Net, location still matters. But many people are also global-looking.

All of this is part of the setup for some of my reflections on a GNoCS. A kind of prelude/prologue. While my basic idea is very much a “pie in the sky,” I do have more precise notions about what the future may look like and the conditions in which some social changes might happen. At this point, I realize that these thoughts will be part of future blogposts, including some which might be closer to science-fiction than to this type semi- (or pseudo-) scholarly rambling.

But I might still flesh out a few notes.

Demographically, cities may matter more now than ever as the majority of the Globe’s population is urban. At least, the continued urbanization trend may fit well with a city-focused post-national model.

Some metropolitan areas have become so large as to connect with one another, constituting a kind of urban continuum. Contrary to boundaries between “nation-states,” divisions between cities can be quite blurry. In fact, a same location can be connected to dispersed centres of activity and people living in the same place can participate in more than one local sphere. Rotterdam-Amsterdam, Tokyo-Kyoto, Boston-NYC…

Somewhat counterintuitvely, urban areas tend to work relatively as the source of solutions to problems in the natural environment. For instance, some mayors have taken a lead in terms of environmental initiatives, not waiting for their national governments. And such issues as public transportations represent core competencies for municipal governments.

While transborder political entities like the European Union (EU), the African Union (AU), and the North American Free-Trade Agreement (NAFTA) are enmeshed in the national logic, they fit well with notions of globalized decentralization. As the mayor of a small Swiss town was saying on the event of Switzerland’s official 700th anniversary, we can think about «l’Europe des régions» (“Europe of regions”), beyond national borders.

Speaking of Switzerland, the confederacy/confederation model fits rather well with a network structure, perhaps more than with the idea of a “nation-state.” It also seems to go well with some forms of participatory democracy (as opposed to representative democracy). Not to mean that Switzerland or any other confederation/confederacy works as a participatory democracy. But these notions can help situate this GNoCS.

While relatively rare and unimportant “on the World Stage,” micro-states and micro-nations represent interesting cases in view of post-nationalist entities. For one thing, they may help dispel the belief that any political apart from the “nation-state” is a “reversal” to feudalism or even (Greek) Antiquity. The very existence of those entities which are “the exceptions to the rule” make it possible to “think outside of the national box.”

Demographically at the opposite end of the spectrum from microstates and micronations, the notion of a China-India union (or even a collaboration between China, India, Brazil, and Russia) may sound crazy in the current state of national politics but it would go well with a restructuring of the Globe, especially if this “New World Order” goes beyond currency-based trade.

Speaking of currency, the notion of the International Monetary Fund having its own currency is quite striking as a sign of a major shift from the “nation-state” logic. Of course, the IMF is embedded in “national” structures, but it can shift the focus away from “individual countries.”

The very notion of “democracy” has been on many lips, over the years. Now may be the time to pay more than lipservice to a notion of “Global Democracy,” which would transcend national boundaries (and give equal rights to all people across the Globe). Chances are that representative democracy may still dominate but a network structure connecting a large number of localized entities can also fit in other systems including participatory democracy, consensus culture, republicanism, and even the models of relatively egalitarian systems that some cultural anthropologists have been constructing over the years.

I still have all sorts of notes about examples and issues related to this notion of a GNoCS. But that will do for now.