Category Archives: online publishing

Bean Counters and Ecologists

[So many things in my drafts, but this one should be quick.]

Recently met someone who started describing their restaurant after calling it a “café”. The “pitch” revolved around ethical practices, using local products, etc. As both a coffee geek and ethnographer, my simple question was: “Which coffee do you use?” Turns out, they’re importing coffee from a multinational corporation. “Oh, but, they’re lending us an expensive espresso machine for free! And they have fair-trade coffee!”

Luckily, we didn’t start talking about “fair trade”. And this person was willing to reflect upon the practices involved, including about the analogy with Anheuser-Busch or Coca-Cola. We didn’t get further into the deeper consequences of the resto’s actions, but the “seed” has been planted.

Sure, it’s important to focus on your financials and there’s nothing preventing a business from being both socially responsible and profitable. It just requires a shift in mindset. Small, lean, nimble businesses are more likely to do it than big, multinational corporate empires…

…which leads me to Google.

Over the years since its IPO, Google has attracted its share of praise and criticism. Like any big, multinational corporate empire. In any sector.

Within the tech sector, the Goog‘ is often compared with Microsoft, Facebook, Twitter, and Apple. All of these corporate entities have been associated in some people’s minds with some specific issue, from child labour and failure to protect users’ privacy to anticompetitive practices (the tech equivalent of free fridges and espresso machines). The issues are distinct and tech enthusiast spend a large amount of time discussing which one is worse. Meanwhile, we’re forgetting a number of larger issues.

Twitter is an interesting example, here. The service took its value from being at the centre of an ecosystem. As with any ecosystem, numerous interactions among many different members produce unexpected and often remarkable results. As the story goes, elements like hashtags and “@-replies” were invented by users and became an important part of the system. Third-party developers were instrumental in Twitter’s reach outside of its original confines. Though most of the original actors have since left the company, the ecosystem has maintained itself over the years.

When Twitter started changing the rules concerning its API, it shook the ecosystem. Sure, the ecosystem will maintain itself, in the end. But it’s nearly impossible to predict how it will change. For people at Twitter, it must have been obvious that the first changes was a warning shot to scare away those they didn’t want in their ecosystem. But, to this day, there are people who depend on Twitter, one way or another.

Google Reader offers an interesting case. The decision to kill it might have been myopic and its death might have a domino effect.

The warning shot was ambiguous, but the “writing was on the wall”. Among potential consequences of the move, the death of RSS readers was to be expected. One might also expect users of feedreaders to be displeased. In the end, the ecosystem will maintain itself.

Chances are, feedreading will be even more marginalized than it’s been and something else might replace it. Already, many people have been switching from feedreading to using Twitter as a way to gather news items.

What’s not so well-understood is the set of indirect consequences, further down the line. Again, domino effect. Some dominoes are falling in the direction of news outlets which have been slow to adapt to the ways people create and “consume” news items. Though their ad-driven models may sound similar to Google’s, and though feedreading might not be a significant source of direct revenue, the death of feedreaders may give way to the birth of new models for news production and “consumption” which might destabilize them even further. Among the things I tag as #FoJ (“Future of Journalism”) are several pieces of a big puzzle which seems misunderstood by news organizations.

There are other big dominoes which might fall from the death of Google Reader. Partly because RSS itself is part of a whole ecosystem. Dave Winer and Aaron Swartz have been major actors in the technical specifications of RSS. But Chris Lydon and people building on calendar syndication are also part of the ecosystem. In business-speak, you might call them “stakeholders”. But thinking about the ecosystem itself leads to a deeper set of thoughts, beyond the individuals involved. In the aftermath of Aaron Swartz’s premature death, it may be appropriate to point out that the ecosystem is more than the sum of its parts.

As I said on a service owned by another widely-criticized corporate empire:

Many of us keep saying that Google needs to listen to its social scientists. It also needs to understand ecology.

The Magazine and Social Media

Megaphone red
Megaphone red by Adamantios (via Wikimedia Commons, (GFDL, CC-BY-SA)

The following is my App Store review of The Magazine, a Newsstand offering by Instapaper developer Marco Arment.

Though I like Marco Arment’s work and there’s nothing specifically wrong about this implementation of the magazine model, I don’t find the magazine model particularly useful, at this point. And, make no mistake. The Magazine is indeed a magazine.

Oh, sure, this format overcomes several of the limitations set by advertising-based models and hierarchical boards. But it maintains something of the magazine logic: a tight bundle of a few articles authored by people connected through the same “editorial intent”. It’s not a conversation with the public. In this first issue, it’s not even a conversation among co-authors.

The “linked list” aspect of the “Fireball Format” (from John Gruber’s Daring Fireball media property) is described in one of the pieces in this first issue. Other distinguishing factors of the “Fireball Format” aren’t discussed in that same piece. They include a “no comment” policy which has become rather common among high-profile blogs. Unlike most blogs of the pioneer era in social media, these blogs don’t allow readers to comment directly.

A justification for this policy is that comments can be posted elsewhere. And since most of these bloggers are active on microblogging platforms like App.net and Twitter, there’s a chance that a comment might be noticed by those authors. What’s missing, though, is the sense of belonging which bloggers created among themselves before MySpace.

In other words, now that there are large social networking services online, the social aspects of blogging have been deemphasized and authorial dimensions have come to prominence. Though Arment dislikes the word, blog authors have become “brands”. It still works when these authors are in conversation with one another, when there’s a likelihood of a “followup” (FU in 5by5 parlance), when authors are responsive.

None of that interaction potential seems to be part of the core model for The Magazine. You can scream at your iOS device all you want, Jason Snell will probably not respond to you in a future edition of The Magazine. You can attempt dialogue on Twitter, but any conversation you may succeed in starting there is unlikely to have any impact on The Magazine. You’re talking with authors, now, not with members of a community.

With The Magazine, the transition from social to authorial is almost complete. Not only are posts set apart from the conversation but the editorial act of bundling posts together brings back all the problems media scholars have been pointing out for the past several decades. The issue at stake isn’t merely the move to online delivery. It’s the structure of authority and the one-to-many broadcast-style transmission. We’ve taken a step back.

So, while The Magazine has certain technical advantages over old school magazines like The Daily and Wired, it represents a step away from social media and towards mass media. Less critical thinking, more pedestals.

A new model could emerge using the infrastructure and business model that Arment built. But it’d require significant work outside of the application. The Feature might contribute something to this new model, especially if the way posts are bundled together became more flexible.

So, all in all, I consider The Magazine to be a step in the wrong direction by someone whose work I respect.

Good thing we still have podcasts.

Timeline of Apple’s Online Services

[I’d like people’s help in completing a timeline of Apple’s online services.]

[Update: deleted the WordPress shortcode, which seems not to be working.]

As kind of a followup to yesterday’s post about some early rumours and speculations about the iPhone, I thought about posting some info about Apple’s online services. Part of the reason is that Asymco’s Horace Dediu has frequently talked about what we could call “Apple’s data play”, for instance in this post about the iCloud data centre in North Carolina. I was also thinking about Mike Davidson’s comments about Apple’s presence (and “dominance”) in such diverse fields as hardware, software, licensing, and commerce. The trigger for this post, though, was from this Steve Jobs comment, which appeared in a recent NYT piece about the Apple Maps fiasco:

The MobileMe launch clearly demonstrates that we have more to learn about Internet services

The overall context for this quote as well as a number of discussions about Apple is the consensus that Apple does a poor job with online services. MobileMe and iTunes Ping are often used in these discussions and it seems clear to most people (including Apple executives and insiders, it sounds like) that the “computer company turned consumer electronics vendor” has a lot to learn about online services.

The reason I find this so interesting is that Apple seems insistent on pushing at least some of its online services. A bit less of a “betting the farm” strategy as Google’s “Emerald Sea” initiative, but an intriguing strategy for such a large and still-successful company. Dediu’s frequent reference to Clay Christensen’s concept of “Disruptive Innovation” might apply, here. Apple might be “disrupting itself into” an online services company, at least in part.

There are several things I find intriguing about this strategy.

As opposed to most other enterprises’ “online plays”, Apple’s model tends not to be based on ad revenues. The divide between Google and Apple couldn’t be stronger when we talk about ad-supported free/freemium services as opposed to paid services or services attached to other purposes. It’s likely an irreconcilable difference between fans of  both teams.

Online services are clearly not Apple’s strong suit. It often sounds like Apple is missing a “magic touch” with online services, the same way other companies are said to lack Apple’s design sense. This is more similar to Google+ given the consensus that “Google doesn’t know how to do ‘social’”. But it’s still surprising.

Though Apple may not have a “knack” for online services, it’s been trying to do it for quite a while. I keep thinking about eWorld as a precursor to the whole thing. It’s one thing for a company to try its hand at something new or to pivot into a strong business. It’s another thing entirely to shift more energies into something which has so far proven to be mostly a lost cause.

Adding to my thoughts on this was a podcast conversation (I think between John Siracusa and Dan Benjamin, though it might have been between Marco Arment and John Gruber) during which comments were made about those Apple employees working on online services.

So, basically, Apple’s online services have been on my mind. But I couldn’t find an exhaustive list. Tried Wikipedia but it doesn’t really separate online services from other things Apple does. And I ended up thinking about what would define “online services” in Apple’s case. Everything Apple does which incurs some bandwidth costs would be my working definition. Basically, it’s something to do with Apple investing in data centres and such. Some of these seem like very small costs (hosting data about podcasts, instead of the podcasts themselves, for instance). Given Apple’s size, these costs and the infrastructure behind all of this can be quite big.

So I started listing some of these services and organizing them in a sort of timeline, first in MultiMarkdown format in nvAlt, then in a Google Spreadsheet. I then discovered Vérité.CO’s Timeline.JS which takes a Google Spreadsheet and makes it into a visual timeline.

A few notes:

  • It’s a quick draft and I didn’t really check any of the data points.
  • In most cases, I only added months and, in the case of “AppleLink”, I only put years.
  • I took most dates from diverse Wikipedia pages, not necessarily backtracking on the whole process.
  • On at least one occasion, there was a discrepancy between two dates.
  • Sometimes, I took the date of the service’s announcement while I used an actual launch date for other services.
  • I only added a couple of pictures to show that it can be done. Many of the relevant pix are likely to be under copyright or to constitute a trademark.
  • I tried to be as exhaustive as I could be, but I’m sure I forgot stuff.
  • Some things may not sound like they qualify as part of “Apple’s online offering” but I think they’re still relevant. My rule of thumb is that if it goes to Apple’s servers, it’s an online service.
  • I separated some services from “suites” like iCloud or iTools, partly because some of those services haven’t been kept, which is important to see in a timeline. There are several services missing, here.
  • None of this timeline is meant to be editorial. I was just curious about what Apple has been doing online since the 1980s. The reason I care can be found in my earlier notes. I consider myself neither an “Apple fanboi” nor an “Apple hater”. I just find the situation revealing of something happening in the tech world, which has an impact on the Geek Niche.

So, here goes.

Here’s the Google Spreadsheet (editable by anyone):

Apple Online Services

Here’s the timeline through an embed code:

Here’s the embed code:

<iframe src='http://embed.verite.co/timeline/?source=0AjnWdp-FPwEKdHVqOXhWVlZuZjZYajN5QnExcExuVmc&font=Bevan-PotanoSans&maptype=toner&lang=en&hash_bookmark=true&height=650' width='100%' height='650' frameborder='0'>

 

Using WordPress as a Syllabus Database: Learning is Fun

(More screenshots in a previous post on this blog.)

Worked on a WordPress project all night, the night before last. Was able to put together a preliminary version of a syllabus database that I’ve been meaning to build for an academic association with which I’m working.

There are some remaining bugs to solve but, I must say, I’m rather pleased with the results so far. In fact, I’ve been able to solve the most obvious bugs rather quickly, last night.

More importantly, I’ve learnt a lot. And I think I can build a lot of things on top of that learning experience.

Part of the inspiration comes from Kyle Jones’s blogpost about a “staff directory”. In addition, Justin Tadlock has had a large (and positive) impact on my learning process, either through his WordPress-related blogposts about custom post types and his work on the Hybrid Theme (especially through the amazing support forums). Not to mention WordCamp Montrealofficial documentationplugin pagestutorials, and a lot of forum– and blogposts about diverse things surrounding WordPress (including CSS).

I got a lot of indirect help and I wouldn’t have been able to go very far in my project without that help. But, basically, it’s been a learning experience for me as an individual. I’m sure more skilled people would have been able to whip this up in no time.

Thing is, it’s been fun. Close to Csíkszentmihályi’s notion of “flow”. (Philippe’s a friend of mine who did research on flow and videogames. He’s the one who first introduced me to “flow”, in this sense.)

So, how did I achieve this? Well, through both plugins and theme files.

To create this database, I’ve originally been using three plugins from More Plugins: More Fields, More Taxonomies, and More Types. Had also done so in my previous attempt at a content database. At the time, these plugins helped me in several ways. But, with the current WordPress release (3.2.1), the current versions of these plugins (2.0.5.2, 1.0.1, and 1.1.1b1, respectively) are a bit buggy.

In fact, I ended up coding my custom taxonomies “from scratch”, after running into apparent problems with the More Taxonomies plugin. Eventually did the same thing with my “Syllabus” post type, replacing More Types. Wasn’t very difficult and it solved some rather tricky bugs.

Naïvely, I thought that the plugins’ export function would actually create that code, so I’d be able to put it in my own files and get rid of that plugin. But it’s not the case. Doh! Unfortunately, the support forums don’t seem so helpful either, with many questions left unanswered. So I wouldn’t really recommend these plugins apart from their pedagogical value.

The plugins were useful in helping me get around some “conceptual” issues, but it seems safer and more practical to code things from scratch, at least with taxonomies and custom post types. For “custom metaboxes”, I’m not sure I’ll have as easy a time replacing More Fields as I did replacing More Taxonomies and More Types. (More Fields helps create custom fields in the post editing interface.)

Besides the More Plugins, I’m only using two other plugins: Jonathan Christopher’s Attachments and the very versatile google doc embedder (gde) by Kevin Davis.

Attachments provides an easy way to attach files to a post and, importantly, its plugin page provides usable notes about implementation which greatly helped me in my learning process. I think I could code in some of that plugin’s functionality, now that I get a better idea of how WordPress attachments work. But it seems not to be too buggy so I’ll probably keep it.

As its name does not imply, gde can embed any file from a rather large array of file types: Adobe Reader (PDF), Microsoft Office (doc/docx, ppt/pptx/pps, xsl/xslx), and iWork Pages, along with multipage image files (tiff, Adobe Illustrator, Photoshop, SVG, EPS/PS…). The file format support comes from Google Docs Viewer (hence the plugin name).

In fact, I just realized that GDV supports zip and RAR archives. Had heard (from Gina Trapani) of that archive support in Gmail but didn’t realize it applied to GDV. Tried displaying a zip file through gde, last night, and it didn’t work. Posted something about this on the plugin’s forum and “k3davis” already fixed this, mentioning me in the 2.2 release notes.

Allowing the display of archives might be very useful, in this case. It’s fairly easily to get people to put files in a zip archive and upload it. In fact, several mail clients do all of this automatically, so there’s probably a way to get documents through emailed zip files and display the content along with the syllabus.

So, a cool plugin became cooler.

Syllabus Database (archive)

GDE Error: Unable to load requested profile.

As it so happens, gde is already installed on the academic site for which I’m building this very same syllabus database. In that case, I’ve been using gde to embed PDF files (for instance, in this page providing web enhancements page for an article in the association’s journal). So I knew it could be useful in terms of displaying course outlines and such, within individual pages of the syllabus database.

What I wasn’t sure I could do is programmatically embed files added to a syllabus page. In other words, I knew I could display these files using some shortcode on appropriate files’ URLs (including those of attached files). What I wasn’t sure how to do (and had a hard time figuring out) is how to send these URLs from a field in the database: I knew how to manually enter the code, but I didn’t know how to automatically display the results of the code when a link is entered in the right place.

The reason this matters is that I would like “normal human beings” (i.e., noncoders and, mostly, nongeeks) to enter the relevant information for their syllabi. One of WordPress’s advantages is the fact that, despite its power, it’s very easy to get nongeeks to do neat things with it. I’d like the syllabus database to be this type of neat thing.

The Attachmentsplugin helps, but still isn’t completely ideal. It does allow for drag-and-drop upload and it does provide a minimalist interface for attaching uploaded files to blogposts.

First Attach Button (Screenshot)
Screenshot of First “Attach” Button

In the first case, it’s just a matter of clicking the Attach button and dropping a file in the appropriate field. In the second case, it’s a matter of clicking another Attachbutton.

Second Attach Button (Screenshot)
Screenshot of the Second “Attach” Button

The problem is between these two Attach buttons.

File Uploaded Screenshot
Screenshot of the Uploaded File

The part of the process between uploading the file and finding the Attach button takes several nonobvious  steps. After the file has been uploaded, the most obvious buttons are Insert into Post and Save all changes, neither of which sounds particularly useful in this context. But Save all changes is the one which should be clicked.

To get to the second Attach button, I first need to go to the Media Library a second time. Recently uploaded images are showing.

Images Only Screenshot
Screenshot of the Media Library Only Showing Images

For other types of files, I then click All Types, which shows a reverse chronological list of all recently uploaded files (older files can be found through the Search Media field). I then click on the Show link associated with a given file (most likely, the most recent upload, which is the first in the list).

Second Attach Button (Screenshot)
Screenshot of the second “Attach” Button

Then, finally, the final Attach button shows up.

Clicking it, the file is attached to the current post, which was the reason behind the whole process. Thanks to both gde and Attachments, that file is then displayed along with the rest of the syllabus entry.

It only takes a matter of seconds to minutes, to attach a file (depending on filesize, connection speed, etc.). Not that long. And the media library can be very useful in many ways. But I just imagine myself explaining the process to instructors and other people submitting syllabi for inclusion the the database.

Far from ideal.

A much easier process is the one of adding files by pasting a file URL in a field. Which is exactly what I’ve added as a possibility for a syllabus’s main document (say, the PDF version of the syllabus).

Course Data Screenshot
Screenshot of the Course Data Box

Passing that URL to gde, I can automatically display the document in the document page, as I’m doing with attachments from the media library.  The problem with this, obviously, is that it requires a public URL for the document. The very same “media library” can be used to upload documents. In fact, copying the URL from an uploaded file is easier than finding the “Attach” button as explained previously. But it makes the upload a separate process on the main site. A process which can be taught fairly easily, but a process which isn’t immediately obvious.

I might make use of a DropBox account for just this kind of situation. It’s also a separate process, but it’s one which may be easier for some people.

In the end, I’ll have to see with users what makes the most sense for them.

In the past, I’ve used plugins like  Contact Form 7 (CF7), by Takayuki Miyoshi, and Fast Secure Contact Form (FSCF)  by Mike Challis to try and implement something similar. A major advantage is that they allow for submissions by users who aren’t logged in. This might be a dealmaking feature for either FSCF or CF7, as I don’t necessarily want to create accounts for everyone who might submit a syllabus. Had issues with user registration, in the past. Like attachments, onboarding remains an issue for a lot of people. Also, thanks to yet other plugins like Michael Simpson’s Contact Form to Database (CFDB), it should be possible to make form submissions into pending items in the syllabus database. I’ll be looking into this.

Another solution might be Gravity Forms. Unlike the plugins I’ve mentioned so far, it’s a commercial product. But it sounds like it might offer some rather neat features which may make syllabus submission a much more interesting process. However, it’s meant for a very different use case, which has more to do with “lead data management” and other business-focused usage. I could innovate through its use. But there might be more appropriate solutions.

As is often the case with WordPress, the “There’s a plugin for that” motto can lead to innovation.  Even documenting the process (by blogging it) can be a source of neat ideas.

A set of ideas I’ve had, for this syllabus database, came from looking into the Pods CMS Framework for WordPress. Had heard about Pods CMS through the WordCast Conversations podcast. For several reasons, it sent me on an idea spree and, for days, I was taking copious notes about what could be done. Not only about this syllabus database but about a full “learning object repository” built on top of WordPress. The reason I want to use WordPress is that, not only am I a “fanboi” of Automattic (the organization behind WordPress) but I readily plead guilty to using WordPress as a Golden Hammer. There are multiple ways to build a learning object repository. (Somehow, I’m convinced that some of my Web developing friends that Ruby on Rails is the ideal solution.) But I’ve got many of my more interesting ideas through looking into Pods CMS, a framework for WordPress and I don’t know the first thing about RoR.

Overall, Pods CMS sounds like a neat approach. Its pros and cons make it sound like an interesting alternative to WordPress’s custom post types for certain projects, as well as a significant shift from the main ways WordPress is used. During WordCamp Montreal, people I asked about it were wary of Pods. I eventually thought I would wait for version 2.0 to come out before investing significant effort in it.

In the meantime, what I’ve built is a useful base knowledge of how to use WordPress as a content database.

Can’t wait to finish adding features and fixing bugs, so I can release it to the academic organization. I’m sure they’ll enjoy it.

Even if they don’t ever use it, I’ve gained a lot of practical insight into how to do such things. It may be obvious to others but it does wonders to my satisfaction levels.

I’m truly in flow!

Scriptocentrism and the Freedom to Think

As a comment on my previous blogpost on books, a friend sent me (through Facebook) a link to a blogpost about a petition to Amazon with the following statement:

The freedom to read is tantamount to the freedom to think.

As this friend and I are both anthros+africanists, I’m reacting (perhaps a bit strongly) to that statement.

Given my perspective, I would dare say that I find this statement (brought about by DbD)… ethnocentric.

There, I said it.

And I’ll try to back it up in this blogpost in order to spark even more discussion.

We won’t exhaust this topic any time soon, but I feel there’s a lot we can do about it which has rarely been done.

I won’t use the textbook case of “Language in the Inner City,” but it could help us talk about who decides, in a given social context, what is important. We both come from a literacy-focused background, so we may have to take a step back. Not sure if Bourdieu has commented on Labov, especially in terms of what all this means for “education,” but I’d even want to bring in Ivan Illich, at some point.

Hunters with whom I’ve been working, in Mali, vary greatly in terms of literacy. Some of them have a strong university background and one can even write French legalese (he’s a judge). Others (or some of the same) have gone to Koranic school long enough that can read classical Arabic. Some have the minimal knowledge of Arabic which suffices, for them, to do divination. Many of them have a very low level of functional literacy. There’s always someone around them who can read and write, so they’re usually not out of the loop and it’s not like the social hierarchy stereotypical of the Catholic Church during the Middle Ages in Europe. It’s a very different social context which can hardly be superimposed with the history of writing and the printing press in Europe.

In terms of “freedom to thinik,” I really wouldn’t say that they’re lacking. Of course, “free thinker” has a specific meaning in liberal societies with a European background. But even this meaning can be applied to many people I’ve met in Mali.

And I go back to the social context. Those with the highest degree of functional literacy aren’t necessarily those with the highest social status. And unlike Harlem described by Labov, it’s a relatively independent context from the one in which literacy is a sine qua non. Sure, it’s a neocolonial context and Euro-Americans keep insisting that literacy in Latin script is “the most important thing ever” if they are to become a true liberal democracy. Yet, internally, it’s perfectly possible for someone to think freely, get recognition, and help other people to think without going through the written medium.

Many of those I know who have almost nonexistent skills in the written medium also have enough power (in a Weberian sense) that they get others to do the reading and writing for them. And because there are many social means to ensure that communication has worked appropriately, these “scribes” aren’t very likely to use this to take anything away from those for whom they read and write.

In Switzerland, one of my recent ancestors was functionally illiterate. Because of this, she “signed away” most of her wealth. Down the line, I’m one of her very few heirs. So, in a way, I lost part of my inheritance due to illiteracy.

Unless the switch to a European model for notarial services becomes complete, a case like this is unlikely to occur among people I know in Mali. If it does happen, it’s clearly not a failure of the oral system but a problem with this kind of transition. It’s somewhat similar to the situation with women in diverse parts of the continent during the period of direct colonialism: the fact that women have lost what powers they had (say, in a matrilineal/matrilocal society) has to do with the switch to a hierarchical system which put the emphasis on new factors which excluded the type of influence women had.

In other words, I fully understand the connections between liberalism and literacy and I’ve heard enough about the importance of the printing press and journalism in these liberal societies to understand what role reading has played in those contexts. I simply dispute the notion that these connections should be universal.

Yes, I wish the “Universal Declaration of Human Rights” (including the (in)famous Article 26, which caused so many issues) were more culturally aware.

I started reading Deschooling Society a few weeks ago. In terms of “insight density,” it’s much higher than the book which prompted this discussion. While reading the first chapter, I constructed a number of ideas which I personally find useful.

I haven’t finished reading the book. Yet. I might eventually finish it. But much of what I wanted to get from that book, I was able to get from diverse sources. Including that part of the book I did read, sequentially. But, also, everything which has been written about Illich since 1971. And I’ll be interested in reading comments by the reading group at Wikiversity.

Given my background, I have as many “things to say” about the issues surrounding schooling as what I’ve read. If I had the time, I could write as much on what I’ve read from that book and it’d probably bring me a lot of benefits.

I’ve heard enough strong reactions against this attitude I’m displaying that I can hear it, already: “how can you talk about a book you haven’t read.” And I sincerely think these people miss an important point. I wouldn’t go so far as to say that their reading habits are off (that’d be mean), especially since those are well-adapted to certain contexts, including what I call scriptocentrism. Not that these people are scriptocentric. But their attitude “goes well with” scriptocentrism.

Academia, despite being to context for an enormous amount of writing and reading, isn’t displaying that kind of scriptocentrism. Sure, a lot of what we do needs to be written (although, it’s often surprising how much insight goes unwritten in the work of many an academic). And we do get evaluated through our writing. Not to mention that we need to write in a very specific mode, which almost causes a diglossia.

But we simply don’t feel forced to “read the whole text.”

A colleague has described this as the “dirty little secret” of academia. And one which changes many things for students, to the point that it almost sounds as if it remains a secret so as to separate students into categories of “those who get it” and “the mass.”

It doesn’t take a semester to read a textbook so there are students who get the impression that they can simply read the book in a weekend and take the exams. These students may succeed, depending on the course. In fact, they may get really good grades. But they run into a wall if they want to go on with a career making any use of knowledge construction skills.

Bill Reimer has interesting documents about “better reading.” It’s a PowerPoint presentation accompanied by exercises in a PDF format. (No, I won’t discuss format here.)

I keep pointing students to those documents for a simple reason: Reimer isn’t advocating reading every word in sequence. His “skim then focus” advice might be the one piece which is harder to get through to people but it’s tremendously effective in academic contexts. It’s also one which is well-adapted to the kind of online reading I’m thinking about. And not necessarily that good for physical books. Sure, you can efficiently flip pages in a book. But skimming a text on paper is more likely to be about what stands out visually than about the structure of the text. Especially with book-length texts. The same advice holds with physical books, of course. After all, this kind of advice originally comes from that historical period which I might describe as the “heyday of books”: the late 20th Century. But I’d say that the kind of “better reading” Reimer describes is enhanced in the context of online textuality. Not just the “Read/Write Web” but Instant Messaging, email, forums, ICQ, wikis, hypertext, Gopher, even PowerPoint…

Much of this has to do with different models of human communication. The Shannon/Weaver crowd have a linear/directional model, based on information processing. Codec and modem. Something which, after Irvine’s Shadow Conversations, I tend to call “the football theory of communication.” This model might be the best-known one, especially among those who study in departments of communication along with other would-be journalists. Works well for a “broadcast” medium with mostly indirect interaction (books, television, radio, cinema, press conferences, etc.). Doesn’t work so well for the backchannel-heavy “smalltalk”  stuff of most human communication actually going on in this world.

Some cognitivists (including Chomsky) have a schema-based model. Constructivists (from Piaget on) have an elaborate model based on knowledge. Several linguistic anthropologists (including yours truly but also Judith Irvine, Richard Bauman, and Dell Hymes) have a model which gives more than lipservice to the notion of performance. And there’s a functional model of any human communication in Jakobson’s classic text on verbal communication. It’s a model which can sound as if it were linear/bidirectional but it’s much broader than this. His six “functions of verbal communication” do come from six elements of the communication process (channel, code, form, context, speaker, listener). But each of these elements embeds a complex reality and Jakobson’s model seems completely compatible with a holistic approach to human communication. In fact, Jakobson has had a tremendous impact on a large variety of people, including many key figures in linguistic anthropology along with Lévi-Strauss and, yes, even Chomsky.

(Sometimes, I wish more people knew about Jakobson. Oh, wait! Since Jakobson was living in the US, I need to americanize this statement: “Jakobson is the most underrated scholar ever.”)

All these models do (or, in my mind, should) integrate written communication. Yet scriptocentrism has often led us far away from “texts as communication” and into “text as an object.” Scriptocentrism works well with modernity. Going away from scriptocentrism is a way to accept our postmodern reality.

I Hate Books

In a way, this is a followup to a discussion happening on Facebook after something I posted (available publicly on Twitter): “(Alexandre) wishes physical books a quick and painfree death. / aime la connaissance.”

As I expected, the reactions I received were from friends who are aghast: how dare I dismiss physical books? Don’t I know no shame?

Apparently, no, not in this case.

And while I posted it as a quip, it’s the result of a rather long reflection. It’s not that I’m suddenly anti-books. It’s that I stopped buying several of the “pro-book” arguments a while ago.

Sure, sure. Books are the textbook case of technlogy which needs no improvement. eBooks can’t replace the experience of doing this or that with a book. But that’s what folkloristics defines as a functional shift. Like woven baskets which became objects of nostalgia, books are being maintained as the model for a very specific attitude toward knowledge construction based on monolithic authored texts vetted by gatekeepers and sold as access to information.

An important point, here, is that I’m not really thinking about fiction. I used to read two novel-length works a week (collections of short stories, plays…), for a period of about 10 years (ages 13 to 23). So, during that period, I probably read about 1,000 novels, ranging from Proust’s Recherche to Baricco’s Novecentoand the five books of Rabelais’s Pantagruel series. This was after having read a fair deal of adolescent and young adult fiction. By today’s standards, I might be considered fairly well-read.

My life has changed a lot, since that time. I didn’t exactly stop reading fiction but my move through graduate school eventually shifted my reading time from fiction to academic texts. And I started writing more and more, online and offline.
In the same time, the Web had also been making me shift from pointed longform texts to copious amounts of shortform text. Much more polyvocal than what Bakhtin himself would have imagined.

(I’ve also been shifting from French to English, during that time. But that’s almost another story. Or it’s another part of the story which can reamin in the backdrop without being addressed directly at this point. Ask, if you’re curious.)
The increase in my writing activity is, itself, a shift in the way I think, act, talk… and get feedback. See, the fact that I talk and write a lot, in a variety of circumstances, also means that I get a lot of people to play along. There’s still a risk of groupthink, in specific contexts, but one couldn’t say I keep getting things from the same perspective. In fact, the very Facebook conversation which sparked this blogpost is an example, as the people responding there come from relatively distant backgrounds (though there are similarities) and were not specifically queried about this. Their reactions have a very specific value, to me. Sure, it comes in the form of writing. But it’s giving me even more of something I used to find in writing: insight. The stuff you can’t get through Google.

So, back to books.

I dislike physical books. I wish I didn’t have to use them to read what I want to read. I do have a much easier time with short reading sessions on a computer screen that what would turn into rather long periods of time holding a book in my hands.

Physical books just don’t do it for me, anymore. The printing press is, like, soooo 1454!

Yes, books had “a good run.” No, nothing replaces them. That’s not the way it works. Movies didn’t replace theater, television didn’t replace radio, automobiles didn’t replace horses, photographs didn’t replace paintings, books didn’t replace orality. In fact, the technology itself doesn’t do much by itself. But social contexts recontextualize tools. If we take technology to be the set of both tools and the knowledge surrounding it, technology mostly goes through social processes, since tool repertoires and corresponding knowledge mostly shift in social contexts, not in their mere existence. Gutenberg’s Bible was a “game-changer” for social, as well as technical reasons.

And I do insist on orality. Journalists and other “communication is transmission of information” followers of Shannon&Weaver tend to portray writing as the annihilation of orality. How long after the invention of writing did Homer transfer an oral tradition to the writing media? Didn’t Albert Lord show the vitality of the epic well into the 20th Century? Isn’t a lot of our knowledge constructed through oral means? Is Internet writing that far, conceptually, from orality? Is literacy a simple on/off switch?

Not only did I maintain an interest in orality through the most book-focused moments of my life but I probably care more about orality now than I ever did. So I simply cannot accept the idea that books have simply replaced the human voice. It doesn’t add up.

My guess is that books won’t simply disappear either. There should still be a use for “coffee table books” and books as gifts or collectables. Records haven’t disappeared completely and CDs still have a few more days in dedicated stores. But, in general, we’re moving away from the “support medium” for “content” and more toward actual knowledge management in socially significant contexts.

In these contexts, books often make little sense. Reading books is passive while these contexts are about (hyper-)/(inter-)active.

Case in point (and the reason I felt compelled to post that Facebook/Twitter quip)…
I hear about a “just released” French book during a Swiss podcast. Of course, it’s taken a while to write and publish. So, by the time I heard about it, there was no way to participate in the construction of knowledge which led to it. It was already “set in stone” as an “opus.”

Looked for it at diverse bookstores. One bookstore could eventually order it. It’d take weeks and be quite costly (for something I’m mostly curious about, not depending on for something really important).

I eventually find it in the catalogue at BANQ. I reserve it. It wasn’t on the shelves, yet, so I had to wait until it was. It took from November to February. I eventually get a message that I have a couple of days to pick up my reservation but I wasn’t able to go. So it went back on the “just released” shelves. I had the full call number but books in that section aren’t in their call number sequence. I spent several minutes looking back and forth between eight shelves to eventually find out that there were four more shelves in the “humanities and social sciences” section. The book I was looking was on one of those shelves.

So, I was able to borrow it.

Phew!

In the metro, I browse through it. Given my academic reflex, I look for the back matter first. No bibliography, no index, a ToC with rather obscure titles (at random: «Taylor toujours à l’œuvre»/”Taylor still at work,” which I’m assuming to be a reference to continuing taylorism). The book is written by two separate dudes but there’s no clear indication of who wrote what. There’s a preface (by somebody else) but no “acknowledgments” section, so it’s hard to see who’s in their network. Footnotes include full URLs to rather broad sites as well as “discussion with <an author’s name>.” The back cover starts off with references to French popular culture (including something about “RER D,” which would be difficult to search). Information about both authors fits in less than 40 words (including a list of publication titles).

The book itself is fairly large print, ways almost a pound (422g, to be exact) for 327 pages (including front and back matter). Each page seems to be about 50 characters per line, about 30 lines per page. So, about half a million characters or 3500 tweets (including spaces). At 5+1 characters per word, about 80,000 words (I have a 7500-words blogpost, written in an afternoon). At about 250 words per minute, about five hours of reading. This book is listed at 19€ (about 27CAD).
There’s no direct way to do any “postprocessing” with the text: no speech synthesis for visually impaired, concordance analysis, no machine translation, even a simple search for occurences of “Sarkozy” is impossible. Not to mention sharing quotes with students or annotating in an easy-to-retrieve fashion (à la Diigo).

Like any book, it’s impossible to read in the dark and I actually have a hard time to find a spot where I can read with appropriate lighting.

Flipping through the book, I get the impression that there’s some valuable things to spark discussions, but there’s also a whole lot of redundancy with frequent discussions on the topic (the Future of Journalism, or #FoJ, as a matter of fact). My guesstimate is that, out of 5 hours of reading, I’d get at most 20 pieces of insight that I’d have exactly no way to find elsewhere. Comparable books to which I listened as audiobooks, recently, had much less. In other words, I’d have at most 20 tweets worth of things to say from the book. Almost a 200:1 compression.
Direct discussion with the authors could produce much more insight. The radio interviews with these authors already contained a few insight hints, which predisposed me to look for more. But, so many months later, without the streams of thought which animated me at the time, I end up with something much less valuable than what I wanted to get, back in November.

Bottomline: Books aren’t necessarily “broken” as a tool. They just don’t fit my life, anymore.

Installing BuddyPress on a Webhost

[Jump here for more technical details.]

A few months ago, I installed BuddyPress on my Mac to try it out. It was a bit of an involved process, so I documented it:

WordPress MU, BuddyPress, and bbPress on Local Machine « Disparate.

More recently, I decided to get a webhost. Both to run some tests and, eventually, to build something useful. BuddyPress seems like a good way to go at it, especially since it’s improved a lot, in the past several months.

In fact, the installation process is much simpler, now, and I ran into some difficulties because I was following my own instructions (though adapting the process to my webhost). So a new blogpost may be in order. My previous one was very (possibly too) detailed. This one is much simpler, technically.

One thing to make clear is that BuddyPress is a set of plugins meant for WordPress µ (“WordPress MU,” “WPMU,” “WPµ”), the multi-user version of the WordPress blogging platform. BP is meant as a way to make WPµ more “social,” with such useful features as flexible profiles, user-to-user relationships, and forums (through bbPress, yet another one of those independent projects based on WordPress).

While BuddyPress depends on WPµ and does follow a blogging logic, I’m thinking about it as a social platform. Once I build it into something practical, I’ll probably use the blogging features but, in a way, it’s more of a tool to engage people in online social activities. BuddyPress probably doesn’t work as a way to “build a community” from scratch. But I think it can be quite useful as a way to engage members of an existing community, even if this engagement follows a blogger’s version of a Pareto distribution (which, hopefully, is dissociated from elitist principles).

But I digress, of course. This blogpost is more about the practical issue of adding a BuddyPress installation to a webhost.

Webhosts have come a long way, recently. Especially in terms of shared webhosting focused on LAMP (or PHP/MySQL, more specifically) for blogs and content-management. I don’t have any data on this, but it seems to me that a lot of people these days are relying on third-party webhosts instead of relying on their own servers when they want to build on their own blogging and content-management platforms. Of course, there’s a lot more people who prefer to use preexisting blog and content-management systems. For instance, it seems that there are more bloggers on WordPress.com than on other WordPress installations. And WP.com blogs probably represent a small number of people in comparison to the number of people who visit these blogs. So, in a way, those who run their own WordPress installations are a minority in the group of active WordPress bloggers which, itself, is a minority of blog visitors. Again, let’s hope this “power distribution” not a basis for elite theory!

Yes, another digression. I did tell you to skip, if you wanted the technical details!

I became part of the “self-hosted WordPress” community through a project on which I started work during the summer. It’s a website for an academic organization and I’m acting as the organization’s “Web Guru” (no, I didn’t choose the title). The site was already based on WordPress but I was rebuilding much of it in collaboration with the then-current “Digital Content Editor.” Through this project, I got to learn a lot about WordPress, themes, PHP, CSS, etc. And it was my first experience using a cPanel- (and Fantastico-)enabled webhost (BlueHost, at the time). It’s also how I decided to install WordPress on my local machine and did some amount of work from that machine.

But the local installation wasn’t an ideal solution for two reasons: a) I had to be in front of that local machine to work on this project; and b) it was much harder to show the results to the person with whom I was collaborating.

So, in the Fall, I decided to get my own staging server. After a few quick searches, I decided HostGator, partly because it was available on a monthly basis. Since this staging server was meant as a temporary solution, HG was close to ideal. It was easy to set up as a PayPal “subscription,” wasn’t that expensive (9$/month), had adequate support, and included everything that I needed at that point to install a current version of WordPress and play with theme files (after importing content from the original site). I’m really glad I made that decision because it made a number of things easier, including working from different computers, and sending links to get feedback.

While monthly HostGator fees were reasonable, it was still a more expensive proposition than what I had in mind for a longer-term solution. So, recently, a few weeks after releasing the new version of the organization’s website, I decided to cancel my HostGator subscription. A decision I made without any regret or bad feeling. HostGator was good to me. It’s just that I didn’t have any reason to keep that account or to do anything major with the domain name I was using on HG.

Though only a few weeks elapsed since I canceled that account, I didn’t immediately set out to transition to a new webhost. I didn’t go from HostGator to another webhost.

But having my own webhost still remained at the back of my mind as something which might be useful. For instance, while not really making a staging server necessary, a new phase in the academic website project brought up a sandboxing idea. Also, I went to a “WordPress Montreal” meeting and got to think about further WordPress development/deployment, including using BuddyPress for my own needs (both as my own project and as a way to build my own knowledge of the platform) instead of it being part of an organization’s project. I was also thinking about other interesting platforms which necessitate a webhost.

(More on these other platforms at a later point in time. Bottom line is, I’m happy with the prospects.)

So I wanted a new webhost. I set out to do some comparison shopping, as I’m wont to do. In my (allegedly limited) experience, finding the ideal webhost is particularly difficult. For one thing, search results are cluttered with a variety of “unuseful” things such as rants, advertising, and limited comparisons. And it’s actually not that easy to give a new webhost a try. For one thing, these hosting companies don’t necessarily have the most liberal refund policies you could imagine. And, switching a domain name between different hosts and registrars is a complicated process through which a name may remain “hostage.” Had I realized what was involved, I might have used a domain name to which I have no attachment or actually eschewed the whole domain transition and just try the webhost without a dedicated domain name.

Doh!
Live and learn. I sure do. Loving almost every minute of it.

At any rate, I had a relatively hard time finding my webhost.

I really didn’t need “bells and whistles.” For instance, all the AdSense, shopping cart, and other business-oriented features which seem to be publicized by most webhosting companies have no interest, to me.

I didn’t even care so much about absolute degree of reliability or speed. What I’m to do with this host is fairly basic stuff. The core idea is to use my own host to bypass some limitations. For instance, WordPress.com doesn’t allow for plugins yet most of the WordPress fun has to do with plugins.

I did want an “unlimited” host, as much as possible. Not because expect to have huge resource needs but I just didn’t want to have to monitor bandwidth.

I thought that my needs would be basic enough that any cPanel-enabled webhost would fit. As much as I could see, I needed FTP access to something which had PHP 5 and MySQL 5. I expected to install things myself, without use of the webhost’s scripts but I also thought the host would have some useful scripts. Although I had already registered the domain I wanted to use (through Name.com), I thought it might be useful to have a free domain in the webhosting package. Not that domain names are expensive, it’s more of a matter of convenience in terms of payment or setup.

I ended up with FatCow. But, honestly, I’d probably go with a different host if I were to start over (which I may do with another project).

I paid 88$ for two years of “unlimited” hosting, which is quite reasonable. And, on paper, FatCow has everything I need (and I bunch of things I don’t need). The missing parts aren’t anything major but have to do with minor annoyances. In other words, no real deal-breaker, here. But there’s a few things I wish I had realized before I committed on FatCow with a domain name I actually want to use.

Something which was almost a deal-breaker for me is the fact that FatCow requires payment for any additional subdomain. And these aren’t cheap: the minimum is 5$/month for five subdomains, up to 25$/month for unlimited subdomains! Even at a “regular” price of 88$/year for the basic webhosting plan, the “unlimited subdomains” feature (included in some webhosting plans elsewhere) is more than three times more expensive than the core plan.

As I don’t absolutely need extra subdomains, this is mostly a minor irritant. But it’s one reason I’ll probably be using another webhost for other projects.

Other issues with FatCow are probably not enough to motivate a switch.

For instance, the PHP version installed on FatCow (5.2.1) is a few minor releases behind the one needed by some interesting web applications. No biggie, especially if PHP is updated in a relatively reasonable timeframe. But still makes for a slight frustration.

The MySQL version seems recent enough, but it uses non-standard tools to manage it, which makes for some confusion. Attempting to create some MySQL databases with obvious names (say “wordpress”) fails because the database allegedly exists (even though it doesn’t show up in the MySQL administration). In the same vein, the URL of the MySQL is <username>.fatcowmysql.com instead of localhost as most installers seem to expect. Easy to handle once you realize it, but it makes for some confusion.

In terms of Fantastico-like simplified installation of webapps, FatCow uses InstallCentral, which looks like it might be its own Fantastico replacement. InstallCentral is decent enough as an installation tool and FatCow does provide for some of the most popular blog and CMS platforms. But, in some cases, the application version installed by FatCow is old enough (2005!)  that it requires multiple upgrades to get to a current version. Compared to other installation tools, FatCow’s InstallCentral doesn’t seem really efficient at keeping track of installed and released versions.

Something which is partly a neat feature and partly a potential issue is the way FatCow handles Apache-related security. This isn’t something which is so clear to me, so I might be wrong.

Accounts on both BlueHost and HostGator include a public_html directory where all sorts of things go, especially if they’re related to publicly-accessible content. This directory serves as the website’s root, so one expects content to be available there. The “index.html” or “index.php” file in this directory serves as the website’s frontpage. It’s fairly obvious, but it does require that one would understand a few things about webservers. FatCow doesn’t seem to create a public_html directory in a user’s server space. Or, more accurately, it seems that the root directory (aka ‘/’) is in fact public_html. In this sense, a user doesn’t have to think about which directory to use to share things on the Web. But it also means that some higher-level directories aren’t available. I’ve already run into some issues with this and I’ll probably be looking for a workaround. I’m assuming there’s one. But it’s sometimes easier to use generally-applicable advice than to find a custom solution.

Further, in terms of access control… It seems that webapps typically make use of diverse directories and .htaccess files to manage some forms of access controls. Unix-style file permissions are also involved but the kind of access needed for a web app is somewhat different from the “User/Group/All” of Unix filesystems. AFAICT, FatCow does support those .htaccess files. But it has its own tools for building them. That can be a neat feature, as it makes it easier, for instance, to password-protect some directories. But it could also be the source of some confusion.

There are other issues I have with FatCow, but it’s probably enough for now.

So… On to the installation process… 😉

It only takes a few minutes and is rather straightforward. This is the most verbose version of that process you could imagine…

Surprised? 😎

Disclaimer: I’m mostly documenting how I did it and there are some things about which I’m unclear. So it may not work for you. If it doesn’t, I may be able to help but I provide no guarantee that I will. I’m an anthropologist, not a Web development expert.

As always, YMMV.

A few instructions here are specific to FatCow, but the general process is probably valid on other hosts.

I’m presenting things in a sequence which should make sense. I used a slightly different order myself, but I think this one should still work. (If it doesn’t, drop me a comment!)

In these instructions, straight quotes (“”) are used to isolate elements from the rest of the text. They shouldn’t be typed or pasted.

I use “example.com” to refer to the domain on which the installation is done. In my case, it’s the domain name I transfered to FatCow from another registrar but it could probably be done without a dedicated domain (in which case it would be “<username>.fatcow.com” where “<username>” is your FatCow username).

I started with creating a MySQL database for WordPress MU. FatCow does have phpMyAdmin but the default tool in the cPanel is labeled “Manage MySQL.” It’s slightly easier to use for creating new databases than phpMyAdmin because it creates the database and initial user (with confirmed password) in a single, easy-to-understand dialog box.

So I created that new database, user, and password, noting down this information. Since that password appears in clear text at some point and can easily be changed through the same interface, I used one which was easy to remember but wasn’t one I use elsewhere.
Then, I dowloaded the following files to my local machine in order to upload them to my FatCow server space. The upload can be done through either FTP or FatCow’s FileManager. I tend to prefer FTP (via CyberDuck on the Mac or FileZilla on PC). But the FileManager does allow for easy uploads.
(Wish it could be more direct, using the HTTP links directly instead of downloading to upload. But I haven’t found a way to do it through either FTP or the FileManager.)
At any rate, here are the four files I transfered to my FatCow space, using .zip when there’s a choice (the .tar.gz “tarball” versions also work but require a couple of extra steps).
  1. WordPress MU (wordpress-mu-2.9.1.1.zip, in my case)
  2. Buddymatic (buddymatic.0.9.6.3.1.zip, in my case)
  3. EarlyMorning (only one version, it seems)
  4. EarlyMorning-BP (only one version, it seems)

Only the WordPress MU archive is needed to install BuddyPress. The last three files are needed for EarlyMorning, a BuddyPress theme that I found particularly neat. It’s perfectly possible to install BuddyPress without this specific theme. (Although, doing so, you need to install a BuddyPress-compatible theme, if only by moving some folders to make the default theme available, as I explained in point 15 in that previous tutorial.) Buddymatic itself is a theme framework which includes some child themes, so you don’t need to install EarlyMorning. But installing it is easy enough that I’m adding instructions related to that theme.

These files can be uploaded anywhere in my FatCow space. I uploaded them to a kind of test/upload directory, just to make it clear, for me.

A major FatCow idiosyncrasy is its FileManager (actually called “FileManager Beta” in the documentation but showing up as “FileManager” in the cPanel). From my experience with both BlueHost and HostGator (two well-known webhosting companies), I can say that FC’s FileManager is quite limited. One thing it doesn’t do is uncompress archives. So I have to resort to the “Archive Gateway,” which is surprisingly slow and cumbersome.

At any rate, I used that Archive Gateway to uncompress the four files. WordPress µ first (in the root directory or “/”), then both Buddymatic and EarlyMorning in “/wordpress-mu/wp-content/themes” (you can chose the output directory for zip and tar files), and finally EarlyMorning-BP (anywhere, individual files are moved later). To uncompress each file, select it in the dropdown menu (it can be located in any subdirectory, Archive Gateway looks everywhere), add the output directory in the appropriate field in the case of Buddymatic or EarlyMorning, and press “Extract/Uncompress”. Wait to see a message (in green) at the top of the window saying that the file has been uncompressed successfully.

Then, in the FileManager, the contents of the EarlyMorning-BP directory have to be moved to “/wordpress-mu/wp-content/themes/earlymorning”. (Thought they could be uncompressed there directly, but it created an extra folder.) To move those files in the FileManager, I browse to that earlymorning-bp directory, click on the checkbox to select all, click on the “Move” button (fourth from right, marked with a blue folder), and add the output path: /wordpress-mu/wp-content/themes/earlymorning

These files are tweaks to make the EarlyMorning theme work with BuddyPress.

Then, I had to change two files, through the FileManager (it could also be done with an FTP client).

One change is to EarlyMorning’s style.css:

/wordpress-mu/wp-content/themes/earlymorning/style.css

There, “Template: thematic” has to be changed to “Template: buddymatic” (so, “the” should be changed to “buddy”).

That change is needed because the EarlyMorning theme is a child theme of the “Thematic” WordPress parent theme. Buddymatic is a BuddyPress-savvy version of Thematic and this changes the child-parent relation from Thematic to Buddymatic.

The other change is in the Buddymatic “extensions”:

/wordpress-mu/wp-content/themes/buddymatic/library/extensions/buddypress_extensions.php

There, on line 39, “$bp->root_domain” should be changed to “bp_root_domain()”.

This change is needed because of something I’d consider a bug but that a commenter on another blog was kind enough to troubleshoot. Without this modification, the login button in BuddyPress wasn’t working because it was going to the website’s root (example.com/wp-login.php) instead of the WPµ installation (example.com/wordpress-mu/wp-login.php). I was quite happy to find this workaround but I’m not completely clear on the reason it works.

Then, something I did which might not be needed is to rename the “wordpress-mu” directory. Without that change, the BuddyPress installation would sit at “example.com/wordpress-mu,” which seems a bit cryptic for users. In my mind, “example.com/<name>,” where “<name>” is something meaningful like “social” or “community” works well enough for my needs. Because FatCow charges for subdomains, the “<name>.example.com” option would be costly.

(Of course, WPµ and BuddyPress could be installed in the site’s root and the frontpage for “example.com” could be the BuddyPress frontpage. But since I think of BuddyPress as an add-on to a more complete site, it seems better to have it as a level lower in the site’s hierarchy.)

With all of this done, the actual WPµ installation process can begin.

The first thing is to browse to that directory in which WPµ resides, either “example.com/wordpress-mu” or “example.com/<name>” with the “<name>” you chose. You’re then presented with the WordPress µ Installation screen.

Since FatCow charges for subdomains, it’s important to choose the following option: “Sub-directories (like example.com/blog1).” It’s actually by selecting the other option that I realized that FatCow restricted subdomains.

The Database Name, username and password are the ones you created initially with Manage MySQL. If you forgot that password, you can actually change it with that same tool.

An important FatCow-specific point, here, is that “Database Host” should be “<username>.fatcowmysql.com” (where “<username>” is your FatCow username). In my experience, other webhosts use “localhost” and WPµ defaults to that.

You’re asked to give a name to your blog. In a way, though, if you think of BuddyPress as more of a platform than a blogging system, that name should be rather general. As you’re installing “WordPress Multi-User,” you’ll be able to create many blogs with more specific names, if you want. But the name you’re entering here is for BuddyPress as a whole. As with <name> in “example.com/<name>” (instead of “example.com/wordpress-mu”), it’s a matter of personal opinion.

Something I noticed with the EarlyMorning theme is that it’s a good idea to keep the main blog’s name relatively short. I used thirteen characters and it seemed to fit quite well.

Once you’re done filling in this page, WPµ is installed in a flash. You’re then presented with some information about your installation. It’s probably a good idea to note down some of that information, including the full paths to your installation and the administrator’s password.

But the first thing you should do, as soon as you log in with “admin” as username and the password provided, is probably to the change that administrator password. (In fact, it seems that a frequent advice in the WordPress community is to create a new administrator user account, with a different username than “admin,” and delete the “admin” account. Given some security issues with WordPress in the past, it seems like a good piece of advice. But I won’t describe it here. I did do it in my installation and it’s quite easy to do in WPµ.

Then, you should probably enable plugins here:

example.com/<name>/wp-admin/wpmu-options.php#menu

(From what I understand, it might be possible to install BuddyPress without enabling plugins, since you’re logged in as the administrator, but it still makes sense to enable them and it happens to be what I did.)

You can also change a few other options, but these can be set at another point.

One option which is probably useful, is this one:

Allow new registrations Disabled
Enabled. Blogs and user accounts can be created.
Only user account can be created.

Obviously, it’s not necessary. But in the interest of opening up the BuddyPress to the wider world without worrying too much about a proliferation of blogs, it might make sense. You may end up with some fake user accounts, but that shouldn’t be a difficult problem to solve.

Now comes the installation of the BuddyPress plugin itself. You can do so by going here:

example.com/<name>/wp-admin/plugin-install.php

And do a search for “BuddyPress” as a term. The plugin you want was authored by “The BuddyPress Community.” (In my case, version 1.1.3.) Click the “Install” link to bring up the installation dialog, then click “Install Now” to actually install the plugin.

Once the install is done, click the “Activate” link to complete the basic BuddyPress installation.

You now have a working installation of BuddyPress but the BuddyPress-savvy EarlyMorning isn’t enabled. So you need to go to “example.com/<name>/wp-admin/wpmu-themes.php” to enable both Buddymatic and EarlyMorning. You should then go to “example.com/<name>/wp-admin/themes.php” to activate the EarlyMorning theme.

Something which tripped me up because it’s now much easier than before is that forums (provided through bbPress) are now, literally, a one-click install. If you go here:

example.com/<name>/wp-admin/admin.php?page=bb-forums-setup

You can set up a new bbPress install (“Set up a new bbPress installation”) and everything will work wonderfully in terms of having forums fully integrated in BuddyPress. It’s so seamless that I wasn’t completely sure it had worked.

Besides this, I’d advise that you set up a few widgets for the BuddyPress frontpage. You do so through an easy-to-use drag-and-drop interface here:

example.com/<name>/wp-admin/widgets.php

I especially advise you to add the Twitter RSS widget because it seems to me to fit right in. If I’m not mistaken, the EarlyMorning theme contains specific elements to make this widget look good.

After that, you can just have fun with your new BuddyPress installation. The first thing I did was to register a new user. To do so, I logged out of my admin account,  and clicked on the Sign Up button. Since I “allow new registrations,” it’s a very simple process. In fact, this is one place where I think that BuddyPress shines. Something I didn’t explain is that you can add a series of fields for that registration and the user profile which goes with it.

The whole process really shouldn’t take very long. In fact, the longest parts have probably to do with waiting for Archive Gateway.

The rest is “merely” to get people involved in your BuddyPress installation. It can happen relatively easily, if you already have a group of people trying to do things together online. But it can be much more complicated than any software installation process… 😉

Actively Reading: "Teach Naked" sans PowerPoint

Some Diigo comments on a Chronicle piece on moving lectures out of the classroom. (Or, if you ask the piece’s author and some commenters, on PowerPoint as a source of boredom.)

I’d like to transform some of my own comments in a standalone blog entry, especially given the discussions Pamthropologist and I have been having through comments on her blog and mine. (And I just noticed Pamthropologist had written her own blogpost about this piece…) As I’m preparing for the Fall semester, I tend to think a lot about learning and teaching but I also get a bit less time.

Semi-disclaimer: John Bentley, instructional developer and programme coordinator at Concordia’s CTLS pointed me to this piece. John used to work for the Open University and the BBC. Together, John and I are currently developing a series of workshops on the use of online tools in learning and teaching. We’ve been discussing numerous dimensions of the connection between learning, teaching, and online tools. Our current focus is on creating communities of learners. One thing that I find especially neat about this collaboration is that our perspectives and spheres of expertise are quite different. Makes for interesting and thoughtful discussions.

‘Teach Naked’ Effort Strips Computers From Classrooms – Technology – The Chronicle of Higher Education

  • Not to be too snarky but… I can’t help but feel this is typical journalism. Take a complex issue, get a diverse array of comments on it, boil it down to an overly simplistic point about some polarizing question (PPT: is it evil?). Tadaa! You got an article and you’ve discouraged critical thinking.Sorry. I’m bad. I really shouldn’t go there.But I guess I’m disappointed in myself. When I first watched the video interview, I was reacting fairly strongly against Bowen. After reading (very actively!) the whole piece, I now realize that Jeff Young is the one who set the whole thing up.The problem with this is that I should know better. Right?Well, ok, I wasn’t that adamantly opposed to Bowen. I didn’t shout at my computer screen or anything. But watching the video interview again, after reading the piece, I notice that I interpret as much more open a discussion than the setup made it sound like. In other words, I went from thinking that Bowen was imposing a radical view on members of his faculty to hearing Bowen proposing ideas about ways to cope with social changes surrounding university education.The statement about most on-campus lectures being bad is rather bold, but it’s nothing we haven’t heard and it’s a reasonable comment to make in such a context. The stronger statement against PPT is actually weakened by Bowen himself in two ways: he explicitly talks about using PPT online and he frames his comment in comparison with podcasts. It then sounds like his problem isn’t with PPT itself. It’s with the use of PPT in the classroom by comparison to both podcasts and PPTs online. He may be wrong about the relative merits of podcasts, online “presentations,” and classroom lectures using PPT. But his opinion is much less radical than what I originally thought.Still, there’s room for much broader discussion of what classroom lectures and PPT presentations imply in teaching. Young’s piece and several Diigo comments on it focus on the value of PPT either in the abstract or through appropriate use. But there’s a lot more ground to cover, including such apparently simple issues as the effort needed to create compelling “presentation content” or students’ (and future employers’) expectations about PPT presentations.
  • Mr. Bowen wants to discourage professors from using PowerPoint, because they often lean on the slide-display program as a crutch rather using it as a creative tool.
    • damn you got there first! comment by dean groom
    • I think the more important point that’s being made by the article – is something that many of us in edtech world realised very quickly – that being able to teach well is a prerequisite to being able to effectively and creatively engage technology to help others learn…Powerpoint is probably the most obvious target because oif its ubiquity – but I suspect that there will also be a backlash when the masses start adopting other technologies… they’ll be misused just as effectively as PPT is.When we can assume that all university lecturers/tutors are effective teachers then the argument will be moot… until then we’ll continue to see death by powerpoint and powerpointlessness…I’m a drama teacher and love the idea of active rooms filled with proactive engaged learners… and if we have proactive engaged learners we can more effectively deploy technology in the mix…The world of teaching and learning is far from perfect and expectations seem to be geared towards a paradigm that says : “professors should tell me every last thing I need to know in order to get good grades and if students sat still and shut up long enough they might just learn something useful.”I even had one “lecturer” recently tell me “I’m a subject specialist, why do I need to know about pedagogy?” – sadly he was serious. comment by Kim FLINTOFF
    • On the subject specialist uninterested in pedagogy…It’s not an uncommon perspective, in university teaching. In fact, it might be more common among French-speakers, as most of those I’ve heard say something like this were French-speakers.I reacted quite negatively when I first heard some statement about university teachers not needing pedagogy. Don’t they care about learning?But… Isn’t there a point to be made about “non-pedagogy?”Not trying to be contrarian, here. Not playing devil’s advocate. Nor am I going on the kind of “anti-anti” PoMo mode which seems not to fit too well in English-speaking communities. I’m just thinking about teacher-less learning. And a relativist’s attitude to not judge before I know more. After all, can we safely assume that courses given by someone with such a reluctant attitude to learning pedagogy are inherently bad?There are even some people out there who take constructivism and constructionism to such an extreme that they’d say teachers aren’t needed. To an extent, the OLPC project has been going in that direction. “Students will teach themselves. We don’t need to train teachers or to engage with them in building this project.”There’s also a lot of discussion about learning outside of formal institutions. Including “on-the-job training” but also all sorts of learning strategies which don’t rely on the teacher/student (mentee, apprentice, pupil…) hierarchy. For instance, actual learning occurs in a large set of online activities. Enthusiastic people learn about things that passion them by reading about the subject, participating in online discussions, presenting their work for feedback, etc. Oftentimes, there is a hierarchy in terms of prestige, but it’s mostly negotiated through actions and not set in advance. More like “achieved status” than “ascribed status” (to use a convenient distinction from SOC101 courses). As this kind of training not infrequently leads to interesting careers, we’d be remiss to ignore the trend.Speaking of trends… It’s quite clear that many universities tend toward a more consumer-based approach. Students register and pay tuition to get “credentials” (good grades and impressive degrees). The notion that they might be there to do the actual learning is going by the wayside. In some professional contexts, people are quite explicit about how little they learnt in classrooms. It makes for difficult teaching contexts (especially at prestigious universities in the US), but it’s also something with which people learn to cope.My personal attitude is that “learning happens despite teachers.” I still think teachers make a difference, that we should learn about learners and learning, that pedagogy matters a whole lot. In fact, I’m passionate about pedagogy and I do what I can to improve my teaching.Yet the bottomline is: do people learn? If they do, does it matter what pedagogical training the teacher has? This isn’t a rhetorical question. comment by Alexandre Enkerli
  • A study published in the April issue of British Educational Research Journal
    • http://www.informaworld.com/smpp/content~db=all~content=a902053143 comment by Alexandre Enkerli
  • PowerPoint was one of the dullest methods they saw.
    • Can somebody post links to especially good PowerPoint files? comment by Bill Chapman
    • I don’t think this is really about PPT, but more about blind use of technology. It’s not the software to blame but the user.Also if you’re looking for great PPT examples, check out slideshare.net comment by Dean Shareski
    • Looking forward to reading what their criteria are for boredom.And the exact justification they give for lectures needing not to be boring.Or if they discuss the broad implications of lecturing, as opposed to the many other teaching methods that we use.Now, to be honest, I do use PPT in class. In fact, my PPT slides are the very example of what many people would consider boring: text outlines transformed into bullet points. Usually black on white, without images.But, overall, students seem to find me engaging. In student evaluations, I do get the occasional comment about the course being boring, but that’s also about the book and the nature of what we discuss.I upload these PPT files to Slideshare before going to class. In seminars, I use the PPT file to outline some topics, themes, and questions brought up by students and I upload the updated file after class.The PPT files on Slideshare are embedded into Moodle and serve as “course notes,” in conjunction with the audio recordings from the class meetings. These slides may include material which wasn’t covered in class.During “lecture,” I often spend extend periods of time discussing things with the class as a whole, leaving a slide up as a reminder of the general topic. Going from a bullet point to an extended discussion has the benefit of providing context for the discussion. When I started teaching, several students were saying that I’m “disorganized.” I still get a few comments like that but they’re much less frequent. And I still go on tangents, based on interactions with the group.Once in a while, I refrain from using PPT altogether. Which can lead to interesting challenges, in part because of student expectations and the fact that the screen becomes an indicator that “teaching is going on.”Perhaps a more important point: I try to lecture as little as possible. My upper-level courses are rapidly transformed into seminars. Even in large classes, the last class meetings of the semester involve just a few minutes of lecturing.This may all sound like a justification for my teaching method. But it’s also a reaction to the frequent discussions about PPT as evil. I do hate PPT, but I still use it.If only Google Wave could be released soon, we could use it to replace PPT. Wikis and microblogging tools are good and well, but they’re not as efficient in terms of real-time collaboration on complex material. comment by Alexandre Enkerli
  • seminars, practical sessions, and group discussions
  • In other words, tech-free classrooms were the most engaging.
    • Does it follow so directly? It’s quite easy to integrate technology with “seminars, practical sessions, and group discussions.” comment by Alexandre Enkerli
  • better than many older classroom technologies, like slate chalkboards or overhead transparencies
    • Which seems to support a form of technological determinism or, at least, a notion of a somewhat consistent improvement in the use of tools, if not in the tools themselves. comment by Alexandre Enkerli
  • But technology has hardly revolutionized the classroom experience for most college students, despite millions of dollars in investment and early predictions that going digital would force professors to rethink their lectures and would herald a pedagogical renaissance.
    • If so, then it’s only because profs aren’t bringing social technologies into their classrooms. Does the author of this article understand what’s current in ed tech? comment by Shelly Blake-Plock
    • the problem here is that in higher education, student satisfaction drives a service mentality – and students WANT summised PPTs and the want PODCASTS. Spoooon feeeeeed me – for I am paying. comment by dean groom
    • A rather broad statement which might be difficult to support with evidence.If we look at “classroom experience” in different contexts, we do notice large differences. Not necessarily in a positive sense. Technology is an integral part of all sorts of changes happening in, around, and away from the classroom.It would be quite different if that sentence said: “But institutional programs based on the adoption of specific tools in the classroom have hardly revolutionized…” It’s still early to assess the effectiveness of these programs, especially if we think about lifelong learning and about ongoing social changes related to technology use. But the statement would make more sense if it were more directly tied to specific programs instead of being a blanket critique of “technology” (left undefined). comment by Alexandre Enkerli
  • dream of shaking up college instruction
    • One of the most interesting parts of the interview with Bowen has to do with the notion that this isn’t, in fact, about following a dream. It’s about remaining relevant in a changing world. There’s a lot about Bowen’s perspective which sounds quite strange, to me. But the notion that universities should “wake up and smell the coffee” is something I wish were the object of more discussion in academic circles. comment by Alexandre Enkerli
  • Here’s the kicker, though: The biggest resistance to Mr. Bowen’s ideas has come from students, some of whom have groused about taking a more active role during those 50-minute class periods.
    • Great points, here. Let’s wish more students were involved in this conversation. It’s not just “about” them.One thing we should probably not forget about student populations is that they’re diverse. Chances are, some students in Meadows are delighted by the discussion focus. Others may be puzzled. It’s likely an adaptation for most of them. And it doesn’t sound like they were ever consulted about those changes. comment by Alexandre Enkerli
  • lecture model is pretty comfortable
    • And, though many of us are quick to criticize it, it’s difficult to avoid in the current systems of formal education in which we work. comment by Alexandre Enkerli
  • cool gadgets
    • The easiest way to dismiss the social role of technology is to call tools “gadgets.” But are these tools really just gadgets? In fact, some tools which are put to good use really aren’t that cool or even new. Are we discussing them enough? Are we aware of how they fit in the grand scheme of things?An obvious example would be cellphones. Some administrators and teachers perceive them as a nuisance. Rather few people talk about educational opportunities with cellphones, even though they already are used by people in different parts of the World to empower themselves and to learn. Negroponte has explicltly dimissed the educational potential of cellphones but the World isn’t waiting for approval from designers. comment by Alexandre Enkerli
  • seasoned performer,
    • There’s a larger point to be about performance in teaching. Including through a reference to Dick Bauman’s “Verbal Art as Performance” or other dimensions of Performance Theory.There’s also a more “mundane” point about a kind of conflict in universities between academic material and performance. In French-speaking universities, at least, it’s not uncommon to hear teachers talk about the necessity to be a “performer” as something of a distraction in teaching. Are teachers in front of the class to entertain students or is the classroom an environment in which to think and learn about difficult concepts? The consumer approach to universities, pushed in part by administrators who run universities like businesses, tends to emphasize the “entertainment paradigm,” hence the whole “boredom” issue.Having said all of this, Bowen’s own attitude goes beyond this simplistic “entertainment paradigm.” In fact, it sounds like he’s specifically not advocating for lectures to become a series of TEDtalks. Judging from the interview, it sounds like he might say that TEDtalk-style presentation should be put online and classroom-time should be devoted to analyzing those presentations.I do consider myself a performer, as I’ve been playing saxophone in a rather broad range of circumstances, from outdoor stages at festivals to concert halls. And my experience as a performer does influence the way I teach large classes. At the same time, it probably makes more salient the distinction between teaching and performing. comment by Alexandre Enkerli
  • The goateed administrator sported a suit jacket over a dark T-shirt
    • Though I’d be the first one to say that context is key, I fail to see what Bowen’s clothes contribute to the discussion. comment by Alexandre Enkerli
  • philosophical argument about the best way to engage students, he grounded it
  • information delivery common in today’s classroom lectures should be recorded and delivered to students as podcasts or online videos before class sessions
    • Fully agreed. Especially if we throw other things in the mix such as journal articles and collaboratively-created learning material. comment by Alexandre Enkerli
  • short online multiple-choice tests.
    • I don’t think he’s using the mc tests with an essessment focus rather an engagement focus – noit necessarily the most sophisticated but done playfully and creatively it can be a good first step to getting reluctatnt students to engage in first instance… comment by Kim FLINTOFF
    • I would also “defend” the use of MCTs in this context. Especially if the stakes are relatively low, the questions are well-crafted, and students do end up engaging.Like PPT, MCTs have some advantages, including because of student expectations.But, of course, it’s rather funny to hear Bowen talk about shaking things up and find out that he uses such tools. Still, the fact that these tests are online (and, one would think, taken outside of class time) goes well with Bowen’s main point about class time vs. tech-enabled work outside of class. comment by Alexandre Enkerli
  • Introduce issues of debate within the discipline and get the students to weigh in based on the knowledge they have from those lecture podcasts, Mr. Bowen says.
    • This wouldn’t be too difficult to do in social sciences and there are scenarios in which it would work wonderfully for lab sciences (if we think of “debate” as something similar to “discussion” sections in scientific articles).At the same time, some people do react negatively to such approaches based not on discipline but on “responsibilities of the university.” Some people even talk about responsibilities toward students’ parents! comment by Alexandre Enkerli
  • But if the student believes they can contribute, they’re a whole lot more motivated to enter the discourse, and to enter the discipline.
    • Sounds a bit like some of the “higher” positions in William Perry’s scheme. comment by Alexandre Enkerli
  • don’t be boring
    • Is boredom induced exclusively by the teacher? Can a student bored during a class meeting still be motivated and engaged in the material at another point? Should we apply the same principle to the readings we assign? Is there a way to efficiently assess the “boredom factor” of an academic article? How can we convince academic publishers that fighting boredom isn’t necessarily done through the addition of pretty pictures? comment by Alexandre Enkerli
  • you need a Ph.D. to figure it out
    • While I agree that these panels are difficult to use and could afford a redesign, the joke about needing a PhD sounds a bit strange in context. comment by Alexandre Enkerli
  • plug in their laptops
    • There’s something of a more general move toward getting people to use their own computers in the workplace. In fact, classroom computers are often so restricted as to be quite cumbersome to use in teaching. comment by Alexandre Enkerli
  • allow students to work in groups more easily
    • Not a bad idea. A good number of classrooms are structured in a way that makes it very hard to get students to do group work. Of course, it’s possible to do group work in any setting, but it’s remarkable how some of these seemingly trivial matters as the type of desk used can be enough to discourage some teachers from using certain teaching strategies. comment by Alexandre Enkerli
  • The classroom computers were old and needed an upgrade when Mr. Bowen arrived, so ditching them instead saved money.
    • Getting into the core of the issue. The reason it’s so important to think about “new ways” to do things isn’t necessarily that “old ways” weren’t conducive to learning. It’s because there are increased pressures on the system and some seem to perceive that cost-cutting and competition from online learning, making the issue so pressing. comment by Alexandre Enkerli
  • eliminate one staff position for a technician
    • Sounds sad, especially since support staff is already undervalued. But, at the same time, it does sound like relatively rational cost-cutting. One would just wish that they replaced that position with, say, teaching support. comment by Alexandre Enkerli
  • gave every professor a laptop
    • Again, this is a rather common practise outside of universities. Knowing how several colleagues think, this may also function as a way to “keep them happy.” comment by Alexandre Enkerli
  • support so they could create their own podcasts and videos.
    • This is where the tech support position which was cut could be useful. Recording and podcasting aren’t difficult to set up or do. But it’s an area where support can mean more than answering questions about which button to press. In fact, podcasting projects are an ideal context for collaboration between tech, teach, and research. comment by Alexandre Enkerli
  • lugging their laptops to class,
    • It can be an issue, especially if there wasn’t a choice in the type of laptop which could be used. comment by Alexandre Enkerli
  • She’s made podcasts for her course on “Critical Scholarship in Communication” that feature interviews she recorded with experts in the field.
    • One cool thing about these podcasting projects is that people can build upon them, one semester after the other. Interviews with practitioners do help provide a multiplicity of voices. And, yes, getting students to produce their own content is often a good way to go, especially if the topic is somehow related to the activity. Getting students in applied communication to create material does sound appropriate. comment by Alexandre Enkerli
  • they come in actually much more informed
    • Sounds effective. Especially since Bowen’s approach seems to be oriented toward pre-class preparation. comment by Alexandre Enkerli
  • if they had been assigned a reading.
    • There’s a lot to be said about this. One reason this method may be more efficient than reading assignments could have to do with the particularities of written language, especially the very formal style of those texts we often assign as readings. Not that students shouldn’t read, of course! But there’s a case to be made for some readings being replaced by oral sources, especially those which have to do with people’s experience in a field. Reading primary source material, integrating some reference texts, and using oral material can all be part of an appropriate set of learning strategies. comment by Alexandre Enkerli
  • created podcast lectures
    • An advantage of such “lecturecasts,” “profcasts,” and “slidecasts” is that they’re relatively easy to build and can be tightly structured. It’s not the end-all of learning material, but it’s a better substitute for classroom lectures than one might think.Still, there’s room for improvement in the technology itself. For instance, it’d be nice to have Revver-style comments in the timeline. comment by Alexandre Enkerli
  • shows movie clips from his laptop
    • This one is slightly surprising because one would expect that these clips could easily be shown, online, ahead of class. It might have to do with the chilling effect of copyright regulation or Heffernan’s strategy of getting “fresh” feedback. There would have been good questions to ask Heffernan in preparation for this piece. comment by Alexandre Enkerli
  • “Strangely enough, the people who are most resistant to this model are the students, who are used to being spoon-fed material that is going to be quote unquote on the test,” says Mr. Heffernan. “Students have been socialized to view the educational process as essentially passive. The only way we’re going to stop that is by radically refiguring the classroom in precisely the way José wants to do it.”
    • This interpretation sounds a tiny bit lopsided. After all, aren’t there students who were already quite active and engaged in the “old system” who have expressed some resistance to the “new system?” Sounds likely to me. But maybe my students are different.One fascinating thing is the level of agreement, among teachers, about the necessity to have students who aren’t passive. I certainly share this opinion but there are teachers in this World who actually do prefer students who are somewhat passive and… “obedient.” comment by Alexandre Enkerli
  • The same sequence of events
    • That part is quite significant: Bowen was already a reformer and already had gone through the process. In this case, he sounds more like one of those CEOs who are hired to save a company from a difficult situation. He originally sounded more like someone who wanted to impose specific views about teaching. comment by Alexandre Enkerli
  • ‘I paid for a college education and you’re not going to lecture?'”
    • A fairly common reaction, in certain contexts. A combination of the infamous “sense of entitlement,” the “customer-based approach to universities,” and student expectations about the way university teaching is supposed to go.One version I’ve had in student evaluations is that the student felt like s/he was hearing too much from other students instead of from me. It did pain me, because of the disconnect between what I was trying to do and that student’s notion of what university courses are supposed to bring her/him. comment by Alexandre Enkerli
  • PowerPoint lecture
    • As a commenter to my blog was saying about lectures in general, some of us (myself included) have been building a straw man. We may have negative attitudes toward certain teaching strategies, including some lecturing techniques. But that shouldn’t prevent us from discussing a wide array of teaching and learning strategies.In this case, it’s remarkable that despite the radical nature of Bowen’s reform, we learn that there are teachers who record PPT-based presentations. It then sounds like the issue isn’t so much about using PPT as it is about what is done in the classroom as opposed to what is done during the rest of the week.Boring or not, PPT lectures, even some which aren’t directly meant to engage students, can still find their place in the “teaching toolbox.” A dogmatic anti-PPT stance (such as the one displayed by this journalist) is unlikely to foster conversations about tools and learning. Based on the fact that teachers are in fact doing PPT lectures to be used outside the classroom, one ends up seeing Bowen’s perspective as much more open than that of the Chronicle’s editorial staff. comment by Alexandre Enkerli
  • Sandi Mann, the British researcher who led the recent study on student attitudes toward teaching, argues that boredom has serious implications in an educational setting.
    • Unsurprising perspective. Wonder if it had any impact on Mann’s research results. Makes the research sound more oriented than one might hope. comment by Alexandre Enkerli
  • according to some studies
  • low-cost online alternatives to the traditional campus experience
    • This could have been the core issue discussed in an article about Bowen. Especially if we are to have a thoughtful conversation about the state of higher education in a changing context. Justification for high tuition fees, the latent functions of “college life,” the likely outcome of “competing with free,” the value of the complete learning experience as opposed to the value of information transmission… comment by Alexandre Enkerli
  • give away videos
    • This is the “competing with free” part, to which record companies have been oblivious for so long but which makes OCW appear like quite a forward-looking proposition. comment by Alexandre Enkerli
  • colleges must make sure their in-person teaching really is superior to those alternatives
    • It’s both a free-market argument, which goes so well with the customer-based approach to learning, and a plea to consider learning in a broader way than the mere transmission of information from authoritative source to passive mass. An old saw, for sure, but one which surprisingly hasn’t been heard by everyone. comment by Alexandre Enkerli
  • add value
    • This might be appropriate language to convince trustees. At some institutions, this might be more important than getting students’ or teachers’ approval. comment by Alexandre Enkerli
  • not being online
    • Although, they do have an online presence. The arguments used have more to do with blended learning than with exclusively face-to-face methods. comment by Alexandre Enkerli
  • might need to stay a low-tech zone to survive.
    • Rubbish there is no reason to dumb down learning; and he obviously is not teaching 2500 students at one time. PPT is not the problem here, and this really is a collection of facile arguements that are not ironically substantiated. Lowering his overhead does not increase student learning – wheres the evidence? comment by dean groom
    • Come to think of it, it sounds like the argument was made more forcefully by Young than by Bowen himself. Bowen is certainly quite vocal but the “need… to survive” sounds a tad bit stronger than Bowen’s project.What’s funny is that the video made Bowen sound almost opinionated. The article makes Young sound like he has his own axe to grind comment by Alexandre Enkerli

Sharing Tool Wishlist

The following is an edited version of a wishlist I had been keeping on the side. The main idea is to define what would be, in my mind, the “ultimate social bookmarking system.” Which, obviously, goes way beyond social bookmarking. In a way, I even conceive of it as the ultimate tool for sharing online content. Yes, it’s that ambitious. Will it ever exist? Probably not. Should it exist? I personally think so. But I may be alone in this. Surely, you’ll tell me that I am indeed alone, which is fine. As long as you share your own wishlist items.

The trigger for my posting this is that someone contacted me, asking for what I’d like in a social bookmarking system. I find this person’s move quite remarkable, as a thoughtful strategy. Not only because this person contacted me directly (almost flattering), but because such a request reveals an approach to listening and responding to people’s needs that I find lacking in some software development circles.

This person’s message served as a prompt for my blogging this, but I’ve been meaning to blog this for a while. In fact, my guess is that I created a first version of this wishlist in 2007 after having it on my mind for a while before that. As such, it represents a type of “diachronic” or “longitudinal” view of social bookmarking and the way it works in the broader scheme of social media.

Which also means that I wrote this before I heard about Google Wave. In fact, I’m still unclear about Google Wave and I’ll need to blog about that. Not that I expect Wave to fulfill all the needs I set up for a sharing tool, but I get the impression that Google is finally putting some cards on the table.

The main part of this post is in outline form. I often think through outlines, especially with such a type of notes. I fully realize that it may not be that clear, as a structure, for other people to understand. Some of these bullet points cover a much broader issue than what they look like. But the overall idea might be fairly obvious to grasp, even if it may sound crazy to other people.

I’m posting this to the benefit of anyone who may wish to build the killer app for social media. Of course, it’s just one man’s opinion. But it’s my entitled opinion.

Concepts

What do we share online?

  • “Link”
  • “Page”
  • Identified content
  • Text
    • Narrative
    • Contact information
    • Event description
  • Contact information
  • Event invitation
  • Image
  • Recording
  • Structured content
  • Snippet
  • Access to semi-private content
  • Site’s entry point

Selective sharing

Private
  • Archiving
  • Cloud access
Individually shared
  • “Check this out”
  • Access to address book
  • Password protection
  • Specialization/expertise
  • Friendship
Group shared
  • Shared interests (SIG)
  • Collaboration (task-based)
Shared through network
  • Define identity in network
  • Semi-public
Public
  • Publishing
  • Processed
  • Reading lists

Notetaking

  • Active reading
  • Anchoring text
  • Ad hoc list of bookmarks
  • “Empty URL”
    • Create container/page
    • Personal notes

Todos

  • To read
  • To blog
  • To share
  • To update
  • Projects
    • GTD
    • Contexts
  • Add to calendar (recognized as event)

Outlining/Mindmapping

  • Manage lists of links
  • Prioritize
  • Easily group

Social aspects of sharing

  • Gift economy
  • Personal interaction
  • Trust
  • Hype
  • Value
  • Customized

Cloud computing

  • Webware
  • “Online disk”
  • Without download
  • Touch devices
  • Edit online

Personal streaming

  • Activities through pages
  • Logging
  • Flesh out personal profile

Tagging

  • “Folksonomy”
  • Enables non-hierarchical structure
  • Semantic fields
  • Related tags
  • Can include hierarchy
  • Tagclouds define concept map

Required Features

Crossplatform, crossbrowser

  • Browser-specific tools
  • Bookmarklets
  • Complete access through cloud
Keyboard shortcuts
  • Quick add (to account)
  • Vote
  • Bookmark all tabs (à la Flock)
  • Quick tags

Related pages

Recommended
  • Based on social graph
  • Based on tags
  • Based on content
  • Based on popularity
  • Pointing to this page

Quickly enter links

  • Add in place (while editing)
  • Similar to “spell as you type”
  • Incremental search
  • Add full link (title, URL, text, metadata)

Archiving

  • Prevent linkrot
  • Prepare for post-processing (offline reading, blogging…)
  • Enable bulk processing
  • Maintain version history
  • Internet Archive

Automatic processing

  • Tags
  • Summary
  • Wordcount
  • Reading time
  • Language(s)
  • Page structure analysis
  • Geotagging
  • Vote

Thread following

  • Blog comments
  • Forum comments
  • Trackbacks
  • Pings

Exporting

All
  • Archiving
  • Prepare for import
  • Maintain hierarchy
Selected
  • Tag
  • Category
  • Recently used
  • Shared
  • Site homepage
  • Blogroll
  • Blogs
Formats
  • Other services
  • HTML
  • RSS
  • OPML
  • Widget
Features
  • Comments
  • Tags
  • Statistics
  • Content

Offline processing

  • Browser-based
  • Device based
  • Offline archiving
  • Include content
  • Synchronization

Microblogging support

  • Laconi.ca/Identi.ca
  • Twitter
  • Ping.fm
  • Jaiku

Fixed/Static URL

  • Prevent linkrot
  • Maintain list for same page
  • Short URLs
  • Automatically generated
  • Expansion on mouseover
  • Statistics

Authentication

  • Use of resources
  • Identify
  • Privacy
  • Unnecessary for basic processing
  • Sticks (no need to login frequently)
  • Access to contacts and social graph
  • Multiple accounts
    • Personal/professional
    • Contexts
    • Group accounts
  • Premium accounts
    • Server space
    • Usage statistics
    • Promotion
  • Support
    • OpenID
      • As group login
    • Google Accounts
    • Facebook Connect
    • OAuth

Integration

  • Web history
  • Notebook
  • Blogging platform
  • Blog editor
  • Microblogging platform
  • Logbook
  • General purpose content editor
  • Toolbar
  • URL shortening
  • Address book
  • Social graph
  • Personal profile
  • Browser
    • Bookmarks
    • History
    • Autocomplete
  • Analytics
  • Email
  • Search
    • Online
    • Offline

Related Tools

  • Diigo
  • WebCitation
  • Ping.fm
  • BackType
  • Facebook share
  • Blog This
  • Link This
  • Share this
  • Digg
  • Plum
  • Spurl
  • CoComments
  • MyBlogLog
  • TwtVite
  • Twistory
  • Windows Live Writer
  • Magnolia
  • Stumble Upon
  • Delicious
  • Google Reader
  • Yahoo Pipes
  • Google Notebook
  • Zoho Notebook
  • Google Browser Sync
  • YouTube
  • Flock
  • Zotero

Relevant Blogposts

Social Networks and Microblogging

Microblogging (Laconica, Twitter, etc.) is still a hot topic. For instance, during the past few episodes of This Week in Tech, comments were made about the preponderance of Twitter as a discussion theme: microblogging is so prominent on that show that some people complain that there’s too much talk about Twitter. Given the centrality of Leo Laporte’s podcast in geek culture (among Anglos, at least), such comments are significant.

The context for the latest comments about TWiT coverage of Twitter had to do with Twitter’s financials: during this financial crisis, Twitter is given funding without even asking for it. While it may seem surprising at first, given the fact that Twitter hasn’t publicized a business plan and doesn’t appear to be profitable at this time, 

Along with social networking, microblogging is even discussed in mainstream media. For instance, Médialogues (a media critique on Swiss national radio) recently had a segment about both Facebook and Twitter. Just yesterday, Comedy Central’s The Daily Show with Jon Stewart made fun of compulsive twittering and mainstream media coverage of Twitter (original, Canadian access).

Clearly, microblogging is getting some mindshare.

What the future holds for microblogging is clearly uncertain. Anything can happen. My guess is that microblogging will remain important for a while (at least a few years) but that it will transform itself rather radically. Chances are that other platforms will have microblogging features (something Facebook can do with status updates and something Automattic has been trying to do with some WordPress themes). In these troubled times, Montreal startup Identi.ca received some funding to continue developing its open microblogging platform.  Jaiku, bought by Google last year, is going open source, which may be good news for microblogging in general. Twitter itself might maintain its “marketshare” or other players may take over. There’s already a large number of third-party tools and services making use of Twitter, from Mahalo Answers to Remember the Milk, Twistory to TweetDeck.

Together, these all point to the current importance of microblogging and the potential for further development in that sphere. None of this means that microblogging is “The Next Big Thing.” But it’s reasonable to expect that microblogging will continue to grow in use.

(Those who are trying to grok microblogging, Common Craft’s Twitter in Plain English video is among the best-known descriptions of Twitter and it seems like an efficient way to “get the idea.”)

One thing which is rarely mentioned about microblogging is the prominent social structure supporting it. Like “Social Networking Systems” (LinkedIn, Facebook, Ning, MySpace…), microblogging makes it possible for people to “connect” to one another (as contacts/acquaintances/friends). Like blogs, microblogging platforms make it possible to link to somebody else’s material and get notifications for some of these links (a bit like pings and trackbacks). Like blogrolls, microblogging systems allow for lists of “favourite authors.” Unlike Social Networking Systems but similar to blogrolls, microblogging allow for asymmetrical relations, unreciprocated links: if I like somebody’s microblogging updates, I can subscribe to those (by “following” that person) and publicly show my appreciation of that person’s work, regardless of whether or not this microblogger likes my own updates.

There’s something strangely powerful there because it taps the power of social networks while avoiding tricky issues of reciprocity, “confidentiality,” and “intimacy.”

From the end user’s perspective, microblogging contacts may be easier to establish than contacts through Facebook or Orkut. From a social science perspective, microblogging links seem to approximate some of the fluidity found in social networks, without adding much complexity in the description of the relationships. Subscribing to someone’s updates gives me the role of “follower” with regards to that person. Conversely, those I follow receive the role of “following” (“followee” would seem logical, given the common “-er”/”-ee” pattern). The following and follower roles are complementary but each is sufficient by itself as a useful social link.

Typically, a microblogging system like Twitter or Identi.ca qualifies two-way connections as “friendship” while one-way connections could be labelled as “fandom” (if Andrew follows Betty’s updates but Betty doesn’t follow Andrew’s, Andrew is perceived as one of Betty’s “fans”). Profiles on microblogging systems are relatively simple and public, allowing for low-involvement online “presence.” As long as updates are kept public, anybody can connect to anybody else without even needing an introduction. In fact, because microblogging systems send notifications to users when they get new followers (through email and/or SMS), subscribing to someone’s update is often akin to introducing yourself to that person. 

Reciprocating is the object of relatively intense social pressure. A microblogger whose follower:following ratio is far from 1:1 may be regarded as either a snob (follower:following much higher than 1:1) or as something of a microblogging failure (follower:following much lower than 1:1). As in any social context, perceived snobbery may be associated with sophistication but it also carries opprobrium. Perry Belcher  made a video about what he calls “Twitter Snobs” and some French bloggers have elaborated on that concept. (Some are now claiming their right to be Twitter Snobs.) Low follower:following ratios can result from breach of etiquette (for instance, ostentatious self-promotion carried beyond the accepted limit) or even non-human status (many microblogging accounts are associated to “bots” producing automated content).

The result of the pressure for reciprocation is that contacts are reciprocated regardless of personal relations.  Some users even set up ways to automatically follow everyone who follows them. Despite being tricky, these methods escape the personal connection issue. Contrary to Social Networking Systems (and despite the term “friend” used for reciprocated contacts), following someone on a microblogging service implies little in terms of friendship.

One reason I personally find this fascinating is that specifying personal connections has been an important part of the development of social networks online. For instance, long-defunct SixDegrees.com (one of the earliest Social Networking Systems to appear online) required of users that they specified the precise nature of their relationship to users with whom they were connected. Details escape me but I distinctly remember that acquaintances, colleagues, and friends were distinguished. If I remember correctly, only one such personal connection was allowed for any pair of users and this connection had to be confirmed before the two users were linked through the system. Facebook’s method to account for personal connections is somewhat more sophisticated despite the fact that all contacts are labelled as “friends” regardless of the nature of the connection. The uniform use of the term “friend” has been decried by many public commentators of Facebook (including in the United States where “friend” is often applied to any person with whom one is simply on friendly terms).

In this context, the flexibility with which microblogging contacts are made merits consideration: by allowing unidirectional contacts, microblogging platforms may have solved a tricky social network problem. And while the strength of the connection between two microbloggers is left unacknowledged, there are several methods to assess it (for instance through replies and republished updates).

Social contacts are the very basis of social media. In this case, microblogging represents a step towards both simplified and complexified social contacts.

Which leads me to the theme which prompted me to start this blogpost: event-based microblogging.

I posted the following blog entry (in French) about event-based microblogging, back in November.

Microblogue d’événement

I haven’t received any direct feedback on it and the topic seems to have little echoes in the social media sphere.

During the last PodMtl meeting on February 18, I tried to throw my event-based microblogging idea in the ring. This generated a rather lengthy between a friend and myself. (Because I don’t want to put words in this friend’s mouth, who happens to be relatively high-profile, I won’t mention this friend’s name.) This friend voiced several objections to my main idea and I got to think about this basic notion a bit further. At the risk of sounding exceedingly opinionated, I must say that my friend’s objections actually comforted me in the notion that my “event microblog” idea makes a lot of sense.

The basic idea is quite simple: microblogging instances tied to specific events. There are technical issues in terms of hosting and such but I’m mostly thinking about associating microblogs and events.

What I had in mind during the PodMtl discussion has to do with grouping features, which are often requested by Twitter users (including by Perry Belcher who called out Twitter Snobs). And while I do insist on events as a basis for those instances (like groups), some of the same logic applies to specific interests. However, given the time-sensitivity of microblogging, I still think that events are more significant in this context than interests, however defined.

In the PodMtl discussion, I frequently referred to BarCamp-like events (in part because my friend and interlocutor had participated in a number of such events). The same concept applies to any event, including one which is just unfolding (say, assassination of Guinea-Bissau’s president or bombings in Mumbai).

Microblogging users are expected to think about “hashtags,” those textual labels preceded with the ‘#’ symbol which are meant to categorize microblogging updates. But hashtags are problematic on several levels.

  • They require preliminary agreement among multiple microbloggers, a tricky proposition in any social media. “Let’s use #Bissau09. Everybody agrees with that?” It can get ugly and, even if it doesn’t, the process is awkward (especially for new users).
  • Even if agreement has been reached, there might be discrepancies in the way hashtags are typed. “Was it #TwestivalMtl or #TwestivalMontreal, I forgot.”
  • In terms of language economy, it’s unsurprising that the same hashtag would be used for different things. Is “#pcmtl” about Podcamp Montreal, about personal computers in Montreal, about PCM Transcoding Library…?
  • Hashtags are frequently misunderstood by many microbloggers. Just this week, a tweep of mine (a “peep” on Twitter) asked about them after having been on Twitter for months.
  • While there are multiple ways to track hashtags (including through SMS, in some regions), there is no way to further specify the tracked updates (for instance, by user).
  • The distinction between a hashtag and a keyword is too subtle to be really useful. Twitter Search, for instance, lumps the two together.
  • Hashtags take time to type. Even if microbloggers aren’t necessarily typing frantically, the time taken to type all those hashtags seems counterproductive and may even distract microbloggers.
  • Repetitively typing the same string is a very specific kind of task which seems to go against the microblogging ethos, if not the cognitive processes associated with microblogging.
  • The number of character in a hashtag decreases the amount of text in every update. When all you have is 140 characters at a time, the thirteen characters in “#TwestivalMtl” constitute almost 10% of your update.
  • If the same hashtag is used by a large number of people, the visual effect can be that this hashtag is actually dominating the microblogging stream. Since there currently isn’t a way to ignore updates containing a certain hashtag, this effect may even discourage people from using a microblogging service.

There are multiple solutions to these issues, of course. Some of them are surely discussed among developers of microblogging systems. And my notion of event-specific microblogs isn’t geared toward solving these issues. But I do think separate instances make more sense than hashtags, especially in terms of specific events.

My friend’s objections to my event microblogging idea had something to do with visibility. It seems that this friend wants all updates to be visible, regardless of the context. While I don’t disagree with this, I would claim that it would still be useful to “opt out” of certain discussions when people we follow are involved. If I know that Sean is participating in a PHP conference and that most of his updates will be about PHP for a period of time, I would enjoy the possibility to hide PHP-related updates for a specific period of time. The reason I talk about this specific case is simple: a friend of mine has manifested some frustration about the large number of updates made by participants in Podcamp Montreal (myself included). Partly in reaction to this, he stopped following me on Twitter and only resumed following me after Podcamp Montreal had ended. In this case, my friend could have hidden Podcamp Montreal updates and still have received other updates from the same microbloggers.

To a certain extent, event-specific instances are a bit similar to “rooms” in MMORPG and other forms of real-time many-to-many text-based communication such as the nostalgia-inducing Internet Relay Chat. Despite Dave Winer’s strong claim to the contrary (and attempt at defining microblogging away from IRC), a microblogging instance could, in fact, act as a de facto chatroom. When such a structure is needed. Taking advantage of the work done in microblogging over the past year (which seems to have advanced more rapidly than work on chatrooms has, during the past fifteen years). Instead of setting up an IRC channel, a Web-based chatroom, or even a session on MSN Messenger, users could use their microblogging platform of choice and either decide to follow all updates related to a given event or simply not “opt-out” of following those updates (depending on their preferences). Updates related to multiple events are visible simultaneously (which isn’t really the case with IRC or chatrooms) and there could be ways to make event-specific updates more prominent. In fact, there would be easy ways to keep real-time statistics of those updates and get a bird’s eye view of those conversations.

And there’s a point about event-specific microblogging which is likely to both displease “alpha geeks” and convince corporate users: updates about some events could be “protected” in the sense that they would not appear in the public stream in realtime. The simplest case for this could be a company-wide meeting during which backchannel is allowed and even expected “within the walls” of the event. The “nothing should leave this room” attitude seems contradictory to social media in general, but many cases can be made for “confidential microblogging.” Microblogged conversations can easily be archived and these archives could be made public at a later date. Event-specific microblogging allows for some control of the “permeability” of the boundaries surrounding the event. “But why would people use microblogging instead of simply talking to another?,” you ask. Several quick answers: participants aren’t in the same room, vocal communication is mostly single-channel, large groups of people are unlikely to communicate efficiently through oral means only, several things are more efficiently done through writing, written updates are easier to track and archive…

There are many other things I’d like to say about event-based microblogging but this post is already long. There’s one thing I want to explain, which connects back to the social network dimension of microblogging.

Events can be simplistically conceived as social contexts which bring people together. (Yes, duh!) Participants in a given event constitute a “community of experience” regardless of the personal connections between them. They may be strangers, ennemies, relatives, acquaintances, friends, etc. But they all share something. “Participation,” in this case, can be relatively passive and the difference between key participants (say, volunteers and lecturers in a conference) and attendees is relatively moot, at a certain level of analysis. The key, here, is the set of connections between people at the event.

These connections are a very powerful component of social networks. We typically meet people through “events,” albeit informal ones. Some events are explicitly meant to connect people who have something in common. In some circles, “networking” refers to something like this. The temporal dimension of social connections is an important one. By analogy to philosophy of language, the “first meeting” (and the set of “first impressions”) constitute the “baptism” of the personal (or social) connection. In social media especially, the nature of social connections tends to be monovalent enough that this “baptism event” gains special significance.

The online construction of social networks relies on a finite number of dimensions, including personal characteristics described in a profile, indirect connections (FOAF), shared interests, textual content, geographical location, and participation in certain activities. Depending on a variety of personal factors, people may be quite inclusive or rather exclusive, based on those dimensions. “I follow back everyone who lives in Austin” or “Only people I have met in person can belong to my inner circle.” The sophistication with which online personal connections are negotiated, along such dimensions, is a thing of beauty. In view of this sophistication, tools used in social media seem relatively crude and underdeveloped.

Going back to the (un)conference concept, the usefulness of having access to a list of all participants in a given event seems quite obvious. In an open event like BarCamp, it could greatly facilitate the event’s logistics. In a closed event with paid access, it could be linked to registration (despite geek resistance, closed events serve a purpose; one could even imagine events where attendance is free but the microblogging backchannel incurs a cost). In some events, everybody would be visible to everybody else. In others, there could be a sort of ACL for diverse types of participants. In some cases, people could be allowed to “lurk” without being seen while in others radically transparency could be enforced. For public events with all participants visible, lists of participants could be archived and used for several purposes (such as assessing which sessions in a conference are more popular or “tracking” event regulars).

One reason I keep thinking about event-specific microblogging is that I occasionally use microblogging like others use business cards. In a geek crowd, I may ask for someone’s Twitter username in order to establish a connection with that person. Typically, I will start following that person on Twitter and find opportunities to communicate with that person later on. Given the possibility for one-way relationships, it establishes a social connection without requiring personal involvement. In fact, that person may easily ignore me without the danger of a face threat.

If there were event-specific instances from microblogging platforms, we could manage connections and profiles in a more sophisticated way. For instance, someone could use a barebones profile for contacts made during an impersonal event and a full-fledged profile for contacts made during a more “intimate” event. After noticing a friend using an event-specific business card with an event-specific email address, I got to think that this event microblogging idea might serve as a way to fill a social need.

 

More than most of my other blogposts, I expect comments on this one. Objections are obviously welcomed, especially if they’re made thoughtfully (like my PodMtl friend made them). Suggestions would be especially useful. Or even questions about diverse points that I haven’t addressed (several of which I can already think about).

So…

 

What do you think of this idea of event-based microblogging? Would you use a microblogging instance linked to an event, say at an unconference? Can you think of fun features an event-based microblogging instance could have? If you think about similar ideas you’ve seen proposed online, care to share some links?

 

Thanks in advance!