Category Archives: podcasts

Wearable Hub: Getting the Ball Rolling

Statement

After years of hype, wearable devices are happening. What wearable computing lacks is a way to integrate devices into a broader system.

Disclaimer/Disclosure/Warning

  • For the past two months or so, I’ve been taking notes about this “wearable hub” idea (started around CES’s time, as wearable devices like the Pebble and Google Glass were discussed with more intensity). At this point, I have over 3000 words in notes, which probably means that I’d have enough material for a long essay. This post is just a way to release a few ideas and to “think aloud” about what wearables may mean.
  • Some of these notes have to do with the fact that I started using a few wearable devices to monitor my activities, after a health issue pushed me to start doing some exercise.
  • I’m not a technologist nor do I play one on this blog. I’m primarily an ethnographer, with diverse interests in technology and its implications for human beings. I do research on technological appropriation and some of the course I teach relate to the social dimensions of technology. Some of the approaches to technology that I discuss in those courses relate to constructionism and Actor-Network Theory.
  • I consider myself a “geek ethnographer” in the sense that I take part in geek culture (and have come out as a geek) but I’m also an outsider to geekdom.
  • Contrary to the likes of McLuhan, Carr, and Morozov, my perspective on technology and society is non-deterministic. The way I use them, “implication” and “affordance” aren’t about causal effects or, even, about direct connections. I’m not saying that society is causing technology to appear nor am I proposing a line from tools to social impacts. Technology and society are in a complex system.
  • Further, my approach isn’t predictive. I’m not saying what will happen based on technological advances nor am I saying what technology will appear. I’m thinking about the meaning of technology in an intersubjective way.
  • My personal attitude on tools and gadgets is rather ambivalent. This becomes clear as I go back and forth between techno-enthusiastic contexts (where I can almost appear like a Luddite) and techno-skeptical contexts (where some might label me as a gadget freak). I integrate a number of tools in my life but I can be quite wary about them.
  • I’m not wedded to the ideas I’m putting forth, here. They’re just broad musings of what might be. More than anything, I hope to generate thoughtful discussion. That’s why I start this post with a broad statement (not my usual style).
  • Of course, I know that other people have had similar ideas and I know that a concept of “wearable hub” already exists. It’s obvious enough that it’s one of these things which can be invented independently.

From Wearables to Hubs

Back in the 1990s, “wearable computing” became something of a futuristic buzzword, often having to do with articles of clothing. There have been many experiments and prototypes converging on an idea that we would, one day, be able to wear something resembling a full computer. Meanwhile, “personal digital assistants” became something of a niche product and embedded systems became an important dimension of car manufacturing.

Fast-forward to 2007, when a significant shift in the use of smartphones occurred. Smartphones existed before that time, but their usages, meanings, and positions in the public discourse changed quite radically around the time of the iPhone’s release. Not that the iPhone itself “caused a smartphone revolution” or that smartphone adoption suddenly reached a “tipping point”. I conceive of this shift as a complex interplay between society and tools. Not only more Kuhn than Popper, but more Latour than Kurzweil.

Smartphones, it may be argued, “happened”.

Without being described as “wearable devices”, smartphones started playing some of the functions people might have assigned to wearable devices. The move was subtle enough that Limor Fried recently described it as a realization she’s been having. Some tech enthusiasts may be designing location-aware purses and heads-up displays in the form of glasses. Smartphones are already doing a lot of the things wearables were supposed to do. Many people “wear” smartphones at most times during their waking lives and these Internet-connected devices are full of sensors. With the proliferation of cases, one might even perceive some of them as fashion accessories, like watches and sunglasses.

Where smartphones become more interesting, in terms of wearable computing, is as de facto wearable hubs.

My Wearable Devices

Which brings me to mention the four sensors I’ve been using more extensively during the past two months:

Yes, these all have to do with fitness (and there’s quite a bit of overlap between them). And, yes, I started using them a few days after the New Year. But it’s not about holiday gifts or New Year’s resolutions. I’ve had some of these devices for a while and decided to use them after consulting with a physician about hypertension. Not only have they helped me quite a bit in solving some health issues, but these devices got me to think.

(I carry several other things with me at most times. Some of my favourites include Tenqa REMXD Bluetooth headphones and the LiveScribe echo smartpen.)

One aspect is that they’re all about the so-called “quantified self”. As a qualitative researcher, I tend to be skeptical of quants. In this case, though, the stats I’m collecting about myself fit with my qualitative approach. Along with quantitative data from these devices, I’ve started collecting qualitative data about my life. The next step is to integrate all those data points automatically.

These sensors are also connected to “gamification”, a tendency I find worrisome, preferring playfulness. Though game mechanics are applied to the use of these sensors, I choose to rely on my intrinsic motivation, not paying much attention to scores and badges.

But the part which pushed me to start taking the most notes was that all these sensors connect with my iOS ()and Android) devices. And this is where the “wearable hub” comes into play. None of these devices is autonomous. They’re all part of my personal “arsenal”, the equipment I have on my me on most occasions. Though there are many similarities between them, they still serve different purposes, which are much more limited than those “wearable computers” might have been expected to serve. Without a central device serving as a type of “hub”, these sensors wouldn’t be very useful. This “hub” needs not be a smartphone, despite the fact that, by default, smartphones are taken to be the key piece in this kind of setup.

In my personal scenario, I do use a smartphone as a hub. But I also use tablets. And I could easily use an existing device of another type (say, an iPod touch), or even a new type of device meant to serve as a wearable hub. Smartphones’ “hub” affordances aren’t exclusive.

From Digital Hub to Wearable Hub

Most of the devices which would likely serve as hubs for wearable sensors can be described as “Post-PC”. They’re clearly “personal” and they’re arguably “computers”. Yet they’re significantly different from the “Personal Computers” which have been so important at the end of last century (desktop and laptop computers not used as servers, regardless of the OS they run).

Wearability is a key point, here. But it’s not just a matter of weight or form factor. A wearable hub needs to be wireless in at least two important ways: independent from a power source and connected to other devices through radio waves. The fact that they’re worn at all times also implies a certain degree of integration with other things carried throughout the day (wallets, purses, backpacks, pockets…). These devices may also be more “personal” than PCs because they may be more apparent and more amenable to customization than PCs.

Smartphones fit the bill as wearable hubs. Their form factors and battery life make them wearable enough. Bluetooth (or ANT+, Nike+, etc.) has been used to pair them wirelessly with sensors. Their connectivity to GPS and cellular networking as well as their audio and visual i/o can have interesting uses (mapping a walk, data updates during a commute, voice feedback…). And though they’re far from ubiquitous, smartphones have become quite common in key markets.

Part of the reason I keep thinking about “hubs” has to do with comments made in 2001 by then Apple CEO Steve Jobs about the “digital lifestyle” age in “PC evolution” (video of Jobs’s presentation; as an anthropologist, I’ll refrain from commenting on the evolutionary analogies):

We believe the PC, or more… importantly, the Mac can become the “digital hub” of our emerging digital lifestyle, with the ability to add tremendous value to … other digital devices.

… like camcorders, portable media players, cellphones, digital cameras, handheld organizers, etc. (Though they weren’t mentioned, other peripherals like printers and webcams also connect to PCs.)

The PC was thus going to serve as a hub, “not only adding value to these devices but interconnecting them, as well”.

At the time, key PC affordances which distinguished them from those other digital devices:

  • Big screen affording more complex user interfaces
  • Large, inexpensive hard disk storage
  • Burning DVDs and CDs
  • Internet connectivity, especially broadband
  • Running complex applications (including media processing software like the iLife suite)

Though Jobs pinpointed iLife applications as the basis for this “digital hub” vision, it sounds like FireWire was meant to be an even more important part of this vision. Of course, USB has supplanted FireWire in most use cases. It’s interesting, then, to notice that Apple only recently started shipping Macs with USB 3. In fact, DVD burning is absent from recent Macs. In 2001, the Mac might have been at the forefront of this “digital lifestyle” age. In 2013, the Mac has moved away from its role as “digital hub”.

In the meantime, the iPhone has become one of the best known examples of what I’m calling “wearable hubs”. It has a small screen and small, expensive storage (by today’s standards). It also can’t burn DVDs. But it does have nearly-ubiquitous Internet connectivity and can run fairly complex applications, some of which are adapted from the iLife suite. And though it does have wired connectivity (through Lightning or the “dock connector”), its main hub affordances have to do with Bluetooth.

It’s interesting to note that the same Steve Jobs, who used the “digital hub” concept to explain that the PC wasn’t dead in 2001, is partly responsible for popularizing the concept of “post-PC devices” six years later. One might perceive hypocrisy in this much delayed apparent flip-flop. On the other hand, Steve Jobs’s 2007 comments (video) were somewhat nuanced, as to the role of post-PC devices. What’s more interesting, though, is to think about the implications of the shift between two views of digital devices, regardless of Apple’s position through that shift.

Some post-PC devices (including the iPhone, until quite recently) do require a connection to a PC. In this sense, a smartphone might maintain its position with regards to the PC as digital hub. Yet, some of those devices are used independently of PCs, including by some people who never owned PCs.

Post-Smartphone Hubs

It’s possible to imagine a wearable hub outside of the smartphone (and tablet) paradigm. While smartphones are a convenient way to interconnect wearables, their hub-related affordances still sound limited: they lack large displays and their storage space is quite expensive. Their battery life may also be something to consider in terms of serving as hubs. Their form factors make some sense, when functioning as phones. Yet they have little to do with their use as hubs.

Part of the realization, for me, came from the fact that I’ve been using a tablet as something of an untethered hub. Since I use Bluetooth headphones, I can listen to podcasts and music while my tablet is in my backpack without being entangled in a cable. Sounds trivial but it’s one of these affordances I find quite significant. Delegating music playing functions to my tablet relates in part to battery life and use of storage. The tablet’s display has no importance in this scenario. In fact, given some communication between devices, my smartphone could serve as a display for my tablet. So could a “smartwatch” or “smartglasses”.

The Body Hub

Which led me to think about other devices which would work as wearable hubs. I originally thought about backpackable and pocketable devices.

But a friend had a more striking idea:

Under Armour’s Recharge Energy Suit may be an extreme version of this, one which would fit nicely among things Cathi Bond likes to discuss with Nora Young on The Sniffer. Nora herself has been discussing wearables on her blog as well as on her radio show. Sure, part of this concept is quite futuristic. But a sensor mesh undershirt is a neat idea for several reasons.

  • It’s easy to think of various sensors it may contain.
  • Given its surface area, it could hold enough battery power to supplement other devices.
  • It can be quite comfortable in cold weather and might even help diffuse heat in warmer climates.
  • Though wearable, it needs not be visible.
  • Thieves would probably have a hard time stealing it.
  • Vibration and haptic feedback on the body can open interesting possibilities.

Not that it’s the perfect digital hub and I’m sure there are multiple objections to a connected undershirt (including issues with radio signals). But I find the idea rather fun to think, partly because it’s so far away from the use of phones, glasses, and watches as smart devices.

Another thing I find neat, and it may partly be a coincidence, is the very notion of a “mesh”.

The Wearable Mesh

Mesh networking is a neat concept, which generates more hype than practical uses. As an alternative to WiFi access points and cellular connectivity, it’s unclear that it may “take the world by storm”. But as a way to connect personal devices, it might have some potential. After all, as Bernard Benhamou recently pointed out on France Culture’s Place de la toile, the Internet of Things may not require always-on full-bandwith connectivity. Typically, wearable sensors use fairly little bandwidth or only use it for limited amounts of time. A wearable mesh could connect wearable devices to one another while also exchanging data through the Internet itself.

Or with local devices. Smart cities, near field communication, and digital appliances occupy interesting positions among widely-discussed tendencies in the tech world. They may all have something to do with wearable devices. For instance, data exchanged between transit systems and their users could go through wearable devices. And while mobile payment systems can work through smartphones and other cellphones, wallet functions can also be fulfilled by other wearable devices.

Alternative Futures

Which might provide an appropriate segue into the ambivalence I feel toward the “wearable hub” concept I’m describing. Though I propose these ideas as if I were enthusiastic about them, they all give me pause. As a big fan of critical thinking, I like to think about “what might be” to generate questions and discussions exposing a diversity of viewpoints about the future.

Mass media discussions about these issues tend to focus on such things as privacy, availability, norms, and usefulness. Google Glass has generated quite a bit of buzz about all four. Other wearables may mainly raise issues for one or two of these broad dimensions. But the broad domain of wearable computing raises a lot more issues.

Technology enthusiasts enjoy discussing issues through the dualism between dystopia and utopia. An obvious issue with this dualism is that humans disagree about the two categories. Simply put, one person’s dystopia can be another person’s utopia, not to mention the nuanced views of people who see complex relationships between values and social change.

In such a context, a sociologist’s reflex may be to ask about the implications of these diverse values and opinions. For instance:

  • How do people construct these values?
  • Who decides which values are more important?
  • How might social groups cope with changes in values?

Discussing these issues and more, in a broad frame, might be quite useful. Some of the trickiest issues are raised after some changes in technology have already happened. From writing to cars, any technological context has unexpected implications. An ecological view of these implications could broaden the discussion.

I tend to like the concept of the “drift-off moment”, during which listeners (or readers) start thinking about the possibilities afforded a new tool (or concept). In the context of a sales pitch, the idea is that these possibilities are positive, a potential buyer is thinking about the ways she might use a newfangled device. But I also like the deeper process of thinking about all sorts of implications, regardless of their value.

So…

What might be the implications of a wearable hub?

Bean Counters and Ecologists

[So many things in my drafts, but this one should be quick.]

Recently met someone who started describing their restaurant after calling it a “café”. The “pitch” revolved around ethical practices, using local products, etc. As both a coffee geek and ethnographer, my simple question was: “Which coffee do you use?” Turns out, they’re importing coffee from a multinational corporation. “Oh, but, they’re lending us an expensive espresso machine for free! And they have fair-trade coffee!”

Luckily, we didn’t start talking about “fair trade”. And this person was willing to reflect upon the practices involved, including about the analogy with Anheuser-Busch or Coca-Cola. We didn’t get further into the deeper consequences of the resto’s actions, but the “seed” has been planted.

Sure, it’s important to focus on your financials and there’s nothing preventing a business from being both socially responsible and profitable. It just requires a shift in mindset. Small, lean, nimble businesses are more likely to do it than big, multinational corporate empires…

…which leads me to Google.

Over the years since its IPO, Google has attracted its share of praise and criticism. Like any big, multinational corporate empire. In any sector.

Within the tech sector, the Goog‘ is often compared with Microsoft, Facebook, Twitter, and Apple. All of these corporate entities have been associated in some people’s minds with some specific issue, from child labour and failure to protect users’ privacy to anticompetitive practices (the tech equivalent of free fridges and espresso machines). The issues are distinct and tech enthusiast spend a large amount of time discussing which one is worse. Meanwhile, we’re forgetting a number of larger issues.

Twitter is an interesting example, here. The service took its value from being at the centre of an ecosystem. As with any ecosystem, numerous interactions among many different members produce unexpected and often remarkable results. As the story goes, elements like hashtags and “@-replies” were invented by users and became an important part of the system. Third-party developers were instrumental in Twitter’s reach outside of its original confines. Though most of the original actors have since left the company, the ecosystem has maintained itself over the years.

When Twitter started changing the rules concerning its API, it shook the ecosystem. Sure, the ecosystem will maintain itself, in the end. But it’s nearly impossible to predict how it will change. For people at Twitter, it must have been obvious that the first changes was a warning shot to scare away those they didn’t want in their ecosystem. But, to this day, there are people who depend on Twitter, one way or another.

Google Reader offers an interesting case. The decision to kill it might have been myopic and its death might have a domino effect.

The warning shot was ambiguous, but the “writing was on the wall”. Among potential consequences of the move, the death of RSS readers was to be expected. One might also expect users of feedreaders to be displeased. In the end, the ecosystem will maintain itself.

Chances are, feedreading will be even more marginalized than it’s been and something else might replace it. Already, many people have been switching from feedreading to using Twitter as a way to gather news items.

What’s not so well-understood is the set of indirect consequences, further down the line. Again, domino effect. Some dominoes are falling in the direction of news outlets which have been slow to adapt to the ways people create and “consume” news items. Though their ad-driven models may sound similar to Google’s, and though feedreading might not be a significant source of direct revenue, the death of feedreaders may give way to the birth of new models for news production and “consumption” which might destabilize them even further. Among the things I tag as #FoJ (“Future of Journalism”) are several pieces of a big puzzle which seems misunderstood by news organizations.

There are other big dominoes which might fall from the death of Google Reader. Partly because RSS itself is part of a whole ecosystem. Dave Winer and Aaron Swartz have been major actors in the technical specifications of RSS. But Chris Lydon and people building on calendar syndication are also part of the ecosystem. In business-speak, you might call them “stakeholders”. But thinking about the ecosystem itself leads to a deeper set of thoughts, beyond the individuals involved. In the aftermath of Aaron Swartz’s premature death, it may be appropriate to point out that the ecosystem is more than the sum of its parts.

As I said on a service owned by another widely-criticized corporate empire:

Many of us keep saying that Google needs to listen to its social scientists. It also needs to understand ecology.

The Magazine and Social Media

Megaphone red
Megaphone red by Adamantios (via Wikimedia Commons, (GFDL, CC-BY-SA)

The following is my App Store review of The Magazine, a Newsstand offering by Instapaper developer Marco Arment.

Though I like Marco Arment’s work and there’s nothing specifically wrong about this implementation of the magazine model, I don’t find the magazine model particularly useful, at this point. And, make no mistake. The Magazine is indeed a magazine.

Oh, sure, this format overcomes several of the limitations set by advertising-based models and hierarchical boards. But it maintains something of the magazine logic: a tight bundle of a few articles authored by people connected through the same “editorial intent”. It’s not a conversation with the public. In this first issue, it’s not even a conversation among co-authors.

The “linked list” aspect of the “Fireball Format” (from John Gruber’s Daring Fireball media property) is described in one of the pieces in this first issue. Other distinguishing factors of the “Fireball Format” aren’t discussed in that same piece. They include a “no comment” policy which has become rather common among high-profile blogs. Unlike most blogs of the pioneer era in social media, these blogs don’t allow readers to comment directly.

A justification for this policy is that comments can be posted elsewhere. And since most of these bloggers are active on microblogging platforms like App.net and Twitter, there’s a chance that a comment might be noticed by those authors. What’s missing, though, is the sense of belonging which bloggers created among themselves before MySpace.

In other words, now that there are large social networking services online, the social aspects of blogging have been deemphasized and authorial dimensions have come to prominence. Though Arment dislikes the word, blog authors have become “brands”. It still works when these authors are in conversation with one another, when there’s a likelihood of a “followup” (FU in 5by5 parlance), when authors are responsive.

None of that interaction potential seems to be part of the core model for The Magazine. You can scream at your iOS device all you want, Jason Snell will probably not respond to you in a future edition of The Magazine. You can attempt dialogue on Twitter, but any conversation you may succeed in starting there is unlikely to have any impact on The Magazine. You’re talking with authors, now, not with members of a community.

With The Magazine, the transition from social to authorial is almost complete. Not only are posts set apart from the conversation but the editorial act of bundling posts together brings back all the problems media scholars have been pointing out for the past several decades. The issue at stake isn’t merely the move to online delivery. It’s the structure of authority and the one-to-many broadcast-style transmission. We’ve taken a step back.

So, while The Magazine has certain technical advantages over old school magazines like The Daily and Wired, it represents a step away from social media and towards mass media. Less critical thinking, more pedestals.

A new model could emerge using the infrastructure and business model that Arment built. But it’d require significant work outside of the application. The Feature might contribute something to this new model, especially if the way posts are bundled together became more flexible.

So, all in all, I consider The Magazine to be a step in the wrong direction by someone whose work I respect.

Good thing we still have podcasts.

Timeline of Apple’s Online Services

[I’d like people’s help in completing a timeline of Apple’s online services.]

[Update: deleted the WordPress shortcode, which seems not to be working.]

As kind of a followup to yesterday’s post about some early rumours and speculations about the iPhone, I thought about posting some info about Apple’s online services. Part of the reason is that Asymco’s Horace Dediu has frequently talked about what we could call “Apple’s data play”, for instance in this post about the iCloud data centre in North Carolina. I was also thinking about Mike Davidson’s comments about Apple’s presence (and “dominance”) in such diverse fields as hardware, software, licensing, and commerce. The trigger for this post, though, was from this Steve Jobs comment, which appeared in a recent NYT piece about the Apple Maps fiasco:

The MobileMe launch clearly demonstrates that we have more to learn about Internet services

The overall context for this quote as well as a number of discussions about Apple is the consensus that Apple does a poor job with online services. MobileMe and iTunes Ping are often used in these discussions and it seems clear to most people (including Apple executives and insiders, it sounds like) that the “computer company turned consumer electronics vendor” has a lot to learn about online services.

The reason I find this so interesting is that Apple seems insistent on pushing at least some of its online services. A bit less of a “betting the farm” strategy as Google’s “Emerald Sea” initiative, but an intriguing strategy for such a large and still-successful company. Dediu’s frequent reference to Clay Christensen’s concept of “Disruptive Innovation” might apply, here. Apple might be “disrupting itself into” an online services company, at least in part.

There are several things I find intriguing about this strategy.

As opposed to most other enterprises’ “online plays”, Apple’s model tends not to be based on ad revenues. The divide between Google and Apple couldn’t be stronger when we talk about ad-supported free/freemium services as opposed to paid services or services attached to other purposes. It’s likely an irreconcilable difference between fans of  both teams.

Online services are clearly not Apple’s strong suit. It often sounds like Apple is missing a “magic touch” with online services, the same way other companies are said to lack Apple’s design sense. This is more similar to Google+ given the consensus that “Google doesn’t know how to do ‘social’”. But it’s still surprising.

Though Apple may not have a “knack” for online services, it’s been trying to do it for quite a while. I keep thinking about eWorld as a precursor to the whole thing. It’s one thing for a company to try its hand at something new or to pivot into a strong business. It’s another thing entirely to shift more energies into something which has so far proven to be mostly a lost cause.

Adding to my thoughts on this was a podcast conversation (I think between John Siracusa and Dan Benjamin, though it might have been between Marco Arment and John Gruber) during which comments were made about those Apple employees working on online services.

So, basically, Apple’s online services have been on my mind. But I couldn’t find an exhaustive list. Tried Wikipedia but it doesn’t really separate online services from other things Apple does. And I ended up thinking about what would define “online services” in Apple’s case. Everything Apple does which incurs some bandwidth costs would be my working definition. Basically, it’s something to do with Apple investing in data centres and such. Some of these seem like very small costs (hosting data about podcasts, instead of the podcasts themselves, for instance). Given Apple’s size, these costs and the infrastructure behind all of this can be quite big.

So I started listing some of these services and organizing them in a sort of timeline, first in MultiMarkdown format in nvAlt, then in a Google Spreadsheet. I then discovered Vérité.CO’s Timeline.JS which takes a Google Spreadsheet and makes it into a visual timeline.

A few notes:

  • It’s a quick draft and I didn’t really check any of the data points.
  • In most cases, I only added months and, in the case of “AppleLink”, I only put years.
  • I took most dates from diverse Wikipedia pages, not necessarily backtracking on the whole process.
  • On at least one occasion, there was a discrepancy between two dates.
  • Sometimes, I took the date of the service’s announcement while I used an actual launch date for other services.
  • I only added a couple of pictures to show that it can be done. Many of the relevant pix are likely to be under copyright or to constitute a trademark.
  • I tried to be as exhaustive as I could be, but I’m sure I forgot stuff.
  • Some things may not sound like they qualify as part of “Apple’s online offering” but I think they’re still relevant. My rule of thumb is that if it goes to Apple’s servers, it’s an online service.
  • I separated some services from “suites” like iCloud or iTools, partly because some of those services haven’t been kept, which is important to see in a timeline. There are several services missing, here.
  • None of this timeline is meant to be editorial. I was just curious about what Apple has been doing online since the 1980s. The reason I care can be found in my earlier notes. I consider myself neither an “Apple fanboi” nor an “Apple hater”. I just find the situation revealing of something happening in the tech world, which has an impact on the Geek Niche.

So, here goes.

Here’s the Google Spreadsheet (editable by anyone):

Apple Online Services

Here’s the timeline through an embed code:

Here’s the embed code:

<iframe src='http://embed.verite.co/timeline/?source=0AjnWdp-FPwEKdHVqOXhWVlZuZjZYajN5QnExcExuVmc&font=Bevan-PotanoSans&maptype=toner&lang=en&hash_bookmark=true&height=650' width='100%' height='650' frameborder='0'>

 

Early iPhone Rumours

[The Lar.me/2ke link originally pointed to Mike Davidson’s 2005 piece. More explanations here.]

[Update, a bit later… Added some thoughts, links, and tags…]

While listening to the Critical Path podcast on 5by5 with Asymco’s Horace Dediu, I got stuck on Dediu’s comment that there weren’t iPhone rumours when Google acquired Android. After a quick search, I ended up on this 2005 piece by Mike Davidson (written eight months before the Google purchase), so I tweeted to @Asymco with a link to Davidson’s post. Several people, including Dediu himself, tell me that this wouldn’t qualify as a rumour (though my own definition of rumour probably differs from theirs). Still, I’ve received some comments about how insightful this piece was. It was partly based on a November 2004 piece by Russell Beattie, which was itself a partial reaction to a short Ross Mayfield post about a “WiFi iPod”. In comments on Davidson’s piece, Ste Grainer mentioned a Robert X. Cringely piece about a Mac Media Centre.

I later found a NYT piece from 2002 which contained an actual rumour about the “iPhone”, including the name:

industry analysts see evidence that Apple is contemplating what inside the company is being called an ”iPhone.”

This, I think, would qualify as a rumour in most people’s definitions, though it didn’t include “leaked prototypes”.

But back to this Davidson piece, which might have been more insightful than the NYT’s one or even Beattie’s…

In hindsight, Davidson’s piece was both prescient of what would actually happen and telling in what didn’t happen. He talked about satellite radio, Plays for Sure, and WiMAX none of which panned out as planned. Also, Davidson surmised some things about Apple’s “content play” which were both less ambitious and more impactful (on Apple’s bottomline) than what actually happened. Apple’s 2007 move against DRM might have been surprising to the 2005 Davidson. And it’s funny to think back to an era when high prices for flash storage made it prohibitive to build a mobile device… 😉

Basically, though, Davidson was speculating about an integrated device which would replace several devices at once:

It won’t be long before the cell phone is your camera, your music player, your organizer, your portable web client, your remote control, and your digital wallet

[We could argue about Android’s NFC play being closer to the digital wallet ideal than Apple’s passbook. The other parts are closer to a Treo anyway…]

In the abstract at least (and in Steve Jobs’s way of describing it), the iPhone has been this integrated communicating device about which people had been talking for years. So, kudos to Mike Davidson for predicting this a while in advance. He was neither the first nor the last, but he painted an interesting portrait.

Now, there are other parts to this story, I think. Given the fact that work on what would become iOS devices (iPad first, we’re told) hadn’t begun when Charles Wolf told the New York Times about a device called “iPhone” internally at Apple, I get the impression that the rumours predated much of the actual development work leading to the device. Speculation happened later still. It seems to relate to a number of things demonstrated by STS generally and SCOT specifically. Namely that technological development is embedded in a broader social process.

I also find interesting some side notions in all of these pieces. For instance, ideas about the impact the device might have on people’s usage. Or the fact that the move from the Treo to the iPhone ends up being quite significant, in retrospect. Even Davidson’s points about headphones and retail stores seem to relate to current things. So does the existence of the iPod touch and Apple TV in Apple’s lineup, addressing Mayfield and Cringely, respectively.

I also end up reflecting upon the shift from the “digital hub” strategy (peaking around 2007 or so) to the one revealed with iCloud, “Back to the Mac” and, yes, even Apple Maps. Dediu devotes much time to his mentor Clay Christensen’s notion of “disruptive innovation” and spent part of this latest Critcal Path episode talking about the risks behind Apple not being disruptive enough.

All of this makes me think…

Not that I have a very clear idea of what might happen but, recently, I’ve been thinking about the broader picture. Including the Maps kerfuffle. The importance of social disruption. Apple’s financial state and market presence. The so-called “Post-PC” era in relation to other “post-” notions (post-industrialism, post-colonialism, post-nationalism, post-modernism…). The boring nature of the Google/Apple conflict. The recent financial crisis. The tech world’s emphasis on Apple. The future of academia and education. The iconicity of Steve Jobs…

As Mike Wesch has been saying:

We’ll need to rethink a few things…

Ethnic Diversity and Post-Nationalism

I normally don’t enjoy Quora. But I was just asked an anonymous question there which made me react. It’s close to the kind of question I get in my intro-level courses in sociology or anthropology, so I like to “do my job” of elucidating these issues.

Here’s the question:

Can there be such a thing as too much diversity?
Up until recently the rule for all immigrants was “When in Rome do as the Romans do.” This appears to have been replaced by “We’re not going to integrate but live as we did back home.”

Is it possible that at some point diversity becomes a detriment that divides society? Just look at how segregated some cities have become

Here’s my answer:

Funnily enough, I’m preparing an exam on material where this very issue appears. Unfortunately, this material isn’t online.
One of sociology’s core perspectives, functionalism, had “extreme diversity” among the conditions under which social order breaks down. The idea, there, was that it went against society’s integration, since the model was based on well-delimited groups.
That theory has been challenged multiple times. For one thing, very few groups have been that well-integrated. The modern notion of “what The Romans were” comes from a biased view and a limited understanding of what went on at the time. In fact, an episode of the Entitled Opinions podcast contains useful discussions of the very issue.

Same thing can be said about a number of other societies, including contemporary ones.
And this is where things get interesting. We’re probably living a transition from a period marked by the Nation-State (19th and 20th Centuries) to a period marked by fluid groupings, including social networks.
In the Nation-State (contemporary Somalia and Japan, along with the fiction of 19th Century France and possibly a short period of time in Ancient Rome), ethnic homogeneity is presumed and ethnicity is managed through a very complex bureaucratic system related to citizenship. The way ethnic groups are treated then is based on what Benedict Anderson called “Imagined Communities”.
In more fluid systems, which include most of human history, diversity is taken for granted and social integration comes from other dimensions of social life.
In the current context, we have an unusual mixture of rigid Nation-State identities in parallel with the reality of transnationalism, postnationalism, Globalization, and blurred boundaries.
So, to answer the question: is it so clear what the limits of the group are? If so, what are those limits based on? If not, why would diversity be a problem?

For those interested in fluid boundaries, a classic work is Norwegian anthropologist Fredrik Barth’s “Ethnic Groups and Boundaries”.

Future of Learning Content

If indeed Apple plans to announce not just more affordable textbook options for students, but also more interactive, immersive ebook experiences…

Forecasting next week’s Apple education event (Dan Moren and Lex Friedman for Macworld)

I’m still in catchup mode (was sick during the break), but it’s hard to let this pass. It’s exactly the kind of thing I like to blog about: wishful thinking and speculation about education. Sometimes, my crazy predictions are fairly accurate. But my pleasure at blogging these things has little to do with the predictions game. I’m no prospectivist. I just like to build wishlists.

In this case, I’ll try to make it short. But I’m having drift-off moments just thinking about the possibilities. I do have a lot to say about this but we’ll see how things go.

Overall, I agree with the three main predictions in that MacWorld piece: Apple might come out with eBook creation tools, office software, and desktop reading solutions. I’m interested in all of these and have been thinking about the implications.

That MacWorld piece, like most media coverage of textbooks, these days, talks about the weight of physical textbooks as a major issue. It’s a common refrain and large bookbags/backpacks have symbolized a key problem with “education”. Moren and Friedman finish up with a zinger about lecturing. Also a common complaint. In fact, I’ve been on the record (for a while) about issues with lecturing. Which is where I think more reflection might help.

For one thing, alternative models to lecturing can imply more than a quip about the entertainment value of teaching. Inside the teaching world, there’s a lot of talk about the notion that teaching is a lot more than providing access to content. There’s a huge difference between reading a book and taking a class. But it sounds like this message isn’t heard and that there’s a lot of misunderstanding about the role of teaching.

It’s quite likely that Apple’s announcement may make things worse.

I don’t like textbooks but I do use them. I’m not the only teacher who dislikes textbook while still using them. But I feel the need to justify myself. In fact, I’ve been on the record about this. So, in that context, I think improvements in textbooks may distract us from a bigger issue and even lead us in the wrong direction. By focusing even more on content-creation, we’re commodifying education. What’s more, we’re subsuming education to a publishing model. We all know how that’s going. What’s tragic, IMHO, is that textbook publishers themselves are going in the direction of magazines! If, ten years from now, people want to know when we went wrong with textbook publishing, it’ll probably be a good idea for them to trace back from now. In theory, magazine-style textbooks may make a lot of sense to those who perceive learning to be indissociable from content consumption. I personally consider these magazine-style textbooks to be the most egregious of aberrations because, in practice, learning is radically different from content consumption.

So… If, on Thursday, Apple ends up announcing deals with textbook publishers to make it easier for them to, say, create and distribute free ad-supported magazine-style textbooks, I’ll be going through a large range of very negative emotions. Coming out of it, I might perceive a silverlining in the fact that these things can fairly easily be subverted. I like this kind of technological subversion and it makes me quite enthusiastic.

In fact, I’ve had this thought about iAd producer (Apple’s tool for creating mobile ads). Never tried it but, when I heard about it, it sounded like something which could make it easy to produce interactive content outside of mobile advertising. I don’t think the tool itself is restricted to Apple’s iAd, but I could see how the company might use the same underlying technology to create some content-creation tool.

“But,” you say, “you just said that you think learning isn’t about content.” Quite so. I’m not saying that I think these tools should be the future of learning. But creating interactive content can be part of something wider, which does relate to learning.

The point isn’t that I don’t like content. The point is that I don’t think content should be the exclusive focus of learning. To me, allowing textbook publishers to push more magazine-style content more easily is going in the wrong direction. Allowing diverse people (including learners and teachers) to easily create interactive content might in fact be a step in the right direction. It’s nothing new, but it’s an interesting path.

In fact, despite my dislike of a content emphasis in learning, I’m quite interested in “learning objects”. In fact, I did a presentation about them during the Spirit of Inquiry conference at Concordia, a few years ago (PDF).

A neat (but Flash-based) example of a learning object was introduced to me during that same conference: Mouse Party. The production value is quite high, the learning content seems relatively high, and it’s easily accessible.

But it’s based on Flash.

Which leads me to another part of the issue: formats.

I personally try to avoid Flash as much as possible. While a large number of people have done amazing things with Flash, it’s my sincere (and humble) opinion that Flash’s time has come and gone. I do agree with Steve Jobs on this. Not out of fanboism (I’m no Apple fanboi), not because I have something against Adobe (I don’t), not because I have a vested interested in an alternative technology. I just think that mobile Flash isn’t going anywhere and that. Even on the desktop, I think Flash-free is the way to go. Never installed Flash on my desktop computer, since I bought it in July. I do run Chrome for the occasional Flash-only video. But Flash isn’t the only video format out there and I almost never come across interesting content which actually relies on something exclusive to Flash. Flash-based standalone apps (like Rdio and Machinarium) are a different issue as Flash was more of a development platform for them and they’re available as Flash-free apps on Apple’s own iOS.

I wouldn’t be surprised if Apple’s announcements had something to do with a platform for interactive content as an alternative to Adobe Flash. In fact, I’d be quite enthusiastic about that. Especially given Apple’s mobile emphasis. We might be getting further in “mobile computing for the rest of us”.

Part of this may be related to HTML5. I was quite enthusiastic when Tumult released its “Hype” HTML5-creation tool. I only used it to create an HTML5 version of my playfulness talk. But I enjoyed it and can see a lot of potential.

Especially in view of interactive content. It’s an old concept and there are many tools out there to create interactive content (from Apple’s own QuickTime to Microsoft PowerPoint). But the shift to interactive content has been slower than many people (including educational technologists) would have predicted. In other words, there’s still a lot to be done with interactive content. Especially if you think about multitouch-based mobile devices.

Which eventually brings me back to learning and teaching.

I don’t “teach naked”, I do use slides in class. In fact, my slides are mostly bullet points, something presentation specialists like to deride. Thing is, though, my slides aren’t really meant for presentation and, while they sure are “content”, I don’t really use them as such. Basically, I use them as a combination of cue cards, whiteboard, and coursenotes. Though I may sound defensive about this, I’m quite comfortable with my use of slides in the classroom.

Yet, I’ve been looking intently for other solutions.

For instance, I used to create outlines in OmniOutliner that I would then send to LaTeX to produce both slides and printable outlines (as PDFs). I’ve thought about using S5, but it doesn’t really fit in my workflow. So I end up creating Keynote files on my Mac, uploading them (as PowerPoint) before class, and using them in the classroom using my iPad. Not ideal, but rather convenient.

(Interestingly enough, the main thing I need to do today is create PowerPoint slides as ancillary material for a textbook.)

In all of these cases, the result isn’t really interactive. Sure, I could add buttons and interactive content to the slides. But the basic model is linear, not interactive. The reason I don’t feel bad about it is that my teaching is very interactive (the largest proportion of classtime is devoted to open discussions, even with 100-plus students). But I still wish I could have something more appropriate.

I have used other tools, especially whiteboarding and mindmapping ones. Basically, I elicit topics and themes from students and we discuss them in a semi-structured way. But flow remains an issue, both in terms of workflow and in terms of conversation flow.

So if Apple were to come up with tools making it easy to create interactive content, I might integrate them in my classroom work. A “killer feature” here is if interaction could be recorded during class and then uploaded as an interactive podcast (à la ProfCast).

Of course, content-creation tools might make a lot of sense outside the classroom. Not only could they help distribute the results of classroom interactions but they could help in creating learning material to be used ahead of class. These could include the aforementioned learning objects (like Mouse Party) as well as interactive quizzes (like Hot Potatoes) and even interactive textbooks (like Moglue) and educational apps (plenty of these in the App Store).

Which brings me back to textbooks, the alleged focus of this education event.

One of my main issues with textbooks, including online ones, is usability. I read pretty much everything online, including all the material for my courses (on my iPad) but I find CourseSmart and its ilk to be almost completely unusable. These online textbooks are, in my experience, much worse than scanned and OCRed versions of the same texts (in part because they don’t allow for offline access but also because they make navigation much more difficult than in GoodReader).

What I envision is an improvement over PDFs.

Part of the issue has to do with PDF itself. Despite all its benefits, Adobe’s “Portable Document Format” is the relic of a bygone era. Sure, it’s ubiquitous and can preserve formatting. It’s also easy to integrate in diverse tools. In fact, if I understand things correctly, PDF replaced Display PostScript as the basis for Quartz 2D, a core part of Mac OS X’s graphics rendering. But it doesn’t mean that it can’t be supplemented by something else.

Part of the improvement has to do with flexibility. Because of its emphasis on preserving print layouts, PDF tends to enforce print-based ideas. This is where EPUB is at a significant advantage. In a way, EPUB textbooks might be the first step away from the printed model.

From what I can gather, EPUB files are a bit like Web archives. Unlike PDFs, they can be reformatted at will, just like webpages can. In fact, iBooks and other EPUB readers (including Adobe’s, IIRC) allow for on-the-fly reformatting, which puts the reader in control of a much greater part of the reading experience. This is exactly the kind of thing publishers fail to grasp: readers, consumers, and users want more control on the experience. EPUB textbooks would thus be easier to read than PDFs.

EPUB is the basis for Apple’s iBooks and iBookstore and people seem to be assuming that Thursday’s announcement will be about iBooks. Makes sense and it’d be nice to see an improvement over iBooks. For one thing, it could support EPUB 3. There are conversion tools but, AFAICT, iBooks is stuck with EPUB 2.0. An advantage there is that EPUBs can possibly include scripts and interactivity. Which could make things quite interesting.

Interactive formats abound. In fact, PDFs can include some interactivity. But, as mentioned earlier, there’s a lot of room for improvement in interactive content. In part, creation tools could be “democratized”.

Which gets me thinking about recent discussions over the fate of HyperCard. While I understand John Gruber’s longstanding position, I find room for HyperCard-like tools. Like some others, I even had some hopes for ATX-based TileStack (an attempt to bring HyperCard stacks back to life, online). And I could see some HyperCard thinking in an alternative to both Flash and PDF.

“Huh?”, you ask?

Well, yes. It may sound strange but there’s something about HyperCard which could make sense in the longer term. Especially if we get away from the print model behind PDFs and the interaction model behind Flash. And learning objects might be the ideal context for this.

Part of this is about hyperlinking.  It’s no secret that HyperCard was among HTML precursors. As the part of HTML which we just take for granted, hyperlinking is among the most undervalued features of online content. Sure, we understand the value of sharing links on social networking systems. And there’s a lot to be said about bookmarking. In fact, I’ve been thinking about social bookmarking and I have a wishlist about sharing tools, somewhere. But I’m thinking about something much more basic: hyperlinking is one of the major differences between online and offline wriiting.

Think about the differences between, say, a Wikibook and a printed textbook. My guess is that most people would focus on the writing style, tone, copy-editing, breadth, reviewing process, etc. All of these are relevant. In fact, my sociology classes came up with variations on these as disadvantages of the Wikibook over printed textbooks. Prior to classroom discussion about these differences, however, I mentioned several advantages of the Wikibook:

  • Cover bases
  • Straightforward
  • Open Access
  • Editable
  • Linked

(Strangely enough, embedded content from iWork.com isn’t available and I can’t log into my iWork.com account. Maybe it has to do with Thursday’s announcement?)

That list of advantages is one I’ve been using since I started to use this Wikibook… excerpt for the last one. And this is one which hit me, recently, as being more important than the others.

So, in class, I talked about the value of links and it’s been on my mind quite a bit. Especially in view of textbooks. And critical thinking.

See, academic (and semi-academic) writing is based on references, citations, quotes. English-speaking academics are likely to be the people in the world of publishing who cite the most profusely. It’s not rare for a single paragraph of academic writing in English to contain ten citations or more, often stringed in parentheses (Smith 1999, 2005a, 2005b; Smith and Wesson 1943, 2010). And I’m not talking about Proust-style paragraphs either. I’m convinced that, with some quick searches, I could come up with a paragraph of academic writing which has less “narrative content” than citation.

Textbooks aren’t the most egregious example of what I’d consider over-citing. But they do rely on citations quite a bit. As I work more specifically on textbook content, I notice even more clearly the importance of citations. In fact, in my head, I started distinguishing some patterns in textbook content. For instance, there are sections which mostly contain direct explanations of key concepts while other sections focus on personal anecdotes from the authors or extended quotes from two sides of the debate. But one of the most obvious sections are summaries from key texts.

For instance (hypothetical example):

As Nora Smith explained in her 1968 study Coming Up with Something to Say, the concept of interpretation has a basis in cognition.

Smith (1968: 23) argued that Pierce’s interpretant had nothing to do with theatre.

These citations are less conspicuous than they’d be in peer-reviewed journals. But they’re a central part of textbook writing. One of their functions should be to allow readers (undergraduate students, mostly) to learn more about a topic. So, when a student wants to know more about Nora Smith’s reading of Pierce, she “just” have to locate Smith’s book, go to the right page, scan the text for the read for the name “Pierce”, and read the relevant paragraph. Nothing to it.

Compare this to, say, a blogpost. I only cite one text, here. But it’s linked instead of being merely cited. So readers can quickly know more about the context for what I’m discussing before going to the library.

Better yet, this other blogpost of mine is typical of what I’ve been calling a linkfest, a post containing a large number of links. Had I put citations instead of links, the “narrative” content of this post would be much less than the citations. Basically, the content was a list of contextualized links. Much textbook content is just like that.

In my experience, online textbooks are citation-heavy and take almost no benefit from linking. Oh, sure, some publisher may replace citations with links. But the result would still not be the same as writing meant for online reading because ex post facto link additions are quite different from link-enhanced writing. I’m not talking about technological determinism, here. I’m talking about appropriate tool use. Online texts can be quite different from printed ones and writing for an online context could benefit greatly from this difference.

In other words, I care less about what tools publishers are likely to use to create online textbooks than about a shift in the practice of online textbooks.

So, if Apple comes out with content-creation tools on Thursday (which sounds likely), here are some of my wishes:

  • Use of open standards like HTML5 and EPUB (possibly a combination of the two).
  • Completely cross-platform (should go without saying, but Apple’s track record isn’t that great, here).
  • Open Access.
  • Link library.
  • Voice support.
  • Mobile creation tools as powerful as desktop ones (more like GarageBand than like iWork).
  • HyperCard-style emphasis on hyperlinked structures (à la “mini-site” instead of web archives).
  • Focus on rich interaction (possibly based on the SproutCore web framework).
  • Replacement for iWeb (which is being killed along with MobileMe).
  • Ease creation of lecturecasts.
  • Deep integration with iTunes U.
  • Combination of document (à la Pages or Word), presentation (à la Keynote or PowerPoint), and standalone apps (à la The Elements or even Myst).
  • Full support for course management systems.
  • Integration of textbook material and ancillary material (including study guides, instructor manuals, testbanks, presentation files, interactive quizzes, glossaries, lesson plans, coursenotes, etc.).
  • Outlining support (more like OmniOutliner or even like OneNote than like Keynote or Pages).
  • Mindmapping support (unlikely, but would be cool).
  • Whiteboard support (both in-class and online).
  • Collaboration features (à la Adobe Connect).
  • Support for iCloud (almost a given, but it opens up interesting possibilities).
  • iWork integration (sounds likely, but still in my wishlist).
  • Embeddable content (à la iWork.com).
  • Stability, ease of use, and low-cost (i.e., not Adobe Flash or Acrobat).
  • Better support than Apple currently provides for podcast production and publishing.
  • More publisher support than for iBooks.
  • Geared toward normal users, including learners and educators.

The last three are probably where the problem lies. It’s likely that Apple has courted textbook publishers and may have convinced them that they should up their game with online textbooks. It’s clear to me that publishers risk to fall into oblivion if they don’t wake up to the potential of learning content. But I sure hope the announcement goes beyond an agreement with publishers.

Rumour has it that part of the announcement might have to do with bypassing state certification processes, in the US. That would be a big headline-grabber because the issue of state certification is something of wedge issue. Could be interesting, especially if it means free textbooks (though I sure hope they won’t be ad-supported). But that’s much less interesting than what could be done with learning content.

User-generated content” may be one of the core improvements in recent computing history, much of which is relevant for teaching. As fellow anthro Mike Wesch has said:

We’ll  need to rethink a few things…

And Wesch sure has been thinking about learning.

Problem is, publishers and “user-generated content” don’t go well together. I’m guessing that it’s part of the reason for Apple’s insufficient support for “user-generated content”. For better or worse, Apple primarily perceives its users as consumers. In some cases, Apple sides with consumers to make publishers change their tune. In other cases, it seems to be conspiring with publishers against consumers. But in most cases, Apple fails to see its core users as content producers. In the “collective mind of Apple”, the “quality content” that people should care about is produced by professionals. What normal users do isn’t really “content”. iTunes U isn’t an exception, those of us who give lectures aren’t Apple’s core users (even though the education market as a whole has traditionally being an important part of Apple’s business). The fact that Apple courts us underlines the notion that we, teachers and publishers (i.e. non-students), are the ones creating the content. In other words, Apple supports the old model of publishing along with the old model of education. Of course, they’re far from alone in this obsolete mindframe. But they happen to have several of the tools which could be useful in rethinking education.

Thursday’s events is likely to focus on textbooks. But much more is needed to shift the balance between publishers and learners. Including a major evolution in podcasting.

Podcasting is especially relevant, here. I’ve often thought about what Apple could do to enhance podcasting for learning. Way beyond iTunes U. Into something much more interactive. And I don’t just mean “interactive content” which can be manipulated seamless using multitouch gestures. I’m thinking about the back-and-forth of learning and teaching, the conversational model of interactivity which clearly distinguishes courses from mere content.

iCloud Dreams

Got lots more to blog, including something about “received knowledge”. And a list of things I love about Google. (I’m also getting started on “logical punctuation”, as you may already be noticing…)

But, at the risk of attracting trolls and Apple haters, I thought I’d post some notes from a daydreaming session. In some ways, it’s easier to write than the rest. And it’s more “time-sensitive”, in that my thoughts will likely sound very silly, very soon.

But I don’t care.

So, yes, this post is about iCloud, which will be officially unveiled in a few hours. No, it doesn’t mean that I expect anything specific from iCloud or that I trust Apple to deliver something awesome.

Contrary to what some people seem to think, I’m no Apple fanboi. I use a number of Apple products and I find several of them to be close to the ideal in my workflow, but I don’t have any sort of deep involvement in “the Cult of Mac”, Apple Inc., AAPL, or even Apple-focused development. I use the tools and like them, but I don’t think Apple will save us any more than will Facebook, Dell, Google, Amazon, Twitter, HP, or Microsoft.

[Automattic, on the other hand… 😉 ]

So, back to iCloud…

According to many, “cloud computing” (whatever that means) is a domain in which Apple has been relatively weak. I tend to share that opinion, despite the fact that a number of tools that I use have to do with either “the cloud”, Apple, or both. What might give trolls and haters some ammo is that I do have a MobileMe subscription. But there’s a lot I dislike about it and the only features I really find valuable are “over-the-air” syncing (henceforth “OTA”) and “Find My iPhone”. And since I use GSync on my iPod touch, MobileMe’s OTA isn’t that incredibly important. Depending on what iCloud may be, my MobileMe renewal (which comes up in a few days) could be a very hard sell. I don’t regret having it as it did help me retrieve my iPad. But it’s rather expensive if it’s the only thing it does. (Then again, so is insurance of any kind, but I digress…)

So, I’m no MobileMe poweruser. Why would I care about iCloud?

In some ways, I don’t. Or, at least, I didn’t. Until very recently, though I saw rumours about Apple’s new “cloud services”, I was only vaguely intrigued about it. I did think that it might solve my MobileMe issue. But I treated these rumours with a lot of skepticism and a rather low level of interest.

Yet, today, iCloud has been giving me a drift-off moment. Like Android did, at some point.

It’s not that I have predictions to make about iCloud. I’m not even speculating, really. But it got me to think. And, I admit, I enjoy thinking.

Without further ado (about nothing), my fanciful thoughts stemming from a short daydreaming session about iCloud…

The main thing people seem to be expecting  (based on rumoured negotiations with music publishers) is a music streaming service similar to Music Beta by Google or a digital file storage service similar to Amazon Cloud Drive. Both of these are quite neat and I could see myself using something like this. But it’s not exactly what makes me dream. While iTunes integration might make Apple’s version of a music streaming service somewhat more useful than the others. Besides, rumours have it that, through agreements with the recording industry, iCloud might sync music without requiring long uploads. It’s quite possible that this only works with tracks purchased on iTunes, which would upset those whose expectations are high, but could already be useful to some.

Where I’m beginning to drift off, though, is when I start thinking about OTA for podcasts. It’s been high up on my wishlist, as a feature, and you might say that it’s a pet peeve with iOS devices for podcatching. Having to sync my iPod touch to my main desktop just to have my podcast list up-to-date is a major hassle. Sure, there are apps which sync podcasts OTA. Problem is, they can’t add podcasts to the native iOS media player, which is a dealbreaker in my case. (As absurd as it may sound to others, one reason this is a dealbreaker is that I now listen to everything at doublespeed. Hey, it’s my podcast library and I listen to it as I want, ok?)

So, OTA podcasts would constitute a significant enhancement to my experience. Nothing absolutely required and possibly not that significant for others, but it’d really help me in more ways than one could imagine.

Thing is, syncing my iPod touch isn’t just about podcasts, even though podcatching is my main motivation to sync. After all, I don’t listen to podcasts yet I still sync my iPad. So, what else? Well, backing up is the main other thing, and it might be one of the core reason for Apple’s implicit insistence on syncing. That’d be classic Apple. Data loss can be such a big problem that they’d “do what they can” to prevent users from losing data. Far from perfect, in my experience (I ended up having some problems when I lost my “iTunes Library” file). And quite annoying when it meant that the sync would take a very long time to finish at precisely the point when I’m trying to leave home. But a classic Apple move, even in the way Apple haters may mean it.

So OTA synchronization of the whole iOS device, and not just podcasts or music, would be a definite plus, in this perspective. If it does end up coming with iCloud, it’d provide support to the idea that the tethering of iOS devices to desktop computers is really about ensuring that users back up their devices…

…and stay up to date. Firmware updates aren’t that frequent, but they’re probably a major part of the equation for Apple.

But not so much for me. If OTA podcasts were available, I’d still sync my iOS devices on occasion, through whatever means necessary. In fact, were I to use an Android device, a backup app would be essential, to me. So still not much dreaming from the backup aspect of iCloud.

Although… Sync is much broader than preventing device-specific data loss and making sure your device has the latest firmware.

For one thing, it does encompass some of the aforementioned OTA functionalities in MobileMe. Useful, but still not dreamworthy.

We get a bit closer to a “dream come true” if we talk about Xmarks, a bookmark-sync service originally meant for Firefox.  Sure, it sounds incredibly prosaic. But OTA bookmarks would open up a wide range of possibilities. This is about a qualitative difference from going OTA. In the case of backups, it’s about avoiding an annoyance but, arguably, it’s not really about changing something major about our behaviour. (Then again, maybe it is, with people who don’t back their devices up.) Point is, with something as simple as bookmarks, OTA is “disruptive”. At least, it gets me to daydream. One reason is that:

…no matter how fundamental they have been for the Web, links and bookmarks have yet to find their full value.

Hmm… Ok, perhaps a bit hyperbolic… So let me rephrase…

There’s still a lot to be done with URLs and, as simple as they are, I love thinking about links. Maybe I’m just obsessed with URLs.

As it so happens, I have a full list of thoughts about “link processing” and I’ve already blogged about related topics (on more than one occasion, in different contexts, going back to relatively early blogposts). And I even think social science can help.

I mean, think about it! There’s so much you can do, with links! Much of it is obvious, but I’d argue, rarely discussed. For instance, it’s very clear that we can post links pretty much anywhere. Doing so, we’re sharing their “content”. (In a semiotic sense, links are indices. I wish we can move from the “semantic Web” to the “semiotic Web”. But that’s another issue.) Sharing a link is the basic act of the social Web. It’s so obvious and frequent that it seems not to require discussion”.

Another obvious thing about links: we can measure the number of times they’re followed. In 2011, more than thirty years after hypertext has been introduced as a stable concept, much of the Web’s finances still relies on “clickthroughs”. Seems important.

And there’s a lot of processing which can be done with URLs: shortening them, adding them to “to do” lists, checking them for validity, keeping them in link libraries, archiving their “content”, showing them as external or internal links, preventing them from “rotting away”, showing the wordcount or reading time of the item they “target”, display them as QR codes, abuse them, etc.

As you can notice, it’s easy to get me on a tangent simply thinking about URLs. What’s this have t’do with iCloud, you ask? Probably not much, in terms of the actual service which will be announced at Moscone. But I’ve been dreaming about iCloud as a way to integrate Diigo, Instapaper, Delicious, reddit, digg, Slashdot, StumbleUpon, Spurl, The NethernetXmarks

Hey, I told you I was dreaming! Something as simple as managing, processing, sharing, and archiving links in iCloud could lead to just about anything, in my imagination.

And speaking of Xmarks… It’s now owned by Lastpass, a company which focus on password management. IMHO, some Lastpass-like features could make their way in diverse products, including iCloud. Is this far-fetched? Possibly. But secure handling of passwords can be a major issue in both of Apple’s new operating systems (Mac OS X Lion and iOS5). From “keychains” to SSO, there’s a lot of work to be done which relates to password management, in my mind.

Which leads me to think about authentication in general and the rumours about “deep Twitter integration in iOS 5”. (Not directly related to iCloud, but who knows?) Again, something which can send me (and others) on drift-off moments. What if this integration suddenly made iOS devices more useful in terms of social networking services? Something to ponder, if one has a propensity for pondering.

At the same time, given the relative lack of activity on iTunes Ping, I wouldn’t bet on Twitter integration having that major an impact by itself. Not unlike Google, Apple has a hard time making a mark on the social Web. Now, if Twitter integration does connect to everything else Apple does, it could lead to interesting things. A full-fledged online identity? Access to contacts for not only messaging and photo sharing but for collaboration, group management, and media sharing? Not betting on any of this, but it could be fun. Again, not specific to iCloud, but quite related to “The Cloud”. If Twitter integration is deep enough, in iOS 5, it’d be possible to use iOS devices for “cloud computing”, getting further into the “post-PC era”.

An iCloud feature which is expected by several people, is something like an OTA version of the “iTunes file sharing” feature in iOS. Several apps (especially Apple’s own apps) use iTunes and a USB cable to share files. It was a welcome addition to iTunes 9.1 but it’s rather inconvenient. So many other apps rely on Dropbox for file sharing.

Which leads me to dream about iCloud as a replacement for Dropbox. Sounds extremely unlikely that it’ll have the full Dropbox feature set, especially if one thinks about the “Pro 50” and “Pro 100” plans on Dropbox. But I dream of the day when Apple’s iDisk will compete with Dropbox. Not that I’m convinced it ever will. But it’d make Apple’s devices all the more useful if it did.

Something similar, which isn’t frequently discussed directly, in connection with iCloud rumours, but which would rock: Mozy– or Carbonite-style backup, for Mac OS X machines. Sounds very unlikely that Apple will ever offer something like this but, as crazy as it may sound, the connection between Time Capsule and iCloud would be great if it went that far. From a user’s perspective, the similarities between Time Machine backup and “backing up in the cloud” (à la Mozy/Carbonite) are quite obvious. The advantages of both are clear. And while no hardware announcement is supposed to make its way to the WWDC 2011 keynote, I’d give the Time Capsule some consideration if it provided me with the equivalent of what I currently have with Mozy. Not to mention that Mozy has already sparked some drift-off moments, in me, before they announced their new plans. What if I could have a single service which combines features from Mozy, Time Machine, Dropbox, and YouSendIt?

I even think about the possibilities in terms of web hosting. As it stands, MobileMe does allow for some Web publishing through the iWeb application in its iLife suite. But iWeb has never been a major effort for Apple and it hasn’t been seen a significant update in quite a while. What if iCloud could become a true webhost just like, say… iWeb.com? (Semi-disclaimer: I won a free account with iWeb.com, last Fall, and I host some sites there. I also know some of the people who work there…)

Yet again, I don’t expect this to happen. It’s not speculation, on my part. It’s a daydream.

The reason this makes me dream is that I find all these things to be related and I wish they were integrated more seamlessly. Something about which Apple haters may not care much is the type of integration represented by iTunes. As clunky as iTunes may be, in some respects, it’s quite a success in terms of integrating a lot of different things. In fact, it probably overextended its reach a bit too much and we need to replace it. Apple needs to replace iTunes and we should also replace iTunes in our lives.

Like Gruber, I end up thinking about iCloud in relation to iTunes more than in relation to MobileMe. But I also dream about the ideal cloud service, which would not only sync and backup files between iOS devices, hundreds of millions of iTunes store accounts, and Macs, but replace several of the services for which I’m paying monthly fees.

Here’s to dreaming…

Other parts of this crazy, iCloud-infused daydream, in notes form:

Trusting People: Tummel Edition

Tummelvison 56: Howard Rheingold on crap detection, collaborative learning, and online community | Tummelvision

Reminds me of my own blogpost about “Trusting People” from five years ago, but with more emphasis on the political aspect. There was also some notion of reciprocity involved, which connects to well-known anthropological concepts.  Most directly, Rheingold alluded to Marcel Mauss’s approach to gifts (which also influenced Austin Hill’s idea about the social economy as a gift economy). And, as he did last week, Marks described the cultural dimension of the ultimatum experiment (which also has to do with reciprocity). All of which reminds me of my own little pun about the “attention economy”:

To get attention, you need to pay attention.

(I know, it’s overly simplistic. But, sometimes, I like to play with buzzphrases.)

Anyhoo…

It’s nice to hear much of this. Makes for conceptual connections which can, in turn, make for human connections.

The segment about couchsurfing was particularly interesting, in this respect. Including the acknowledgement that it may be gendered (like most things in social life). Mostly heard about couchsurfing from men so it was useful to hear a woman share her perspective on it.

The more I listen to TummelVision, the more I feel like tummlers are “my kind of people.” (No, not in an exclusive sense. In an inclusive sense reminiscent of communitas, sense of belonging, in-group, etc. Sure, there’s an out-group, an eventual sense of exclusion. But the feeling involved is about sharing, connecting, accepting, welcoming, opening…)

Academics and Their Publics

Misunderstood by Raffi Asdourian
Misunderstood by Raffi Asdourian

Academics are misunderstood.

Almost by definition.

Pretty much any academic eventually feels that s/he is misunderstood. Misunderstandings about some core notions in about any academic field are involved in some of the most common pet peeves among academics.

In other words, there’s nothing as transdisciplinary as misunderstanding.

It can happen in the close proximity of a given department (“colleagues in my department misunderstand my work”). It can happen through disciplinary boundaries (“people in that field have always misunderstood our field”). And, it can happen generally: “Nobody gets us.”

It’s not paranoia and it’s probably not self-victimization. But there almost seems to be a form of “onedownmanship” at stake with academics from different disciplines claiming that they’re more misunderstood than others. In fact, I personally get the feeling that ethnographers are more among the most misunderstood people around, but even short discussions with friends in other fields (including mathematics) have helped me get the idea that, basically, we’re all misunderstood at the same “level” but there are variations in the ways we’re misunderstood. For instance, anthropologists in general are mistaken for what they aren’t based on partial understanding by the general population.

An example from my own experience, related to my decision to call myself an “informal ethnographer.” When you tell people you’re an anthropologist, they form an image in their minds which is very likely to be inaccurate. But they do typically have an image in their minds. On the other hand, very few people have any idea about what “ethnography” means, so they’re less likely to form an opinion of what you do from prior knowledge. They may puzzle over the term and try to take a guess as to what “ethnographer” might mean but, in my experience, calling myself an “ethnographer” has been a more efficient way to be understood than calling myself an “anthropologist.”

This may all sound like nitpicking but, from the inside, it’s quite impactful. Linguists are frequently asked about the number of languages they speak. Mathematicians are taken to be number freaks. Psychologists are perceived through the filters of “pop psych.” There are many stereotypes associated with engineers. Etc.

These misunderstandings have an impact on anyone’s work. Not only can it be demoralizing and can it impact one’s sense of self-worth, but it can influence funding decisions as well as the use of research results. These misunderstandings can underminine learning across disciplines. In survey courses, basic misunderstandings can make things very difficult for everyone. At a rather basic level, academics fight misunderstandings more than they fight ignorance.

The  main reason I’m discussing this is that I’ve been given several occasions to think about the interface between the Ivory Tower and the rest of the world. It’s been a major theme in my blogposts about intellectuals, especially the ones in French. Two years ago, for instance, I wrote a post in French about popularizers. A bit more recently, I’ve been blogging about specific instances of misunderstandings associated with popularizers, including Malcolm Gladwell’s approach to expertise. Last year, I did a podcast episode about ethnography and the Ivory Tower. And, just within the past few weeks, I’ve been reading a few things which all seem to me to connect with this same issue: common misunderstandings about academic work. The connections are my own, and may not be so obvious to anyone else. But they’re part of my motivations to blog about this important issue.

In no particular order:

But, of course, I think about many other things. Including (again, in no particular order):

One discussion I remember, which seems to fit, included comments about Germaine Dieterlen by a friend who also did research in West Africa. Can’t remember the specifics but the gist of my friend’s comment was that “you get to respect work by the likes of Germaine Dieterlen once you start doing field research in the region.” In my academic background, appreciation of Germaine Dieterlen’s may not be unconditional, but it doesn’t necessarily rely on extensive work in the field. In other words, while some parts of Dieterlen’s work may be controversial and it’s extremely likely that she “got a lot of things wrong,” her work seems to be taken seriously by several French-speaking africanists I’ve met. And not only do I respect everyone but I would likely praise someone who was able to work in the field for so long. She’s not my heroine (I don’t really have heroes) or my role-model, but it wouldn’t have occurred to me that respect for her wasn’t widespread. If it had seemed that Dieterlen’s work had been misunderstood, my reflex would possibly have been to rehabilitate her.

In fact, there’s  a strong academic tradition of rehabilitating deceased scholars. The first example which comes to mind is a series of articles (PDF, in French) and book chapters by UWO linguistic anthropologist Regna Darnell.about “Benjamin Lee Whorf as a key figure in linguistic anthropology.” Of course, saying that these texts by Darnell constitute a rehabilitation of Whorf reveals a type of evaluation of her work. But that evaluation comes from a third person, not from me. The likely reason for this case coming up to my mind is that the so-called “Sapir-Whorf Hypothesis” is among the most misunderstood notions from linguistic anthropology. Moreover, both Whorf and Sapir are frequently misunderstood, which can make matters difficulty for many linguistic anthropologists talking with people outside the discipline.

The opposite process is also common: the “slaughtering” of “sacred cows.” (First heard about sacred cows through an article by ethnomusicologist Marcia Herndon.) In some significant ways, any scholar (alive or not) can be the object of not only critiques and criticisms but a kind of off-handed dismissal. Though this often happens within an academic context, the effects are especially lasting outside of academia. In other words, any scholar’s name is likely to be “sullied,” at one point or another. Typically, there seems to be a correlation between the popularity of a scholar and the likelihood of her/his reputation being significantly tarnished at some point in time. While there may still be people who treat Darwin, Freud, Nietzsche, Socrates, Einstein, or Rousseau as near divinities, there are people who will avoid any discussion about anything they’ve done or said. One way to put it is that they’re all misunderstood. Another way to put it is that their main insights have seeped through “common knowledge” but that their individual reputations have decreased.

Perhaps the most difficult case to discuss is that of Marx (Karl, not Harpo). Textbooks in introductory sociology typically have him as a key figure in the discipline and it seems clear that his insight on social issues was fundamental in social sciences. But, outside of some key academic contexts, his name is associated with a large series of social events about which people tend to have rather negative reactions. Even more so than for Paul de Man or  Martin Heidegger, Marx’s work is entangled in public opinion about his ideas. Haven’t checked for examples but I’m quite sure that Marx’s work is banned in a number of academic contexts. However, even some of Marx’s most ardent opponents are likely to agree with several aspects of Marx’s work and it’s sometimes funny how Marxian some anti-Marxists may be.

But I digress…

Typically, the “slaughtering of sacred cows” relates to disciplinary boundaries instead of social ones. At least, there’s a significant difference between your discipline’s own “sacred cows” and what you perceive another discipline’s “sacred cows” to be. Within a discipline, the process of dismissing a prior scholar’s work is almost œdipean (speaking of Freud). But dismissal of another discipline’s key figures is tantamount to a rejection of that other discipline. It’s one thing for a physicist to show that Newton was an alchemist. It’d be another thing entirely for a social scientist to deconstruct James Watson’s comments about race or for a theologian to argue with Darwin. Though discussions may have to do with individuals, the effects of the latter can widen gaps between scholarly disciplines.

And speaking of disciplinarity, there’s a whole set of issues having to do with discussions “outside of someone’s area of expertise.” On one side, comments made by academics about issues outside of their individual areas of expertise can be very tricky and can occasionally contribute to core misunderstandings. The fear of “talking through one’s hat” is quite significant, in no small part because a scholar’s prestige and esteem may greatly decrease as a result of some blatantly inaccurate statements (although some award-winning scholars seem not to be overly impacted by such issues).

On the other side, scholars who have to impart expert knowledge to people outside of their discipline  often have to “water down” or “boil down” their ideas and, in effect, oversimplifying these issues and concepts. Partly because of status (prestige and esteem), lowering standards is also very tricky. In some ways, this second situation may be more interesting. And it seems unavoidable.

How can you prevent misunderstandings when people may not have the necessary background to understand what you’re saying?

This question may reveal a rather specific attitude: “it’s their fault if they don’t understand.” Such an attitude may even be widespread. Seems to me, it’s not rare to hear someone gloating about other people “getting it wrong,” with the suggestion that “we got it right.”  As part of negotiations surrounding expert status, such an attitude could even be a pretty rational approach. If you’re trying to position yourself as an expert and don’t suffer from an “impostor syndrome,” you can easily get the impression that non-specialists have it all wrong and that only experts like you can get to the truth. Yes, I’m being somewhat sarcastic and caricatural, here. Academics aren’t frequently that dismissive of other people’s difficulties understanding what seem like simple concepts. But, in the gap between academics and the general population a special type of intellectual snobbery can sometimes be found.

Obviously, I have a lot more to say about misunderstood academics. For instance, I wanted to address specific issues related to each of the links above. I also had pet peeves about widespread use of concepts and issues like “communities” and “Eskimo words for snow” about which I sometimes need to vent. And I originally wanted this post to be about “cultural awareness,” which ends up being a core aspect of my work. I even had what I might consider a “neat” bit about public opinion. Not to mention my whole discussion of academic obfuscation (remind me about “we-ness and distinction”).

But this is probably long enough and the timing is right for me to do something else.

I’ll end with an unverified anecdote that I like. This anecdote speaks to snobbery toward academics.

[It’s one of those anecdotes which was mentioned in a course I took a long time ago. Even if it’s completely fallacious, it’s still inspiring, like a tale, cautionary or otherwise.]

As the story goes (at least, what I remember of it), some ethnographers had been doing fieldwork  in an Australian cultural context and were focusing their research on a complex kinship system known in this context. Through collaboration with “key informants,” the ethnographers eventually succeeded in understanding some key aspects of this kinship system.

As should be expected, these kinship-focused ethnographers wrote accounts of this kinship system at the end of their field research and became known as specialists of this system.

After a while, the fieldworkers went back to the field and met with the same people who had described this kinship system during the initial field trip. Through these discussions with their “key informants,” the ethnographers end up hearing about a radically different kinship system from the one about which they had learnt, written, and taught.

The local informants then told the ethnographers: “We would have told you earlier about this but we didn’t think you were able to understand it.”