Category Archives: Tech

Free from Freelance

For my Happiness Anniversary, this year, I got myself a brand new job.

Ok, it was two days late and a job isn’t really a gift. But it’s the thought that counts.

We’ll see how things go, but the position (Learning Technology Advisor) is right in line with things I already enjoy doing. Such as enabling technological appropriation in learning contexts. And holding thoughtful group discussions on interesting issues. And trying out new tools. And discussing learning objects and learning objectives. All things I’d probably do, regardless of my employment status.

So the work itself is likely to be very satisfying.

It’s been a while since I’ve had a fulltime dayjob. Years. Not that I haven’t been employed fulltime during that period. I did cumulate quite a few hours of work, most years. But they were part of different jobs, contracts, contexts. Which means that very significant a part of my “bandwidth” had to do with professional development. It also meant that my status tended to fluctuate. Teaching part-time was a large part of it, but a major distinction between part-time workers and fulltime ones relates to identity, status, recognition. For instance, sharing an office with a few colleagues is quite different from having your own.

I start my new job Tuesday, so I’ll know more by then (such as the office situation). But I’m already getting different interactions with people, such as this one teacher who says that we now have good reasons to be even better friends.

What’s funny is that the onset of my 2008 Happiness Phase coincided with my shift to freelancing. Had been doing several different things before that, mostly revolving around teaching and learning. But, from that point on, I allowed myself to take on contracts as a freelancer. I was no longer a Ph.D. candidate trying to squeeze in some work opportunity in view of an academic job. I was in control of my professional life, despite all the difficulties associated with freelancing.

It was a nice run. Ebbs and flows. Had the opportunity to try out many different things, sometimes within the same period of time. Landed a part-time position at a startup/community organization where the fit wasn’t great. Struggled to find a balance between acting as my own self and looking for new opportunities at every occasion. Had slow periods which made me question things. Coped with health issues in ways which would have been impossible while working fulltime. Invested time and money in all sorts of things to improve my life as a self-employed individual…

Overall, I learnt a lot. Much of it will be useful in fulltime work.

Though the job is fulltime, it’s based on a renewable contract. When, during the job interview, the HR advisor asked me for my thoughts on this situation, my whole freelance experience was behind me. No, it’s not an issue. I’ll manage even if it’s not renewed. But I’m starting a new life.

Something else about this new life connects to 2008. It’s in a Cegep.

Cegeps are Quebec public colleges for both vocational and pre-university education. I care deeply enough about the Cegep system to defend it. More than once. It’s occasionally under attack by politicians who try to stir things up. But it’s a part of post-secondary education in Quebec which makes it unique. Having taught in diverse places, I find that it makes a significant improvement in university life here. It also enables the kinds of training and learning that  people really need, as “adulteens” (very young adults who are also “teen-aged”). In the past week, even before settling down in my new position, I got to see some impressive things happening in Cegeps. I sincerely think that cegeps are an example to follow, not an anomaly. Similar systems exist elsewhere (from “gymnasium” and “international baccalaureate” to “prep schools” and “community colleges”). But Quebec’s Cegep network is its own very specific thing, fully adapted to its own cultural and social context.

Surely, I’ll have a lot more to say about Cegeps as I work in one.

The connection to 2008 is much more personal. At the time, I was going through a difficult transition in my life. Questioning all sorts of things. Growing dissatisfied with the model for university careers (especially tenure-track professorships and what they entail). Thinking of “what I could do with my life”…

…when it suddenly hit me: I could work in a Cegep.

Can still remember the overwhelming feeling of comfort I experienced when that thought hit me. It was so obvious! So fitting! Sure, there’d be some difficulties, but nothing impossible. I was ready, then, to embark in a Cegep career.

It’s not what happened, right away. I came back to Quebec from Texas and applied to a few things in Cegeps. Was getting other contracts, including teaching contracts at Concordia (where I started teaching in 2006). Never abandoned the idea of working in a Cegep but “life had other plans”, at the time.

I did do several things which got me closer to the Cegep system. Including participations in every MoodleDay event at Dawson. And workshops with Cegep institutions. I even participated in a living lab on educational innovation with the very organization which just hired me (lab summary in French). Without really noticing it, I was preparing myself to join the Cegep World.

One obvious possibility was to add Cegep courses to my part-time teaching load or eventually becoming fulltime as a Cegep teacher. Cegep teaching has clear advantages over university teaching. Simply put, Cegep teachers are allowed to care about learning. University professors who care about pedagogical issues bump into lots of hurdles. Since I care a whole lot about teaching (and I can still do actual research without a tenured position), it sounded like the right place for me. Friends and acquaintances who work in Cegeps kept telling me things which made the fit even more obvious. Though grades do matter in Cegeps, the obsession with grades is much less of an issue in Cegeps than in universities. I care enough about this that I co-organized a public conversation on grades, back in November 2013. Of course, the Cegep population is quite different (and often younger) than the university population. Having taught in the US where people enter university or college directly from high school, I didn’t think it’d be an issue.

But teaching Cegep students directly wasn’t the only option. Having accumulated some expertise on post-secondary learning through 15 years of teaching experience, I was starting to think about being a learning advisor of some sort. This is finally happening, officially.

I’ve often acted as an informal advisor for people. Even during my M.Sc., I would discuss a Ph.D. student’s research in pretty much the way an advisor would. Not the advisor who focuses on logistics and rules and citation impact. But the person who challenges you to rethink a research question or brings you to think of your whole project in a completely new way. Since then, I’ve done the same thing numerous times without ever having an official title to go with it. I’d occasionally get a bit of (informal) credit for it, but I wasn’t aiming for that. I just enjoy helping people in this way.

I also became something of a mentor to some people. For instance, in  view of a pilot project at Concordia, I was able to mentor two teaching assistants who were holding classes in parallel with me. The mentoring included tips and tricks about classroom management along with deeper things on the meaning of university learning. It was still a limited scope, but it was in line with things I wanted to do.

What was even more fitting, given my new position, is that I became the “go-to person” for several things having to do with technology in learning and teaching. This all started in 2007 when, through the Spirit of Inquiry conference, I started collaborating with Concordia’s Centre for Teaching and Learning Services. Created workshops, did screencasts, experimented with some solutions, answered informal questions… Without being employed directly by CTLS, I think it’s fair to say (as several people have been saying) that I was playing a key role in terms of learning technology at Concordia.

There’s a pattern, here. From diverse activities as a freelancer, I now get to merge things to be Learning Technology Advisor at Vitrine technologie-éducation.

Good times!

Wearable Hub: Getting the Ball Rolling


After years of hype, wearable devices are happening. What wearable computing lacks is a way to integrate devices into a broader system.


  • For the past two months or so, I’ve been taking notes about this “wearable hub” idea (started around CES’s time, as wearable devices like the Pebble and Google Glass were discussed with more intensity). At this point, I have over 3000 words in notes, which probably means that I’d have enough material for a long essay. This post is just a way to release a few ideas and to “think aloud” about what wearables may mean.
  • Some of these notes have to do with the fact that I started using a few wearable devices to monitor my activities, after a health issue pushed me to start doing some exercise.
  • I’m not a technologist nor do I play one on this blog. I’m primarily an ethnographer, with diverse interests in technology and its implications for human beings. I do research on technological appropriation and some of the course I teach relate to the social dimensions of technology. Some of the approaches to technology that I discuss in those courses relate to constructionism and Actor-Network Theory.
  • I consider myself a “geek ethnographer” in the sense that I take part in geek culture (and have come out as a geek) but I’m also an outsider to geekdom.
  • Contrary to the likes of McLuhan, Carr, and Morozov, my perspective on technology and society is non-deterministic. The way I use them, “implication” and “affordance” aren’t about causal effects or, even, about direct connections. I’m not saying that society is causing technology to appear nor am I proposing a line from tools to social impacts. Technology and society are in a complex system.
  • Further, my approach isn’t predictive. I’m not saying what will happen based on technological advances nor am I saying what technology will appear. I’m thinking about the meaning of technology in an intersubjective way.
  • My personal attitude on tools and gadgets is rather ambivalent. This becomes clear as I go back and forth between techno-enthusiastic contexts (where I can almost appear like a Luddite) and techno-skeptical contexts (where some might label me as a gadget freak). I integrate a number of tools in my life but I can be quite wary about them.
  • I’m not wedded to the ideas I’m putting forth, here. They’re just broad musings of what might be. More than anything, I hope to generate thoughtful discussion. That’s why I start this post with a broad statement (not my usual style).
  • Of course, I know that other people have had similar ideas and I know that a concept of “wearable hub” already exists. It’s obvious enough that it’s one of these things which can be invented independently.

From Wearables to Hubs

Back in the 1990s, “wearable computing” became something of a futuristic buzzword, often having to do with articles of clothing. There have been many experiments and prototypes converging on an idea that we would, one day, be able to wear something resembling a full computer. Meanwhile, “personal digital assistants” became something of a niche product and embedded systems became an important dimension of car manufacturing.

Fast-forward to 2007, when a significant shift in the use of smartphones occurred. Smartphones existed before that time, but their usages, meanings, and positions in the public discourse changed quite radically around the time of the iPhone’s release. Not that the iPhone itself “caused a smartphone revolution” or that smartphone adoption suddenly reached a “tipping point”. I conceive of this shift as a complex interplay between society and tools. Not only more Kuhn than Popper, but more Latour than Kurzweil.

Smartphones, it may be argued, “happened”.

Without being described as “wearable devices”, smartphones started playing some of the functions people might have assigned to wearable devices. The move was subtle enough that Limor Fried recently described it as a realization she’s been having. Some tech enthusiasts may be designing location-aware purses and heads-up displays in the form of glasses. Smartphones are already doing a lot of the things wearables were supposed to do. Many people “wear” smartphones at most times during their waking lives and these Internet-connected devices are full of sensors. With the proliferation of cases, one might even perceive some of them as fashion accessories, like watches and sunglasses.

Where smartphones become more interesting, in terms of wearable computing, is as de facto wearable hubs.

My Wearable Devices

Which brings me to mention the four sensors I’ve been using more extensively during the past two months:

Yes, these all have to do with fitness (and there’s quite a bit of overlap between them). And, yes, I started using them a few days after the New Year. But it’s not about holiday gifts or New Year’s resolutions. I’ve had some of these devices for a while and decided to use them after consulting with a physician about hypertension. Not only have they helped me quite a bit in solving some health issues, but these devices got me to think.

(I carry several other things with me at most times. Some of my favourites include Tenqa REMXD Bluetooth headphones and the LiveScribe echo smartpen.)

One aspect is that they’re all about the so-called “quantified self”. As a qualitative researcher, I tend to be skeptical of quants. In this case, though, the stats I’m collecting about myself fit with my qualitative approach. Along with quantitative data from these devices, I’ve started collecting qualitative data about my life. The next step is to integrate all those data points automatically.

These sensors are also connected to “gamification”, a tendency I find worrisome, preferring playfulness. Though game mechanics are applied to the use of these sensors, I choose to rely on my intrinsic motivation, not paying much attention to scores and badges.

But the part which pushed me to start taking the most notes was that all these sensors connect with my iOS ()and Android) devices. And this is where the “wearable hub” comes into play. None of these devices is autonomous. They’re all part of my personal “arsenal”, the equipment I have on my me on most occasions. Though there are many similarities between them, they still serve different purposes, which are much more limited than those “wearable computers” might have been expected to serve. Without a central device serving as a type of “hub”, these sensors wouldn’t be very useful. This “hub” needs not be a smartphone, despite the fact that, by default, smartphones are taken to be the key piece in this kind of setup.

In my personal scenario, I do use a smartphone as a hub. But I also use tablets. And I could easily use an existing device of another type (say, an iPod touch), or even a new type of device meant to serve as a wearable hub. Smartphones’ “hub” affordances aren’t exclusive.

From Digital Hub to Wearable Hub

Most of the devices which would likely serve as hubs for wearable sensors can be described as “Post-PC”. They’re clearly “personal” and they’re arguably “computers”. Yet they’re significantly different from the “Personal Computers” which have been so important at the end of last century (desktop and laptop computers not used as servers, regardless of the OS they run).

Wearability is a key point, here. But it’s not just a matter of weight or form factor. A wearable hub needs to be wireless in at least two important ways: independent from a power source and connected to other devices through radio waves. The fact that they’re worn at all times also implies a certain degree of integration with other things carried throughout the day (wallets, purses, backpacks, pockets…). These devices may also be more “personal” than PCs because they may be more apparent and more amenable to customization than PCs.

Smartphones fit the bill as wearable hubs. Their form factors and battery life make them wearable enough. Bluetooth (or ANT+, Nike+, etc.) has been used to pair them wirelessly with sensors. Their connectivity to GPS and cellular networking as well as their audio and visual i/o can have interesting uses (mapping a walk, data updates during a commute, voice feedback…). And though they’re far from ubiquitous, smartphones have become quite common in key markets.

Part of the reason I keep thinking about “hubs” has to do with comments made in 2001 by then Apple CEO Steve Jobs about the “digital lifestyle” age in “PC evolution” (video of Jobs’s presentation; as an anthropologist, I’ll refrain from commenting on the evolutionary analogies):

We believe the PC, or more… importantly, the Mac can become the “digital hub” of our emerging digital lifestyle, with the ability to add tremendous value to … other digital devices.

… like camcorders, portable media players, cellphones, digital cameras, handheld organizers, etc. (Though they weren’t mentioned, other peripherals like printers and webcams also connect to PCs.)

The PC was thus going to serve as a hub, “not only adding value to these devices but interconnecting them, as well”.

At the time, key PC affordances which distinguished them from those other digital devices:

  • Big screen affording more complex user interfaces
  • Large, inexpensive hard disk storage
  • Burning DVDs and CDs
  • Internet connectivity, especially broadband
  • Running complex applications (including media processing software like the iLife suite)

Though Jobs pinpointed iLife applications as the basis for this “digital hub” vision, it sounds like FireWire was meant to be an even more important part of this vision. Of course, USB has supplanted FireWire in most use cases. It’s interesting, then, to notice that Apple only recently started shipping Macs with USB 3. In fact, DVD burning is absent from recent Macs. In 2001, the Mac might have been at the forefront of this “digital lifestyle” age. In 2013, the Mac has moved away from its role as “digital hub”.

In the meantime, the iPhone has become one of the best known examples of what I’m calling “wearable hubs”. It has a small screen and small, expensive storage (by today’s standards). It also can’t burn DVDs. But it does have nearly-ubiquitous Internet connectivity and can run fairly complex applications, some of which are adapted from the iLife suite. And though it does have wired connectivity (through Lightning or the “dock connector”), its main hub affordances have to do with Bluetooth.

It’s interesting to note that the same Steve Jobs, who used the “digital hub” concept to explain that the PC wasn’t dead in 2001, is partly responsible for popularizing the concept of “post-PC devices” six years later. One might perceive hypocrisy in this much delayed apparent flip-flop. On the other hand, Steve Jobs’s 2007 comments (video) were somewhat nuanced, as to the role of post-PC devices. What’s more interesting, though, is to think about the implications of the shift between two views of digital devices, regardless of Apple’s position through that shift.

Some post-PC devices (including the iPhone, until quite recently) do require a connection to a PC. In this sense, a smartphone might maintain its position with regards to the PC as digital hub. Yet, some of those devices are used independently of PCs, including by some people who never owned PCs.

Post-Smartphone Hubs

It’s possible to imagine a wearable hub outside of the smartphone (and tablet) paradigm. While smartphones are a convenient way to interconnect wearables, their hub-related affordances still sound limited: they lack large displays and their storage space is quite expensive. Their battery life may also be something to consider in terms of serving as hubs. Their form factors make some sense, when functioning as phones. Yet they have little to do with their use as hubs.

Part of the realization, for me, came from the fact that I’ve been using a tablet as something of an untethered hub. Since I use Bluetooth headphones, I can listen to podcasts and music while my tablet is in my backpack without being entangled in a cable. Sounds trivial but it’s one of these affordances I find quite significant. Delegating music playing functions to my tablet relates in part to battery life and use of storage. The tablet’s display has no importance in this scenario. In fact, given some communication between devices, my smartphone could serve as a display for my tablet. So could a “smartwatch” or “smartglasses”.

The Body Hub

Which led me to think about other devices which would work as wearable hubs. I originally thought about backpackable and pocketable devices.

But a friend had a more striking idea:

Under Armour’s Recharge Energy Suit may be an extreme version of this, one which would fit nicely among things Cathi Bond likes to discuss with Nora Young on The Sniffer. Nora herself has been discussing wearables on her blog as well as on her radio show. Sure, part of this concept is quite futuristic. But a sensor mesh undershirt is a neat idea for several reasons.

  • It’s easy to think of various sensors it may contain.
  • Given its surface area, it could hold enough battery power to supplement other devices.
  • It can be quite comfortable in cold weather and might even help diffuse heat in warmer climates.
  • Though wearable, it needs not be visible.
  • Thieves would probably have a hard time stealing it.
  • Vibration and haptic feedback on the body can open interesting possibilities.

Not that it’s the perfect digital hub and I’m sure there are multiple objections to a connected undershirt (including issues with radio signals). But I find the idea rather fun to think, partly because it’s so far away from the use of phones, glasses, and watches as smart devices.

Another thing I find neat, and it may partly be a coincidence, is the very notion of a “mesh”.

The Wearable Mesh

Mesh networking is a neat concept, which generates more hype than practical uses. As an alternative to WiFi access points and cellular connectivity, it’s unclear that it may “take the world by storm”. But as a way to connect personal devices, it might have some potential. After all, as Bernard Benhamou recently pointed out on France Culture’s Place de la toile, the Internet of Things may not require always-on full-bandwith connectivity. Typically, wearable sensors use fairly little bandwidth or only use it for limited amounts of time. A wearable mesh could connect wearable devices to one another while also exchanging data through the Internet itself.

Or with local devices. Smart cities, near field communication, and digital appliances occupy interesting positions among widely-discussed tendencies in the tech world. They may all have something to do with wearable devices. For instance, data exchanged between transit systems and their users could go through wearable devices. And while mobile payment systems can work through smartphones and other cellphones, wallet functions can also be fulfilled by other wearable devices.

Alternative Futures

Which might provide an appropriate segue into the ambivalence I feel toward the “wearable hub” concept I’m describing. Though I propose these ideas as if I were enthusiastic about them, they all give me pause. As a big fan of critical thinking, I like to think about “what might be” to generate questions and discussions exposing a diversity of viewpoints about the future.

Mass media discussions about these issues tend to focus on such things as privacy, availability, norms, and usefulness. Google Glass has generated quite a bit of buzz about all four. Other wearables may mainly raise issues for one or two of these broad dimensions. But the broad domain of wearable computing raises a lot more issues.

Technology enthusiasts enjoy discussing issues through the dualism between dystopia and utopia. An obvious issue with this dualism is that humans disagree about the two categories. Simply put, one person’s dystopia can be another person’s utopia, not to mention the nuanced views of people who see complex relationships between values and social change.

In such a context, a sociologist’s reflex may be to ask about the implications of these diverse values and opinions. For instance:

  • How do people construct these values?
  • Who decides which values are more important?
  • How might social groups cope with changes in values?

Discussing these issues and more, in a broad frame, might be quite useful. Some of the trickiest issues are raised after some changes in technology have already happened. From writing to cars, any technological context has unexpected implications. An ecological view of these implications could broaden the discussion.

I tend to like the concept of the “drift-off moment”, during which listeners (or readers) start thinking about the possibilities afforded a new tool (or concept). In the context of a sales pitch, the idea is that these possibilities are positive, a potential buyer is thinking about the ways she might use a newfangled device. But I also like the deeper process of thinking about all sorts of implications, regardless of their value.


What might be the implications of a wearable hub?

Bean Counters and Ecologists

[So many things in my drafts, but this one should be quick.]

Recently met someone who started describing their restaurant after calling it a “café”. The “pitch” revolved around ethical practices, using local products, etc. As both a coffee geek and ethnographer, my simple question was: “Which coffee do you use?” Turns out, they’re importing coffee from a multinational corporation. “Oh, but, they’re lending us an expensive espresso machine for free! And they have fair-trade coffee!”

Luckily, we didn’t start talking about “fair trade”. And this person was willing to reflect upon the practices involved, including about the analogy with Anheuser-Busch or Coca-Cola. We didn’t get further into the deeper consequences of the resto’s actions, but the “seed” has been planted.

Sure, it’s important to focus on your financials and there’s nothing preventing a business from being both socially responsible and profitable. It just requires a shift in mindset. Small, lean, nimble businesses are more likely to do it than big, multinational corporate empires…

…which leads me to Google.

Over the years since its IPO, Google has attracted its share of praise and criticism. Like any big, multinational corporate empire. In any sector.

Within the tech sector, the Goog‘ is often compared with Microsoft, Facebook, Twitter, and Apple. All of these corporate entities have been associated in some people’s minds with some specific issue, from child labour and failure to protect users’ privacy to anticompetitive practices (the tech equivalent of free fridges and espresso machines). The issues are distinct and tech enthusiast spend a large amount of time discussing which one is worse. Meanwhile, we’re forgetting a number of larger issues.

Twitter is an interesting example, here. The service took its value from being at the centre of an ecosystem. As with any ecosystem, numerous interactions among many different members produce unexpected and often remarkable results. As the story goes, elements like hashtags and “@-replies” were invented by users and became an important part of the system. Third-party developers were instrumental in Twitter’s reach outside of its original confines. Though most of the original actors have since left the company, the ecosystem has maintained itself over the years.

When Twitter started changing the rules concerning its API, it shook the ecosystem. Sure, the ecosystem will maintain itself, in the end. But it’s nearly impossible to predict how it will change. For people at Twitter, it must have been obvious that the first changes was a warning shot to scare away those they didn’t want in their ecosystem. But, to this day, there are people who depend on Twitter, one way or another.

Google Reader offers an interesting case. The decision to kill it might have been myopic and its death might have a domino effect.

The warning shot was ambiguous, but the “writing was on the wall”. Among potential consequences of the move, the death of RSS readers was to be expected. One might also expect users of feedreaders to be displeased. In the end, the ecosystem will maintain itself.

Chances are, feedreading will be even more marginalized than it’s been and something else might replace it. Already, many people have been switching from feedreading to using Twitter as a way to gather news items.

What’s not so well-understood is the set of indirect consequences, further down the line. Again, domino effect. Some dominoes are falling in the direction of news outlets which have been slow to adapt to the ways people create and “consume” news items. Though their ad-driven models may sound similar to Google’s, and though feedreading might not be a significant source of direct revenue, the death of feedreaders may give way to the birth of new models for news production and “consumption” which might destabilize them even further. Among the things I tag as #FoJ (“Future of Journalism”) are several pieces of a big puzzle which seems misunderstood by news organizations.

There are other big dominoes which might fall from the death of Google Reader. Partly because RSS itself is part of a whole ecosystem. Dave Winer and Aaron Swartz have been major actors in the technical specifications of RSS. But Chris Lydon and people building on calendar syndication are also part of the ecosystem. In business-speak, you might call them “stakeholders”. But thinking about the ecosystem itself leads to a deeper set of thoughts, beyond the individuals involved. In the aftermath of Aaron Swartz’s premature death, it may be appropriate to point out that the ecosystem is more than the sum of its parts.

As I said on a service owned by another widely-criticized corporate empire:

Many of us keep saying that Google needs to listen to its social scientists. It also needs to understand ecology.

Timeline of Apple’s Online Services

[I’d like people’s help in completing a timeline of Apple’s online services.]

[Update: deleted the WordPress shortcode, which seems not to be working.]

As kind of a followup to yesterday’s post about some early rumours and speculations about the iPhone, I thought about posting some info about Apple’s online services. Part of the reason is that Asymco’s Horace Dediu has frequently talked about what we could call “Apple’s data play”, for instance in this post about the iCloud data centre in North Carolina. I was also thinking about Mike Davidson’s comments about Apple’s presence (and “dominance”) in such diverse fields as hardware, software, licensing, and commerce. The trigger for this post, though, was from this Steve Jobs comment, which appeared in a recent NYT piece about the Apple Maps fiasco:

The MobileMe launch clearly demonstrates that we have more to learn about Internet services

The overall context for this quote as well as a number of discussions about Apple is the consensus that Apple does a poor job with online services. MobileMe and iTunes Ping are often used in these discussions and it seems clear to most people (including Apple executives and insiders, it sounds like) that the “computer company turned consumer electronics vendor” has a lot to learn about online services.

The reason I find this so interesting is that Apple seems insistent on pushing at least some of its online services. A bit less of a “betting the farm” strategy as Google’s “Emerald Sea” initiative, but an intriguing strategy for such a large and still-successful company. Dediu’s frequent reference to Clay Christensen’s concept of “Disruptive Innovation” might apply, here. Apple might be “disrupting itself into” an online services company, at least in part.

There are several things I find intriguing about this strategy.

As opposed to most other enterprises’ “online plays”, Apple’s model tends not to be based on ad revenues. The divide between Google and Apple couldn’t be stronger when we talk about ad-supported free/freemium services as opposed to paid services or services attached to other purposes. It’s likely an irreconcilable difference between fans of  both teams.

Online services are clearly not Apple’s strong suit. It often sounds like Apple is missing a “magic touch” with online services, the same way other companies are said to lack Apple’s design sense. This is more similar to Google+ given the consensus that “Google doesn’t know how to do ‘social’”. But it’s still surprising.

Though Apple may not have a “knack” for online services, it’s been trying to do it for quite a while. I keep thinking about eWorld as a precursor to the whole thing. It’s one thing for a company to try its hand at something new or to pivot into a strong business. It’s another thing entirely to shift more energies into something which has so far proven to be mostly a lost cause.

Adding to my thoughts on this was a podcast conversation (I think between John Siracusa and Dan Benjamin, though it might have been between Marco Arment and John Gruber) during which comments were made about those Apple employees working on online services.

So, basically, Apple’s online services have been on my mind. But I couldn’t find an exhaustive list. Tried Wikipedia but it doesn’t really separate online services from other things Apple does. And I ended up thinking about what would define “online services” in Apple’s case. Everything Apple does which incurs some bandwidth costs would be my working definition. Basically, it’s something to do with Apple investing in data centres and such. Some of these seem like very small costs (hosting data about podcasts, instead of the podcasts themselves, for instance). Given Apple’s size, these costs and the infrastructure behind all of this can be quite big.

So I started listing some of these services and organizing them in a sort of timeline, first in MultiMarkdown format in nvAlt, then in a Google Spreadsheet. I then discovered Vérité.CO’s Timeline.JS which takes a Google Spreadsheet and makes it into a visual timeline.

A few notes:

  • It’s a quick draft and I didn’t really check any of the data points.
  • In most cases, I only added months and, in the case of “AppleLink”, I only put years.
  • I took most dates from diverse Wikipedia pages, not necessarily backtracking on the whole process.
  • On at least one occasion, there was a discrepancy between two dates.
  • Sometimes, I took the date of the service’s announcement while I used an actual launch date for other services.
  • I only added a couple of pictures to show that it can be done. Many of the relevant pix are likely to be under copyright or to constitute a trademark.
  • I tried to be as exhaustive as I could be, but I’m sure I forgot stuff.
  • Some things may not sound like they qualify as part of “Apple’s online offering” but I think they’re still relevant. My rule of thumb is that if it goes to Apple’s servers, it’s an online service.
  • I separated some services from “suites” like iCloud or iTools, partly because some of those services haven’t been kept, which is important to see in a timeline. There are several services missing, here.
  • None of this timeline is meant to be editorial. I was just curious about what Apple has been doing online since the 1980s. The reason I care can be found in my earlier notes. I consider myself neither an “Apple fanboi” nor an “Apple hater”. I just find the situation revealing of something happening in the tech world, which has an impact on the Geek Niche.

So, here goes.

Here’s the Google Spreadsheet (editable by anyone):

Apple Online Services

Here’s the timeline through an embed code:

Here’s the embed code:

<iframe src='' width='100%' height='650' frameborder='0'>


Actively Reading: Organic Ideas for Startups

Been using Diigo as a way to annotate online texts. In this case, I was as interested in the tone as in the text itself. At the same time, I kept thinking about things which seem to be missing from Diigo.
Organic Startup Ideas

One thing I like about this text is its tone. There’s an honesty, an ingenuity that I find rare in this type of writing.

  • startup ideas
    • The background is important, in terms of the type of ideas about which we’re constructing something.
  • what do you wish someone would make for you?
    • My own itch has to do with Diigo, actually. There’s a lot I wish Diigo would make for me. I may be perceived as an annoyance, but I think my wishlist may lead to something bigger and possibly quite successful.
    • The difference between this question and the “scratch your own itch” principle seems significant, and this distinction may have some implications in terms of success: we’re already talking about others, not just running ideas in our own head.
  • what do you wish someone would make for you?
    • It’s somewhat different from the well-known “scratch your own itch” principle. In this difference might be located something significant. In a way, part of the potential for this version to lead to success comes from the fact that it’s already connected with others, instead of being about running ideas in your own mind.
  • grow organically
    • The core topic of the piece, put in a comparative context. The comparison isn’t the one people tend to make and one may argue about the examples used. But the concept of organic ideas is fascinating and inspiring.
  • you decide, from afar,
    • What we call, in anthropology, the “armchair” approach. Also known as “backbenching.” For this to work, you need to have a deep knowledge of the situation, which is part of the point in this piece. Nice that it’s not demonizing this position but putting it in context.
  • Apple
    was the first type
    • One might argue that it was a hybrid case. Although, it does sound like the very beginnings of Apple weren’t about “thinking from afar.”
  • class of users other than you
    • Since developers are part of a very specific “class” of people, this isn’t insignificant a way to phrase this.
  • They still rely on this principle today, incidentally.
    The iPhone is the phone Steve Jobs wants.
    • Apple tends to be perceived in a different light. According to many people, it’s the “textbook example” of a company where decisions are made without concerns for what people need. “Steve Jobs uses a top-down approach,” “They don’t even use focus groups,” “They don’t let me use their tools the way I want to use them.” But we’re not talking about the same distinction between top-down and bottom-up. Though “organic ideas” seem to imply that it’s a grassroots/bottom-up phenomenon, the core distinction isn’t about the origin of the ideas (from the “top,” in both cases) but on the reasoning behind these ideas.
  • We didn’t need this software ourselves.
    • Sounds partly like a disclaimer but this approach is quite common and “there’s nothing wrong with it.”
  • comparatively old
    • Age and life experience make for an interesting angle. It’s not that this strategy needs people of a specific age to work. It’s that there’s a connection between one’s experience and the way things may pan out.
  • There is no sharp line between the two types of ideas,
    • Those in the “engineering worldview” might go nuts, at this point. I can hear the claims of “hand waving.” But we’re talking about something complex, here, not a merely complicated problem.
  • Apple type
    • One thing to note in the three examples here: they’re all made by pairs of guys. Jobs and Woz, Gates and Allen, Page and Brin. In many cases, the formula might be that one guy (or gal, one wishes) comes up with ideas knowing that the other can implement them. Again, it’s about getting somebody else to build it for you, not about scratching your own itch.
  • Bill Gates was writing something he would use
    • Again, Gates may not be the most obvious example, since he’s mostly known for another approach. It’s not inaccurate to say he was solving his own problem, at the time, but it may not be that convincing as an example.
  • Larry and Sergey when they wrote the first versions of Google.
    • Although, the inception of the original ideas was academic in context. They weren’t solving a search problem or thinking about monetization. They were discovering the power of CitationRank.
  • generally preferable
    • Nicely relativistic.
  • It takes experience
    to predict what other people will want.
    • And possibly a lot more. Interesting that he doesn’t mention empirical data.
  • young founders
    • They sound like a fascinating group to observe. They do wonders when they open up to others, but they seem to have a tendency to impose their worldviews.
  • I’d encourage you to focus initially on organic ideas
    • Now, this advice sounds more like the “scratch your own itch” advocation. But there’s a key difference in that it’s stated as part of a broader process. It’s more of a “walk before you run” or “do your homework” piece of advice, not a “you can’t come up with good ideas if you just think about how people will use your tool.”
  • missing or broken
    • It can cover a lot, but it’s couched in terms of the typical “problem-solving” approach at the centre of the engineering worldview. Since we’re talking about developing tools, it makes sense. But there could be a broader version, admitting for dreams, inspiration, aspiration. Not necessarily of the “what would make you happy?” kind, although there’s a lot to be said about happiness and imagination. You’re brainstorming, here.
  • immediate answers
    • Which might imply that there’s a second step. If you keep asking yourself the same question, you may be able to get a very large number of ideas. The second step could be to prioritize them but I prefer “outlining” as a process: you shuffle things together and you group some ideas to get one which covers several. What’s common between your need for a simpler way to code on the Altair and your values? Why do you care so much about algorithms instead of human encoding?
  • You may need to stand outside yourself a bit to see brokenness
    • Ah, yes! “Taking a step back,” “distancing yourself,” “seeing the forest for the trees”… A core dimension of the ethnographic approach and the need for a back-and-forth between “inside” and “outside.” There’s a reflexive component in this “being an outsider to yourself.” It’s not only psychological, it’s a way to get into the social, which can lead to broader success if it’s indeed not just about scratching your own itch.
  • get used to it and take it for granted
    • That’s enculturation, to you. When you do things a certain way simply because “we’ve always done them that way,” you may not create these organic ideas. But it’s a fine way to do your work. Asking yourself important questions about what’s wrong with your situation works well in terms of getting new ideas. But, sometimes, you need to get some work done.
  • a Facebook
    • Yet another recontextualized example. Zuckerberg wasn’t trying to solve that specific brokenness, as far as we know. But Facebook became part of what it is when Zuck began scratching that itch.
  • organic startup ideas usually don’t
    seem like startup ideas at first
    • Which gets us to the pivotal importance of working with others. Per this article, VCs and “angel investors,” probably. But, in the case of some of cases cited, those we tend to forget, like Paul Allen, Narendra, and the Winklevosses.
  • end up making
    something of value to a lot of people
    • Trial and error, it’s an iterative process. So you must recognize errors quickly and not invest too much effort in a specific brokenness. Part of this requires maturity.
  • something
    other people dismiss as a toy
    • The passage on which Gruber focused and an interesting tidbit. Not that central, come to think of it. But it’s important to note that people’s dismissive attitude may be misled, that “toys” may hide tools, that it’s probably a good idea not to take all feedback to heart…
  • At this point, when someone comes to us with
    something that users like but that we could envision forum trolls
    dismissing as a toy, it makes us especially likely to invest.
  • the best source of organic ones
    • Especially to investors. Potentially self-serving… in a useful way.
  • they’re at the forefront of technology
    • That part I would dispute, actually. Unless we talk about a specific subgroup of young founders and a specific set of tools. Young founders tend to be oblivious to a large field in technology, including social tools.
  • they’re in a position to discover
    valuable types of fixable brokenness first
    • The focus on fixable brokenness makes sense if we’re thinking exclusively through the engineering worldview, but it’s at the centre of some failures like the Google Buzz launch.
  • you still have to work hard
    • Of the “inspiration shouldn’t make use forget perspiration” kind. Makes for a more thoughtful approach than the frequent “all you need to do…” claims.
  • I’d encourage anyone
    starting a startup to become one of its users, however unnatural it
    • Not merely an argument for dogfooding. It’s deeper than that. Googloids probably use Google tools but they didn’t actually become users. They’re beta testers with a strong background in troubleshooting. Not the best way to figure out what users really want or how the tool will ultimately fail.
  • It’s hard to compete directly with open source software
    • Open Source as competition isn’t new as a concept, but it takes time to seep in.
  • there has to be some part
    you can charge for
    • The breach through which old-school “business models” enter with little attention paid to everything else. To the extent that much of the whole piece might crumble from pressure built up by the “beancounter” worldview. Good thing he acknowledges it.

Installing BuddyPress 1.2 on FatCow: Quick Edition

I recently posted a rambling version of instructions about how to install BuddyPress 1.1.3 on FatCow:

Installing BuddyPress on a Webhost « Disparate.

BuddyPress 1.2 was just released, with some neat new features including the ability to run on a standard (non-WPµ) version of WordPress and a new way to handle templates. They now have three-step instructions on how to install BuddyPress. Here’s my somewhat more verbose take (but still reasonably straightforward and concise). A few things are FatCow-specific but everything should be easy to adapt for any decent webhost. (In fact, it’d likely be easier elsewhere.)

  1. Create database in FatCow’s Manage MySQL
  2. Download WP 2.9.2 from the download page.
  3. Uncompress WP  using FatCow’s Archive Gateway
  4. Rename “wordpress” to <name> (not necessary, but useful, I find; “community” or “commons” would make sense for <name>)
  5. Go to <full domain>/<name> (say, “” if you used “commons” for <name> and your domain were “”)
  6. Click “Create a Configuration file”
  7. Click “Let’s Go”
  8. Enter database information from database created in MySQL
  9. Change “localhost” to “<username>” (where <username> is your FatCow username)
  10. Click “Submit”
  11. Click “Run the Install”
  12. Fill in blog title and email, decide on crawling
  13. Click on “Install WordPress”
  14. Copy generated password
  15. Enter login details
  16. Click “Log In”
  17. Click “Yes, Take me to my profile page”
  18. Add “New password” (twice)
  19. Click “Update Profile”
  20. Click “Plugins”
  21. Click “Add New”
  22. In the searchbox, type “BuddyPress”
  23. Find BuddyPress 1.2 (created by “The BuddyPress Community”)
  24. Click “Install”
  25. Click “Install Now”
  26. Click “Activate Plugin”
  27. Click “update your permalink structure”
  28. Choose an option (say, “Day and Name”)
  29. Click “Save Changes”
  30. Click “Appearance”
  31. Click “Activate” under BuddyPress default

You now have a full BuddyPress installation. “Social Networking in a Box”

You can do a number of things, now. Including visiting your BuddyPress installation by clicking on the “Visit Site” link at the top of the page.

But there are several options in BuddyPress which should probably be set if you want to do anything with the site. For instance, you can setup forums through the bbPress installation which is included in BuddyPress. So, if you’re still in the WordPress dashboard, you can do the following:

  1. Click “BuddyPress”
  2. Click “Forum Setup”
  3. Click “Set up a new bbPress installation”
  4. Click “Complete Installation”

This way, any time you create a new group, you’ll be able to add a forum to it. To do so:

  1. Click “Visit Site”
  2. Click “Groups”
  3. Click “Create a Group”
  4. Fill in the group name and description.
  5. Click “Create Group and Continue”
  6. Select whether or not you want to enable the discussion forum, choose the privacy options, and click “Next Step”
  7. Choose an avatar (or leave the default one) and click “Next Step”
  8. Click “Finish”

You now have a fully functioning group, with discussion forum.

There’s a lot of things you can now do, including all sorts of neat plugins, change the theme, etc. But this installation is fully functional and fun. So I’d encourage you to play with it, especially if you already have a group of users. The way BuddyPress is set up, you can do all sorts of things on the site itself. So you can click “Visit Site” and start creating profiles, groups, etc.

One thing with the BuddyPress Default Theme, though, is that it doesn’t make it obvious how you can come back to the Dashboard. You can do so by going to “<full domain>/<name>/wp-admin/” (adding “/wp-admin/” to your BuddyPress address). Another way, which is less obvious, but also works is to go to a blog comment (there’s one added by defaul) and click on “Edit.”

Why I Need an iPad

I’m one of those who feel the iPad is the right tool for the job.

This is mostly meant as a reply to this blogthread. But it’s also more generally about my personal reaction to Apple’s iPad announcement.

Some background.

I’m an ethnographer and a teacher. I read a fair deal, write a lot of notes, and work in a variety of contexts. These days, I tend to spend a good amount of time in cafés and other public places where I like to work without being too isolated. I also commute using public transit, listen to lots of podcast, and create my own. I’m also very aural.

I’ve used a number of PDAs, over the years, from a Newton MessagePad 130 (1997) to a variety of PalmOS devices (until 2008). In fact, some people readily associated me with PDA use.

As soon as I learnt about the iPod touch, I needed one. As soon as I’ve heard about the SafariPad, I wanted one. I’ve been an intense ‘touch user since the iPhone OS 2.0 release and I’m a happy camper.

(A major reason I never bought an iPhone, apart from price, is that it requires a contract.)

In my experience, the ‘touch is the most appropriate device for all sorts of activities which are either part of an other activity (reading during a commute) or are simply too short in duration to constitute an actual “computer session.” You don’t “sit down to work at your ‘touch” the way you might sit in front of a laptop or desktop screen. This works great for “looking up stufff” or “checking email.” It also makes a lot of sense during commutes in crowded buses or metros.

In those cases, the iPod touch is almost ideal. Ubiquitous access to Internet would be nice, but that’s not a deal-breaker. Alternative text-input methods would help in some cases, but I do end up being about as fast on my ‘touch as I was with Graffiti on PalmOS.

For other tasks, I have a Mac mini. Sure, it’s limited. But it does the job. In fact, I have no intention of switching for another desktop and I even have an eMachines collecting dust (it’s too noisy to make a good server).

What I miss, though, is a laptop. I used an iBook G3 for several years and loved it. For a little while later, I was able to share a MacBook with somebody else and it was a wonderful experience. I even got to play with the OLPC XO for a few weeks. That one was not so pleasant an experience but it did give me a taste for netbooks. And it made me think about other types of iPhone-like devices. Especially in educational contexts. (As I mentioned, I’m a teacher)

I’ve been laptop-less for a while, now. And though my ‘touch replaces it in many contexts, there are still times when I’d really need a laptop. And these have to do with what I might call “mobile sessions.”

For instance: liveblogging a conference or meeting. I’ve used my ‘touch for this very purpose on a good number of occasions. But it gets rather uncomfortable, after a while, and it’s not very fast. A laptop is better for this, with a keyboard and a larger form factor. But the iPad will be even better because of lower risks of RSI. A related example: just imagine TweetDeck on iPad.

Possibly my favourite example of a context in which the iPad will be ideal: presentations. Even before learning about the prospect of getting iWork on a tablet, presentations were a context in which I really missed a laptop.

Sure, in most cases, these days, there’s a computer (usually a desktop running XP) hooked to a projector. You just need to download your presentation file from Slideshare, show it from Prezi, or transfer it through USB. No biggie.

But it’s not the extra steps which change everything. It’s the uncertainty. Even if it’s often unfounded, I usually get worried that something might just not work, along the way. The slides might not show the same way as you see it because something is missing on that computer or that computer is simply using a different version of the presentation software. In fact, that software is typically Microsoft PowerPoint which, while convenient, fits much less in my workflow than does Apple Keynote.

The other big thing about presentations is the “presenter mode,” allowing you to get more content than (or different content from) what the audience sees. In most contexts where I’ve used someone else’s computer to do a presentation, the projector was mirroring the computer’s screen, not using it as a different space. PowerPoint has this convenient “presenter view” but very rarely did I see it as an available option on “the computer in the room.” I wish I could use my ‘touch to drive presentations, which I could do if I installed software on that “computer in the room.” But it’s not something that is likely to happen, in most cases.

A MacBook solves all of these problems. and it’s an obvious use for laptops. But how, then, is the iPad better? Basically because of interface. Switching slides on a laptop isn’t hard, but it’s more awkward than we realize. Even before watching the demo of Keynote on the iPad, I could simply imagine the actual pleasure of flipping through slides using a touch interface. The fit is “natural.”

I sincerely think that Keynote on the iPad will change a number of things, for me. Including the way I teach.

Then, there’s reading.

Now, I’m not one of those people who just can’t read on a computer screen. In fact, I even grade assignments directly from the screen. But I must admit that online reading hasn’t been ideal, for me. I’ve read full books as PDF files or dedicated formats on PalmOS, but it wasn’t so much fun, in terms of the reading process. And I’ve used my ‘touch to read things through Stanza or ReadItLater. But it doesn’t work so well for longer reading sessions. Even in terms of holding the ‘touch, it’s not so obvious. And, what’s funny, even a laptop isn’t that ideal, for me, as a reading device. In a sense, this is when the keyboard “gets in the way.”

Sure, I could get a Kindle. I’m not a big fan of dedicated devices and, at least on paper, I find the Kindle a bit limited for my needs. Especially in terms of sources. I’d like to be able to use documents in a variety of formats and put them in a reading list, for extended reading sessions. No, not “curled up in bed.” But maybe lying down in a sofa without external lighting. Given my experience with the ‘touch, the iPad is very likely the ideal device for this.

Then, there’s the overall “multi-touch device” thing. People have already been quite creative with the small touchscreen on iPhones and ‘touches, I can just imagine what may be done with a larger screen. Lots has been said about differences in “screen real estate” in laptop or desktop screens. We all know it can make a big difference in terms of what you can display at the same time. In some cases, two screens isn’t even a luxury, for instance when you code and display a page at the same time (LaTeX, CSS…). Certainly, the same qualitative difference applies to multitouch devices. Probably even more so, since the display is also used for input. What Han found missing in the iPhone’s multitouch was the ability to use both hands. With the iPad, Han’s vision is finding its space.

Oh, sure, the iPad is very restricted. For instance, it’s easy to imagine how much more useful it’d be if it did support multitasking with third-party apps. And a front-facing camera is something I was expecting in the first iPhone. It would just make so much sense that a friend seems very disappointed by this lack of videoconferencing potential. But we’re probably talking about predetermined expectations, here. We’re comparing the iPad with something we had in mind.

Then, there’s the issue of the competition. Tablets have been released and some multitouch tablets have recently been announced. What makes the iPad better than these? Well, we could all get in the same OS wars as have been happening with laptops and desktops. In my case, the investment in applications, files, and expertise that I have made in a Mac ecosystem rendered my XP years relatively uncomfortable and me appreciate returning to the Mac. My iPod touch fits right in that context. Oh, sure, I could use it with a Windows machine, which is in fact what I did for the first several months. But the relationship between the iPhone OS and Mac OS X is such that using devices in those two systems is much more efficient, in terms of my own workflow, than I could get while using XP and iPhone OS. There are some technical dimensions to this, such as the integration between iCal and the iPhone OS Calendar, or even the filesystem. But I’m actually thinking more about the cognitive dimensions of recognizing some of the same interface elements. “Look and feel” isn’t just about shiny and “purty.” It’s about interactions between a human brain, a complex sensorimotor apparatus, and a machine. Things go more quickly when you don’t have to think too much about where some tools are, as you’re working.

So my reasons for wanting an iPad aren’t about being dazzled by a revolutionary device. They are about the right tool for the job.