Category Archives: technology

Selling Myself Long

Been attending sessions by Meri Aaron Walker about online methods to get paid for our expertise. Meri coaches teachers about those issues.

MAWSTOOLBOX.COM

There’s also a LearnHub “course”: Jumpstart Your Online Teaching Career.

Some notes, on my own thinking about monetization of expertise. Still draft-like, but RERO is my battle cry.

Some obstacles to my selling expertise:

  • My “oral personality.”
  • The position on open/free knowledge in academia and elsewhere.
  • My emphasis on friendship and personal rapport.
  • My abilities as an employee instead of a “boss.”
  • Difficulty in assessing the value of my expertise.
  • The fact that other people have the same expertise that I think I have.
  • High stakes (though this can be decreased, in some contexts).
  • My distaste for competition/competitiveness.
  • Difficulty at selling and advertising myself (despite my social capital).
  • Being a creative generalist instead of a specialist.

Despite all these obstacles, I have been thinking about selling my services online.

One reason is that I really do enjoy teaching. As I keep saying, teaching is my hobby (when I get paid, it’s to learn how to interact with other learners and to set up learning contexts).

In fact, I enjoy almost everything in teaching (the major exception being grading/evaluating). From holding office hours and lecturing to facilitating discussions and answering questions through email. Teaching, for me, is deeply satisfying and I think that learning situations which imply the role of a teacher still make a lot of sense. I also like more informal learning situations and I even try to make my courses more similar to informal teaching. But I still find specific value in a “teaching and learning” system.

Some people seem to assume that teaching a course is the same thing as “selling expertise.” My perspective on learning revolves to a large extent on the difference between teaching and “selling expertise.” One part is that I find a difference between selling a product or process and getting paid in a broader transaction which does involve exchange about knowledge but which isn’t restricted to that exchange. Another part is that I don’t see teachers as specialists imparting their wisdom to eager masses. I see knowledge as being constructed in diverse situations, including formal and informal learning. Expertise is often an obstacle in the kind of teaching I’m interested in!

Funnily enough, I don’t tend to think of expertise as something that is easily measurable or transmissible. Those who study expertise have ways to assess something which is related to “being an expert,” especially in the case of observable skills (many of those are about “playing,” actually: chess, baseball, piano…). My personal perspective on expertise tends to be broader, more fluid. Similar to experience, but with more of a conscious approach to learning.

There also seems to be a major difference between “breadth of expertise” and “topics you can teach.” You don’t necessarily need to be very efficient at some task to help someone learn to do it. In fact, in some cases, being proficient in a domain is an obstacle to teaching in that domain, since expertise is so ingrained as to be very difficult to retrieve consciously.

This is close to “do what I say, not what I do.” I even think that it can be quite effective to actually instruct people without direct experience of these instructions. Similar to consulting, actually. Some people easily disagree with this point and some people tease teachers about “doing vs. teaching.” But we teachers do have a number of ways to respond, some of them snarkier than others. And though I disagree with several parts of his attitude, I quite like this short monologue by Taylor Mali about What Teachers Make.

Another reason I might “sell my expertise” is that I genuinely enjoy sharing my expertise. I usually provide it for free, but I can possibly relate to the value argument. I don’t feel so tied to social systems based on market economy (socialist, capitalist, communist…) but I have to make do.

Another link to “selling expertise” is more disciplinary. As an ethnographer, I enjoy being a “cultural translator.” of sorts. And, in some cases, my expertise in some domains is more of a translation from specialized speech into laypeople’s terms. I’m actually not very efficient at translating utterances from one language to another. But my habit of navigating between different “worlds” makes it possible for me to bridge gaps, cross bridges, serve as mediator, explain something fairly “esoteric” to an outsider. Close to popularization.

So, I’ve been thinking about what can be paid in such contexts which give prominence to expertise. Tutoring, homework help, consulting, coaching, advice, recommendation, writing, communicating, producing content…

And, finally, I’ve been thinking about my domains of expertise. As a “Jack of All Trades,” I can list a lot of those. My level of expertise varies greatly between them and I’m clearly a “Master of None.” In fact, some of them are merely from personal experience or even anecdotal evidence. Some are skills I’ve been told I have. But I’d still feel comfortable helping others with all of them.

I’m funny that way.

Domains of  Expertise

French

  • Conversation
  • Reading
  • Writing
  • Culture
  • Literature
  • Regional diversity
  • Chanson appreciation

Bamanan (Bambara)

  • Greetings
  • Conversation

Social sciences

  • Ethnographic disciplines
  • Ethnographic field research
  • Cultural anthropology
  • Linguistic anthropology
  • Symbolic anthropology
  • Ethnomusicology
  • Folkloristics

Semiotics

Language studies

  • Language description
  • Social dimensions of language
  • Language change
  • Field methods

Education

  • Critical thinking
  • Lifelong learning
  • Higher education
  • Graduate school
  • Graduate advising
  • Academia
  • Humanities
  • Social sciences
  • Engaging students
  • Getting students to talk
  • Online teaching
  • Online tools for teaching

Course Management Systems (Learning Management Systems)

  • Oncourse
  • Sakai
  • WebCT
  • Blackboard
  • Moodle

Social networks

  • Network ethnography
  • Network analysis
  • Influence management

Web platforms

  • Facebook
  • MySpace
  • Ning
  • LinkedIn
  • Twitter
  • Jaiku
  • YouTube
  • Flickr

Music

  • Cultural dimensions of music
  • Social dimensions of music
  • Musicking
  • Musical diversity
  • Musical exploration
  • Classical saxophone
  • Basic music theory
  • Musical acoustics
  • Globalisation
  • Business models for music
  • Sound analysis
  • Sound recording

Beer

  • Homebrewing
  • Brewing techniques
  • Recipe formulation
  • Finding ingredients
  • Appreciation
  • Craft beer culture
  • Brewing trends
  • Beer styles
  • Brewing software

Coffee

  • Homeroasting
  • Moka pot brewing
  • Espresso appreciation
  • Coffee fundamentals
  • Global coffee trade

Social media

Blogging

  • Diverse uses of blogging
  • Writing tricks
  • Workflow
  • Blogging platforms

Podcasts

  • Advantages of podcasts
  • Podcasts in teaching
  • Filming
  • Finding podcasts
  • Embedding content

Technology

  • Trends
  • Geek culture
  • Equipment
  • Beta testing
  • Troubleshooting Mac OS X

Online Life

Communities

  • Mailing-lists
  • Generating discussions
  • Entering communities
  • Building a sense of community
  • Diverse types of communities
  • Community dynamics
  • Online communities

Food

  • Enjoying food
  • Cooking
  • Baking
  • Vinaigrette
  • Pizza dough
  • Bread

Places

  • Montreal, Qc
  • Lausanne, VD
  • Bamako, ML
  • Bloomington, IN
  • Moncton, NB
  • Austin, TX
  • South Bend, IN
  • Fredericton, NB
  • Northampton, MA

Pedestrianism

  • Carfree living
  • Public transportation
  • Pedestrian-friendly places

Tools I Use

  • PDAs
  • iPod
  • iTunes
  • WordPress.com
  • Skype
  • Del.icio.us
  • Diigo
  • Blogger (Blogspot)
  • Mac OS X
  • Firefox
  • Flock
  • Internet Explorer
  • Safari
  • Gmail
  • Google Calendar
  • Google Maps
  • Zotero
  • Endnote
  • RefWorks
  • Zoho Show
  • Wikipedia
  • iPod touch
  • SMS
  • Outlining
  • PowerPoint
  • Slideshare
  • Praat
  • Audacity
  • Nero Express
  • Productivity software

Effective Web searches

Socialization

  • Social capital
  • Entering the field
  • Creating rapport
  • Event participation
  • Event hosting

Computer Use

  • Note-taking
  • Working with RSS feeds
  • Basic programing concepts
  • Data manipulations

Research Methods

  • Open-ended interviewing
  • Qualitative data analysis

Personal

  • Hedonism
  • Public speaking
  • GERD
  • Strabismus
  • Moving
  • Cultural awareness

The Need for Social Science in Social Web/Marketing/Media (Draft)

[Been sitting on this one for a little while. Better RERO it, I guess.]

Sticking My Neck Out (Executive Summary)

I think that participants in many technology-enthusiastic movements which carry the term “social” would do well to learn some social science. Furthermore, my guess is that ethnographic disciplines are very well-suited to the task of teaching participants in these movements something about social groups.

Disclaimer

Despite the potentially provocative title and my explicitly stating a position, I mostly wish to think out loud about different things which have been on my mind for a while.

I’m not an “expert” in this field. I’m just a social scientist and an ethnographer who has been observing a lot of things online. I do know that there are many experts who have written many great books about similar issues. What I’m saying here might not seem new. But I’m using my blog as a way to at least write down some of the things I have in mind and, hopefully, discuss these issues thoughtfully with people who care.

Also, this will not be a guide on “what to do to be social-savvy.” Books, seminars, and workshops on this specific topic abound. But my attitude is that every situation needs to be treated in its own context, that cookie-cutter solutions often fail. So I would advise people interested in this set of issues to train themselves in at least a little bit of social science, even if much of the content of the training material seems irrelevant. Discuss things with a social scientist, hire a social scientist in your business, take a course in social science, and don’t focus on advice but on the broad picture. Really.

Clarification

Though they are all different, enthusiastic participants in “social web,” “social marketing,” “social media,” and other “social things online” do have some commonalities. At the risk of angering some of them, I’m lumping them all together as “social * enthusiasts.” One thing I like about the term “enthusiast” is that it can apply to both professional and amateurs, to geeks and dabblers, to full-timers and part-timers. My target isn’t a specific group of people. I just observed different things in different contexts.

Links

Shameless Self-Promotion

A few links from my own blog, for context (and for easier retrieval):

Shameless Cross-Promotion

A few links from other blogs, to hopefully expand context (and for easier retrieval):

Some raw notes

  • Insight
  • Cluefulness
  • Openness
  • Freedom
  • Transparency
  • Unintended uses
  • Constructivism
  • Empowerment
  • Disruptive technology
  • Innovation
  • Creative thinking
  • Critical thinking
  • Technology adoption
  • Early adopters
  • Late adopters
  • Forced adoption
  • OLPC XO
  • OLPC XOXO
  • Attitudes to change
  • Conservatism
  • Luddites
  • Activism
  • Impatience
  • Windmills and shelters
  • Niche thinking
  • Geek culture
  • Groupthink
  • Idea horizon
  • Intersubjectivity
  • Influence
  • Sphere of influence
  • Influence network
  • Social butterfly effect
  • Cog in a wheel
  • Social networks
  • Acephalous groups
  • Ego-based groups
  • Non-hierarchical groups
  • Mutual influences
  • Network effects
  • Risk-taking
  • Low-stakes
  • Trial-and-error
  • Transparency
  • Ethnography
  • Epidemiology of ideas
  • Neural networks
  • Cognition and communication
  • Wilson and Sperber
  • Relevance
  • Global
  • Glocal
  • Regional
  • City-State
  • Fluidity
  • Consensus culture
  • Organic relationships
  • Establishing rapport
  • Buzzwords
  • Viral
  • Social
  • Meme
  • Memetic marketplace
  • Meta
  • Target audience

Let’s Give This a Try

The Internet is, simply, a network. Sure, technically it’s a meta-network, a network of networks. But that is pretty much irrelevant, in social terms, as most networks may be analyzed at different levels as containing smaller networks or being parts of larger networks. The fact remains that the ‘Net is pretty easy to understand, sociologically. It’s nothing new, it’s just a textbook example of something social scientists have been looking at for a good long time.

Though the Internet mostly connects computers (in many shapes or forms, many of them being “devices” more than the typical “personal computer”), the impact of the Internet is through human actions, behaviours, thoughts, and feelings. Sure, we can talk ad nauseam about the technical aspects of the Internet, but these topics have been covered a lot in the last fifteen years of intense Internet growth and a lot of people seem to be ready to look at other dimensions.

The category of “people who are online” has expanded greatly, in different steps. Here, Martin Lessard’s description of the Internet’s Six Cultures (Les 6 cultures d’Internet) is really worth a read. Martin’s post is in French but we also had a blog discussion in English, about it. Not only are there more people online but those “people who are online” have become much more diverse in several respects. At the same time, there are clear patterns on who “online people” are and there are clear differences in uses of the Internet.

Groups of human beings are the very basic object of social science. Diversity in human groups is the very basis for ethnography. Ethnography is simply the description of (“writing about”) human groups conceived as diverse (“peoples”). As simple as ethnography can be, it leads to a very specific approach to society which is very compatible with all sorts of things relevant to “social * enthusiasts” on- and offline.

While there are many things online which may be described as “media,” comparing the Internet to “The Mass Media” is often the best way to miss “what the Internet is all about.” Sure, the Internet isn’t about anything (about from connecting computers which, in turn, connect human beings). But to get actual insight into the ‘Net, one probably needs to free herself/himself of notions relating to “The Mass Media.” Put bluntly, McLuhan was probably a very interesting person and some of his ideas remain intriguing but fallacies abound in his work and the best thing to do with his ideas is to go beyond them.

One of my favourite examples of the overuse of “media”-based concepts is the issue of influence. In blogging, podcasting, or selling, the notion often is that, on the Internet as in offline life, “some key individuals or outlets are influential and these are the people by whom or channels through which ideas are disseminated.” Hence all the Technorati rankings and other “viewer statistics.” Old techniques and ideas from the times of radio and television expansion are used because it’s easier to think through advertising models than through radically new models. This is, in fact, when I tend to bring back my explanation of the “social butterfly effect“: quite frequently, “influence” online isn’t through specific individuals or outlets but even when it is, those people are influential through virtue of connecting to diverse groups, not by the number of people they know. There are ways to analyze those connections but “measuring impact” is eventually missing the point.

Yes, there is an obvious “qual. vs. quant.” angle, here. A major distinction between non-ethnographic and ethnographic disciplines in social sciences is that non-ethnographic disciplines tend to be overly constrained by “quantitative analysis.” Ultimately, any analysis is “qualitative” but “quantitative methods” are a very small and often limiting subset of the possible research and analysis methods available. Hence the constriction and what some ethnographers may describe as “myopia” on the part of non-ethnographers.

Gone Viral

The term “viral” is used rather frequently by “social * enthusiasts” online. I happen to think that it’s a fairly fitting term, even though it’s used more by extension than by literal meaning. To me, it relates rather directly to Dan Sperber’s “epidemiological” treatment of culture (see Explaining Culture) which may itself be perceived as resembling Dawkins’s well-known “selfish gene” ideas made popular by different online observers, but with something which I perceive to be (to use simple semiotic/semiological concepts) more “motivated” than the more “arbitrary” connections between genetics and ideas. While Sperber could hardly be described as an ethnographer, his anthropological connections still make some of his work compatible with ethnographic perspectives.

Analysis of the spread of ideas does correspond fairly closely with the spread of viruses, especially given the nature of contacts which make transmission possible. One needs not do much to spread a virus or an idea. This virus or idea may find “fertile soil” in a given social context, depending on a number of factors. Despite the disadvantages of extending analogies and core metaphors too far, the type of ecosystem/epidemiology analysis of social systems embedded in uses of the term “viral” do seem to help some specific people make sense of different things which happen online. In “viral marketing,” the type of informal, invisible, unexpected spread of recognition through word of mouth does relate somewhat to the spread of a virus. Moreover, the metaphor of “viral marketing” is useful in thinking about the lack of control the professional marketer may have on how her/his product is perceived. In this context, the term “viral” seems useful.

The Social

While “viral” seems appropriate, the even more simple “social” often seems inappropriately used. It’s not a ranty attitude which makes me comment negatively on the use of the term “social.” In fact, I don’t really care about the use of the term itself. But I do notice that use of the term often obfuscates what is the obvious social character of the Internet.

To a social scientist, anything which involves groups is by definition “social.” Of course, some groups and individuals are more gregarious than others, some people are taken to be very sociable, and some contexts are more conducive to heightened social interactions. But social interactions happen in any context.
As an example I used (in French) in reply to this blog post, something as common as standing in line at a grocery store is representative of social behaviour and can be analyzed in social terms. Any Web page which is accessed by anyone is “social” in the sense that it establishes some link, however tenuous and asymmetric, between at least two individuals (someone who created the page and the person who accessed that page). Sure, it sounds like the minimal definition of communication (sender, medium/message, receiver). But what most people who talk about communication seem to forget (unlike Jakobson), is that all communication is social.

Sure, putting a comment form on a Web page facilitates a basic social interaction, making the page “more social” in the sense of “making that page easier to use explicit social interaction.” And, of course, adding some features which facilitate the act of sharing data with one’s personal contacts is a step above the contact form in terms of making certain type of social interaction straightforward and easy. But, contrary to what Google Friend Connect implies, adding those features doesn’t suddenly make the site social. The site itself isn’t really social and, assuming some people visited it, there was already a social dimension to it. I’m not nitpicking on word use. I’m saying that using “social” in this way may blind some people to social dimensions of the Internet. And the consequences can be pretty harsh, in some cases, for overlooking how social the ‘Net is.

Something similar may be said about the “Social Web,” one of the many definitions of “Web 2.0” which is used in some contexts (mostly, the cynic would say, “to make some tool appear ‘new and improved'”). The Web as a whole was “social” by definition. Granted, it lacked the ease of social interaction afforded such venerable Internet classics as Usenet and email. But it was already making some modes of social interaction easier to perceive. No, this isn’t about “it’s all been done.” It’s about being oblivious to the social potential of tools which already existed. True, the period in Internet history known as “Web 2.0” (and the onset of the Internet’s sixth culture) may be associated with new social phenomena. But there is little evidence that the association is causal, that new online tools and services created a new reality which suddenly made it possible for people to become social online. This is one reason I like Martin Lessard’s post so much. Instead of postulating the existence of a brand new phenomenon, he talks about the conditions for some changes in both Internet use and the form the Web has taken.

Again, this isn’t about terminology per se. Substitute “friendly” for “social” and similar issues might come up (friendship and friendliness being disconnected from the social processes which underline them).

Adoptive Parents

Many “social * enthusiasts” are interested in “adoption.” They want their “things” to be adopted. This is especially visible among marketers but even in social media there’s an issue of “getting people on board.” And some people, especially those without social science training, seem to be looking for a recipe.

Problem is, there probably is no such thing as a recipe for technology adoption.

Sure, some marketing practises from the offline world may work online. Sometimes, adapting a strategy from the material world to the Internet is very simple and the Internet version may be more effective than the offline version. But it doesn’t mean that there is such a thing as a recipe. It’s a matter of either having some people who “have a knack for this sort of things” (say, based on sensitivity to what goes on online) or based on pure luck. Or it’s a matter of measuring success in different ways. But it isn’t based on a recipe. Especially not in the Internet sphere which is changing so rapidly (despite some remarkably stable features).

Again, I’m partial to contextual approaches (“fully-customized solutions,” if you really must). Not just because I think there are people who can do this work very efficiently. But because I observe that “recipes” do little more than sell “best-selling books” and other items.

So, what can we, as social scientists, say about “adoption?” That technology is adopted based on the perceived fit between the tools and people’s needs/wants/goals/preferences. Not the simple “the tool will be adopted if there’s a need.” But a perception that there might be a fit between an amorphous set of social actors (people) and some well-defined tools (“technologies”). Recognizing this fit is extremely difficult and forcing it is extremely expensive (not to mention completely unsustainable). But social scientists do help in finding ways to adapt tools to different social situations.

Especially ethnographers. Because instead of surveys and focus groups, we challenge assumptions about what “must” fit. Our heads and books are full of examples which sound, in retrospect, as common sense but which had stumped major corporations with huge budgets. (Ask me about McDonald’s in Brazil or browse a cultural anthropology textbook, for more information.)

Recently, while reading about issues surrounding the OLPC’s original XO computer, I was glad to read the following:

John Heskett once said that the critical difference between invention and innovation was its mass adoption by users. (Niti Bhan The emperor has designer clothes)

Not that this is a new idea, for social scientists. But I was glad that the social dimension of technology adoption was recognized.

In marketing and design spheres especially, people often think of innovation as individualized. While some individuals are particularly adept at leading inventions to mass adoption (Steve Jobs being a textbook example), “adoption comes from the people.” Yes, groups of people may be manipulated to adopt something “despite themselves.” But that kind of forced adoption is still dependent on a broad acceptance, by “the people,” of even the basic forms of marketing. This is very similar to the simplified version of the concept of “hegemony,” so common in both social sciences and humanities. In a hegemony (as opposed to a totalitarian regime), no coercion is necessary because the logic of the system has been internalized by people who are affected by it. Simple, but effective.

In online culture, adept marketers are highly valued. But I’m quite convinced that pre-online marketers already knew that they had to “learn society first.” One thing with almost anything happening online is that “the society” is boundless. Country boundaries usually make very little sense and the social rules of every local group will leak into even the simplest occasion. Some people seem to assume that the end result is a cultural homogenization, thereby not necessitating any adaptation besides the move from “brick and mortar” to online. Others (or the same people, actually) want to protect their “business models” by restricting tools or services based on country boundaries. In my mind, both attitudes are ineffective and misleading.

Sometimes I Feel Like a Motherless Child

I think the Cluetrain Manifesto can somehow be summarized through concepts of freedom, openness, and transparency. These are all very obvious (in French, the book title is something close to “the evident truths manifesto”). They’re also all very social.

Social scientists often become activists based on these concepts. And among social scientists, many of us are enthusiastic about the social changes which are happening in parallel with Internet growth. Not because of technology. But because of empowerment. People are using the Internet in their own ways, the one key feature of the Internet being its lack of centralization. While the lack of centralized control may be perceived as a “bad thing” by some (social scientists or not), there’s little argument that the ‘Net as a whole is out of the control of specific corporations or governments (despite the large degree of consolidation which has happened offline and online).

Especially in the United States, “freedom” is conceived as a basic right. But it’s also a basic concept in social analysis. As some put it: “somebody’s rights end where another’s begin.” But social scientists have a whole apparatus to deal with all the nuances and subtleties which are bound to come from any situation where people’s rights (freedom) may clash or even simply be interpreted differently. Again, not that social scientists have easy, ready-made answers on these issues. But we’re used to dealing with them. We don’t interpret freedom as a given.

Transparency is fairly simple and relates directly to how people manage information itself (instead of knowledge or insight). Radical transparency is giving as much information as possible to those who may need it. Everybody has a “right to learn” a lot of things about a given institution (instead of “right to know”), when that institution has a social impact. Canada’s Access to Information Act is quite representative of the move to transparency and use of this act has accompanied changes in the ways government officials need to behave to adapt to a relatively new reality.

Openness is an interesting topic, especially in the context of the so-called “Open Source” movement. Radical openness implies participation by outsiders, at least in the form of verbal feedback. The cluefulness of “opening yourself to your users” is made obvious in the context of successes by institutions which have at least portrayed themselves as open. What’s in my mind unfortunate is that many institutions now attempt to position themselves on the openness end of the “closed/proprietary to open/responsive” scale without much work done to really open themselves up.

Communitas

Mottoes, slogans, and maxims like “build it and they will come,” “there’s a sucker born every minute,” “let them have cake,” and “give them what they want” all fail to grasp the basic reality of social life: “they” and “we” are linked. We’re all different and we’re all connected. We all take parts in groups. These groups are all associated with one another. We can’t simply behave the same way with everyone. Identity has two parts: sense of belonging (to an “in-group”) and sense of distinction (from an “out-group”). “Us/Them.”

Within the “in-group,” if there isn’t any obvious hierarchy, the sense of belonging can take the form that Victor Turner called “communitas” and which happens in situations giving real meaning to the notion of “community.” “Community of experience,” “community of practise.” Eckert and Wittgenstein brought to online networks. In a community, contacts aren’t always harmonious. But people feel they fully belong. A network isn’t the same thing as a community.

The World Is My Oyster

Despite the so-called “Digital Divide” (or, more precisely, the maintenance online of global inequalities), the ‘Net is truly “Global.” So is the phone, now that cellphones are accomplishing the “leapfrog effect.” But this one Internet we have (i.e., not Internet2 or other such specialized meta-network) is reaching everywhere through a single set of compatible connections. The need for cultural awareness is increased, not alleviated by online activities.

Release Early, Release Often

Among friends, we call it RERO.

The RERO principle is a multiple-pass system. Instead of waiting for the right moment to release a “perfect product” (say, a blogpost!), the “work in progress” is provided widely, garnering feedback which will be integrated in future “product versions.” The RERO approach can be unnerving to “product developers,” but it has proved its value in online-savvy contexts.

I use “product” in a broad sense because the principle applies to diverse contexts. Furthermore, the RERO principle helps shift the focus from “product,” back into “process.”

The RERO principle may imply some “emotional” or “psychological” dimensions, such as humility and the acceptance of failure. At some level, differences between RERO and “trial-and-error” methods of development appear insignificant. Those who create something should not expect the first try to be successful and should recognize mistakes to improve on the creative process and product. This is similar to the difference between “rehearsal” (low-stakes experimentation with a process) and “performance” (with responsibility, by the performer, for evaluation by an audience).

Though applications of the early/often concept to social domains are mostly satirical, there is a social dimension to the RERO principle. Releasing a “product” implies a group, a social context.

The partial and frequent “release” of work to “the public” relates directly to openness and transparency. Frequent releases create a “relationship” with human beings. Sure, many of these are “Early Adopters” who are already overrepresented. But the rapport established between an institution and people (users/clients/customers/patrons…) can be transfered more broadly.

Releasing early seems to shift the limit between rehearsal and performance. Instead of being able to do mistakes on your own, your mistakes are shown publicly and your success is directly evaluated. Yet a somewhat reverse effect can occur: evaluation of the end-result becomes a lower-stake rating at different parts of the project because expectations have shifted to the “lower” end. This is probably the logic behind Google’s much discussed propensity to call all its products “beta.”

While the RERO principle does imply a certain openness, the expectation that each release might integrate all the feedback “users” have given is not fundamental to releasing early and frequently. The expectation is set by a specific social relationship between “developers” and “users.” In geek culture, especially when users are knowledgeable enough about technology to make elaborate wishlists, the expectation to respond to user demand can be quite strong, so much so that developers may perceive a sense of entitlement on the part of “users” and grow some resentment out of the situation. “If you don’t like it, make it yourself.” Such a situation is rather common in FLOSS development: since “users” have access to the source code, they may be expected to contribute to the development project. When “users” not only fail to fulfil expectations set by open development but even have the gumption to ask developers to respond to demands, conflicts may easily occur. And conflicts are among the things which social scientists study most frequently.

Putting the “Capital” Back into “Social Capital”

In the past several years, ”monetization” (transforming ideas into currency) has become one of the major foci of anything happening online. Anything which can be a source of profit generates an immediate (and temporary) “buzz.” The value of anything online is measured through typical currency-based economics. The relatively recent movement toward ”social” whatever is not only representative of this tendency, but might be seen as its climax: nowadays, even social ties can be sold directly, instead of being part of a secondary transaction. As some people say “The relationship is the currency” (or “the commodity,” or “the means to an end”). Fair enough, especially if these people understand what social relationships entail. But still strange, in context, to see people “selling their friends,” sometimes in a rather literal sense, when social relationships are conceived as valuable. After all, “selling the friend” transforms that relationship, diminishes its value. Ah, well, maybe everyone involved is just cynical. Still, even their cynicism contributes to the system. But I’m not judging. Really, I’m not. I’m just wondering
Anyhoo, the “What are you selling anyway” question makes as much sense online as it does with telemarketers and other greed-focused strangers (maybe “calls” are always “cold,” online). It’s just that the answer isn’t always so clear when the “business model” revolves around creating, then breaking a set of social expectations.
Me? I don’t sell anything. Really, not even my ideas or my sense of self. I’m just not good at selling. Oh, I do promote myself and I do accumulate social capital. As social butterflies are wont to do. The difference is, in the case of social butterflies such as myself, no money is exchanged and the social relationships are, hopefully, intact. This is not to say that friends never help me or never receive my help in a currency-friendly context. It mostly means that, in our cases, the relationships are conceived as their own rewards.
I’m consciously not taking the moral high ground, here, though some people may easily perceive this position as the morally superior one. I’m not even talking about a position. Just about an attitude to society and to social relationships. If you will, it’s a type of ethnographic observation from an insider’s perspective.

Makes sense?

Touch Thoughts: Apple's Handheld Strategy

I’m still on the RDF.
Apple‘s March 6, 2008 event was about enterprise and development support for its iPhone and iPod touch lines of handheld devices. Lots to think about.

(For convenience’s sake, I’ll lump together the iPod touch and the iPhone under the name “Touch,” which seems consistent with Apple’s “Cocoa Touch.”)

Been reading a fair bit about this event. Interesting reactions across the board.

My own thoughts on the whole thing.
I appreciate the fact that Phil Schiller began the “enterprise” section of the event with comments about a university. Though universities need not be run like profit-hungry corporations, linking Apple’s long-standing educational focus with its newly invigorated enterprise focus makes sense. And I had a brief drift-off moment as I was thinking about Touch products in educational contexts.

I’m surprised at how enthusiastic I get about the enterprise features. Suddenly, I can see Microsoft’s Exchange make sense.

I get the clear impression that even more things will come into place at the end of June than has been said by Apple. Possibly new Touch models or lines. Probably the famous 3G iPhone. Apple-released apps. Renewed emphasis on server technology (XServe, Mac OS X Server, XSan…). New home WiFi products (AirPort, Time Capsule, Apple TV…). New partnerships. Cool VC-funded startups. New features on the less aptly named “iTunes” store.

Though it was obvious already, the accelerometer is an important feature. It seems especially well-adapted to games and casual gamers like myself are likely to enjoy games this feature makes possible. It can also lead to very interesting applications. In fact, the “Etch and Sketch” demo was rather convincing as a display of some core Touch features. These are exactly the features which help sell products.
Actually, I enjoyed the “wow factor” of the event’s demos. I’m convinced that it will energize developers and administrators, whether or not they plan on using Touch products. Some components of Apple’s Touch strategy are exciting enough that the more problematic aspects of this strategy may matter a bit less. Those of us dreaming about Android, OpenMoko, or even a revived NewtonOS can still find things to get inspired by in Apple’s roadmap.

What’s to come, apart from what was announced? No idea. But I do daydream about all of this.
I’m especially interested in the idea of Apple Touch as “mainstream, WiFi, mobile platform.” There’s a lot of potential for Apple-designed, WiFi-enabled handhelds. Whether or not they include a cellphone.
At this point, Apple only makes five models of Touch products: three iPod touches and two iPhones. Flash memory is the main differentiating factor within a line. It makes it relatively easy to decide which device to get but some product diversity could be interesting. While some people expect/hope that Apple will release radically new form factors for Touch devices (e.g., a tablet subnotebook), it’s quite likely that other features will help distinguish Apple’s Touch hardware.
Among features I’d like to get through software, add-ons, or included in a Touch product? Number of things, some alluded to in the “categories” for this post. Some of these I had already posted.

  • Quality audio recording (to make it the ideal fieldwork audio tool).
  • eBook support (to compete with Amazon’s Kindle).
  • Voice support (including continuous dictation, voice interface…).
  • Enhanced support for podcasting (interacting with podcasts, sending audio/video responses…)
  • Video conferencing (been thinking about this for a while).
  • GPS (location will be big).
  • Mesh networking (a neat feature of OLPC’s XO).
  • Mobile WiMAX (unlikely, but it could be neat).
  • Battery pack (especially for long trips in remote regions).
  • Add-on flash memory (unlikely, but it could be useful, especially for backup).
  • Offline storage of online content (likely, but worth noting).
  • Inexpensive model (especially for “emerging markets”).
  • Access to 3G data networks without cellular “voice plan” (unlikely, but worth a shot).
  • Alternative input methods (MessagEase, Graffiti, adaptive keyboard, speech recognition…).
  • Use as Mac OS X “host” (kind of like a user partition).
  • Bluetooth/WiFi data transfer (no need for cables and docks).
  • MacBook Touch (unlikely, especially with MacBook Air, but it could be fun).
  • Automatic cell to VoIP-over-WiFi switching (saving cell minutes).

Of course, there are many obvious ones which will likely be implemented in software. I’m already impressed by the Omni Group’s pledge to develop a Touch version of their flagship GTD app.

The Geek Niche (Draft)

As explained before, I am not a “visual” thinker. Unlike some other people, I don’t draw witty charts all the time. However, I do occasionally think visually. In this case, I do “see” Venn diagrams and other cutesy graphics. What I’m seeing is the proportion of “geeks” in the world. And, to be honest, it’s relatively clear for me. I may be completely off, but I still see it clearly.

Of course, much of it is about specifying what we mean by “geek.” Which isn’t easy for someone used to looking at culture’s near-chaotic intricacy and intricacies. At this point, I’m reluctant to define too clearly what I mean by “geek” because some people (self-professed geeks, especially) are such quick nitpickers that anything I say about the term is countered by more authorized definitions. I even expect comments to this blog entry to focus on how inaccurate my perception of geeks is, regardless of any other point I make.

Ah, well…

My intention isn’t to stereotype a group of people. And I don’t want to generalize. I just try to describe a specific situation which I find very interesting. In and of itself, the term “geek” carries a lot of baggage, much of which is problematic for anyone who is trying to understand an important part of the world in which we all live. But the term is remarkably useful as a way to package an ethos, a style, a perspective, an approach, a worldview, a personality type. Among those who could be called “geeks” are very diverse people. There might not even a single set of criteria to define who should legitimately be called a “geek.” But “geekness” is now a reference for some actions, behaviors, markets, and even language varieties. Describing “geeks” as a group makes some sense, even if some people get very sensitive about the ways geeks are described.

For the record, I don’t really consider myself a geek. At the same time, I do enjoy geekness and I consider myself geek-friendly. It’s just that I’m not an actual insider to the geek coterie.

Thinking demographically has some advantages in terms of simplification. Simple is reassuring, especially in geek culture. So, looking at geek demographics on a broad scale…

First, the upper demographic limit for geekery. At the extreme, the Whole Wide World. What’s geeky about The World?

Number of things, actually. Especially in terms of some key technologies. Those technologies some people call “the tech world.” Consumer electronics, digital gadgets, computers…

An obvious tech factor for the upper limit of geekness is the ‘Net. The Internet is now mainstream. Not that everyone, everywhere truly lives online but the ‘Net is having a tremendous impact on the world as a whole. And Internet penetration is shaping up, in diverse parts of the world. This type of effect goes well with a certain type of “low-level geekness.” Along with widespread online communication, a certain approach to the world has become more prominent. A techno-enthusiastic and troubleshooting approach I often associate with engineering. Not that all engineers uses this type of approach or that everyone who uses this type of approach is an engineer. But, in my mind, it’s an “engineering worldview” similar to an updated set of mechanistic metaphors.

Another obvious example of widespread geek-friendly technology is the cellphone. Obvious because extremely widespread (apparently, close to half of the human population of the planet is cellphoned). Yet, cellphones are the geekiest technology item available. What makes them geeky, in my eyes, is the way they’re embedded in a specific social dynamic emphasizing efficiency, mobility, and “always-on connectivity” along with work/life, group/individual, and public/private dichotomies.

The world’s geekiness can also be observed through other lenses, more concerned with the politic and the social drive of human behavior. Meritocracies, relatively non-judgemental ethics, post-national democracies, neo-liberal libertarianism, neo-Darwinian progress-mindedness, networked identities… Figures on populations “affected” by these geeky dimensions of socio-political life are hard to come by and it’s difficult to tell apart these elements from simple “Westernization.” But it’s easy to conceive of a geeky version of the world in which all of these elements are linked. In a way, it’s as if the world were dominated by geekdom.

Which brings me to the lower demographic limit for geekiness: How many “true geeks” are there? What’ are the figures for the “alpha geek” population?

My honest guesstimate? Five to ten million worldwide, concentrated in a relatively small number of urban areas in North America and Eurasia. I base this range on a number of hunches I got throughout the years. In fact, my impression is that there are about two million people in (or “oriented toward”) the United States who come close enough to the geek stereotype to qualify as “alpha geeks.” Haven’t looked at academic literature on the subject but judging from numbers of early adopters in “geeky tech,” looking at FLOSS movements, thinking about desktop Linux, listening to the “tech news” I don’t think this figure is so far off. On top of these U.S. geeks are “worldwide geeks” who are much harder to count. Especially since geekness itself is a culture-specific concept. But, for some reason, I get the impression that those outside the United States who would be prototypical geeks number something like five million people, plus or minus two million.

All this surely sounds specious. In fact, I’m so not a quant dude, I really don’t care about the exact figure. But my feeling, here, is that this ultra-geeky population is probably comparable to a large metropolitan area.

Of course, geeks are dispersed throughout the world. Though there are “geek meccas” like Bangalore and the San Francisco Bay Area, geeks are often modern cosmopolitans. They are typically not “of a place” and they navigate through technology institutions rather than through native locales. Thanks to telecommuting, some geeks adopt a glocal lifestyle making connections outside of their local spheres yet constructing local realities, at least in their minds. In some cases, übergeeks are resolute loners who consciously try to avoid being tied to local circles.

Thanks in part to the “tech industry” connections of geek society, geek-friendly regions compete with one another on the world stage.

Scattered geeks have an impact on local communities and this impact can be disproportionately large in comparison to the size of the geek population.

Started this post last week, after listening to Leo Laporte’s  TWiT “netcast.” 

The TWiT Netcast Network with Leo Laporte

 …

I wanted to finish this post but never got a round tuit. I wanted to connect this post with a few things about the connection between “geek culture” in the computer/tech industry and the “craft beer” and “coffee geek” movements. There was also the obvious personal connection to the subject. I’m now a decent ethnographic insider-outsider to geek culture. Despite (thanks to) the fact that, as a comment-spammer was just saying, I’m such a n00b.

Not to mention that I wanted to expand upon JoCo‘s career, attitude, and character (discussed during the TWiT podcast). And that was before I learned that JoCo himself was coming to Austin during but not through the expensive South by Southwest film/music/interactive festivals.

If I don’t stop myself, I even get the urge to talk about the politics of geek groups, especially in terms of idealism

This thoughtful blogpost questioning the usefulness of the TED conference makes me want to push the “publish” button, even though this post isn’t ready. My comments about TED aren’t too dissimilar to many of the things which have appeared in the past couple of days. But I was going to focus on the groupthink, post-Weberian neo-liberalism, Well/Wired/GBN links, techy humanitarianism, etc.

 

Ah, well… 

Guess I should just RERO it and hope for the best. Maybe I’ll be able to leave those topics behind. RSN

TBC

Free As In Beer: The Case for No-Cost Software

To summarize the situation:

  1. Most of the software for which I paid a fee, I don’t really use.
  2. Most of the software I really use, I haven’t paid a dime for.
  3. I really like no-cost software.
  4. You might want to call me “cheap” but, if you’re developing “consumer software,” you may need to pay attention to the way people like me think about software.

No, I’m not talking about piracy. Piracy is wrong on a very practical level (not to mention legal and moral issues). Piracy and anti-piracy protection are in a dynamic that I don’t particularly enjoy. In some ways, forms of piracy are “ruining it for everyone.” So this isn’t about pirated software.

I’m not talking about “Free/Libre/Open Source Software” (FLOSS) either. I tend to relate to some of the views held by advocates of “Free as in Speech” or “Open” developments but I’ve had issues with FLOSS projects, in the past. I will gladly support FLOSS in my own ways but, to be honest, I ended up losing interest in some of the most promising projects out there. Not saying they’re not worth it. After all, I do rely on many of those projects But in talking about “no-cost software,” I’m not talking about Free, Libre, or Open Source development. At least, not directly.

Basically, I was thinking about the complex equation which, for any computer user, determines the cash value of a software application. Most of the time, this equation is somehow skewed. And I end up frustrated when I pay for software and almost giddy when I find good no-cost software.

An old but representative example of my cost-software frustration: QuickTime Pro. I paid for it a number of years ago, in preparation for a fieldwork trip. It seemed like a reasonable thing to do, especially given the fact that I was going to manipulate media files. When QuickTime was updated, my license stopped working. I was basically never able to use the QuickTime Pro features. And while it’s not a huge amount of money, the frustration of having paid for something I really didn’t need left me surprisingly bitter. It was a bad decision at that time so I’m now less likely to buy software unless I really need it and I really know how I will use it.

There’s an interesting exception to my frustration with cost-software: OmniOutliner (OO). I paid for it and have used it extensively for years. When I was “forced” to switch to Windows XP, OO was possibly the piece of software I missed the most from Mac OS X. And as soon as I was able to come back to the Mac, it’s one of the first applications I installed. But, and this is probably an important indicator, I don’t really use it anymore. Not because it lacks features I found elsewhere. But because I’ve had to adapt my workflow to OO-less conditions. I still wish there were an excellent cross-platform outliner for my needs. And, no, Microsoft OneNote isn’t it.

Now, I may not be a typical user. If the term weren’t so self-aggrandizing, I’d probably call myself a “Power User.” And, as I keep saying, I am not a coder. Therefore, I’m neither the prototypical “end user” nor the stereotypical “code monkey.” I’m just someone spending inordinate amounts of time in front of computers.

One dimension of my computer behavior which probably does put me in a special niche is that I tend to like trying out new things. Even more specifically, I tend to get overly enthusiastic about computer technology to then become disillusioned by said technology. Call me a “dreamer,” if you will. Call me “naïve.” Actually, “you can call me anything you want.” Just don’t call me to sell me things. 😉

Speaking of pressure sales. In a way, if I had truckloads of money, I might be a good target for software sales. But I’d be the most demanding user ever. I’d require things to work exactly like I expect them to work. I’d be exactly what I never am in real life: a dictator.

So I’m better off as a user of no-cost software.

I still end up making feature requests, on occasion. Especially with Open Source and other open development projects. Some developers might think I’m just complaining as I’m not contributing to the code base or offering solutions to a specific usage problem. Eh.

Going back to no-cost software. The advantage isn’t really that we, users, spend less money on the software distribution itself. It’s that we don’t really need to select the perfect software solution. We can just make do with what we have. Which is a huge “value-add proposition” in terms of computer technology, as counter-intuitive as this may sound to some people.

To break down a few no-cost options.

  • Software that came with your computer. With an Eee PC, iPhone, XO, or Mac, it’s actually an important part of the complete computing experience. Sure, there are always ways to expand the software offering. But the included software may become a big part of the deal. After all, the possibilities are already endless. Especially if you have ubiquitous Internet access.
  • Software which comes through a volume license agreement. This often works for Microsoft software, at least at large educational institutions. Even if you don’t like it so much, you end up using Microsoft Office because you have it on your computer for free and it does most of the things you want to do.
  • Software coming with a plan or paid service. Including software given by ISPs. These tend not to be “worth it.” Yet the principle (or “business model,” depending on which end of the deal you’re on) isn’t so silly. You already pay for a plan of some kind, you might as well get everything you need from that plan. Nobody (not even AT&T) has done it yet in such a way that it would be to everyone’s advantage. But it’s worth a thought.
  • “Webware” and other online applications. Call it “cloud computing” if you will (it was a buzzphrase, a few days ago). And it changes a lot of things. Not only does it simplify things like backup and migration, but it often makes for a seamless computer experience. When it works really well, the browser effectively disappears and you just work in a comfortable environment where everything you need (content, tools) is “just there.” This category is growing rather rapidly at this point but many tech enthusiasts were predicting its success a number of years ago. Typical forecasting, I guess.
  • Light/demo versions. These are actually less common than they once were, especially in terms of feature differentiation. Sure, you may still play the first few levels of a game in demo version and some “express” or “lite” versions of software are still distributed for free as teaser versions of more complete software. But, like the shareware model, demo and light software may seem to have become much less prominent a part of the typical computer user’s life than just a few years ago.
  • Software coming from online services. I’m mostly thinking about Skype but it’s a software category which would include any program with a desktop component (a “download”) and an online component, typically involving some kind of individual account (free or paid). Part subscription model, part “Webware companion.” Most of Google’s software would qualify (Sketchup, Google Earth…). If the associated “retail software” were free, I wouldn’t hesitate to put WoW in this category.
  • Actual “freeware.” Much freeware could be included in other categories but there’s still an idea of a “freebie,” in software terms. Sometimes, said freeware is distributed in view of getting people’s attention. Sometimes the freeware is just the result of a developer “scratching her/his own itch.” Sometimes it comes from lapsed shareware or even lapsed commercial software. Sometimes it’s “donationware” disguised as freeware. But, if only because there’s a “freeware” category in most software catalogs, this type of no-cost software needs to be mentioned.
  • “Free/Libre/Open Source Software.” Sure, I said earlier this was not what I was really talking about. But that was then and this is now. 😉 Besides, some of the most useful pieces of software I use do come from Free Software or Open Source. Mozilla Firefox is probably the best example. But there are many other worthy programs out there, including BibDesk, TeXShop, and FreeCiv. Though, to be honest, Firefox and Flock are probably the ones I use the most.
  • Pirated software (aka “warez”). While software piracy can technically let some users avoid the cost of purchasing a piece of software, the concept is directly tied with commercial software licenses. (It’s probably not piracy if the software distribution is meant to be open.) Sure, pirates “subvert” the licensing system for commercial software. But the software category isn’t “no-cost.” To me, there’s even a kind of “transaction cost” involved in the piracy. So even if the legal and ethical issues weren’t enough to exclude pirated software from my list of no-cost software options, the very practicalities of piracy put pirated software in the costly column, not in the “no-cost” one.

With all but the last category, I end up with most (but not all) of the software solutions I need. In fact, there are ways in which I’m better served now with no-cost software than I have ever been with paid software. I should probably make a list of these, at some point, but I don’t feel like it.

I mostly felt like assessing my needs, as a computer user. And though there always are many things I wish I could do but currently can’t, I must admit that I don’t really see the need to pay for much software.

Still… What I feel I need, here, is the “ultimate device.” It could be handheld. But I’m mostly thinking about a way to get ideas into a computer-friendly format. A broad set of issues about a very basic thing.

The spark for this blog entry was a reflection about dictation software. Not only have I been interested in speech technology for quite a while but I still bet that speech (recognition/dictation and “text-to-speech”) can become the killer app. I just think that speech hasn’t “come true.” It’s there, some people use it, the societal acceptance for it is likely (given cellphone penetration most anywhere). But its moment hasn’t yet come.

No-cost “text-to-speech” (TTS) software solutions do exist but are rather impractical. In the mid-1990s, I spent fifteen months doing speech analysis for a TTS research project in Switzerland. One of the best periods in my life. Yet, my enthusiasm for current TTS systems has been dampened. I wish I could be passionate about TTS and other speech technology again. Maybe the reason I’m notis that we don’t have a “voice desktop,” yet. But, for this voice desktop (voicetop?) to happen, we need high quality, continuous speech recognition. IOW, we need a “personal dictation device.” So, my latest 2008 prediction: we will get a voice device (smartphone?) which adapts to our voices and does very efficient and very accurate transcription of our speech. (A correlated prediction: people will complain about speech technology for a while before getting used to the continuous stream of public soliloquy.)

Dictation software is typically quite costly and complicated. Most users don’t see a need for dictation software so they don’t see a need for speech technology in computing. Though I keep thinking that speech could improve my computing life, I’ve never purchased a speech processing package. Like OCR (which is also dominated by Nuance, these days) it seems to be the kind of thing which could be useful to everyone but ends up being limited to “vertical markets.” (As it so happens, I did end up being an OCR program at some point and kept hoping my life would improve as the result of being able to transform hardcopies into searchable files. But I almost never used OCR (so my frustration with cost-software continues).)

Ah, well…

Optimism From OLPC

To say the least, I’ve been ambivalent about the One Laptop Per Child project. And I was not alone in my OLPC discomfort.

But now, I feel optimistic. Not about the OLPC project. But because that project is enabling something important.

Continue reading Optimism From OLPC

Uses for PDAs

Been thinking about blogging on my use of Personal Digital Assistants (PDAs) for a little while. Here’s my chance:

There’s simply no market these days for the traditional PDA, as even basic mobile phones can do everything a PDA can do, just with more style. Report: Apple developing OS X minitablet | One More Thing – CNET News.com

Uh-oh! No you didn’t! Well, Steve Jobs made a similar statement a long time ago so it’s not like you’re the first one to say it. But you’re still wrong!

(This blog entry will be choppier than usual. Should have posted this as a comment. But this is getting longer than I expected and I prefer trackbacks anyway.)

Not exactly sure where people in the Bay Area get the impression that there is no market for the traditional PDA. In my mind, the potential market for “the traditional PDA” is underestimated because the ideal PDA has yet to be released. No, the current crop of smartphones aren’t it.

Having said this, I do think cellphones have the brightest future of pretty much any other electronic device type, but I don’t agree that any cellphone currently does what a PDA can do. So, while I think the ideal portable device would likely be a cellphone, I’d like to focus on what a PDA really is.

While it’s clear that PDAs have had a tortuous history since the first Newton and Magic Cap devices were released, other tools haven’t completely obliterated the need for PDAs. Hence the “cult following” for Newton Message Pads and the interest in new generations of PDA-like devices.

One thing to keep in mind is that PDAs are not merely PIMs (personal information managers, typically focusing on contacts and calendars). Instead of a glorified organiser, a PDA is a complete computer with a focus on personal data. And people do care about personal data in computing.

Now, a disclaimer of sorts: I’ve been an active PDA user for a number of years. When I don’t have a PDA, I almost feel like something is missing from my life.

I have been taken by the very concept of PDAs the first time I saw an article on Apple’s Newton devices in a mainstream U.S. newspaper, way back in the early 1990s. I was dreaming of all the possibilities. And longed for my own Newton MessagePad.

I received a MessagePad 130 from Apple a few years later, having done some work for them. Used that Newton for a while and really enjoyed the experience. While human beings find my handwriting extremely difficult to read, my Newton MP130 did a fairly good job at recognising it. And having installed a version of Graffiti, I was able to write rather quickly on the device. The main issue I had with Newton devices was size. The MP was simply too bulky for me to carry around everywhere. I eventually stopped using the MP after a while, but was missing the convenience of my MP130.

I started using that Newton again in 2001, as I was preparing for fieldwork. Because I didn’t have a parallel port on the iBook I was getting for fieldwork, I also bought a Handspring Visor Deluxe. The Visor became a very valuable tool during my fieldwork trip to Mali, in 2002. IIRC, this model had already been discontinued but I had no trouble using it or finding new software for it. I used the Visor to take copious amounts of data which I was able to periodically transfer to my iBook. The fact that the Visor ran on standard batteries was definitely an asset in the field but I did lose data on occasion because, unlike Newton devices, Palm devices didn’t have persistent memory storage.

Coming back from Mali, I bought my first Sony Clié. I pretty much stuck with Cliés ever since and have been quite happy with them. Cliés have a few advantages over other PalmOS devices like MemorySticks and the jog dial. The form factor and screen resolution of an entry-level Clié were much better than those of my old Visor. Sony has discontinued sales of its Clié devices outside of Japan. Some used Cliés go for 30$ on eBay.

So, what do I do with a PDA? Actually, the main thing really is taking notes. Reading notes, research notes, lecture notes, conference notes, etc. I’ve taken notes on coffee I’ve tried, on things I’d like to learn, on moments I want to write more extensively about… Though my fingers are rather small, typing on a small QWERTY keyboard has never been a real option for me. I tried using the keyboard on a Clié NX70V and it wasn’t nearly as efficient as using Graffiti. In fact, I’ve become quite adept at MessagEase. I can usually take elaborate notes in real time and organise them as I wish. Some notes remain as snippets while other notes become part of bigger pieces, including much of what I’ve written in the past ten years.

I also use my PDA for a number of “simpler” things like converting units (volume and temperature, especially), playing games (while waiting for something or while listening to podcasts), setting different timers, planning trips on public transportation systems, etc. I used to try and use more PDA applications than I do now but I still find third party applications an important component of any real PDA.

I always wanted to have a WiFi-enabled PDA. It’s probably the main reason behind my original reaction to the iPod touch launch. With a good input system and a semi-ubiquitous WiFi connection, a WiFi-enabled PDA could be a “dream come true,” for me. Especially in terms of email, blogging, and social networking. Not to mention simple Web queries.

I do have a very clear idea in mind as to what would be my ideal PDA. I don’t need it to be an MP3 player, a gaming console, or a phone. I don’t really want it to have a qwerty keyboard or a still camera. I don’t even care so much about it having a colour screen. But it should have an excellent battery life, a small size, good synchronisation features, third party apps, persistent memory, a very efficient input system, and a user community. I dream of it having a high-quality sound recorder, a webcam (think videoconferencing), large amounts of memory, and a complete set of voice features perfectly tuned to its owner’s voice (like voice activation and speaker-dependent, continuous speech recognition). It could act as the perfect unit to store any kind of personal data as a kind of “smart thumbdrive.” It could be synchronised with almost any other machine without any loss of information. It would probably have GPS and location-enabled features. It could be used to drive other systems or act as the ultimate smartcard. And it should be inexpensive.

I personally think price is one of the main reasons the traditional PDA has had such a hard time building/reaching markets. Inexpensive PDAs tended to miss important features. The most interesting PDAs were as expensive as much more powerful computers. Surely, miniaturisation is costly and it never was possible for any company to release a really inexpensive yet full-featured PDA. So it may be accurate to say that the traditional PDA is too expensive for its potential market. I perceive a huge difference between problems associated with costs and the utter lack of any PDA market.

Price does tend to be a very important factor with computer technology. The OLPC project is a good example of this. While the laptop produced through this project has many other features, the one feature which caught most of the media attention was the expected price for the device, around USD$100. All this time, many people are thinking that the project should have been a cellphone project because cellphone penetration is already high and cellphones are already the perfect leapfrog tool.

So it’s unlikely that I will get my dream PDA any time soon. Chances are, I’ll end up having to use a smartphone with very few of the features I really want my PDA to have. But, as is my impression with the OLPC project, we still need to dream and talk about what these devices can be.

Flock as Blog Editor

Now that Flock has a “spell as you type” feature, I can give it a fair shake as a blog editor.
Flock 0.9 Beta Release Notes | Flock
Thing is, though, it seems to have a problem with my main blog‘s extremely long list of categories. When I tried posting a blog entry with my WordPress.com blog account listed among the others, Flock became unresponsive as it was trying to download all my categories (all 2,751 of them). Maybe I could have waited longer but after a number of minutes without being able to use my computer at all, I decided to let it go.
Still, it made me a bit cranky. So the rest of this post will sound like a rant. It’s more like wishful thinking, though. I like wishful thinking.
I don’t like the way Flock’s blog editor handles link insertion. Sure, like any WYSIWYG editor out there, it has a button on which you can click to add an appropriate URL to text you’ve selected. But there’s no clear shortcut for this button and it could be much more powerful than it currently is.
Qumana has a better way to handle links. For one thing, it automatically inserts the clipboard’s content in the URL section of link insertion dialog box. And since it keeps published blog posts, it makes it easy to copy the “permalink” to another blog entry (for those bloggers, like me, who tend to be self-referential). Can’t remember off the top of my head but I’m pretty sure ecto and Windows Live Writer have similar features.
Ok… Flock does have this drag and drop interface for “media streams” (basically, Flickr or PhotoBucket accounts) and for “Web snippets” (local content, including text and links). Good idea and I guess I could make my “bloggable content” available to me while blogging by adding lots of content to my Flock installation and Flickr account. Makes a lot of sense for those who mostly use blogs as placeholders. But it’s still not the ideal method for blogs which rely on more extensive writing. Or for message writing.

What I want is pretty obvious but I haven’t found it yet, even in dedicated blog editors. I want my blog editor to have access to all of my links (Web history, favourites, social bookmarks…) and make it easy to work with those links while I’m writing. Sure, a “Web Snippets” feature is useful. But it still requires a fair amount of mouse movement to simply insert a link. Call me lazy but I prefer limiting my mouse movement while I’m writing.
My dream editor would integrate all of my social bookmarks, Web histories, and address books in the same interface. I could use a keystroke and start typing to get access to those links and addresses that I use frequently. Why addresses? I want to use the same editor for writing blog posts and email messages. Why not? Messages and posts end up having very similar features anyway. As many sites label it, it’s all about “sharing content.”
(I’m not really into IM but, as logic would have it, the same features should work with IM as well.)

To me, the “killer feature” in modern browsers is that auto-complete in the URL bar. I want to go back to a site I’ve visited recently, I just start typing the URL in the URL bar and the browser shows all related URLs. Same thing in Gmail: start typing an address and Gmail auto-completes it. So simple that nobody ever talks about it. But this simple feature is yet completely absent from blog editors, AFAICT.
Oh, sure, Flock does auto-complete in the search field. In fact, it supports incremental searches, which is really nice. But I need auto-complete for links and addresses.
What makes auto-complete even nicer is that it’s now possible to synchronise browsers through Google’s Browser Sync extension for Firefox (it might work with Flock too). Google also saves a Web History. And Gmail users get easy access to their address books from Google applications like Google Docs. And Google has toolbars for most browsers. So I guess Google could easily implement my dream editor.

Thing with my wishful thinking is that it’s often obvious enough that it becomes concrete very quickly after I say it. For all I know, this feature may exist somewhere and everybody else knows about it. But I’ve been missing it for a while now.
Ah, well.

Blogged with Flock

Tags: , , , , , , , , , , ,

They Dropped The Other Shoe

[Disclaimer: I’m not necessarily an Apple fanboy but I have been an enthusiastic Mac user since 1987 and have owned several Apple products, from an iPod to a QuickTake camera. I also think that technology is having a big impact on arts, media, and entertainment.]

Just watched Apple’s "Showtime" Special Event. Didn’t really read or even listen to anything much about it yet. During that event, Apple CEO Steve Jobs introduced new versions of all the iPod models, a new version of iTunes, and the addition of movies to the iTunes store. In addition, Jobs gave a sneak peak of an upcoming box to link iTunes with televisions and stereo systems.

People are likely to have been disappointed by the announcements. They’re probably saying that Steve Jobs’s famous "Reality Distortion Field" isn’t working, or that he lost his "mojo." They might even wonder about his health. Again…

Not that the new products are really boring, but there tend to be high expectations surrounding Apple announcements. This one is no different as people expected wireless capabilities on iPods and recording capabilities on the new "media centre" box, which was in fact part of the expected new products from Apple.

But this event is significant in another way. Through it, Apple explained their strategy, revealed a number of years ago as the Digital Hub. What some have called "convergence," quite a few years ago. Nothing really new. It’s just coming into full focus.

Though we may never know how much of it unfolded as planned, Apple’s media/tech strategy may appear rather prescient in retrospect. IIRC, it started in 1996, during Gil Amelio’s tenure. Or, more probably, in 1997 during the switch between Amelio and Jobs. Even by, say, 1999, that strategy was still considered a bold move. That was before the first iPod which, itself, was before iTunes, the iTunes Music Store, and most other current media-centric technologies at Apple. It was also at a time when user-generated content was relatively unimportant. In other ways, that was during the "Web 1.0" Internet bubble, before the "Web 2.0" craze for blogs, podcasts, and "social networking."

Apple isn’t the only corporation involved in the changes in the convergence between technology and the world of "content" (arts, media, entertainment). But it has played a key role. Whatever his success as a CEO, Steve Jobs has influenced the direction of change and, to an extent, shape a part of digital life to his own liking. While he’s clearly not clueless, his vision of the link between "content" and technology is quite specific. It does integrate user-generated content of "varying degrees of professionalism" (which he joked about during his presentation) but it gives precedence to the "content industry" (involving such powerful groups and lobbies as WIPO, NAB, MPAA, RIAA, etc.). Jobs’s position at Pixar makes him a part of that industry. Which is quite different from what arts and expressive culture can be.

Jobs invites musicians on stage with him (John Mayer, Wynton Marsalis, John Legend). He respects musicians and he might even appreciate their work. But his view of their work is that they produce content to consumed. For Jobs, music tracks, audiobooks, television episodes, movies, and music videos are all "contents" to be enjoyed by consumers. Now, the consumer can enjoy content "anywhere" as Apple is "in your den, in your living-room, in your car, and in your pocket." But what about public spaces? Concert halls, churches, coffee shops, parks, public libraries, classrooms, etc.? Oh! Apple can be there too! Yeah, of course. But those are not part of the primary vision. In Apple’s view, consumers all have their own iTunes accounts, media libraries, preferences, and content-consuming habits. A nuclear family may count as a unit to a certain extent (as Bob Iger pointed out in his "cameo appearance" during Jobs’s event). But the default mode is private consumption.

And there’s nothing wrong with that. Even the coolest things online are often based on the same model. It’s just that it’s not the only way to do things. Music, for instance, can be performed in public. In fact, it can be a collaborative process. The performers themselves need not be professionals. There’s no need for an audience, even. And there’s no need to see it as "intellectual property." Music is not a product. It’s a process by which human beings organize sound.

Ah, well…

Tags: , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ,

African Ingenuity

Via BoingBoing.

Who says Africans lack business acumen?

(Actually, such methods of empowerment are quite common, throughout Africa. And many Africans are rightfully proud of being able to manage by themselves. When will people from OECD “nations” get this?)