Category Archives: Placeholders

Pretty much any academic eventually feels that s/he is misunderstood. Misunderstandings about some core notions in about any academic field are involved in some of the most common pet peeves among academics.

In other words, there’s nothing as transdisciplinary as misunderstanding.

It can happen in the close proximity of a given department (“colleagues in my department misunderstand my work”). It can happen through disciplinary boundaries (“people in that field have always misunderstood our field”). And, it can happen generally: “Nobody gets us.”

It’s not paranoia and it’s probably not self-victimization. But there almost seems to be a form of “onedownmanship” at stake with academics from different disciplines claiming that they’re more misunderstood than others. In fact, I personally get the feeling that ethnographers are more among the most misunderstood people around, but even short discussions with friends in other fields (including mathematics) have helped me get the idea that, basically, we’re all misunderstood at the same “level” but there are variations in the ways we’re misunderstood. For instance, anthropologists in general are mistaken for what they aren’t based on partial understanding by the general population.

An example from my own experience, related to my decision to call myself an “informal ethnographer.” When you tell people you’re an anthropologist, they form an image in their minds which is very likely to be inaccurate. But they do typically have an image in their minds. On the other hand, very few people have any idea about what “ethnography” means, so they’re less likely to form an opinion of what you do from prior knowledge. They may puzzle over the term and try to take a guess as to what “ethnographer” might mean but, in my experience, calling myself an “ethnographer” has been a more efficient way to be understood than calling myself an “anthropologist.”

These misunderstandings have an impact on anyone’s work. Not only can it be demoralizing and can it impact one’s sense of self-worth, but it can influence funding decisions as well as the use of research results. These misunderstandings can underminine learning across disciplines. In survey courses, basic misunderstandings can make things very difficult for everyone. At a rather basic level, academics fight misunderstandings more than they fight ignorance.

The main reason I’m discussing this is that I’ve been given several occasions to think about the interface between the Ivory Tower and the rest of the world. It’s been a major theme in my blogposts about intellectuals, especially the ones in French. Two years ago, for instance, I wrote a post in French about popularizers. A bit more recently, I’ve been blogging about specific instances of misunderstandings associated with popularizers, including Malcolm Gladwell’s approach to expertise. Last year, I did a podcast episode about ethnography and the Ivory Tower. And, just within the past few weeks, I’ve been reading a few things which all seem to me to connect with this same issue: common misunderstandings about academic work. The connections are my own, and may not be so obvious to anyone else. But they’re part of my motivations to blog about this important issue.

In no particular order:

A thread on a mailing-list about linguistic anthropology. A paleoanthropologist interviewed for a radio show discussed language and cognitive evolution in a way which seemed to some linguistic anthropologists as conveying some misunderstandings about language.

Two blogposts by LibriVox founder Hugh McGuire about “Why Academics Should Blog.” The first post made me react and the second post was in a small part motivated by my reaction. (As an aside, McGuire should be commended for his flexibility of thoughts. His abilirty to adapt his ideas as the result of thoughtful discussion has helped me have less “visceral” reactions.)

One discussion I remember, which seems to fit, included comments about Germaine Dieterlen by a friend who also did research in West Africa. Can’t remember the specifics but the gist of my friend’s comment was that “you get to respect work by the likes of Germaine Dieterlen once you start doing field research in the region.” In my academic background, appreciation of Germaine Dieterlen’s may not be unconditional, but it doesn’t necessarily rely on extensive work in the field. In other words, while some parts of Dieterlen’s work may be controversial and it’s extremely likely that she “got a lot of things wrong,” her work seems to be taken seriously by several French-speaking africanists I’ve met. And not only do I respect everyone but I would likely praise someone who was able to work in the field for so long. She’s not my heroine (I don’t really have heroes) or my role-model, but it wouldn’t have occurred to me that respect for her wasn’t widespread. If it had seemed that Dieterlen’s work had been misunderstood, my reflex would possibly have been to rehabilitate her.

In fact, there’s a strong academic tradition of rehabilitating deceased scholars. The first example which comes to mind is a series of articles (PDF, in French) and book chapters by UWO linguistic anthropologist Regna Darnell.about “Benjamin Lee Whorf as a key figure in linguistic anthropology.” Of course, saying that these texts by Darnell constitute a rehabilitation of Whorf reveals a type of evaluation of her work. But that evaluation comes from a third person, not from me. The likely reason for this case coming up to my mind is that the so-called “Sapir-Whorf Hypothesis” is among the most misunderstood notions from linguistic anthropology. Moreover, both Whorf and Sapir are frequently misunderstood, which can make matters difficulty for many linguistic anthropologists talking with people outside the discipline.

The opposite process is also common: the “slaughtering” of “sacred cows.” (First heard about sacred cows through an article by ethnomusicologist Marcia Herndon.) In some significant ways, any scholar (alive or not) can be the object of not only critiques and criticisms but a kind of off-handed dismissal. Though this often happens within an academic context, the effects are especially lasting outside of academia. In other words, any scholar’s name is likely to be “sullied,” at one point or another. Typically, there seems to be a correlation between the popularity of a scholar and the likelihood of her/his reputation being significantly tarnished at some point in time. While there may still be people who treat Darwin, Freud, Nietzsche, Socrates, Einstein, or Rousseau as near divinities, there are people who will avoid any discussion about anything they’ve done or said. One way to put it is that they’re all misunderstood. Another way to put it is that their main insights have seeped through “common knowledge” but that their individual reputations have decreased.

Perhaps the most difficult case to discuss is that of Marx (Karl, not Harpo). Textbooks in introductory sociology typically have him as a key figure in the discipline and it seems clear that his insight on social issues was fundamental in social sciences. But, outside of some key academic contexts, his name is associated with a large series of social events about which people tend to have rather negative reactions. Even more so than for Paul de Man or Martin Heidegger, Marx’s work is entangled in public opinion about his ideas. Haven’t checked for examples but I’m quite sure that Marx’s work is banned in a number of academic contexts. However, even some of Marx’s most ardent opponents are likely to agree with several aspects of Marx’s work and it’s sometimes funny how Marxian some anti-Marxists may be.

But I digress…

Typically, the “slaughtering of sacred cows” relates to disciplinary boundaries instead of social ones. At least, there’s a significant difference between your discipline’s own “sacred cows” and what you perceive another discipline’s “sacred cows” to be. Within a discipline, the process of dismissing a prior scholar’s work is almost œdipean (speaking of Freud). But dismissal of another discipline’s key figures is tantamount to a rejection of that other discipline. It’s one thing for a physicist to show that Newton was an alchemist. It’d be another thing entirely for a social scientist to deconstruct James Watson’s comments about race or for a theologian to argue with Darwin. Though discussions may have to do with individuals, the effects of the latter can widen gaps between scholarly disciplines.

And speaking of disciplinarity, there’s a whole set of issues having to do with discussions “outside of someone’s area of expertise.” On one side, comments made by academics about issues outside of their individual areas of expertise can be very tricky and can occasionally contribute to core misunderstandings. The fear of “talking through one’s hat” is quite significant, in no small part because a scholar’s prestige and esteem may greatly decrease as a result of some blatantly inaccurate statements (although some award-winning scholars seem not to be overly impacted by such issues).

On the other side, scholars who have to impart expert knowledge to people outside of their discipline often have to “water down” or “boil down” their ideas and, in effect, oversimplifying these issues and concepts. Partly because of status (prestige and esteem), lowering standards is also very tricky. In some ways, this second situation may be more interesting. And it seems unavoidable.

How can you prevent misunderstandings when people may not have the necessary background to understand what you’re saying?

This question may reveal a rather specific attitude: “it’s their fault if they don’t understand.” Such an attitude may even be widespread. Seems to me, it’s not rare to hear someone gloating about other people “getting it wrong,” with the suggestion that “we got it right.” As part of negotiations surrounding expert status, such an attitude could even be a pretty rational approach. If you’re trying to position yourself as an expert and don’t suffer from an “impostor syndrome,” you can easily get the impression that non-specialists have it all wrong and that only experts like you can get to the truth. Yes, I’m being somewhat sarcastic and caricatural, here. Academics aren’t frequently that dismissive of other people’s difficulties understanding what seem like simple concepts. But, in the gap between academics and the general population a special type of intellectual snobbery can sometimes be found.

Obviously, I have a lot more to say about misunderstood academics. For instance, I wanted to address specific issues related to each of the links above. I also had pet peeves about widespread use of concepts and issues like “communities” and “Eskimo words for snow” about which I sometimes need to vent. And I originally wanted this post to be about “cultural awareness,” which ends up being a core aspect of my work. I even had what I might consider a “neat” bit about public opinion. Not to mention my whole discussion of academic obfuscation (remind me about “we-ness and distinction”).

But this is probably long enough and the timing is right for me to do something else.

I’ll end with an unverified anecdote that I like. This anecdote speaks to snobbery toward academics.

[It’s one of those anecdotes which was mentioned in a course I took a long time ago. Even if it’s completely fallacious, it’s still inspiring, like a tale, cautionary or otherwise.]

As the story goes (at least, what I remember of it), some ethnographers had been doing fieldwork in an Australian cultural context and were focusing their research on a complexkinship system known in this context. Through collaboration with “key informants,” the ethnographers eventually succeeded in understanding some key aspects of this kinship system.

As should be expected, these kinship-focused ethnographers wrote accounts of this kinship system at the end of their field research and became known as specialists of this system.

After a while, the fieldworkers went back to the field and met with the same people who had described this kinship system during the initial field trip. Through these discussions with their “key informants,” the ethnographers end up hearing about a radically different kinship system from the one about which they had learnt, written, and taught.

The local informants then told the ethnographers: “We would have told you earlier about this but we didn’t think you were able to understand it.”

[I’m typically not very good at going back to drafts and I don’t have much time to write this. But I can RERO this. It’s an iterative process in any case….]

Been thinking about different things which all relate to the same theme: changing course, seizing opportunities, shifting focus, adapting to new situations, starting over, getting a clean slate… Moving on.

One reason is that I recently decided to end my ethnography podcast. Not that major a decision and rather easy to make. Basically, I had stopped doing it but I had yet to officially end it. I had to make it clear, in my mind, that it’s not part of the things I’m doing, these days. Not that it was a big thing in my life but I had set reminders every month that I had to record a podcast episode. It worked for ten episode (in ten months) but, once I had missed one episode, the reminder was nagging me more than anything else.

In this sense, “moving on” is realistic/pragmatic. Found something similar in Getting Things Done, by David Allen.

It’s also similar to something Larry Lessig called “email bankruptcy,” as a step toward enhanced productivity.

In fact, even financial bankruptcy can relate to this, in some contexts. In Canada, at least, bankruptcy is most adequately described as a solution to a problem, not the problem itself. I’ve known some people who were able to completely rebuild their finances after declaring bankruptcy, sometimes even getting a better credit rating than someone who hadn’t gone bankrupt. I know how strongly some people may react to this concept of bankruptcy (based on principle, resentment, fears, hopes…). It’s an extreme example of what I mean by “moving on.” It goes well with the notion, quite common in North American cultural contexts, that you always deserve a second chance (but that you should do things yourself).

Of course, similar things happen with divorces which, similarly, can often be considered as solutions to a problem rather than the problem itself. No matter how difficult or how bad divorce might be, it’s a way to start over. In some sense, it’s less extreme an example as the bankruptcy one. But it may still generate negative vibes or stir negative emotions.

Because what I’m thinking about has more to do with “turning over a new leaf.” And taking the “leap of faith” which will make you go where you feel more comfortable. I’m especially thinking about all sorts of cases of people who decided to make radical changes in their professional or personal lives, often leaving a lot behind. Whether they were forced to implement such changes or decided to jump because they simply wanted to, all of the cases I remember have had positive outcomes.

It reminds me of a good friend of mine with whom I went through music school, in college. When he finished college, he decided to follow the music path and registered for the conservatory. But, pretty quickly, he realized that it wasn’t for him. Even though he had been intensely “in music” for several years, with days of entering the conservatory, he saw that music wasn’t to remain the central focus of his career. Through a conversation with a high school friend (who later became his wife and the mother of his children), he found out that it wasn’t too late for him to register for university courses. He had been thinking about phys. ed., and thought it might be a nice opportunity to try that path. He’s been a phys. ed. teacher for a number of years. We had lunch together last year and he seems very happy with his career choice. He also sounds like a very dedicated and effective phys. ed. teacher.

In my last podcast episode, I mentioned a few things about my views of this “change of course.” Including what has become something of an expression, for me: “Done with fish.” Comes from the movie Adaptation. The quote is found here (preceded by a bit of profanity). Basically, John Laroche, who was passionately dedicated to fish, decided to completely avoid anything having to do with fish. I can relate to this at some rather deep level.

I’m also thinking about the negative consequences of “sticking with” something which isn’t working, shifting too late or too quickly, implementing changes in inappropriate ways. Plenty of examples there. Most of the ones which come to my mind have to do with business settings. One which would require quite a bit of “explaining” is my perception of Google’s strategy with Wave. Put briefly (with the hope of revisiting this issue), I think Google made bad decisions with Wave, including killing it both too late and too early (no, I don’t see this as a contradiction; but I don’t have time to explain it). They also, I feel, botched a few transitions, in this. And, more importantly, I’d say that they failed to adapt the product to what was needed.

And the trigger for several of my reflections on this “moving on” idea have to do with this kind of adaptation (fun that the movie of that name should be involved, eh?). Twitter could be an inspiration, in this case. Not only did they, like Flickr, start through a switch away from another project, but Twitter was able to transform users’ habits into the basis for some key features. Hashtags and “@replies” are well-known examples. But you could even say that most of the things they’ve been announcing have been related to the way people use their tools.

So, in a way, it’s about the balance between vision and responsiveness. Vision is often discussed and it sounds to some people as a key thing in any “task-based group (from a team to a corporation). But the way a team can switch from one project to the next based on feedback (from users or other stakeholders) seems underrated. Although, there is some talk about the “startup mentality” in many contexts, including Google and Apple. Words which fit this semantic field include: “agile,” “flexible,” “pivot,” “lean,” and “nimble” (the latter word seemed to increase in currency after being used by Barack Obama in a speech).

Anyhoo… Gotta go.

But, just before I go: I am moving on with some things (including my podfade but also a shift away from homebrewing). But the key things in my life are very stable, especially my sentimental life.

Literature Major: “Content’s a tale / Told by an idiot, full of sound and fury, / Signifying nothing.”

Arts Major: “Content Is in the Eye of the Beholder.”

Entertainer: “There’s no content / like show content / like no content I know.”

Journalist: “Content is my job and I deserve to be paid for what I make, the exact same way that a baker is paid for selling bread. What other people called ‘content’ isn’t really content since it hasn’t been vetted by professionals like my editor. So my role is to create content so that my editor can distribute it through exclusive channels. Other people’s content becomes my content when I secure the rights to it through the use of a clearance service. Comments by people I interview only become content after they sign a release. Everything else is noise.”

Economist: “There are four ways to get paid for content: a) subscription; b) advertising; c) private or public sponsorship; d) sale on media. Since advertising and sponsorship are two aspects of the same model and since consumers epend money on either subscription or media sales, there are two basic models.”

Functionalist (Sociology): “Content serves different goals, both manifest and latent.”
Conflict-Theorist (Sociology): “Providing free content is a way for the ruling class to make the audience into a commodity.”

Been using Diigo as a way to annotate online texts. In this case, I was as interested in the tone as in the text itself. At the same time, I kept thinking about things which seem to be missing from Diigo.

One thing I like about this text is its tone. There’s an honesty, an ingenuity that I find rare in this type of writing.

startup ideas

The background is important, in terms of the type of ideas about which we’re constructing something.

what do you wish someone would make for you?

My own itch has to do with Diigo, actually. There’s a lot I wish Diigo would make for me. I may be perceived as an annoyance, but I think my wishlist may lead to something bigger and possibly quite successful.

The difference between this question and the “scratch your own itch” principle seems significant, and this distinction may have some implications in terms of success: we’re already talking about others, not just running ideas in our own head.

what do you wish someone would make for you?

It’s somewhat different from the well-known “scratch your own itch” principle. In this difference might be located something significant. In a way, part of the potential for this version to lead to success comes from the fact that it’s already connected with others, instead of being about running ideas in your own mind.

grow organically

The core topic of the piece, put in a comparative context. The comparison isn’t the one people tend to make and one may argue about the examples used. But the concept of organic ideas is fascinating and inspiring.

you decide, from afar,

What we call, in anthropology, the “armchair” approach. Also known as “backbenching.” For this to work, you need to have a deep knowledge of the situation, which is part of the point in this piece. Nice that it’s not demonizing this position but putting it in context.

Apple
was the first type

One might argue that it was a hybrid case. Although, it does sound like the very beginnings of Apple weren’t about “thinking from afar.”

class of users other than you

Since developers are part of a very specific “class” of people, this isn’t insignificant a way to phrase this.

They still rely on this principle today, incidentally.
The iPhone is the phone Steve Jobs wants.

Apple tends to be perceived in a different light. According to many people, it’s the “textbook example” of a company where decisions are made without concerns for what people need. “Steve Jobs uses a top-down approach,” “They don’t even use focus groups,” “They don’t let me use their tools the way I want to use them.” But we’re not talking about the same distinction between top-down and bottom-up. Though “organic ideas” seem to imply that it’s a grassroots/bottom-up phenomenon, the core distinction isn’t about the origin of the ideas (from the “top,” in both cases) but on the reasoning behind these ideas.

We didn’t need this software ourselves.

Sounds partly like a disclaimer but this approach is quite common and “there’s nothing wrong with it.”

comparatively old

Age and life experience make for an interesting angle. It’s not that this strategy needs people of a specific age to work. It’s that there’s a connection between one’s experience and the way things may pan out.

There is no sharp line between the two types of ideas,

Those in the “engineering worldview” might go nuts, at this point. I can hear the claims of “hand waving.” But we’re talking about something complex, here, not a merely complicated problem.

Apple type

One thing to note in the three examples here: they’re all made by pairs of guys. Jobs and Woz, Gates and Allen, Page and Brin. In many cases, the formula might be that one guy (or gal, one wishes) comes up with ideas knowing that the other can implement them. Again, it’s about getting somebody else to build it for you, not about scratching your own itch.

Bill Gates was writing something he would use

Again, Gates may not be the most obvious example, since he’s mostly known for another approach. It’s not inaccurate to say he was solving his own problem, at the time, but it may not be that convincing as an example.

Larry and Sergey when they wrote the first versions of Google.

Although, the inception of the original ideas was academic in context. They weren’t solving a search problem or thinking about monetization. They were discovering the power of CitationRank.

generally preferable

Nicely relativistic.

It takes experience
to predict what other people will want.

And possibly a lot more. Interesting that he doesn’t mention empirical data.

young founders

They sound like a fascinating group to observe. They do wonders when they open up to others, but they seem to have a tendency to impose their worldviews.

I’d encourage you to focus initially on organic ideas

Now, this advice sounds more like the “scratch your own itch” advocation. But there’s a key difference in that it’s stated as part of a broader process. It’s more of a “walk before you run” or “do your homework” piece of advice, not a “you can’t come up with good ideas if you just think about how people will use your tool.”

missing or broken

It can cover a lot, but it’s couched in terms of the typical “problem-solving” approach at the centre of the engineering worldview. Since we’re talking about developing tools, it makes sense. But there could be a broader version, admitting for dreams, inspiration, aspiration. Not necessarily of the “what would make you happy?” kind, although there’s a lot to be said about happiness and imagination. You’re brainstorming, here.

immediate answers

Which might imply that there’s a second step. If you keep asking yourself the same question, you may be able to get a very large number of ideas. The second step could be to prioritize them but I prefer “outlining” as a process: you shuffle things together and you group some ideas to get one which covers several. What’s common between your need for a simpler way to code on the Altair and your values? Why do you care so much about algorithms instead of human encoding?

You may need to stand outside yourself a bit to see brokenness

Ah, yes! “Taking a step back,” “distancing yourself,” “seeing the forest for the trees”… A core dimension of the ethnographic approach and the need for a back-and-forth between “inside” and “outside.” There’s a reflexive component in this “being an outsider to yourself.” It’s not only psychological, it’s a way to get into the social, which can lead to broader success if it’s indeed not just about scratching your own itch.

get used to it and take it for granted

That’s enculturation, to you. When you do things a certain way simply because “we’ve always done them that way,” you may not create these organic ideas. But it’s a fine way to do your work. Asking yourself important questions about what’s wrong with your situation works well in terms of getting new ideas. But, sometimes, you need to get some work done.

a Facebook

Yet another recontextualized example. Zuckerberg wasn’t trying to solve that specific brokenness, as far as we know. But Facebook became part of what it is when Zuck began scratching that itch.

organic startup ideas usually don’t
seem like startup ideas at first

Which gets us to the pivotal importance of working with others. Per this article, VCs and “angel investors,” probably. But, in the case of some of cases cited, those we tend to forget, like Paul Allen, Narendra, and the Winklevosses.

end up making
something of value to a lot of people

Trial and error, it’s an iterative process. So you must recognize errors quickly and not invest too much effort in a specific brokenness. Part of this requires maturity.

something
other people dismiss as a toy

The passage on which Gruber focused and an interesting tidbit. Not that central, come to think of it. But it’s important to note that people’s dismissive attitude may be misled, that “toys” may hide tools, that it’s probably a good idea not to take all feedback to heart…

At this point, when someone comes to us with
something that users like but that we could envision forum trolls
dismissing as a toy, it makes us especially likely to invest.

the best source of organic ones

Especially to investors. Potentially self-serving… in a useful way.

they’re at the forefront of technology

That part I would dispute, actually. Unless we talk about a specific subgroup of young founders and a specific set of tools. Young founders tend to be oblivious to a large field in technology, including social tools.

they’re in a position to discover
valuable types of fixable brokenness first

The focus on fixable brokenness makes sense if we’re thinking exclusively through the engineering worldview, but it’s at the centre of some failures like the Google Buzz launch.

you still have to work hard

Of the “inspiration shouldn’t make use forget perspiration” kind. Makes for a more thoughtful approach than the frequent “all you need to do…” claims.

I’d encourage anyone
starting a startup to become one of its users, however unnatural it
seems.

Not merely an argument for dogfooding. It’s deeper than that. Googloids probably use Google tools but they didn’t actually become users. They’re beta testers with a strong background in troubleshooting. Not the best way to figure out what users really want or how the tool will ultimately fail.

It’s hard to compete directly with open source software

Open Source as competition isn’t new as a concept, but it takes time to seep in.

there has to be some part
you can charge for

The breach through which old-school “business models” enter with little attention paid to everything else. To the extent that much of the whole piece might crumble from pressure built up by the “beancounter” worldview. Good thing he acknowledges it.

This is mostly meant as a reply to this blogthread. But it’s also more generally about my personal reaction to Apple’s iPad announcement.

Some background.

I’m an ethnographer and a teacher. I read a fair deal, write a lot of notes, and work in a variety of contexts. These days, I tend to spend a good amount of time in cafés and other public places where I like to work without being too isolated. I also commute using public transit, listen to lots of podcast, and create my own. I’m also very aural.

I’ve used a number of PDAs, over the years, from a Newton MessagePad 130 (1997) to a variety of PalmOS devices (until 2008). In fact, some people readily associated me with PDA use.

As soon as I learnt about the iPod touch, I needed one. As soon as I’ve heard about the SafariPad, I wanted one. I’ve been an intense ‘touch user since the iPhone OS 2.0 release and I’m a happy camper.

(A major reason I never bought an iPhone, apart from price, is that it requires a contract.)

In my experience, the ‘touch is the most appropriate device for all sorts of activities which are either part of an other activity (reading during a commute) or are simply too short in duration to constitute an actual “computer session.” You don’t “sit down to work at your ‘touch” the way you might sit in front of a laptop or desktop screen. This works great for “looking up stufff” or “checking email.” It also makes a lot of sense during commutes in crowded buses or metros.

In those cases, the iPod touch is almost ideal. Ubiquitous access to Internet would be nice, but that’s not a deal-breaker. Alternative text-input methods would help in some cases, but I do end up being about as fast on my ‘touch as I was with Graffiti on PalmOS.

For other tasks, I have a Mac mini. Sure, it’s limited. But it does the job. In fact, I have no intention of switching for another desktop and I even have an eMachines collecting dust (it’s too noisy to make a good server).

What I miss, though, is a laptop. I used an iBook G3 for several years and loved it. For a little while later, I was able to share a MacBook with somebody else and it was a wonderful experience. I even got to play with the OLPC XO for a few weeks. That one was not so pleasant an experience but it did give me a taste for netbooks. And it made me think about other types of iPhone-like devices. Especially in educational contexts. (As I mentioned, I’m a teacher)

I’ve been laptop-less for a while, now. And though my ‘touch replaces it in many contexts, there are still times when I’d really need a laptop. And these have to do with what I might call “mobile sessions.”

For instance: liveblogging a conference or meeting. I’ve used my ‘touch for this very purpose on a good number of occasions. But it gets rather uncomfortable, after a while, and it’s not very fast. A laptop is better for this, with a keyboard and a larger form factor. But the iPad will be even better because of lower risks of RSI. A related example: just imagine TweetDeck on iPad.

Possibly my favourite example of a context in which the iPad will be ideal: presentations. Even before learning about the prospect of getting iWork on a tablet, presentations were a context in which I really missed a laptop.

Sure, in most cases, these days, there’s a computer (usually a desktop running XP) hooked to a projector. You just need to download your presentation file from Slideshare, show it from Prezi, or transfer it through USB. No biggie.

But it’s not the extra steps which change everything. It’s the uncertainty. Even if it’s often unfounded, I usually get worried that something might just not work, along the way. The slides might not show the same way as you see it because something is missing on that computer or that computer is simply using a different version of the presentation software. In fact, that software is typically Microsoft PowerPoint which, while convenient, fits much less in my workflow than does Apple Keynote.

The other big thing about presentations is the “presenter mode,” allowing you to get more content than (or different content from) what the audience sees. In most contexts where I’ve used someone else’s computer to do a presentation, the projector was mirroring the computer’s screen, not using it as a different space. PowerPoint has this convenient “presenter view” but very rarely did I see it as an available option on “the computer in the room.” I wish I could use my ‘touch to drive presentations, which I could do if I installed software on that “computer in the room.” But it’s not something that is likely to happen, in most cases.

A MacBook solves all of these problems. and it’s an obvious use for laptops. But how, then, is the iPad better? Basically because of interface. Switching slides on a laptop isn’t hard, but it’s more awkward than we realize. Even before watching the demo of Keynote on the iPad, I could simply imagine the actual pleasure of flipping through slides using a touch interface. The fit is “natural.”

I sincerely think that Keynote on the iPad will change a number of things, for me. Including the way I teach.

Then, there’s reading.

Now, I’m not one of those people who just can’t read on a computer screen. In fact, I even grade assignments directly from the screen. But I must admit that online reading hasn’t been ideal, for me. I’ve read full books as PDF files or dedicated formats on PalmOS, but it wasn’t so much fun, in terms of the reading process. And I’ve used my ‘touch to read things through Stanza or ReadItLater. But it doesn’t work so well for longer reading sessions. Even in terms of holding the ‘touch, it’s not so obvious. And, what’s funny, even a laptop isn’t that ideal, for me, as a reading device. In a sense, this is when the keyboard “gets in the way.”

Sure, I could get a Kindle. I’m not a big fan of dedicated devices and, at least on paper, I find the Kindle a bit limited for my needs. Especially in terms of sources. I’d like to be able to use documents in a variety of formats and put them in a reading list, for extended reading sessions. No, not “curled up in bed.” But maybe lying down in a sofa without external lighting. Given my experience with the ‘touch, the iPad is very likely the ideal device for this.

Then, there’s the overall “multi-touch device” thing. People have already been quite creative with the small touchscreen on iPhones and ‘touches, I can just imagine what may be done with a larger screen. Lots has been said about differences in “screen real estate” in laptop or desktop screens. We all know it can make a big difference in terms of what you can display at the same time. In some cases, two screens isn’t even a luxury, for instance when you code and display a page at the same time (LaTeX, CSS…). Certainly, the same qualitative difference applies to multitouch devices. Probably even more so, since the display is also used for input. What Han found missing in the iPhone’s multitouch was the ability to use both hands. With the iPad, Han’s vision is finding its space.

Oh, sure, the iPad is very restricted. For instance, it’s easy to imagine how much more useful it’d be if it did support multitasking with third-party apps. And a front-facing camera is something I was expecting in the first iPhone. It would just make so much sense that a friend seems very disappointed by this lack of videoconferencing potential. But we’re probably talking about predetermined expectations, here. We’re comparing the iPad with something we had in mind.

Then, there’s the issue of the competition. Tablets have been released and some multitouch tablets have recently been announced. What makes the iPad better than these? Well, we could all get in the same OS wars as have been happening with laptops and desktops. In my case, the investment in applications, files, and expertise that I have made in a Mac ecosystem rendered my XP years relatively uncomfortable and me appreciate returning to the Mac. My iPod touch fits right in that context. Oh, sure, I could use it with a Windows machine, which is in fact what I did for the first several months. But the relationship between the iPhone OS and Mac OS X is such that using devices in those two systems is much more efficient, in terms of my own workflow, than I could get while using XP and iPhone OS. There are some technical dimensions to this, such as the integration between iCal and the iPhone OS Calendar, or even the filesystem. But I’m actually thinking more about the cognitive dimensions of recognizing some of the same interface elements. “Look and feel” isn’t just about shiny and “purty.” It’s about interactions between a human brain, a complex sensorimotor apparatus, and a machine. Things go more quickly when you don’t have to think too much about where some tools are, as you’re working.

So my reasons for wanting an iPad aren’t about being dazzled by a revolutionary device. They are about the right tool for the job.

A bit of context for my tweet (actually, a post from Ping.fm, meant as a status update, thereby giving support in favour of conscious duplication, «n’en déplaise aux partisans de l’action contre la duplication».)

Though I didn’t emphasize the “sometimes” part in that podcast episode, it was an important part of what I wanted to say. In fact, my intention wasn’t to defend draft æsthetics but to note that there seems to be a tendency toward this æsthetic mode. I do situate myself within that mode in many things I do, but it really doesn’t mean that this mode should be the exclusive one used in any context.

That aforequoted tweet was thus a response to my podcast episode on draft æsthetics. “Yes, ‘good enough’ may work, sometimes. But it needs not be applied in all cases.”

As I often get into convoluted discussions with people who seem to think that I condone or defend a position because I take it for myself, the main thing I’d say there is that I’m not only a relativist but I cherish nuance. In other words, my tweet was a way to qualify the core statement I was talking about in my podcast episode (that “good enough” exists, at times). And that statement isn’t necessarily my own. I notice a pattern by which this statement seems to be held as accurate by people. I share that opinion, but it’s not a strongly held belief of mine.

Of course, I digress…

So, the tweet which motivated Vikrama had to do with my approach to “good enough.” In this case, I tend to think about writing but in view of Eric S. Raymond’s approach to “Release Early, Release Often” (RERO). So there is a connection to software development and geek culture. But I think of “good enough” in a broader sense.

The Calder quote remained in my head, after it was mentioned by a colleague who had read it in a local newspaper. One reason it struck me is that I spend some time thinking about artists and engineers, especially in social terms. I spend some time hanging out with engineers but I tend to be more on the “artist” side of what I perceive to be an axis of attitudes found in some social contexts. I do get a fair deal of flack for some of my comments on this characterization and it should be clear that it isn’t meant to imply any evaluation of individuals. But, as a model, the artist and engineer distinction seems to work, for me. In a way, it seems more useful than the distinction between science and art.

An engineer friend with whom I discussed this kind of distinction was quick to point out that, to him, there’s no such thing as “good enough.” He was also quick to point out that engineers can be creative and so on. But the point isn’t to exclude engineers from artistic endeavours. It’s to describe differences in modes of thought, ways of knowing, approaches to reality. And the way these are perceived socially. We could do a simple exercise with terms like “troubleshooting” and “emotional” to be assigned to the two broad categories of “engineer” and “artist.” Chances are that clear patterns would emerge. Of course, many concepts are as important to both sides (“intelligence,” “innovation”…) and they may also be telling. But dichotomies have heuristic value.

Now, to go back to software development, the focus in Vikrama’s Agile Diary post…

What pushed me to post my status update and tweet is in fact related to software development. Contrary to what Vikrama presumes, it wasn’t about a Web application. And it wasn’t even about a single thing. But it did have to do with firmware development and with software documentation.

The first case is that of my Fonera 2.0n router. Bought it in early November and I wasn’t able to connect to its private signal using my iPod touch. I could connect to the router using the public signal, but that required frequent authentication, as annoying as with ISF. Since my iPod touch is my main WiFi device, this issue made my Fonera 2.0n experience rather frustrating.

Of course, I’ve been contacting Fon‘s tech support. As is often the case, that experience was itself quite frustrating. I was told to reset my touch’s network settings which forced me to reauthenticate my touch on a number of networks I access regularly and only solved the problem temporarily. The same tech support person (or, at least, somebody using the same name) had me repeat the same description several times in the same email message. Perhaps unsurprisingly, I was also told to use third-party software which had nothing to do with my issue. All in all, your typical tech support experience.

But my tweet wasn’t really about tech support. It was about the product. Thougb I find the overall concept behind the Fonera 2.0n router very interesting, its implementation seems to me to be lacking. In fact, it reminds me of several FLOSS development projects that I’ve been observing and, to an extent, benefitting from.

This is rapidly transforming into a rant I’ve had in my “to blog” list for a while about “thinking outside the geek box.” I’ll try to resist the temptation, for now. But I can mention a blog thread which has been on my mind, in terms of this issue.

The blogpost refers to a situation in which, according to at least some users (including the blogpost’s author), Firefox uses up more memory than it should and becomes difficult to use. The thread has several comments providing support to statements about the relatively poor performance of Firefox on people’s systems, but it also has “contributions” from an obvious troll, who keeps assigning the problem on the users’ side.

The thing about this is that it’s representative of a tricky issue in the geek world, whereby developers and users are perceived as belonging to two sides of a type of “class struggle.” Within the geek niche, users are often dismissed as “lusers.” Tech support humour includes condescending jokes about “code 6”: “the problem is 6″ from the screen.” The aforementioned Eric S. Raymond wrote a rather popular guide to asking questions in geek circles which seems surprisingly unaware of social and cultural issues, especially from someone with an anthropological background. Following that guide, one should switch their mind to that of a very effective problem-solver (i.e., the engineer frame) to ask questions “the smart way.” Not only is the onus on users, but any failure to comply with these rules may be met with this air of intellectual superiority encoded in that guide. IOW, “Troubleshoot now, ask questions later.”

Of course, many users are “guilty” of all sorts of “crimes” having to do with not reading the documentation which comes with the product or with simply not thinking about the issue with sufficient depth before contacting tech support. And as the majority of the population is on the “user” side, the situation can be described as both a form of marginalization (geek culture comes from “nerd” labels) and a matter of elitism (geek culture as self-absorbed).

This does have something to do with my Fonera 2.0n. With it, I was caught in this dynamic whereby I had to switch to the “engineer frame” in order to solve my problem. I eventually did solve my Fonera authentication problem, using a workaround mentioned in a forum post about another issue (free registration required). Turns out, the “release candidate” version of my Fonera’s firmware does solve the issue. Of course, this new firmware may cause other forms of instability and installing it required a bit of digging. But it eventually worked.

The point is that, as released, the Fonera 2.0n router is a geek toy. It’s unpolished in many ways. It’s full of promise in terms of what it may make possible, but it failed to deliver in terms of what a router should do (route a signal). In this case, I don’t consider it to be a finished product. It’s not necessarily “unstable” in the strict sense that a software engineer might use the term. In fact, I hesitated between different terms to use instead of “stable,” in that tweet, and I’m not that happy with my final choice. The Fonera 2.0n isn’t unstable. But it’s akin to an alpha version released as a finished product. That’s something we see a lot of, these days.

The main other case which prompted me to send that tweet is “CivRev for iPhone,” a game that I’ve been playing on my iPod touch.

I’ve played with different games in the Civ franchise and I even used the FLOSS version on occasion. Not only is “Civilization” a geek classic, but it does connect with some anthropological issues (usually in a problematic view: Civ’s worldview lacks anthro’s insight). And it’s the kind of game that I can easily play while listening to podcasts (I subscribe to a number of th0se).

What’s wrong with that game? Actually, not much. I can’t even say that it’s unstable, unlike some other items in the App Store. But there’s a few things which aren’t optimal in terms of documentation. Not that it’s difficult to figure out how the game works. But the game is complex enough that some documentation is quite useful. Especially since it does change between one version of the game and another. Unfortunately, the online manual isn’t particularly helpful. Oh, sure, it probably contains all the information required. But it’s not available offline, isn’t optimized for the device it’s supposed to be used with, doesn’t contain proper links between sections, isn’t directly searchable, and isn’t particularly well-written. Not to mention that it seems to only be available in English even though the game itself is available in multiple languages (I play it in French).

Nothing tragic, of course. But coupled with my Fonera experience, it contributed to both a slight sense of frustration and this whole reflection about unfinished products.

Critical thinking has been on my mind, recently. For one thing, I oriented an “intro. to sociology” course I teach toward critical skills and methods. To me, it’s a very important part of university education, going much beyond media literacy.

And media literacy is something about which I care a great deal. Seems to me that several journalists have been giving up on trying to help the general population increase and enhance their own media literacy skills. It’s almost as if they were claiming they’re the only ones who can reach a significant level of media literacy. Of course, many of them seem unable to have a critical approach to their own work. I’m with Bourdieu on this one. And I make my problem with journalism known.

As a simple example, I couldn’t help but notice a number of problems with this CBC coverage of a new citizenship guidebook. My approach to this coverage is partly visible in short discussions I’ve had on Aardvark about bylines.

A bit over a week ago, I heard about something interesting related to “making technology work,” on WTP (a technology podcast for PRI/BBC/Discovery The World, a bit like Search Engine from bigger media outlets). It was a special forum discussion related to issues broader than simply finding the right tool for the right task. In fact, it sounded like it could become a broad discussion of issues and challenges going way beyond the troubleshooting/problem-solving approach favoured by some technology enthusiasts. Given my ethnographic background, my interest in geek culture, and my passion for social media, I thought I’d give it a try.

The first thing I noticed was a link to a critique of the OLPC project. I’ve personally been quite critical of that project, writing several blogposts about it. So I had to take a look.

And although I find the critical stance of this piece relatively useful (there was way too much groupthink with the original coverage of the OLPC), I couldn’t help but use my critical sense as I was reading this piece.

Which motivated me to do some Diigo annotations on it. For some reason, there are things that I wanted to highlight which aren’t working and I think I may have lost some annotations in the process. But the following is the result of a relatively simple reading of this piece. True to the draft aesthetics, I made no attempt to be thorough, clean, precise, or clear.

50 percent of staff were being laid off and a major restructuring was under way

The dramatic version which sends the message: OLPC Inc. was in big trouble. (The fact that it’s allegedly a non-profit is relatively irrelevant.)

the project seems nearly dead in the water

A strong statement. Stronger than all those “beleaguered company” ones made about Apple in the mid90s before Jobs went back.

And that may be great news for children in the developing world.

Tadaa! Here’s the twist! The OLPC is dead, long live the Child!

lobbied national governments and international agencies

Right. The target was institutional. Kind of strange for a project which was billed as a way to get tools in the hands of individual children. And possibly one of the biggest downfalls of the project.

Negroponte and other techno-luminati

Oh, snap!
It could sound relatively harmless an appellation. But the context and the piece’s tone make it sound like a rather deep insult.

Innovate

Ah, nice! Not “create” or “build.” But “innovate.” Which is something the project has been remarkably good at. It was able to achieve a number of engineering feats. Despite Negroponte’s repeated claims to the contrary, the OLPC project can be conceived as an engineering project. In fact, it’s probably the most efficient way to shed the most positive light on it. As an engineering project, it was rather successful. As an “education project” (as Negroponte kept calling it), it wasn’t that successful. In fact, it may have delayed a number of things which matter in terms of education.

take control of their education

Self-empowerment, at the individual level. In many ways, it sounds like a very Protestant ideal. And it’s clearly part of the neoliberal agenda (or the neoconservative one, actually). Yet it doesn’t sound strange at all. It sounds naturally good and pure.

technology optimists

Could be neutral in denotation but does connote a form of idealistic technological determinism.

Child

school attendance

“Children who aren’t in school can’t be learning anything, right?”

trending dramatically upward

Fascinating choice of words.

tens of millions of dollars

highly respected center

Formulas such as these are often a way to prevent any form of source criticism. Not sure Wikipedians would consider these “peacock terms,” but they don’t clearly represent a “neutral point of view.”

they don’t seem to be learning much

Nothing which can be measured with our tools, at least. Of course, nothing else matters. But still…

international science exam

Of course, these tend to be ideally suited for most learning contexts…

There’s no question that improving education in the developing world is necessary.

Although, there could be a question or two about this. Not politically expedient, perhaps. But still…

powerful argument

Tools in a rhetorical process.

instinctive appeal

Even the denotative sense is polarized.

precious little evidence

Switching to the “studies have shown” mode. In this mode, lack of proof is proof of lack, critical thinking is somewhat discouraged, and figures are significant by themselves.

circumstantial evidence

The jury isn’t out, on this one.

co-founder of J-PA

Did Esther co-write the article? Honest question.

the technology didn’t work any better than a normal classroom teacher

A very specific point. If the goal of tool use is to improve performance over “regular teaching,” it’s a particular view of technology. One which, itself, is going by the wayside. And which has been a large part of the OLPC worldview.

the goal is improving education for children in the developing world, there are plenty of better, and cheaper, alternatives.

A core belief, orienting the piece. Cost is central. The logic is one of “bang for the buck.”

the teachers simply weren’t using the computers

We’re touching on something, here. People have to actually use the computers for the “concept” to work. Funny that there’s rarely a lot of discussion on how that works. A specific version of “throwing money at a problem” is to “throw technology at” people.

few experimental studies to show a positive impact from the use of computers

Is the number of studies going one way or another the main issue, here? Can’t diverse studies look at different things and be understood as a way to describe a more complex reality than “technology is good and/or bad?”

substituting computers for teachers

Still oriented toward the “time to task” approach. But that’s good enough for cognitive science, which tends to be favourably viewed in educational fields.

supplement

Kept thinking about the well-known Hawthorne effect. In this case, the very idea that providing students with supplementary “care” can be seen as an obvious approach which is most often discussed in the field instead of at the higher levels of decision-making.

The OLPC concept has been pioneered in a number of school districts in the United States over the last decade

From a 2005 project targeting “countries with inconsistent power grids,” we get to a relatively long series of initiatives in individual school districts in the USofA since last century. Telescoping geographical and temporal scales. And, more importantly, assigning the exact same “concept” to diverse projects.

Negroponte has explicitly derided

Not the only thing Negroponte derides. He’s been a professional derider for a while, now.
Negroponte’s personality is part of the subtext of any OLPC-related piece. It’d be interesting to analyse him in view of the “mercurial CEO” type which fascinates a number of people.

It must be said

Acknowledging the fact that there is more to the situation than what this piece is pushing.

academic

In this context, “academic” can have a variety of connotations, many of which are relatively negative.

teachers limited access to the computers

Typically, teachers have relatively little control in terms of students’ access to computers so it sounds likely that the phrase should have read “had limited access.” But, then again, maybe teachers in Hollow’s research were in fact limiting access to computers, which would be a very interesting point to bring and discuss. In fact, part of what is missing in many of those pieces about technology and learning is what access really implies. Typically, most discussions on the subject have to do with time spent alone with such a tool, hence the “one…per child” part of the OLPC approach. But it’s hard to tell if there has been any thought about the benefits of group access to tools or limited access to such tools.
To go even further, there’s a broad critique of the OLPC approach, left unaddressed in this piece, about the emphasis on individual ownership of tools. In the US, it’s usually not ok for neighbours to ask about using others’ lawnmowers and ladders. It’s unsurprising that pushing individual ownership would seem logical to those who design projects from the US.

had not been adequately trained

In the OLPC context, it has been made as a case for the dark side of constructionism. The OLPC project might have been a learning project, but it wasn’t a teaching one. Some explicit comments from project members were doing little to dispel the notion that constructivism isn’t about getting rid of teachers. Even documentation for the OLPC XO contained precious little which could help teachers. Teachers weren’t the target audience. Children and governments were.

not silver bullets

Acknowledging, in an oblique way, that the situation is more complex.

surveys of students

With a clear Hawthorne effect.

parents rolling their eyes

Interesting appeal to parenting experience. Even more than teachers, they’re absent from many of these projects. Not a new pattern. Literacy projects often forget parents and the implications in terms of a generation gap. But what is perhaps more striking is that parents are also invisible in coverage of many of these issues. Contrary to “our” children, children in “those poor countries over there” are “ours to care for,” through development projects, adoptions, future immigration, etc.

evaluation of an OLPC project in Haiti

Sounds more like a pilot project than like field research. But maybe it’s more insightful.

Repeated calls and e-mails to OLPC and Negroponte seeking comment on OLPC did not receive a response

Such statements are “standard procedure” for journalists. But what is striking about this one is where it’s placed in the piece. Not only is it near the end of the argumentation but it’s in a series of comments about alternative views on the OLPC project. Whether or not it was done on purpose, the effect that we get is that there are two main voices, pro and con. Those on the con side can only have arguments in the same line of thought (about the project’s cost and “efficacy,” with possible comments about management). Those on the pro side are put in a defensive position.
In such cases, responsiveness is often key. Though Negroponte has been an effective marketer of his pet project, the fact that he explicitly refuses to respond to criticisms and critiques makes for an even more constrained offense/defense game.

ironic

Strong words, in such a context. Because it’s not the situation which is ironic. It’s a lack of action in a very specific domain.

the Third World

Interesting that the antiquated “Third World” expression comes in two contexts: the alleged target of the OLPC project (with little discussion as to what was meant by that relationship) and as the J-PAL field of expertise.

a leader in

Peacock terms or J-PAL are on the Miller-McCune lovelist?

There are

This is where the piece switches. We’re not talking about the OLPC, anymore. We reduce OLPC to a single goal, which has allegedly not been met, and propose that there are better ways to achieve this goal. Easy and efficient technique, but there still seems to be something missing.

etting children in developing countries into school and helping them learn more while they are there

A more specific goal than it might seem, at first blush.
For a very simple example: how about homeschooling?

proven successful

“We have proof!”

cheap

One might have expected “inexpensive,” here, instead of “cheap.” But, still, the emphasis is on cost.

deworming

Sounds a little bit surprising a switch from computer tech to public health.

50 cents per child per year

$4 per student per year

30 percent increase in lifetime earnings

technology-based approaches to improving student learning in the developing world

Coming back to technology, to an extent, but almost in passing. Technology, here, can still be a saviour. The issue would be to find the key technology to solve that one problem (student learning in the developing world needs calls for improvement). Rather limited in scope, depth, insight.

show more promise than one laptop per child

Perhaps the comment most directly related to opinions. “Showing promise” is closer to “instinctive appeal” but, in this case, it’s a positive. We don’t need to apply critical thinking to something which shows promise. It’s undeniably good. Right?

the J-PAL co-founder

There we are!

$2.20

Remedial education

A study in Kenya

Reference needed.

it didn’t matter

Sounds like a bold statement, as it’s not expressly linked to the scope of the study. It probably did matter. Just not in terms of what was measured. Mattering has to do with significance in general, not just with statistical significance.