Pretty much any academic eventually feels that s/he is misunderstood. Misunderstandings about some core notions in about any academic field are involved in some of the most common pet peeves among academics.

In other words, there’s nothing as transdisciplinary as misunderstanding.

It can happen in the close proximity of a given department (“colleagues in my department misunderstand my work”). It can happen through disciplinary boundaries (“people in that field have always misunderstood our field”). And, it can happen generally: “Nobody gets us.”

It’s not paranoia and it’s probably not self-victimization. But there almost seems to be a form of “onedownmanship” at stake with academics from different disciplines claiming that they’re more misunderstood than others. In fact, I personally get the feeling that ethnographers are more among the most misunderstood people around, but even short discussions with friends in other fields (including mathematics) have helped me get the idea that, basically, we’re all misunderstood at the same “level” but there are variations in the ways we’re misunderstood. For instance, anthropologists in general are mistaken for what they aren’t based on partial understanding by the general population.

An example from my own experience, related to my decision to call myself an “informal ethnographer.” When you tell people you’re an anthropologist, they form an image in their minds which is very likely to be inaccurate. But they do typically have an image in their minds. On the other hand, very few people have any idea about what “ethnography” means, so they’re less likely to form an opinion of what you do from prior knowledge. They may puzzle over the term and try to take a guess as to what “ethnographer” might mean but, in my experience, calling myself an “ethnographer” has been a more efficient way to be understood than calling myself an “anthropologist.”

These misunderstandings have an impact on anyone’s work. Not only can it be demoralizing and can it impact one’s sense of self-worth, but it can influence funding decisions as well as the use of research results. These misunderstandings can underminine learning across disciplines. In survey courses, basic misunderstandings can make things very difficult for everyone. At a rather basic level, academics fight misunderstandings more than they fight ignorance.

The main reason I’m discussing this is that I’ve been given several occasions to think about the interface between the Ivory Tower and the rest of the world. It’s been a major theme in my blogposts about intellectuals, especially the ones in French. Two years ago, for instance, I wrote a post in French about popularizers. A bit more recently, I’ve been blogging about specific instances of misunderstandings associated with popularizers, including Malcolm Gladwell’s approach to expertise. Last year, I did a podcast episode about ethnography and the Ivory Tower. And, just within the past few weeks, I’ve been reading a few things which all seem to me to connect with this same issue: common misunderstandings about academic work. The connections are my own, and may not be so obvious to anyone else. But they’re part of my motivations to blog about this important issue.

In no particular order:

A thread on a mailing-list about linguistic anthropology. A paleoanthropologist interviewed for a radio show discussed language and cognitive evolution in a way which seemed to some linguistic anthropologists as conveying some misunderstandings about language.

Two blogposts by LibriVox founder Hugh McGuire about “Why Academics Should Blog.” The first post made me react and the second post was in a small part motivated by my reaction. (As an aside, McGuire should be commended for his flexibility of thoughts. His abilirty to adapt his ideas as the result of thoughtful discussion has helped me have less “visceral” reactions.)

One discussion I remember, which seems to fit, included comments about Germaine Dieterlen by a friend who also did research in West Africa. Can’t remember the specifics but the gist of my friend’s comment was that “you get to respect work by the likes of Germaine Dieterlen once you start doing field research in the region.” In my academic background, appreciation of Germaine Dieterlen’s may not be unconditional, but it doesn’t necessarily rely on extensive work in the field. In other words, while some parts of Dieterlen’s work may be controversial and it’s extremely likely that she “got a lot of things wrong,” her work seems to be taken seriously by several French-speaking africanists I’ve met. And not only do I respect everyone but I would likely praise someone who was able to work in the field for so long. She’s not my heroine (I don’t really have heroes) or my role-model, but it wouldn’t have occurred to me that respect for her wasn’t widespread. If it had seemed that Dieterlen’s work had been misunderstood, my reflex would possibly have been to rehabilitate her.

In fact, there’s a strong academic tradition of rehabilitating deceased scholars. The first example which comes to mind is a series of articles (PDF, in French) and book chapters by UWO linguistic anthropologist Regna Darnell.about “Benjamin Lee Whorf as a key figure in linguistic anthropology.” Of course, saying that these texts by Darnell constitute a rehabilitation of Whorf reveals a type of evaluation of her work. But that evaluation comes from a third person, not from me. The likely reason for this case coming up to my mind is that the so-called “Sapir-Whorf Hypothesis” is among the most misunderstood notions from linguistic anthropology. Moreover, both Whorf and Sapir are frequently misunderstood, which can make matters difficulty for many linguistic anthropologists talking with people outside the discipline.

The opposite process is also common: the “slaughtering” of “sacred cows.” (First heard about sacred cows through an article by ethnomusicologist Marcia Herndon.) In some significant ways, any scholar (alive or not) can be the object of not only critiques and criticisms but a kind of off-handed dismissal. Though this often happens within an academic context, the effects are especially lasting outside of academia. In other words, any scholar’s name is likely to be “sullied,” at one point or another. Typically, there seems to be a correlation between the popularity of a scholar and the likelihood of her/his reputation being significantly tarnished at some point in time. While there may still be people who treat Darwin, Freud, Nietzsche, Socrates, Einstein, or Rousseau as near divinities, there are people who will avoid any discussion about anything they’ve done or said. One way to put it is that they’re all misunderstood. Another way to put it is that their main insights have seeped through “common knowledge” but that their individual reputations have decreased.

Perhaps the most difficult case to discuss is that of Marx (Karl, not Harpo). Textbooks in introductory sociology typically have him as a key figure in the discipline and it seems clear that his insight on social issues was fundamental in social sciences. But, outside of some key academic contexts, his name is associated with a large series of social events about which people tend to have rather negative reactions. Even more so than for Paul de Man or Martin Heidegger, Marx’s work is entangled in public opinion about his ideas. Haven’t checked for examples but I’m quite sure that Marx’s work is banned in a number of academic contexts. However, even some of Marx’s most ardent opponents are likely to agree with several aspects of Marx’s work and it’s sometimes funny how Marxian some anti-Marxists may be.

But I digress…

Typically, the “slaughtering of sacred cows” relates to disciplinary boundaries instead of social ones. At least, there’s a significant difference between your discipline’s own “sacred cows” and what you perceive another discipline’s “sacred cows” to be. Within a discipline, the process of dismissing a prior scholar’s work is almost œdipean (speaking of Freud). But dismissal of another discipline’s key figures is tantamount to a rejection of that other discipline. It’s one thing for a physicist to show that Newton was an alchemist. It’d be another thing entirely for a social scientist to deconstruct James Watson’s comments about race or for a theologian to argue with Darwin. Though discussions may have to do with individuals, the effects of the latter can widen gaps between scholarly disciplines.

And speaking of disciplinarity, there’s a whole set of issues having to do with discussions “outside of someone’s area of expertise.” On one side, comments made by academics about issues outside of their individual areas of expertise can be very tricky and can occasionally contribute to core misunderstandings. The fear of “talking through one’s hat” is quite significant, in no small part because a scholar’s prestige and esteem may greatly decrease as a result of some blatantly inaccurate statements (although some award-winning scholars seem not to be overly impacted by such issues).

On the other side, scholars who have to impart expert knowledge to people outside of their discipline often have to “water down” or “boil down” their ideas and, in effect, oversimplifying these issues and concepts. Partly because of status (prestige and esteem), lowering standards is also very tricky. In some ways, this second situation may be more interesting. And it seems unavoidable.

How can you prevent misunderstandings when people may not have the necessary background to understand what you’re saying?

This question may reveal a rather specific attitude: “it’s their fault if they don’t understand.” Such an attitude may even be widespread. Seems to me, it’s not rare to hear someone gloating about other people “getting it wrong,” with the suggestion that “we got it right.” As part of negotiations surrounding expert status, such an attitude could even be a pretty rational approach. If you’re trying to position yourself as an expert and don’t suffer from an “impostor syndrome,” you can easily get the impression that non-specialists have it all wrong and that only experts like you can get to the truth. Yes, I’m being somewhat sarcastic and caricatural, here. Academics aren’t frequently that dismissive of other people’s difficulties understanding what seem like simple concepts. But, in the gap between academics and the general population a special type of intellectual snobbery can sometimes be found.

Obviously, I have a lot more to say about misunderstood academics. For instance, I wanted to address specific issues related to each of the links above. I also had pet peeves about widespread use of concepts and issues like “communities” and “Eskimo words for snow” about which I sometimes need to vent. And I originally wanted this post to be about “cultural awareness,” which ends up being a core aspect of my work. I even had what I might consider a “neat” bit about public opinion. Not to mention my whole discussion of academic obfuscation (remind me about “we-ness and distinction”).

But this is probably long enough and the timing is right for me to do something else.

I’ll end with an unverified anecdote that I like. This anecdote speaks to snobbery toward academics.

[It’s one of those anecdotes which was mentioned in a course I took a long time ago. Even if it’s completely fallacious, it’s still inspiring, like a tale, cautionary or otherwise.]

As the story goes (at least, what I remember of it), some ethnographers had been doing fieldwork in an Australian cultural context and were focusing their research on a complexkinship system known in this context. Through collaboration with “key informants,” the ethnographers eventually succeeded in understanding some key aspects of this kinship system.

As should be expected, these kinship-focused ethnographers wrote accounts of this kinship system at the end of their field research and became known as specialists of this system.

After a while, the fieldworkers went back to the field and met with the same people who had described this kinship system during the initial field trip. Through these discussions with their “key informants,” the ethnographers end up hearing about a radically different kinship system from the one about which they had learnt, written, and taught.

The local informants then told the ethnographers: “We would have told you earlier about this but we didn’t think you were able to understand it.”

This is mostly meant as a reply to this blogthread. But it’s also more generally about my personal reaction to Apple’s iPad announcement.

Some background.

I’m an ethnographer and a teacher. I read a fair deal, write a lot of notes, and work in a variety of contexts. These days, I tend to spend a good amount of time in cafés and other public places where I like to work without being too isolated. I also commute using public transit, listen to lots of podcast, and create my own. I’m also very aural.

I’ve used a number of PDAs, over the years, from a Newton MessagePad 130 (1997) to a variety of PalmOS devices (until 2008). In fact, some people readily associated me with PDA use.

As soon as I learnt about the iPod touch, I needed one. As soon as I’ve heard about the SafariPad, I wanted one. I’ve been an intense ‘touch user since the iPhone OS 2.0 release and I’m a happy camper.

(A major reason I never bought an iPhone, apart from price, is that it requires a contract.)

In my experience, the ‘touch is the most appropriate device for all sorts of activities which are either part of an other activity (reading during a commute) or are simply too short in duration to constitute an actual “computer session.” You don’t “sit down to work at your ‘touch” the way you might sit in front of a laptop or desktop screen. This works great for “looking up stufff” or “checking email.” It also makes a lot of sense during commutes in crowded buses or metros.

In those cases, the iPod touch is almost ideal. Ubiquitous access to Internet would be nice, but that’s not a deal-breaker. Alternative text-input methods would help in some cases, but I do end up being about as fast on my ‘touch as I was with Graffiti on PalmOS.

For other tasks, I have a Mac mini. Sure, it’s limited. But it does the job. In fact, I have no intention of switching for another desktop and I even have an eMachines collecting dust (it’s too noisy to make a good server).

What I miss, though, is a laptop. I used an iBook G3 for several years and loved it. For a little while later, I was able to share a MacBook with somebody else and it was a wonderful experience. I even got to play with the OLPC XO for a few weeks. That one was not so pleasant an experience but it did give me a taste for netbooks. And it made me think about other types of iPhone-like devices. Especially in educational contexts. (As I mentioned, I’m a teacher)

I’ve been laptop-less for a while, now. And though my ‘touch replaces it in many contexts, there are still times when I’d really need a laptop. And these have to do with what I might call “mobile sessions.”

For instance: liveblogging a conference or meeting. I’ve used my ‘touch for this very purpose on a good number of occasions. But it gets rather uncomfortable, after a while, and it’s not very fast. A laptop is better for this, with a keyboard and a larger form factor. But the iPad will be even better because of lower risks of RSI. A related example: just imagine TweetDeck on iPad.

Possibly my favourite example of a context in which the iPad will be ideal: presentations. Even before learning about the prospect of getting iWork on a tablet, presentations were a context in which I really missed a laptop.

Sure, in most cases, these days, there’s a computer (usually a desktop running XP) hooked to a projector. You just need to download your presentation file from Slideshare, show it from Prezi, or transfer it through USB. No biggie.

But it’s not the extra steps which change everything. It’s the uncertainty. Even if it’s often unfounded, I usually get worried that something might just not work, along the way. The slides might not show the same way as you see it because something is missing on that computer or that computer is simply using a different version of the presentation software. In fact, that software is typically Microsoft PowerPoint which, while convenient, fits much less in my workflow than does Apple Keynote.

The other big thing about presentations is the “presenter mode,” allowing you to get more content than (or different content from) what the audience sees. In most contexts where I’ve used someone else’s computer to do a presentation, the projector was mirroring the computer’s screen, not using it as a different space. PowerPoint has this convenient “presenter view” but very rarely did I see it as an available option on “the computer in the room.” I wish I could use my ‘touch to drive presentations, which I could do if I installed software on that “computer in the room.” But it’s not something that is likely to happen, in most cases.

A MacBook solves all of these problems. and it’s an obvious use for laptops. But how, then, is the iPad better? Basically because of interface. Switching slides on a laptop isn’t hard, but it’s more awkward than we realize. Even before watching the demo of Keynote on the iPad, I could simply imagine the actual pleasure of flipping through slides using a touch interface. The fit is “natural.”

I sincerely think that Keynote on the iPad will change a number of things, for me. Including the way I teach.

Then, there’s reading.

Now, I’m not one of those people who just can’t read on a computer screen. In fact, I even grade assignments directly from the screen. But I must admit that online reading hasn’t been ideal, for me. I’ve read full books as PDF files or dedicated formats on PalmOS, but it wasn’t so much fun, in terms of the reading process. And I’ve used my ‘touch to read things through Stanza or ReadItLater. But it doesn’t work so well for longer reading sessions. Even in terms of holding the ‘touch, it’s not so obvious. And, what’s funny, even a laptop isn’t that ideal, for me, as a reading device. In a sense, this is when the keyboard “gets in the way.”

Sure, I could get a Kindle. I’m not a big fan of dedicated devices and, at least on paper, I find the Kindle a bit limited for my needs. Especially in terms of sources. I’d like to be able to use documents in a variety of formats and put them in a reading list, for extended reading sessions. No, not “curled up in bed.” But maybe lying down in a sofa without external lighting. Given my experience with the ‘touch, the iPad is very likely the ideal device for this.

Then, there’s the overall “multi-touch device” thing. People have already been quite creative with the small touchscreen on iPhones and ‘touches, I can just imagine what may be done with a larger screen. Lots has been said about differences in “screen real estate” in laptop or desktop screens. We all know it can make a big difference in terms of what you can display at the same time. In some cases, two screens isn’t even a luxury, for instance when you code and display a page at the same time (LaTeX, CSS…). Certainly, the same qualitative difference applies to multitouch devices. Probably even more so, since the display is also used for input. What Han found missing in the iPhone’s multitouch was the ability to use both hands. With the iPad, Han’s vision is finding its space.

Oh, sure, the iPad is very restricted. For instance, it’s easy to imagine how much more useful it’d be if it did support multitasking with third-party apps. And a front-facing camera is something I was expecting in the first iPhone. It would just make so much sense that a friend seems very disappointed by this lack of videoconferencing potential. But we’re probably talking about predetermined expectations, here. We’re comparing the iPad with something we had in mind.

Then, there’s the issue of the competition. Tablets have been released and some multitouch tablets have recently been announced. What makes the iPad better than these? Well, we could all get in the same OS wars as have been happening with laptops and desktops. In my case, the investment in applications, files, and expertise that I have made in a Mac ecosystem rendered my XP years relatively uncomfortable and me appreciate returning to the Mac. My iPod touch fits right in that context. Oh, sure, I could use it with a Windows machine, which is in fact what I did for the first several months. But the relationship between the iPhone OS and Mac OS X is such that using devices in those two systems is much more efficient, in terms of my own workflow, than I could get while using XP and iPhone OS. There are some technical dimensions to this, such as the integration between iCal and the iPhone OS Calendar, or even the filesystem. But I’m actually thinking more about the cognitive dimensions of recognizing some of the same interface elements. “Look and feel” isn’t just about shiny and “purty.” It’s about interactions between a human brain, a complex sensorimotor apparatus, and a machine. Things go more quickly when you don’t have to think too much about where some tools are, as you’re working.

So my reasons for wanting an iPad aren’t about being dazzled by a revolutionary device. They are about the right tool for the job.

This time, I’m heavily editing my comments. So it’s less of a repost than a new blogpost. In some ways, it’s partly a follow-up to my “Ultimate Handheld Device” post (which ended up focusing on spatial positioning).

Given the OLPC context, the angle here is, hopefully, a culturally aware version of “a handheld device for the rest of us.”

Here goes…

I think there’s room in the World for a device category more similar to handhelds than to subnotebooks. Let’s call it “handhelds for the rest of us” (HftRoU). Something between a cellphone, a portable gaming console, a portable media player, and a personal digital assistant. Handheld devices exist which cover most of these features/applications, but I’m mostly using this categorization to think about the future of handhelds in a globalised World.

The “new” device category could serve as the inspiration for a follow-up to the OLPC project. One thing about which I keep thinking, in relation to the “OLPC” project, is that the ‘L’ part was too restrictive. Sure, laptops can be great tools for students, especially if these students are used to (or need to be trained in) working with and typing long-form text. But I don’t think that laptops represent the most “disruptive technology” around. If we think about their global penetration and widespread impact, cellphones are much closer to the leapfrog effect about which we all have been writing.

So, why not just talk about a cellphone or smartphone? Well, I’m trying to think both more broadly and more specifically. Cellphones are already helping people empower themselves. The next step might to add selected features which bring them closer to the OLPC dream. Also, since cellphones are widely distributed already, I think it’s important to think about devices which may complement cellphones. I have some ideas about non-handheld tools which could make cellphones even more relevant in people’s lives. But they will have to wait for another blogpost.

So, to put it simply, “handhelds for the rest of us” (HftRoU) are somewhere between the OLPC XO-1 and Apple’s original iPhone, in terms of features. In terms of prices, I dream that it could be closer to that of basic cellphones which are in the hands of so many people across the globe. I don’t know what that price may be but I heard things which sounded like a third of the price the OLPC originally had in mind (so, a sixth of the current price). Sure, it may take a while before such a low cost can be reached. But I actually don’t think we’re in a hurry.

I guess I’m just thinking of the electronics (and global) version of the Ford T. With more solidarity in mind. And cultural awareness.

Google’s Open Handset Alliance (OHA) may produce something more appropriate to “global contexts” than Apple’s iPhone. In comparison with Apple’s iPhone, devices developed by the OHA could be better adapted to the cultural, climatic, and economic conditions of those people who don’t have easy access to the kind of computers “we” take for granted. At the very least, the OHA has good representation on at least three continents and, like the old OLPC project, the OHA is officially dedicated to openness.

I actually care fairly little about which teams will develop devices in this category. In fact, I hope that new manufacturers will spring up in some local communities and that major manufacturers will pay attention.

I don’t care about who does it, I’m mostly interested in what the devices will make possible. Learning, broadly speaking. Communicating, in different ways. Empowering themselves, generally.

One thing I have in mind, and which deviates from the OLPC mission, is that there should be appropriate handheld devices for all age-ranges. I do understand the focus on 6-12 year-olds the old OLPC had. But I don’t think it’s very productive to only sell devices to that age-range. Especially not in those parts of the world (i.e., almost anywhere) where generation gaps don’t imply that children are isolated from adults. In fact, as an anthropologist, I react rather strongly to the thought that children should be the exclusive target of a project meant to empower people. But I digress, as always.

I don’t tend to be a feature-freak but I have been thinking about the main features the prototypical device in this category should have. It’s not a rigid set of guidelines. It’s just a way to think out loud about technology’s integration in human life.

The OS and GUI, which seem like major advantages of the eeePC, could certainly be of the mobile/handheld type instead of the desktop/laptop type. The usual suspects: Symbian, NewtonOS, Android, Zune, PalmOS, Cocoa Touch, embedded Linux, Playstation Portable, WindowsCE, and Nintendo DS. At a certain level of abstraction, there are so many commonalities between all of these that it doesn’t seem very efficient to invent a completely new GUI/OS “paradigm,” like OLPC’s Sugar was apparently trying to do.

The HftRoU require some form of networking or wireless connectivity feature. WiFi (802.11*), GSM, UMTS, WiMAX, Bluetooth… Doesn’t need to be extremely fast, but it should be flexible and it absolutely cannot be cost-prohibitive. IP might make much more sense than, say, SMS/MMS, but a lot can be done with any kind of data transmission between devices. XO-style mesh networking could be a very interesting option. As VoIP has proven, voice can efficiently be transmitted as data so “voice networks” aren’t necessary.

My sense is that a multitouch interface with an accelerometer would be extremely effective. Yes, I’m thinking of Apple’s Touch devices and MacBooks. As well as about the Microsoft Surface, and Jeff Han’s Perceptive Pixel. One thing all of these have shown is how “intuitive” it can be to interact with a machine using gestures. Haptic feedback could also be useful but I’m not convinced it’s “there yet.”

I’m really not sure a keyboard is very important. In fact, I think that keyboard-focused laptops and tablets are the wrong basis for thinking about “handhelds for the rest of us.” Bear in mind that I’m not thinking about devices for would-be office workers or even programmers. I’m thinking about the broadest user base you can imagine. “The Rest of Us” in the sense of, those not already using computers very directly. And that user base isn’t that invested in (or committed to) touch-typing. Even people who are very literate don’t tend to be extremely efficient typists. If we think about global literacy rates, typing might be one thing which needs to be leapfrogged. After all, a cellphone keypad can be quite effective in some hands and there are several other ways to input text, especially if typing isn’t too ingrained in you. Furthermore, keyboards aren’t that convenient in multilingual contexts (i.e., in most parts of the world). I say: avoid the keyboard altogether, make it available as an option, or use a virtual one. People will complain. But it’s a necessary step.

If the device is to be used for voice communication, some audio support is absolutely required. Even if voice communication isn’t part of it (and I’m not completely convinced it’s the one required feature), audio is very useful, IMHO (I’m an aural guy). In some parts of the world, speakers are much favoured over headphones or headsets. But I personally wish that at least some HftRoU could have external audio inputs/outputs. Maybe through USB or an iPod-style connector.

A voice interface would be fabulous, but there still seem to be technical issues with both speech recognition and speech synthesis. I used to work in that field and I keep dreaming, like Bill Gates and others do, that speech will finally take the world by storm. But maybe the time still hasn’t come.

It’s hard to tell what size the screen should be. There probably needs to be a range of devices with varying screen sizes. Apple’s Touch devices prove that you don’t need a very large screen to have an immersive experience. Maybe some HftRoU screens should in fact be larger than that of an iPhone or iPod touch. Especially if people are to read or write long-form text on them. Maybe the eeePC had it right. Especially if the devices’ form factor is more like a big handheld than like a small subnotebook (i.e., slimmer than an eeePC). One reason form factor matters, in my mind, is that it could make the devices “disappear.” That, and the difference between having a device on you (in your pocket) and carrying a bag with a device in it. Form factor was a big issue with my Newton MessagePad 130. As the OLPC XO showed, cost and power consumption are also important issues regarding screen size. I’d vote for a range of screens between 3.5 inch (iPhone) and 8.9 inch (eeePC 900) with a rather high resolution. A multitouch version of the XO’s screen could be a major contribution.

In terms of both audio and screen features, some consideration should be given to adaptive technologies. Most of us take for granted that “almost anyone” can hear and see. We usually don’t perceive major issues in the fact that “personal computing” typically focuses on visual and auditory stimuli. But if these devices truly are “for the rest of us,” they could help empower visually- or hearing-impaired individuals, who are often marginalized. This is especially relevant in the logic of humanitarianism.

HftRoU needs a much autonomy from a power source as possible. Both in terms of the number of hours devices can be operated without needing to be connected to a power source and in terms of flexibility in power sources. Power management is a major technological issue, with portable, handheld, and mobile devices. Engineers are hard at work, trying to find as many solutions to this issue as they can. This was, obviously, a major area of research for the OLPC. But I’m not even sure the solutions they have found are the only relevant ones for what I imagine HftRoU to be.

GPS could have interesting uses, but doesn’t seem very cost-effective. Other “wireless positioning systems” (à la Skyhook) might reprsent a more rational option. Still, I think positioning systems are one of the next big things. Not only for navigation or for location-based targeting. But for a set of “unintended uses” which are the hallmark of truly disruptive technology. I still remember an article (probably in the venerable Wired magazine) about the use of GPS/GIS for research into climate change. Such “unintended uses” are, in my mind, much closer to the constructionistideal than the OLPC XO’s unified design can ever get.

Though a camera seems to be a given in any portable or mobile device (even the OLPC XO has one), I’m not yet that clear on how important it really is. Sure, people like taking pictures or filming things. Yes, pictures taken through cellphones have had a lasting impact on social and cultural events. But I still get the feeling that the main reason cameras are included on so many devices is for impulse buying, not as a feature to be used so frequently by all users. Also, standalone cameras probably have a rather high level of penetration already and it might be best not to duplicate this type of feature. But, of course, a camera could easily be a differentiating factor between two devices in the same category. I don’t think that cameras should be absent from HftRoU. I just think it’s possible to have “killer apps” without cameras. Again, I’m biased.

Apart from networking/connectivity uses, Bluetooth seems like a luxury. Sure, it can be neat. But I don’t feel it adds that much functionality to HftRoU. Yet again, I could be proven wrong. Especially if networking and other inter-device communication are combined. At some abstract level, there isn’t that much difference between exchanging data across a network and controlling a device with another device.

Yes, I do realize I pretty much described an iPod touch (or an iPhone without camera, Bluetooth, or cellphone fees). I’ve been lusting over an iPod touch since September and it does colour my approach. I sincerely think the iPod touch could serve as an inspiration for a new device type. But, again, I care very little about which company makes that device. I don’t even care about how open the operating system is.

My own thoughts on the whole thing.
I appreciate the fact that Phil Schiller began the “enterprise” section of the event with comments about a university. Though universities need not be run like profit-hungry corporations, linking Apple’s long-standing educational focus with its newly invigorated enterprise focus makes sense. And I had a brief drift-off moment as I was thinking about Touch products in educational contexts.

I’m surprised at how enthusiastic I get about the enterprise features. Suddenly, I can see Microsoft’s Exchange make sense.

I get the clear impression that even more things will come into place at the end of June than has been said by Apple. Possibly new Touch models or lines. Probably the famous 3G iPhone. Apple-released apps. Renewed emphasis on server technology (XServe, Mac OS X Server, XSan…). New home WiFi products (AirPort, Time Capsule, Apple TV…). New partnerships. Cool VC-funded startups. New features on the less aptly named “iTunes” store.

Though it was obvious already, the accelerometer is an important feature. It seems especially well-adapted to games and casual gamers like myself are likely to enjoy games this feature makes possible. It can also lead to very interesting applications. In fact, the “Etch and Sketch” demo was rather convincing as a display of some core Touch features. These are exactly the features which help sell products.
Actually, I enjoyed the “wow factor” of the event’s demos. I’m convinced that it will energize developers and administrators, whether or not they plan on using Touch products. Some components of Apple’s Touch strategy are exciting enough that the more problematic aspects of this strategy may matter a bit less. Those of us dreaming about Android, OpenMoko, or even a revived NewtonOS can still find things to get inspired by in Apple’s roadmap.

What’s to come, apart from what was announced? No idea. But I do daydream about all of this.
I’m especially interested in the idea of Apple Touch as “mainstream, WiFi, mobile platform.” There’s a lot of potential for Apple-designed, WiFi-enabled handhelds. Whether or not they include a cellphone.
At this point, Apple only makes five models of Touch products: three iPod touches and two iPhones. Flash memory is the main differentiating factor within a line. It makes it relatively easy to decide which device to get but some product diversity could be interesting. While some people expect/hope that Apple will release radically new form factors for Touch devices (e.g., a tablet subnotebook), it’s quite likely that other features will help distinguish Apple’s Touch hardware.
Among features I’d like to get through software, add-ons, or included in a Touch product? Number of things, some alluded to in the “categories” for this post. Some of these I had already posted.

Uh-oh! No you didn’t! Well, Steve Jobs made a similar statement a long time ago so it’s not like you’re the first one to say it. But you’re still wrong!

(This blog entry will be choppier than usual. Should have posted this as a comment. But this is getting longer than I expected and I prefer trackbacks anyway.)

Not exactly sure where people in the Bay Area get the impression that there is no market for the traditional PDA. In my mind, the potential market for “the traditional PDA” is underestimated because the ideal PDA has yet to be released. No, the current crop of smartphones aren’t it.

Having said this, I do think cellphones have the brightest future of pretty much any other electronic device type, but I don’t agree that any cellphone currently does what a PDA can do. So, while I think the ideal portable device would likely be a cellphone, I’d like to focus on what a PDA really is.

While it’s clear that PDAs have had a tortuous history since the first Newton and Magic Cap devices were released, other tools haven’t completely obliterated the need for PDAs. Hence the “cult following” for Newton Message Pads and the interest in new generations of PDA-like devices.

One thing to keep in mind is that PDAs are not merely PIMs (personal information managers, typically focusing on contacts and calendars). Instead of a glorified organiser, a PDA is a complete computer with a focus on personal data. And people do care about personal data in computing.

Now, a disclaimer of sorts: I’ve been an active PDA user for a number of years. When I don’t have a PDA, I almost feel like something is missing from my life.

I have been taken by the very concept of PDAs the first time I saw an article on Apple’s Newton devices in a mainstream U.S. newspaper, way back in the early 1990s. I was dreaming of all the possibilities. And longed for my own Newton MessagePad.

I received a MessagePad 130 from Apple a few years later, having done some work for them. Used that Newton for a while and really enjoyed the experience. While human beings find my handwriting extremely difficult to read, my Newton MP130 did a fairly good job at recognising it. And having installed a version of Graffiti, I was able to write rather quickly on the device. The main issue I had with Newton devices was size. The MP was simply too bulky for me to carry around everywhere. I eventually stopped using the MP after a while, but was missing the convenience of my MP130.

I started using that Newton again in 2001, as I was preparing for fieldwork. Because I didn’t have a parallel port on the iBook I was getting for fieldwork, I also bought a Handspring Visor Deluxe. The Visor became a very valuable tool during my fieldwork trip to Mali, in 2002. IIRC, this model had already been discontinued but I had no trouble using it or finding new software for it. I used the Visor to take copious amounts of data which I was able to periodically transfer to my iBook. The fact that the Visor ran on standard batteries was definitely an asset in the field but I did lose data on occasion because, unlike Newton devices, Palm devices didn’t have persistent memory storage.

Coming back from Mali, I bought my first Sony Clié. I pretty much stuck with Cliés ever since and have been quite happy with them. Cliés have a few advantages over other PalmOS devices like MemorySticks and the jog dial. The form factor and screen resolution of an entry-level Clié were much better than those of my old Visor. Sony has discontinued sales of its Clié devices outside of Japan. Some used Cliés go for 30$ on eBay.

So, what do I do with a PDA? Actually, the main thing really is taking notes. Reading notes, research notes, lecture notes, conference notes, etc. I’ve taken notes on coffee I’ve tried, on things I’d like to learn, on moments I want to write more extensively about… Though my fingers are rather small, typing on a small QWERTY keyboard has never been a real option for me. I tried using the keyboard on a Clié NX70V and it wasn’t nearly as efficient as using Graffiti. In fact, I’ve become quite adept at MessagEase. I can usually take elaborate notes in real time and organise them as I wish. Some notes remain as snippets while other notes become part of bigger pieces, including much of what I’ve written in the past ten years.

I also use my PDA for a number of “simpler” things like converting units (volume and temperature, especially), playing games (while waiting for something or while listening to podcasts), setting different timers, planning trips on public transportation systems, etc. I used to try and use more PDA applications than I do now but I still find third party applications an important component of any real PDA.

I always wanted to have a WiFi-enabled PDA. It’s probably the main reason behind my original reaction to the iPod touch launch. With a good input system and a semi-ubiquitous WiFi connection, a WiFi-enabled PDA could be a “dream come true,” for me. Especially in terms of email, blogging, and social networking. Not to mention simple Web queries.

I do have a very clear idea in mind as to what would be my ideal PDA. I don’t need it to be an MP3 player, a gaming console, or a phone. I don’t really want it to have a qwerty keyboard or a still camera. I don’t even care so much about it having a colour screen. But it should have an excellent battery life, a small size, good synchronisation features, third party apps, persistent memory, a very efficient input system, and a user community. I dream of it having a high-quality sound recorder, a webcam (think videoconferencing), large amounts of memory, and a complete set of voice features perfectly tuned to its owner’s voice (like voice activation and speaker-dependent, continuous speech recognition). It could act as the perfect unit to store any kind of personal data as a kind of “smart thumbdrive.” It could be synchronised with almost any other machine without any loss of information. It would probably have GPS and location-enabled features. It could be used to drive other systems or act as the ultimate smartcard. And it should be inexpensive.

I personally think price is one of the main reasons the traditional PDA has had such a hard time building/reaching markets. Inexpensive PDAs tended to miss important features. The most interesting PDAs were as expensive as much more powerful computers. Surely, miniaturisation is costly and it never was possible for any company to release a really inexpensive yet full-featured PDA. So it may be accurate to say that the traditional PDA is too expensive for its potential market. I perceive a huge difference between problems associated with costs and the utter lack of any PDA market.

Price does tend to be a very important factor with computer technology. The OLPC project is a good example of this. While the laptop produced through this project has many other features, the one feature which caught most of the media attention was the expected price for the device, around USD$100. All this time, many people are thinking that the project should have been a cellphone project because cellphone penetration is already high and cellphones are already the perfect leapfrog tool.

So it’s unlikely that I will get my dream PDA any time soon. Chances are, I’ll end up having to use a smartphone with very few of the features I really want my PDA to have. But, as is my impression with the OLPC project, we still need to dream and talk about what these devices can be.

Yeah, everybody’s been talking about the iPhone. It’s last week’s story but it can still generate a fair bit of coverage. People are already thinking about the next models.

Apple has most of the technology to build what would be my dream handheld device but the iPhone isn’t it. Yet.

My wishful thinking for what could in fact be the coolest handheld ever. Of course, the device should have the most often discussed features which the iPhone currently misses (Flash, MMS, chat…). But I’m going much further, here.

Good quality audio recording (as with the recording add-ons for the iPod 5G).

Disk space (say, 80GB).

VoIP support (Skype or other, but as compatible as possible).

Video camera which can face the user (for videoconference).

Full voice interface: speech recognition and text-to-speech for dialing, commands, and text.

Handwriting recognition.

Stylus support.

Data transfer over Bluetooth.

TextEdit.

Adaptive technology for word recognition.

Not tied to cellular provider contract.

UMA Cell-to-WiFi (unlicensed mobile access).

GPS.

iLife support.

Sync with Mac OS X and Windows.

Truly international cellular coverage.

Outliner.

iWork support.

Disk mode.

Multilingual support.

Use as home account on Mac OS X “host.”

FrontRow

USB and Bluetooth printing.

Battery packs with standard batteries.

The key point here isn’t that the iPhone should be a mix between an iPod and a MacBook. I’m mostly thinking about the fact that the “Personal” part of the “PC” and “PDA” concepts has not come to fruition yet. Sure, your PC account has your preferences and some personal data. Your PDA contains your contacts and to-do lists. But you still end up with personal data in different places. Hence the need for Web apps. As we all know, web apps are quite useful but there’s still room for standalone applications, especially on a handheld. It wouldn’t take much for the iPhone to be the ideal tool to serve as a “universal home” where a user can edit and output files. To a musician or podcaster, it could become the ideal portable studio.

But where the logical step needs to be taken is in “personalization.” Apparently, the iPhone’s predictive keyboard doesn’t even learn from the user’s input. Since the iPhone is meant to be used by a single individual, it seems quite strange that it does not, minimally, adapt to typed input. Yet with a device already containing a headset it seems to me that speech technologies could be ideal. Full-text continuous speech recognition already exists and what it requires is exactly what the iPhone could provide: adaptation to a user’s voice and speech patterns. Though it may be awkward for people to use a voice interface in public, cellphones have created a whole group of people who seem to be talking to themselves. 😉

Though very different from speech recognition, text-to-speech could integrate really well with a voice-driven device. Sharing the same “dictionaries” across all applications on the same device, the TTS and SR features could be trained very specifically to a given user. While screens have been important on computers for quite a while, voice-activated computers have been prominent in science-fiction for probably as long. The most common tasks done on computers (writing messages, making appointments, entering data, querying databases…) could all be done quite effectively through a voice interface. And the iPhone could easily serve as a voice interface for other computers.