Over at democracy corner, Manchester Digital is interviewing all of its elected council members. Somehow, I got volunteered to be first interviewee. Here’s my two pence on one of the questions asked: “What do you think is biggest challenge we face as an industry?” (with some extra links)

Firstly, coding and “computational thinking” [1], needs to be understood as something that isn’t just for developers, geeks, coders, techies, boffins or “whizz kids” – as the Manchester Evening News likes to call them. Computational thinking, the ability to understand problems and provide innovative solutions in software and hardware, is a fundamental skill that everyone can learn, starting in primary school. As well as being fun to learn and practice, it is a crucial skill in a wide range of organisations in digital and beyond. Thankfully, the new computing curriculum in UK schools has recognised and addressed this, but it remains to be seen what the long-term impact of the changes in primary & secondary education will be on employers.

Secondly, as an industry, both the digital and technology sectors are seriously hindered by gender imbalance. If only 10-20% of employees are female, then large numbers of talented people are being excluded from the sector – bad news for everyone.

Is that reasonable – or have I missed the point? Are there more pressing issues facing the technology sector? Either way, you can read the rest of the interview at manchesterdigital.com/democracy-corner which will be supplemented with more interviews of council members every week over the next few months.

Back in 2006 the BBC published another impressive application that allowed users to search and browse over 75 years of programme data. The programme was built from metadata, not the actual audio and visual data from the TV and Radio, but the data that comes after-data, information about the programmes from an internal database known as Infax [2,3].

“Thank you for your continued interest in the BBC Programme Catalogue. The BBC is now looking into how this data can be incorporated into its programme information pages.”

You can still get some BBC programme metadata from bbc.co.uk/programmes and bbc.co.uk/archive. Every programme has publicly available metadata but only a fraction of what was in the open catalogue. Although the app has gone, lots of the data must still be there somewhere. Take for example, the opening ceremony for the Olympic Games…

The metadata that is currently available

The metadata for each BBC programme can be found via its own page, so the opening ceremony programme has metadata available in xml and rdf which tells you several things including this synopsis:

“Coverage of the opening ceremony, which officially starts at 9.00, with the eyes of the world focused on the Olympic Stadium as the 30th Olympiad is officially declared open by Her Majesty the Queen. Film director Danny Boyle is set to produce a stunning cultural show ahead of the athletes’ parade, during which over 200 countries are expected to be represented. This is followed by the official opening, the arrival of the torch and the lighting of the cauldron.”

The metadata also tells you that this particular programme was presented by Sue Barker, Huw Edwards, Gary Lineker, Jake Humphrey and Mishal Husain. The Executive Producer was Paul Davies and there’s a bunch of other stuff: date of first broadcast, links to related information and clips but that’s about it.

The metadata that used to be available

The great thing about the open catalogue was that it went into lots more detail than above. So, for the Olympics ceremony, the participants in the programme would have been listed as Danny Boyle, Daniel Craig, Thomas Heatherwick, Elizbaeth II, Paul McCartney, Rowan Atkinson, Bradley Wiggins, Kenneth Brannagh, Steve Redgrave, J.K. Rowling and so on. For each contributor, you could see what other programmes they had been involved in, not just recently broadcast ones, but those going back 75 years. You could also see who had collaborated with who and when their first broadcast was and so on. It didn’t just document the über-famous people either, it went into just as much detail about other people you might not necessarily have heard of like Frank Cottrell Boyce, Callum Airlie and Jordan Duckitt. It was great stuff, but neither the archive or current programmes seem to have this level of detail.

Meta-conclusions

It’s a bit of a mystery where all the lovely BBC metadata went, it’s probably just sitting on some servers somewhere, inaccessible to the outside world. With my licence fee paying hat on, this seems a bit of a waste. I’ve asked everyone I know, including people at the beeb, but have drawn a blank. Most have shrugged their shoulders and pointed to the useful but slightly impoverished /programmes and /archive which is why I’m writing this post on t’interwebs.

Maybe the Olympic task of curating all that data makes it un-sustainable. Perhaps somebody decided there is no point competing with wikipedia where wiki-nerds curate programme data for free? It’s possible you can’t justify serving big metadata without giving the actual data (programmes) too? Maybe there’s a shiny new application in the pipeline to replace the catalogue, currently being worked on or an upgrade to @ArchiveAtBBC & @programmes so they include much more data. Could there be issues with publishing this kind of personal data on the web which meant the whole thing got canned? Nasty copyright issues could probably sink a project like this too. Who knows…

Does anyone reading this know the answers? If you do, I’d love to hear from you.

Karen Loasby (2006). Changing approaches to metadata at bbc.co.uk: From chaos to control and then letting go again, Bulletin of the American Society for Information Science and Technology, 33 (1) 26. DOI: 10.1002/bult.2006.1720330109

The AAAI conference finished last Thursday, here are some highlights and papers that might be worth reading if you are interested in building and / or using a more “intelligent” (and possibly semantic) web in bioinformatics.

Here are the papers or talks I enjoyed the most and hope you might also find them useful or inspiring.

Intelligent agents must be able to handle the complexity and uncertainty of the real world. Logical AI (of which the semantic web is an example) has focused mainly on the former, and statistical AI (e.g. machine learning) on the latter. The two approaches can be united, with significant benefits, some of which are demonstrated by the Alchemy system

Bruce Buchanan gave a talk What Do We Know About Knowledge. A roller-coaster ride through the last 2000+ years of human attempts to understand what knowledge is, how to represent it and why it is powerful

Turing’s dream

Appropriately, the conference which was subtitled Celebrating 50 years of AI finished with two talks by Lenhart K. Schubert and Stuart M. Shieber about the Turing test. The first discussed Turing’s dream and the Knowledge Challenge, the second talk asked Does the Turing Test Demonstrate Intelligence or Not? Now I’m back in Manchester, where Turing once worked, I can’t help wondering, what would Alan make of the current state of AI and the semantic web? I think there are several possibilities, he could be thinking:

EITHER: Fifty odd years later, they’re not still wasting time working on that Turing test are they?!

OR: He is smugly satisifed that he devised a test, that no machine has passed, and perhaps never will, but has provided us with a satisfactory operational definition of “intelligence” ;

…AND What the hell is the “Semantic Web”?

We will never know what Alan Turing would make of todays efforts to make a more intelligent web. However, that won’t stop me speculating that he would be impressed by the current uses of computers (intelligent or otherwise) to drive robots through the desert, perform all sort of computations on proteins and to search for information on this massive distributed global knowledge-base we call the “Web”. Not bad for 50 years of work, here’s to the next 50…