What Do URIs Mean Anyway?

Jul 5, 2011

If you’ve hung around in linked data circles for any amount of time, you’ll probably have come across the httpRange-14 issue. This was an issue placed before the W3C TAG years and years ago which has become a permathread on semantic web and linked data mailing lists. The basic question (or my interpretation of it) is:

Given that URIs can sometimes be used to name things that aren’t on the web (eg the novel Moby Dick) and sometimes things that are (eg the Wikipedia page about Moby Dick), how can you tell, for a given URI, how it’s being used so that you can work out what a statement (say, about its author) means?

TAG F2F, June 2011

Jun 17, 2011

As you may know, I accepted an appointment to the W3C’s Technical Architecture Group earlier this year. Last week was the first face-to-face meeting that I attended, hosted in the Stata Center at MIT. As you can tell from the agenda (which was in fact revised as we went along) it was a packed three days.

What I intend to do here is to briefly report on the major areas that we discussed and give a tiny bit of my own personal take on them. In no way should any of what I write here be judged as revealing the official opinion of the TAG, it’s just me saying what I think, and I’m not going to go into anything in depth because they’re all incredibly gnarly and contentious topics and I’d not only be here all year but also end up in a tar pit.

Schema.org and the Responsibility of Monopoly

Jun 12, 2011

Update: This post has been translated to Italian on the Linked Open Data Italia blog.

In this post about schema.org I’m going to speculate about the economic drivers that affect how search engines use structured metadata on the web. I discuss how the technical features and choices within schema.org may cause wider long-term harm, and the role of open standards as a method for responsible companies to avoid the pitfalls of monopoly.

Lessons for Microdata from schema.org

Jun 10, 2011

There is (obviously, from the way my tweet stream, feed reader and email have filled up) lots to say at many levels about schema.org, a new collaboration between Google, Microsoft and Yahoo! that describes the next phase in search engines’ extraction of semantics from web pages. In this post I’m going to focus on what we can learn from schema.org about the design of microdata and how it might be improved.

Three challenges for alpha.gov.uk

May 20, 2011

The new alpha.gov.uk website was launched recently, as a prototype for the “single Government website” described in Martha Lane Fox’s report Directgov 2010 and Beyond: Revolution Not Evolution. Apparently the real deal could go live “in about a year”.

The site is lovely, a far cry from the standard government fare. But this isn’t exactly surprising: it’s been developed using modern technologies by a top team with a set of design rules far removed from those usually applied to government websites, a budget that’s not exactly tight and using an Agile methodology. These factors mark it out from the majority (though not all) government websites. And this is part of the point, to illustrate the gap between what we have and what a revolution could bring.

There are three challenges where I am and have been particularly interested to see the alpha.gov.uk approach. These are in balancing:

  • simplicity and complexity
  • centralisation and distribution
  • end-user and data re-user

It is not currently clear to me whether alpha.gov.uk has decided an approach on any of these – whether the way the site works currently is the way that they have decided it should work – or whether these are areas that are still up in the air at the moment. I’m hoping it’s the latter.