Tag Archives: information architect

Linking data and journalism: what’s the future?

On Wednesday (September 9), Paul Bradshaw, course director of the MA Online Journalism at Birmingham City University and founder of HelpMeInvestigate.com, chaired a discussion on data and the future of journalism at the first London Linked Data Meetup. This post originally appeared on the OnlineJournalismBlog.

The panel included: Martin Belam (information architect, the Guardian; blogger, Currybet; John O’Donovan (chief architect, BBC News Online); Dan Brickley (Friend of a Friend project; VU University, Amsterdam; SpyPixel Ltd; ex-W3C); Leigh Dodds (Talis).

“Linked Data is about using the web to connect related data that wasn’t previously linked, or using the web to lower the barriers to linking data currently linked using other methods.” (http://linkeddata.org)

I talked about how 2009 was, for me, a key year in data and journalism – largely because it has been a year of crisis in both publishing and government. The seminal point in all of this has been the MPs’ expenses story, which both demonstrated the power of data in journalism, and the need for transparency from government. For example: the government appointment of Sir Tim Berners-Lee, the search for developers to suggest things to do with public data, and the imminent launch of Data.gov.uk around the same issue.

Even before then the New York Times and Guardian both launched APIs at the beginning of the year, MSN Local and the BBC have both been working with Wikipedia and we’ve seen the launch of a number of startups and mashups around data including Timetric, Verifiable, BeVocal, OpenlyLocal, MashTheState, the open source release of Everyblock, and Mapumental.

Q: What are the implications of paywalls for Linked Data?
The general view was that Linked Data – specifically standards like RDF [Resource Description Format] – would allow users and organisations to access information about content even if they couldn’t access the content itself. To give a concrete example, rather than linking to a ‘wall’ that simply requires payment, it would be clearer what the content beyond that wall related to (e.g. key people, organisations, author, etc.)

Leigh Dodds felt that using standards like RDF would allow organisations to more effectively package content in commercially attractive ways, e.g. ‘everything about this organisation’.

Q: What can bloggers do to tap into the potential of Linked Data?
This drew some blank responses, but Leigh Dodds was most forthright, arguing that the onus lay with developers to do things that would make it easier for bloggers to, for example, visualise data. He also pointed out that currently if someone does something with data it is not possible to track that back to the source and that better tools would allow, effectively, an equivalent of pingback for data included in charts (e.g. the person who created the data would know that it had been used, as could others).

Q: Given that the problem for publishing lies in advertising rather than content, how can Linked Data help solve that?
Dan Brickley suggested that OAuth technologies (where you use a single login identity for multiple sites that contains information about your social connections, rather than creating a new ‘identity’ for each) would allow users to specify more specifically how they experience content, for instance: ‘I only want to see article comments by users who are also my Facebook and Twitter friends.’

The same technology would allow for more personalised, and therefore more lucrative, advertising. John O’Donovan felt the same could be said about content itself – more accurate data about content would allow for more specific selling of advertising.

Martin Belam quoted James Cridland on radio: ‘[The different operators] agree on technology but compete on content’. The same was true of advertising but the advertising and news industries needed to be more active in defining common standards.

Leigh Dodds pointed out that semantic data was already being used by companies serving advertising.

Other notes
I asked members of the audience who they felt were the heroes and villains of Linked Data in the news industry. The Guardian and BBC came out well – The Daily Mail were named as repeat offenders who would simply refer to ‘a study’ and not say which, nor link to it.

Martin Belam pointed out that the Guardian is increasingly asking itself ‘how will that look through an API?’ when producing content, representing a key shift in editorial thinking. If users of the platform are swallowing up significant bandwidth or driving significant traffic then that would probably warrant talking to them about more formal relationships (either customer-provider or partners).

A number of references were made to the problem of provenance – being able to identify where a statement came from. Dan Brickley specifically spoke of the problem with identifying the source of Twitter retweets.

Dan also felt that the problem of journalists not linking would be solved by technology. In conversation previously, he also talked of ‘subject-based linking’ and the impact of SKOS [Simple Knowledge Organisation System] and linked data style identifiers. He saw a problem in that, while new articles might link to older reports on the same issue, older reports were not updated with links to the new updates. Tagging individual articles was problematic in that you then had the equivalent of an overflowing inbox.

Finally, here’s a bit of video from the very last question addressed in the discussion (filmed with thanks by @countculture):

Linked Data London 090909 from Paul Bradshaw on Vimeo.

Resources:

#FollowJourn: @currybet/information architect

#FollowJourn: Martin Belam

Who? Internet consultant, information architect and blogger

What? Information architect for the Guardian and also writes the CurryBet blog.

Where? @currybet and http://www.currybet.net/.

Contact? martin dot belam at currybet dot net

Just as we like to supply you with fresh and innovative tips every day, we’re recommending journalists to follow online too. They might be from any sector of the industry: please send suggestions (you can nominate yourself) to judith or laura at journalism.co.uk; or to @journalismnews.

NewsInnovation videos from @newsmatters: featuring @kevglobal, @currybet, @markng, @simonw, @willperrin

The Media Standards Trust has finished uploading content from its NewsInnovation event, held in association with NESTA and the WSRI, earlier this month to its YouTube channel.

[Previous Journalism.co.uk coverage at this link]

We’ll embed the first segment of each session, and for further installments follow the links below each video.

Part 2, Part 3, Part 4, Part 5.

  • Kevin Anderson (@kevglobal) Guardian blogs editor talks about news business models.

Part 2, Part 3, Part 4.

  • Ben Campbell talks about the Media Standards Trust website, Journalisted.

Part 2, Part 3, Part 4.

  • Will Perrin (@willperrin) on digital possibilities for the Chilcot Inquiry into the Iraq War.

Part 2.

  • Simon Willison (@simonw) of The Guardian talks about using the crowd to sift through MPs’ expenses.

Part 2, Part 3, Part 4.

  • Martin Belam (@currybet) information architect at the Guardian on ‘The tyranny of chronology’.

Part 2, Part 3.

BeatBlogging.Org: ‘UK news regulation stands in the way of newsroom convergence’

I’ve provided a guest post for BeatBlogging.org, the US-based site that looks at how to use social networks and other web tools to improve beat reporting. Using examples from various Journalism.co.uk pieces, I argue that it is very difficult to look towards coverged newsroom, under the hybrid regulatory systems with which we operate as UK-based publishers. Thoughts welcomed.

Read it in full over at the site. Here’s an extract:

We talk about converging newsrooms of the future that transcend boundaries between online, print and broadcast, but at a very fundamental level that process is impossible in the United Kingdom.

Martin Belam, information architect for the Guardian, recently emphasized that point in an interview with Journalism.co.uk:

“In a converged media landscape, it seems odd that [BBC’s] Robert Peston’s blog is regulated by the BBC Trust, [Channel 4’s] Jon Snow’s blog is regulated by Ofcom, and [the Guardian’s] Roy Greenslade’s blog is regulated by the PCC.”

Now, Martin was actually wrong on the Jon Snow point: Ofcom does not regulate any television Web sites at all. That is to say, the brands which must adhere to a strict code for television content are completely unregulated online. Ofcom advises consumers to make complaints about online content to their Internet service provider.

The BBC Trust regulates the BBC online; the Press Complaints Commission (PCC) regulates newspapers, magazines and their online content.

And Stephen Fry, who – at the time of writing — is nearing half a million followers on Twitter? Or Guido Fawkes (aka Paul Staines) who has a loyal readership to rival most newspaper commentators? Well, they govern themselves – unless the law gets involved.

When the traditional media sectors go online, they’re regulated by their various bodies, and the ‘online-onlys’ only have the courts to worry about. Press publications have a less strict code than broadcasters, but online, broadcasters have more freedom than the press – though they don’t seem to be exercising it.

In a nutshell, a financial commentator from a newspaper has greater freedom than a financial commentator from a broadcaster, and an independent online-only financial commentator has the greatest freedom of all.

What happens when a bank crashes? Channel 4 and ITV can theoretically report how they like – online. The BBC must always answer to the BBC Trust. The newspapers must comply with the PCC code. Martin Lewis, of the MoneySaving Expert can, if he so chooses, be a law unto himself.

Same news and it’s all online but in very different guises. We might think people know the difference, but do they?

Full post at this link…

#DataJourn part 3: Useful and recent links looking at use of data in journalism

Perhaps we’ll expand this to a Dipity timeline at some point (other ideas?), but for the meantime, here’s a list of a few recent and relevant links relating to CAR and use of data in journalism to get the conversation on Twitter – via #datajourn – going. NB: These are not necessarily in chronological order. Then, the next logical step would be to start looking at examples of where data has been used for specific journalism projects.

Information Architects’ Ning network event sells out in ten minutes

Communication via a Ning network led to tickets for a information architects’ (IA) mini-conference in London ‘selling’ out in just ten minutes.

Information architecture is ‘the emerging art and science of organising large-scale websites,’ increasingly important for media sites.

The Ning network created by Ken Beatson last year, has allowed the UK’s information architects to talk more freely and effectively than via the old mailing list system, Martin Belam, a member of the group and information architect for the Guardian, told Journalism.co.uk.

An event was set up, hosted for free at the Guardian’s offices and sponsored by Axure and Aquent, and after a bit of promotion via its Twitter account (@london_ia), 40 tickets were rapidly snapped up for the event which will take place on April 20. Another 10 will also be released at midday on Friday.

The event will see participants talk for 10-15 minute slots in an informal way.

Martin Belam told Journalism.co.uk that ‘the goal of good information architecture is that people understand information,’ so it suits them to share knowledge and skills in this way. London is one of the biggest centres for information architects, perhaps the biggest outside New York and San Francisco, he said.

An overlap between editorial and technological roles is increasingly important for newspapers, Belam added.

Belam hopes that the event could be rolled out three times a year, with the next one being held in September.

Also see: Q&A with Martin Belam here.

Q&A with an information architect (aka @currybet aka Martin Belam)

Martin Belam, of the CurryBet blog, has recently been appointed as ‘information architect’ for Guardian.co.uk. Journalism.co.uk asked him what he’ll be doing for the site…

For those who don’t know what you do, fill us in your background and the new gig…
[MB] I was at the Hack Day that the Guardian’s technology department ran back in November 2008, and the talent and enthusiasm that day really shone. I’ve really enjoyed the freedom of working as a consultant over the last three years, much of the time based either in Crete or in Austria, but the opportunity of coming to work more permanently for an organisation as forward-thinking as the Guardian is being with initiatives like the Open Platform was too much to resist.

So, an ‘information architect’ what does that mean and what are you doing?
Information Architecture has been defined as ‘the emerging art and science of organising large-scale websites’.

All websites have an inherent information structure – the navigation, the contextual links on a page, whether there are tags describing content and so forth. It is about how people navigate and way-find their way through the information presented on a site.

What I’ll be doing at the Guardian is influencing that structure and functionality as new digital products are developed. It involves working closely with design and editorial teams to produce ‘wireframes’, the blueprints of web design, and also involves being an advocate for the end user – carrying out lots of usability and prototype testing as ideas are developed.

Is it a full-time role?
I’m working four days a week at The Guardian, as I still have some other commitments – for example as contributing editor for FUMSI magazine – although already it feels a bit like cramming a full-time job into just 80 per cent of the time!

It’s not happy times for mainstream media brands: where are they going wrong?
I don’t think it is only mainstream media brands that are suffering from the disruption caused by digital transition, but we do see a lot of focus on this issue for print businesses at the moment. I think one of the things that strikes me, having worked at several big media companies now, including the BBC and Sony, is that you would never set these organisations up in this way in the digital era if you were doing it from scratch.

One of the things that appealed most about joining the Guardian was that the move to Kings Place has brought together the print, online and technical operations in a way that wasn’t physically possible before in the old offices. I’m still very optimistic that there are real opportunities out there for the big media brands that can get their business structures right for the 21st century.

What kind of things do you think could re-enthuse UK readers for their newspapers?
I think our core and loyal readers are still enthusiastic about their papers, but that as an industry we have to face the fact that there is an over-supply of news in the UK, and a lot of it – whether it is on the radio, TV, web or thrust into your hand as a freebie – is effectively free at the point of delivery. I think the future will see media companies who concentrate on playing to their strengths benefit from better serving a narrower target audience.

Do you see print becoming the by rather than primary product for the Guardian – or has that already happened?
I think there might very well be a ‘sweet spot’ in the future where the display quality on network-enabled mobile devices and the ubiquity of data through-the-air means that the newspaper can be delivered primarily in that way, but I don’t see the Guardian’s presses stopping anytime soon. Paper is still a very portable format, and it never loses connection or runs out of batteries.

Your background is in computer programming rather than journalism, will the two increasingly overlap?
I grew up in the generation that had BBC Micros and ZX Spectrums at home, so I used to program a lot as a child, but my degree was actually in History, which in itself is a very journalistic calling. I specialised in the Crusades and the Byzantine Empire, which is all about piecing together evidence from a range of sources of varying degrees of reliability, and synthesizing a coherent narrative and story from there. And, of course, I’ve spent most of this decade blogging, which utilises ‘some’ of the journalist’s skill-set ‘some’ of the time.

Whilst I’d never suggest that journalists need to learn computer programming much beyond a smattering of HTML, I think there is something to be gained from understanding the software engineering mindset. There are a lot of tools and techniques that can really help journalists plough through data to get at the heart of a story, or to use visualisation tools to help tell that story to their audience.

One of the most interesting things about working at the Guardian is the opportunity to work alongside people like Kevin Anderson, Charles Arthur and Simon Willison, who I think really represent that blending of the technical and journalistic cultures.

You’ve spoken out about press regulation before; why do you feel strongly about it?
In a converged media landscape, it seems odd that Robert Peston’s blog is regulated by the BBC Trust, Jon Snow’s blog is regulated by Ofcom, and Roy Greenslade’s blog is regulated by the PCC.

At the moment, I believe that the system works very well for editors, and very well for the ‘great and the good’ who can afford lawyers, but does absolutely nothing for newspaper consumers. If I see something that offends me on TV, I can complain to Ofcom. If I see an advert that offends me in the street, I can complain to ASA. If I see an article in a newspaper that I think is wrong, inaccurate, in bad taste or offensive, unless I am directly involved in the story myself, the PCC dismisses my complaint out of hand without investigating it.

I don’t think that position is sustainable.

The last thing I want to see is some kind of state-sponsored Ofpress quango, which is why I think it is so important that our industry gets self-regulation right – and why I believe that a review of how the PCC works in the digital era is long overdue.