The day has dawned bright and sunny on Open Data Camp 7’s final day. There’s a greta bunch of people present, the coffee is flowing, and it’s time to pitch. Here’s what’s on today’s menu:
Continue reading Open Data Camp 7: Day Two pitches
An Open Data Camp session on helping charities and other low tech bodies create data ecosystem stories improve their impact, led by Pauline Roche.
Liveblogging: prone to error, inaccuracy and howling crimes against grammar and syntax. Post will be updated in the coming days.
Over 80% of charities in this country operate on tiny budgets – often under £10,000 per annum. There are some similarities with, say, libraries, or arts bodies. There are resources out there for them – like 360giving – but they may not know about them, or have the confidence to use them.
Datakind offers a number of resources. They recently worked with the GLA, to help understand the number of refugees and migrants in London. There isn’t good data out there on that. But charities tend to know where they are – so could they provide that information. So they asked – and it would be fair to say that they weren’t keen on the idea. They said that, if they were going to do this, they needed support in working out what to collect, and how. And the GLA was willing to help take that on.
Many of the charities had no idea of the data already available that they could use, nor how data could help their own work. They paired up data experts with subject experts to figure out what was needed, and how to deliver that data.
Continue reading Building a data ecosystem in a low tech envronment
One of the final sessions of the first day of Open Data Camp 7 in London was led by Anneka France from The Rivers Trust. She she had wanted to run the session because she had wanted to get hold of the National Soil Map because she wanted money for an EU-funded project to restore peatland for climate mitigation and flood prevention.
The National Soil Map is covered by a commercial licence, and the charity was quoted £25,000 to get the data it needs. Which it can’t afford. But then Anneka heard about the ‘pillars of power approach’ “which has been used to overthrow governments” and wondered if it could help. Continue reading ODCamp 7: Using a pillars of power approach to opening data
How do you get started with SPARQL, the language for querying linked data? An Open Data Camp 7 session, led by Jen, aimed to help newbies get going.
Liveblogging: prone to error, inaccuracy, and howling crimes against grammar and syntax. Post will be updated in the coming days.
More and more open data platforms are either becoming linked data at their core, or they have offshoots that add it. The data underneath linked data is RDF – and SPARQL is the query language for RDF. Most SparQL endpoint look like a query box with gobbledegook with them – where you are expected to write your own gobbledegook. It’s somewhat intimidating,
In most cases, they also provide an API so you can programmatically query the information – but somebody needs to develop that. SPARQL endpoints give you direct access to all the data. The structure of RDF — the triples — creates a very standardised data format that you can query for whatever you like.
There’s a SPARQL playground where you can experiment with queries. There’s more than one of them, in fact.
You can use the query interface to hone down on the data you want, and then download it as a CSV, or use that as a query to use programatically. The playgrounds help you figure out how to construct queries by showing you the results on a sample dataset. Continue reading SPARQL 101: how to get started with the linked data search query language
Andy Dudfield, who works for “a small charity” called Full Fact, which checks out the claims made by politicians and others, said he had pitched this session because “fact checking is hard to do” and “we are always looking for ways to make it go faster.”
So, he said, “one of the things that we would like to do is to look at using open data for fact checking” and he would like his audience to tell him about any good open data sources for the job. And, also what the caveats were likely to be.
For example, he said, Full Fact might be asked to check a claim that crime had risen: were there good open data crime statistics, and how would the organisation know where they came from, and whether any change was ‘real’ or a statistical artefact? Continue reading ODCamp 7: Just the facts: fact checking with open data
An Open Data Camp 7 session on countering excuses for not publishing open data, led by Jenny Broker. Liveblogging: prone to error, inaccuracy and howling crimes against grammar and syntax. Post will be improved in the coming days.
Excuse: It’s a safety thing – it’s critical and it could be useful to terrorism
Safety is the first thing people will come after you with. For example, in utilities, it’s a very real concern, particularly around the location of assets. Is this a genuine concern, or an easy way of shutting down a conversation? Is this information that’s not already accessible via Google Maps, for example? Crashing critical infrastructure is a genuine risk. The most risky data is already heavily controlled — and is often not even shared within government. That comes with its own problems – issues get missed because staff don’t have access to the full picture.
So, if Google Maps has the data, if we make it more accessible, is there a potential for spotting problems earlier? Well, liability now raises its head. Pretty much all datasets are infested with personal data, so if you published the data, and something happens, you’re liable. Some people don’t want to take that risk. This is another standard way of hiding from open data. Some organisations have developed organised risk assessments for open data – it create a more structured way to talk about risk. Continue reading Dealing with Open Data excuses
After lunch at Open Data Camp 7, consultant Edafe Onerhime explained why she had wanted to raise the issue of ‘decolonialising data’.
She is Nigerian as well as British, she explained, and her research had brought her into contact with the issue of “white by default.” As an example, she said, the census puts ‘white, British’ at the top of its identity options; which affects the data that is collected, because people have to scroll down to find other options.
At the same time, she said, a lot of technology relies on data collection and analysis that is carried out in India and countries and Africa. For example, a lot of tagging of pictures for what is billed as AI is done there. The taggers may not receive much money; yet their work is consumed by affluent, Western consumers.
So, the question: “How can we remove the effects of colonisation on data collection and use?” Campers at the session felt the first issue is to recognise that there is a problem, and the second is to get more voices at the table. Then, the assumptions underlying the collection and use of data need to be interrogated. Continue reading Decolonialising Data
A session from Open Data Camp 7 on delivering started nationally, led by Anne McCrossan. Liveblogged notes. Prone to error, inaccuracy and howling crimes against grammar and syntax. Post will be updated in the coming days.
Do we really have a national data started yet? Where we do have strategies – how well are they being implemented? Can we move things forwards by sharing experiences with each other?
Northern Ireland is on its second data strategy in six years. The first one was all data open by default – but they didn’t really have the delivery mechanisms or incentives to get civil servants to deliver. Hence the reason for a new one so quickly. It has a lot more reporting mechanisms in there, to exert pressure on local authorities, and report that upwards to general government.
Over the first three years, the success stories tended to be with startups and external companies. The frustrations were with the civil service.
What are our desired as a community, and how would they be expressed? Open data strategy in particular has tended to be less a strategy and more a commitment to getting it out. This is partially the result of the movement being kicked off by a coalition of interests. It can be challenging for some political ends.
Do we need outcomes? This is still an emergent space, so it’s hard to know what outcomes you might get. For example, going to the moon gave us Velcro, but it wasn’t part of the strategy… It’s very difficult to know what people will do with open data. So, maybe the strategy should just be delivery.
Continue reading Open Data Strategy Campfire
How do we show the value of opening up data? That was the question asked by Dan Winchester, who runs a company called Get the Data, in the second round of sessions at Open Data Camp 7.
“I am a data publisher and also a data consumer,” he explained. “I create data sets and put them out in the world and people consume that data. But I have very little information about what is created. Is there value: economic, or social from it? That makes it hard for me to know whether I am putting resources into the right place.
“While, as a data consumer, I am seeing benefits from using data, but do the people who are publishing the data aware of those. Am I doing as much as I could do to help them make the case for allocating resources to the data I am benefiting from?”
Continue reading ODCamp 7: The ROI of Open Data
A session on Wardley Maps and value chains, led by Jonathan Kerr. Liveblogged notes – prone to error, inaccuracy and howling crimes against grammar and syntax. Post will be improved in the coming days.
What are Wardley Maps?
A map is a thing that shows a space – where things are in relationship to each other, as well as the overall concept. All models are, by nature, simplifications. You could make an accurate map of France, but it would be unwieldy…
You can map a blue chain by mapping all the components into an end result – but you don’t need to map every single element, just the ones that are important to you.
As you move along the value chain, you increase repeatability:
- Genesis – the original concept. (R&D)
- Custom – you start making it (bespoke elements)
- Product – you start making machines to make the thing. It become replicable
- Utility – available as a service. Think of APIs, charged on per use basis.
This isn’t a linear process – things move in and out of categories. By mapping the parts of the value chain through this grid, you can spot where you have elements of your process which are costing you disproportionate amounts of money.
Continue reading Wardley Maps and Open Data: a discussion