Category Archives: public service sector renewal

Canada launches data.gc.ca – what works and what is broken

Those on twitter will already know that this morning I had the privilege of conducting a press conference with Minister Day about the launch of data.gc.ca – the Federal Government’s Open Data portal. For those wanting to learn more about open data and why it matters, I suggest this and this blog post, and this article – they outline some of the reasons why open data matters.

In this post I want to review what works, and doesn’t work, about data.gc.ca.

What works

Probably the most important thing about data.gc.ca is that it exists. It means that public servants across the Government of Canada who have data they would like to share can now point to a website that is part of government policy. It is an enormous signal of permission from a central agency that will give a number of people who want to share data permission, a process and a vehicle, by which to do this. That, in of itself, is significant.

Indeed, I was informed that already a number of ministries and individuals are starting to approach those operating the portal asking to share their data. This is exactly the type of outcome we as citizens should want.

Moreover, I’ve been told that the government wants to double the number of data sets, and the number of ministries, involved in the site. So the other part that “works” on this site is the commitment to make it bigger. This is also important, as there have been some open data portals that have launched with great fanfare, only to have the site languish as neither new data sets are added and the data sets on the site are not updated and so fall out of date.

What’s a work in progress

The number of “high value” datasets is, relatively speaking, fairly limited. I’m always cautious about this as, I feel, what constitutes high value varies from user to user. That said, there are clearly data sets that will have greater impact on Canadians: budget data, line item spend data by department (as the UK does), food inspection data, product recall data, pretty much everything on the statscan website, Service Canada locations, postal code data and, mailbox location data, business license data, Canada Revenue data on charities and publicly traded companies are all a few that quickly come to mind, clearly I can imagine many, many more…

I think the transparency, tech, innovation, mobile and online services communities will be watching data.gc.ca closely to see what data sets get added. What is great is that the government is asking people what data sets they’d like to see added. I strongly encourage people to let the government know what they’d like to see, especially when it involves data the government is already sharing, but in unhelpful formats.

What doesn’t work

In a word: the license.

The license on data.gc.ca is deeply, deeply flawed. Some might go so far as to say that the license does not make it data open at all – a critique that I think is fair. I would say this: presently the open data license on data.gc.ca effectively kills any possible business innovation, and severally limits the use in non-profit realms.

The first, and most problematic is this line:

“You shall not use the data made available through the GC Open Data Portal in any way which, in the opinion of Canada, may bring disrepute to or prejudice the reputation of Canada.”

What does this mean? Does it mean that any journalist who writes a story, using data from the portal, that is critical of the government, is in violation of the terms of use? It would appear to be the case. From an accountability and transparency perspective, this is a fatal problem.

But it is also problematic from a business perspective. If one wanted to use a data set to help guide citizens around where they might be well, and poorly, served by their government, would you be in violation? The problem here is that the clause is both sufficiently stifling and sufficiently negative that many businesses will see the risk of using this data simply too great.

UPDATE: Thursday March 17th, 3:30pm, the minister called me to inform me that they would be striking this clause from the contract. This is excellent news and Treasury Board deserves credit for moving quickly. It’s also great recognition that this is a pilot (e.g. beta) project and so hopefully, the other problems mentioned here and in the comments below will also be addressed.

It is worth noting that no other open data portal in the world has this clause.

The second challenging line is:

“you shall not disassemble, decompile except for the specific purpose of recompiling for software compatibility, or in any way attempt to reverse engineer the data made available through the GC Open Data Portal or any part thereof, and you shall not merge or link the data made available through the GC Open Data Portal with any product or database for the purpose of identifying an individual, family or household or in such a fashion that gives the appearance that you may have received or had access to, information held by Canada about any identifiable individual, family or household or about an  organization or business.”

While I understand the intent of this line, it is deeply problematic for several reasons. First, many business models rely on identifying individuals, indeed, frequently individuals ask businesses to do this. Google, for example, knows who I am and offers custom services to me based on the data they have about me. It would appear that terms of use would prevent Google from using Government of Canada data to improve its service even if I have given them permission. Moreover, the future of the digital economy is around providing customized services. While this data has been digitized, it effectively cannot be used as part of the digital economy.

More disconcerting is that these terms apply not only to individuals, but also to organizations and businesses. This means that you cannot use the data to “identify” a business. Well, over at Emitter.ca we use data from Environment Canada to show citizens facilities that pollute near them. Since we identify both the facilities and the companies that use them (not to mention the politicians whose ridings these facilities sit in), are we not in violation of the terms of use? In a similar vein, I’ve talked about how government data could have prevented $3B of tax fraud. Sadly, data from this portal would not have changed that since, in order to have found the fraud, you’d have to have identified the charitable organizations involved. Consequently, this requirement manifestly destroys any accountability the data might create.

It is again worth noting that no other open data portal in the world has this clause.

And finally:

4.1 You shall include and maintain on all reproductions of the data made available through the GC Open Data Portal, produced pursuant to section 3 above, the following notice:

Reproduced and distributed with the permission of the Government of Canada.

4.2 Where any of the data made available through the GC Open Data Portal is contained within a Value-Added Product, you shall include in a prominent location on such Value-Added Product the following notice:

This product has been produced by or for (your name – or corporate name, if applicable) and includes data provided by the Government of Canada.

The incorporation of data sourced from the Government of Canada within this product shall not be construed as constituting an endorsement by the Government of Canada of our product.

or any other notice approved in writing by Canada.

The problem here is that this creates what we call the “Nascar effect.” As you use more and more government data, these “prominent” displays of attribution begin to pile up. If I’m using data from 3 different governments, each that requires attribution, pretty soon all your going to see are the attribution statements, and not the map or other information that you are looking for! I outlined this problem in more detail here. The UK Government has handled this issue much, much more gracefully.

Indeed, speaking of the UK Open Government License, I really wish our government had just copied it wholesale. We have a similar government system and legal systems so I see no reason why it would not easily translate to Canada. It is radically better than what is offered on data.gc.ca and, by adopting it, we might begin to move towards a single government license within Commonwealth countries, which would be a real win. Of course, I’d love it if we adopted the PDDL, but the UK Open Government License would be okay to.

In Summary

The launch of data.gc.ca is an important first step. It gives those of us interested in open data and open government a vehicle by which to get more data open and improve the accountability, transparency as well as business and social innovation. That said, there is much work to be done still: getting more data up and, more importantly, addressing the significant concerns around the license. I have spoken to Treasury Board President Stockwell Day about these concerns and he is very interested and engaged by them. My hope is that with more Canadians expressing their concerns, and with better understanding by ministerial and political staff, we can land on the right license and help find ways to improve the website and program. That’s why we to beta launches in the tech world, hopefully it is something the government will be able to do here too.

 

Apologies for any typos, trying to get this out quickly, please let me know if you find any.

How Yelp Could Help Save Millions in Health Care Costs

Okay, before I dive in, a few things.

1) Sorry for the lack of posts last week. Life’s been hectic. Between Code for America, a number of projects and a few articles I’m trying to get through, the blogging slipped. Sorry.

2) I’m presenting on Open Data and Open Government to the Canadian Parliament Access to Information, Privacy and Ethics Committee today – more on that later this week

3) I’m excited about this post

When it comes to opening up government data many of us focus on Governments: we cajole, we pressure, we try to persuade them to open up their data. It’s approach we will continue to have to take for a great deal of the data our tax dollars pay to collect and that government’s continue to not share. There is however another model.

Consider transit data. This data is sought after, intensely useful, and probably the category of data most experimented with by developers. Why is this? Because it has been standardized. Why has it been standardized. Because local government’s (responding to citizen demand) have been desperate to get their transit data integrated with Google Maps (See image).
Screen-shot-2011-01-30-at-10.45.00-PM

It turns out, to get your transit data into Google Maps, Google insists that you submit to them the transit data in a single structured format. Something that has come to be known as the General Transit Feed Specification (GTFS). The great thing about the GTFS is that it isn’t just google that can use it. Anyone can play with data converted into the GTFS. Better still, because the data structure us standardized an application someone develops, or analysis they conduct, can be ported to other cities that share their transit data in a GTFS format (like, say, my home town of Vancouver).

In short, what we have here is a powerful model both for creating open data and standardizing this data across thousands of jurisdictions.

So what does this have to do with Yelp! and Health Care Costs?

For those not in the know Yelp! is a mobile phone location based rating service. I’m particularly a fan of its restaurant locator: it will show you what is nearby and how it has been rated by other users. Handy stuff.

But think bigger.

Most cities in North America inspect restaurants for health violations. This is important stuff. Restaurants with more violations are more likely to transmit diseases and food born illnesses, give people food poisoning and god knows what else. Sadly, in most cases the results of these tests are posted in the most useless place imaginable. The local authorities website.

I’m willing to wager almost anything that the only time anyone visits a food inspection website is after they have been food poisoned. Why? Because they want to know if the jerks have already been cited.

No one checks these agencies websites before choosing a restaurant. Consequently, one of the biggest benefits of the inspection data – shifting market demand to more sanitary options – is lost. And of course, there is real evidence that shows restaurants will improve their sanitation, and people will discriminate against restaurants that get poor ratings from inspectors, when the data is conveniently available. Indeed, in the book Full Disclosure: The Perils and Promise of Transparency Fung, Graham and Weil noted that after Los Angeles required restaurants to post food inspection results, that “Researchers found significant effects in the form of revenue increases for restaurants with high grades and revenue decreases for C-graded (poorly rated) restaurants.” More importantly, the study Fung, Graham and Weil reference also suggested that making the rating system public positively impacted healthcare costs. Again, after inspection results in Los Angeles were posted on restaurant doors (not on some never visited website), the county experienced a reduction in emergency room visits, the most expensive point of contact in the system. As the study notes these were:

an 18.6 percent decline in 1998 (the first year of program operation), a 4.8 percent decline in 1999, and a 5.4 per- cent decline in 2000. This pattern was not observed in the rest of the state.

This is a stunning result.

So, now imagine that rather than just giving contributor generated reviews of restaurants Yelp! actually shared real food inspection data! Think of the impact this would have on the restaurant industry. Suddenly, everyone with a mobile phone and Yelp! (it’s free) could make an informed decision not just about the quality of a restaurant’s food, but also based on its sanitation. Think of the millions (100s of millions?) that could be saved in the United States alone.

All that needs to happen is for a simple first step, Yelp! needs approach one major city – say a New York, or a San Francisco – and work with them to develop a sensible way to share food inspection data. This is what happened with Google Maps and the GTSF, it all started with one city. Once Yelp! develops the feed, call it something generic, like the General Restaurant Inspection Data Feed (GRIDF) and tell the world you are looking for other cities to share the data in that format. If they do, you promise to include it in your platform. I’m willing to bet anything that once one major city has it, other cities will start to clamber to get their food inspection data shared in the GRIDF format. What makes it better still is that it wouldn’t just be Yelp! that could use the data. Any restaurant review website or phone app could use the data – be it Urban Spoon or the New York Times.

The opportunity here is huge. It’s also a win for everyone: Consumers, Health Insurers, Hospitals, Yelp!, Restaurant Inspection Agencies, even responsible Restaurant Owners. It would also be a huge win for Government as platform and open data. Hey Yelp. Call me if you are interested.

A Response to a Ottawa Gov 2.0 Skeptic

So many, many months ago a Peter R. posted this comment (segments copied below) under a post I’d written titled: Prediction, The Digital Economy Strategy Will Fail if it Isn’t Drafted Collaboratively on GCPEDIA. At first blush Peter’s response felt aggressive. I flipped him an email to say hi and he responded in a very friendly manner. I’ve been meaning to respond for months but, life’s been busy.  However, over the break (and my quest to hit inbox 0) I finally carved out some time –  my fear is that this late response will sound like a counter attack – it isn’t intended as such but rather an effort to respond to a genuine question. I thought it would be valuable to post as many of the points may resonate with supporters and detractors of Gov 2.0 alike. Here’s my responses to the various charges:

The momentum, the energy and the excitement behind collaborative/networked/web 2.0/etc is only matched by the momentum, the energy and the excitement that was behind making money off of leveraged debt instruments in the US.

Agreed, there is a great deal of energy and excitement behind collaborative networks, although I don’t think this is sparked – as ensued, by something analogous to bogus debt instruments. People are excited because of the tangible results created by sharing and/or co-production networks like Wikipedia, Mozilla, Flickr Google search and Google Translate (your results improve based on users data) and Ushahidi inspire people because of the tremendous results they are able to achieve with a smaller footprint of resources. I think the question of what these types of networks and this type of collaboration means to government is an important question – that also means that as people experiment their will be failures – but to equate the entire concept of Gov 2.0 and the above cited organizations, tools and websites with financial instruments that repackaged subprime mortgages is, in my mind, fairly problematic.

David, the onus lies squarely with you to prove that policymakers across government are are incapable of finding good policy solutions WITHOUT letting everyone and his twitting brother chime in their two cents.

Actually the onus doesn’t lie squarely with me. This is a silly statement. In fact, for those of us who believe in collaborative technologies such as GCPEDIA or yammer this sentence is the single most revealing point in the Peter’s entire comment. I invite everyone and anyone to add to my rebuttal, or to Peter’s argument. Even those who argue against me would be proving my point – tools like blogs and GCPEDIA allow ideas and issues to be debated with a greater number of people and a wider set of perspectives. The whole notion that any thought or solution lies solely with one person is the type of thinking that leads to bad government (and pretty much bad anything). I personally believe that the best ideas emerge when they are debated and contested – honed by having flaws exposed and repaired. Moreover this has never been more important than today, when more and more issues cross ministerial divides. Indeed, the very fact that we are having this discussion on my blog, and that Peter deemed it worthy of comment, is a powerful counterpoint this statement.

Equally important, I never said policymakers across government are are incapable of finding good policy solutions. This is serious misinterpretation of what said. I did say that the Digital Economy Strategy would fail (and I’ll happily revise, and soften to say, will likely not be meaningful) unless written on GCPEDIA. I still believe this. I don’t believe you can have people writing policy about how to manage an economy who are outside of and/or don’t understand the tools of that economy. I actually think our public servants can find great policy solutions – if we let them. In fact, most public servants I know spend most of their time trying to track down public servants in other ministries or groups to consult them about the policy they are drafting. In short, they spend all their time trying to network, but using tools of the late 20th century (like email), mid 20th century (telephone), or mid 3rd century BC (the meeting) to do it. I just want to give them more efficient tools – digital tools, like those we use in a digital economy – so they can do what they are already doing.

For the last 8 years I’ve worked in government, I can tell you with absolute certainty (and credibility!) that good policy emerges from sound research and strategic instrument choice. Often (select) public consultations are required, but sometimes none at all. Take 3 simple and highly successful policy applications: seat belts laws, carbon tax, banking regulation. Small groups of policymakers have developed policies (or laws, regs, etc) to brilliant effect….sans web 2.0. So why do we need gcpedia now?

Because the world expects you to do more, faster and with less. I find this logic deeply concerning coming from a public servant. No doubts that government developed policies to brilliant effect before the wide adoption of the computer, or even the telephone. So should we get rid of them too? An increasing number of the world’s major corporations are, or are setting up an internal wiki/collaboration platform, a social networking site, even using microblogging services like Yammer to foster internal collaboration. Indeed, these things help us to do research and develop ideas faster, and I think, better. The question isn’t why do we need GCPEDIA now. The question is why aren’t we investing to make GCPEDIA a better platform? The rest of the world is.

I’ll put this another way: tons of excellent policy solutions are waiting in the shared drives of bureaucrats across all governments.

I agree. Let’s at least put it on a wiki where more people can read them, leverage them and, hopefully, implement them. You sitting on a great idea that three other people in the entire public service have read isn’t a recipe for getting it adopted. Nor is it a good use of Canadian tax dollars. Socializing it is. Hence, social media.

Politics — being what it is — doesn’t generate progressive out solutions for various ideological reasons (applies equally to ndp, lib, con). First, tell us what a “failed” digitial economy strategy (DES) looks like. Second, tell us what components need to be included in the DES for it be successful. Third, show us why gcpedia/wikis offer the only viable means to accumulate the necessary policy ingredients.

For the last part – see my earlier post and above. As for what a failed digital economy strategy looks like – it will be one that is irrelevant. It is one that will go ignored by the majority of people who actually work in the digital economy. Of course, an irrelevant policy will be better than a truly bad one which, which I suspect, is also a real risk based on the proceedings of Canada 3.0. (That conference seemed to be about “how do we save analog business that are about to be destroyed by the digital economy” – a link to one of my favourite posts). And of course I have other metrics that matter to me. That all said, after meeting the public servant in charge of the process at Canada 3.0, I was really, really encouraged – she is very smart and gets it.

She also found the idea of writing the policy on GCPEDIA intriguing. I have my doubts that that is how things are proceeding, but it gives me hope.

One Simple Step to Figure out if your Government "gets" Information Technology

Chris Moore has a good post up on his blog at the moment that asks “Will Canadian Cities ever be Strategic?” In it (and it is very much worth reading) he hits on a theme I’ve focused on in many of my talks to government but that I think is also relevant to citizens who care about how technologically sophisticated their government is (which is a metric of how relevant you think you government is going to be in a few years).

If you want to know how serious your government or ministry is about technology there is a first simple step you can take: look at the org chart.

Locate the Chief Information Officer (CIO) or Chief Technology Officer (CTO). Simple question: Is he or she part of the senior executive team?

If yes – there is hope that your government (or the ministry you are looking at) may have some strategic vision for IT.

If no – and this would put you in the bucket with about 80% of local governments, and provincial/federal ministries – then your government almost certainly does not have a strong vision, and it isn’t going to be getting one any time soon.

As Chris also notes and I’ve been banging away at (such as during my keynote to MISA Ontario), in most governments in Canada the CIO/CTO does not sit at the executive table. At the federal level they are frequently Director Generals, (or lower), not Assistant Deputy Minister level roles. At the local level, they often report into someone at the C-level.

This is insanity.

I’m trying to think of a Fortune 500 company – particularly one which operates in the knowledge economy – with this type of reporting structure. The business of government is about managing information… to better regulate, manage resources and/or deliver services. You can’t be talking about how to do that without having the CIO/CTO being part of that conversation.

But that’s what happens almost every single day in many govenrment orgs.

Sadly, it gets worse.

In most organizations the CIO/CTO reports into the Chief Financial Officer (CFO). This really tells you what the organization thinks of technology: It is a cost centre that needs to be contained.

We aren’t going to reinvent government when the person in charge of the infrastructure upon which most of the work is done, the services are delivered and pretty much everything else the org does depends on, isn’t even part of the management team or part of the organizations strategic conversations.

It’s a sad state of affairs and indicative of why our government’s are so slow in engaging in new technology.

Making StatsCan Data Free: Assessing the Cost

Regular readers of my blog will know that I’ve advocated that StatsCan’s data – and particularly its Census data – should be made open (e.g. free, unlicensed, and downloadable in multiple formats). Presently, despite the fact that Canadian tax dollars pay to collect (a sadly diminishing amount, and quality of,) data, it is not open.

The main defense I hear to why StatsCan’s data should not be free is because the department depends on the revenue the data generates.

So exactly how much revenue are we talking about? Thanks to the help of some former public servants I’ve been able to go over the publicly available numbers. The basic assessment – which I encourage people to verify and challenge – turns out not to be a huge a number.

The most interesting figure in StatsCan’s finances is the revenue it generates from its online database (e.g. data downloaded from its website). So how much revenue is it? Well in 2007/2008, it was $559,000.

That’s it. For $559,000 in lost government revenue Canadians could potentially have unlimited access to the Statscan census database their tax dollars paid to collect and organize. I suspect this is a tiny fraction of the value (and tax revenue) that might be generated by economic activity if this data were free.

Worse, the $559,000 is not profit. From what I can tell it is only revenue. Consequently, it doesn’t factor in collection costs StatsCan has to absorb to run and maintain a checkout system on its website, collect credit card info, bill people, etc… I’m willing to bet almost anything that the cost of these functions either exceed $559,000 a year, or come pretty close. So the net cost of making the data free could end up being a less.

StatsCan makes another $763,000 selling Statistics Canada publications (these are 243 data releases of the 29 major economic indicators StatsCan measures and the 5 census releases it does annually – in short these are non-customized reports). So for $1,422,000 Canadians could get access to both the online data statscan has and the reports the organization generates. This is such laughably (or depressingly) small number it begs the question – why are we debating this? (again this is revenue, not profit, so the cost could be much lower)

Of course, the figure that you’ll often hear cited is $100M in revenue. So what accounts for the roughly 100x difference between the above number and the alleged revenue? Well, in 2007/08 StatsCan did make $103,155,000 but this was from value added (e.g. customized) reports. This is very, very different product than the basic data that is available on its website. My sources tell me this is not related to downloaded data.

I think we should concede that if the entire StatsCan’s database were made open and free it would impact some of this revenue. But this would also be a good thing. Why is this? Let’s break it down:

  1. Increase Capacity and Data Literacy: By making a great deal of data open and free, StatsCan would make it easier for competitors to enter the market place. More companies and individuals could analyze the country’s census and other data, and so too could more “ordinary” Canadians than ever would be able to access the database (again, that their tax dollars paid to create). This might include groups like senior high school and university students, non-profits and everyday citizens who wanted to know more about their country. So yes, Statscan would have more competitors, but the country might also benefit from having a more data literate population (and thus potential consumers).
  2. Increase Accessibility of Canadian Data to Marginalized Groups: An increase in the country’s analysis capacity would drop the price for such work. This would make it cheaper and easier for more marginal groups to benefit from this data – charities, religious groups, NGO’s, community organizations, individuals, etc…
  3. Improve Competitiveness: It would also be good for Canadian competitiveness, companies would have to spend less to understand and sell into the Canadian market. This would lower the cost of doing business in Canada – helpful to consumers and the Canadian economy.
  4. StatsCan would not lose all or even most of its business: Those at StatsCan who fear the organization would be overwhelmed by a more open world should remember, not all the data can be shared. Some data – particularly economic data gathered from companies – is sensitive and confidential. As a result there will be some data that StatsCan retains exclusive access to, and thus a monopoly over analysis. More importantly, I suspect that were Statscan data made open the demand for data analysis would grow, so arguably new capacity might end up being devoted to new demand, not existing demand.
  5. It will Reduce the Cost of Government: Finally, the crazy thing about StatsCan is that it sells its data and services to other Ministries and layers of government. This means that governments are paying people to move tax payer money between government ministries and jurisdictions. This is a needless administrative costs that drives up everybody’s taxes and poorly allocates scarce government resources (especially at the local level). Assuming every town and city in Canada pays $50 – 1000 dollars to access statscan data may not seem like much, but in reality, we are really paying that, plus their and StatsCan’s staff time to manage all these transactions, enforce compliance, etc… all of which is probably, far, far more.

So in summary, the cost to Canada of releasing this data will likely be pretty marginal, while the benefits could be enormous.

At best, if costs half a million dollars in forgone revenue. Given the improved access and enormous benefits, this is a pittance to pay.

At worst, StatsCana would lose maybe 20-30 million – this is a real nightmare scenario that assumes much greater competition in the marketplace (again, a lot of assumptions in this scenario). Of course the improved access to data would lead to economic benefits that would far, far, surpass this lost revenue, so the net benefit for the country would be big, but the cost to StatsCan would be real. Obviously, it would be nice if this decline in revenue was offset by improved funding for StatsCan (something a government that was genuinely concerned about Canadian economic competitiveness would jump at doing). However, given the current struggles Statscan faces on the revenue front (cuts across the board) I could see how a worse case scenario would be nerve wracking to the department’s senior public servants, who are also still reeling from the Long Form Census debacle.

Ultimately, however, I think the worse case scenario is unlikely. Moreover, in either scenario the benefits are significant.

Bonus Material:

Possibly the most disconcerting part of the financial reports on StatsCan on Treasury Board’s website was the stakeholder consultation associated with access to statscan’s database. It claimed that:

Usability and client satisfaction survey were conducted with a sample of clients in early 2005. Declared level of satisfaction with service was very high.

This is stunning. I’ve never talked to anyone who has had a satisfactory experience on StatsCan’s website (in contrast to their phone support – which everyone loves). I refer to the statscan site where the place where what you want is always one click away.

I’m willing to bet a great deal that the consultations were with existing long term customers – the type of people that have experience using the website. My suspicion is that if a broader consultation was conducted with potential users (university students, community groups, people like me and you, etc…) the numbers would tank. I dare you to try to use their website. It is virtually unnavigable.

Indeed, had made its website and data more accessible I suspect it the department would engage Canadians and have more stakeholders. This would have been the single most powerful thing it could have done to protect itself from cuts and decisions like the Long Form fiasco.

I know this post may anger a number of people at Statscan. I’m genuinely sorry. I know the staff work hard, are dedicated and are exceedingly skilled and professional. This type of feedback is never flattering – particularly in public. It is because you are so important to the unity, economy and quality of life in our country that it is imperative we hold you to the highest possible bar – not just in the quality of that data your collect (there you already excel) but in the way you serve and engage Canadians. In this, I hope that you get the support you need and deserve.

The False choice: Bilingualism vs. Open Government (and accountability)

Last week a disturbing headline crossed my computer screen:

B.C. RCMP zaps old news releases from its website

2,500 releases deleted because they weren’t translated into French

1) The worst of all possible outcomes

This is a terrible outcome for accountability and open government. When we erase history we diminish accountability and erode our capacity to learn. As of today, Canadians have a poorer sense of what the RCMP has stood for, what it has claimed and what it has tried to do in British Columbia.

Consider this. The Vancouver Airport is a bilingual designated detachment. As of today, all press releases that were not translated were pulled down. This means that any press release related to the national scandal that erupted after Robert Dziekański – the polish immigrant who was tasered five times by the (RCMP) – is now no longer online. Given the shockingly poor performance the RCMP had in managing (and telling the truth about) this issue – this concerns me.

Indeed, I can’t think that anyone thinks this is a good idea.

The BC RCMP does not appear to think it is a good idea. Consider their press officer’s line: “We didn’t have a choice, we weren’t compliant.”

I don’t think there are any BC residents who believe they are better served by this policy.

Nor do I think my fellow francophone citizens believe they are better served by this decision. Now no one – either francophone or anglophone can find these press releases online. (More on this below)

I would be appalled if a similar outcome occurred in Quebec or a francophone community in Manitoba. If the RCMP pulled down all French press releases because they didn’t happen to have English translations, I’d be outraged – even if I didn’t speak French.

That’s because the one thing worse than not having the document in both official languages, is not having access to the document at all. (And having it hidden in some binder in a barracks that I have to call or visit doesn’t event hint of being accessible in the 21st century).

Indeed, I’m willing to bet almost anything that Graham Fraser, the Official Languages Commissioner – who is himself a former journalist – would be deeply troubled by this decision.

2) Guided by Yesterday, Not Preparing for Tomorrow

Of course, what should really anger the Official Languages Commissioner is an attempt to pit open and accountable government against bilingualism. This is a false choice.

I suspect that the current narrative in government is that translating these documents is too expensive. If one relies on government translators, this is probably true. The point is, we no longer have to.

My friend and colleague Luke C. pinged me after I tweeted this story saying “I’d help them automate translating those news releases into french using myGengo. Would be easy.”

Yes, mygengo would make it cheap at 5 cents a word (or 15 if you really want to overkill it). But even smarter would be to approach Google. Google translate – especially between French and English – has become shockingly good. Perfect… no. Of course, this is what the smart and practical people on the ground at the RCMP were doing until the higher ups got scared by a French CBC story that was critical of the practice. A practice that was ended even though it did not violate any policies.

The problem is there isn’t going to be more money to do translation – not in a world of multi-billion dollar deficits and in a province that boasts 63,000 french speakers. But Google translate? It is going to keep getting better and better. Indeed, the more it translates, the better it gets. If the RCMP (or Canadian government) started putting more documents through Google translate and correcting them it would become still more accurate. The best part is… it’s free. I’m willing to bet that if you ran all 2500 of the press releases through Google translate right now… 99% of them would come out legible and of a standard that would be good enough to share. (again, not perfect, but serviceable). Perhaps the CBC won’t be perfectly happy. But I’m not sure the current outcome makes them happy either. And at least we’ll be building a future in which they will be happy tomorrow.

The point here is that this decision reaffirms a false binary: one based on a 20th century assumptions where translations were expensive and laborious. It holds us back and makes our government less effective and more expensive. But worse, it ignores an option that embraces a world of possibilities – the reality of tomorrow. By continuing to automatically translate these documents today we’d continue to learn how to use and integrate this technology now, and push it to get better, faster. Such a choice would serve the interests of both open and accountable governments as well as bilingualism.

Sadly, no one at the head office of the RCMP – or in the federal government – appears to have that vision. So today we are a little more language, information and government poor.

Three asides:

1) I find it fascinating that the media can get mailed a press release that isn’t translated but the public is not allowed to access it on a website until it is – this is a really interesting form of discrimination, one that supports a specific business model and has zero grounding in the law, and indeed may even be illegal given that the media has no special status in Canadian law.

2) Still more fascinating is how the RCMP appears to be completely unresponsive to news stories about inappropriate behavior in its ranks, like say the illegal funding of false research to defend the war on drugs, but one story about language politics causes the organization to change practices that aren’t even in violation of its policies. It us sad to see more evidence still that the RCMP is one of the most broken agencies in the Federal government.

3) Thank you to Vancouver Sun Reporter Chad Skelton for updating me on the Google aspect of this story.

What Governments can Learn about Citizen Engagement from Air Canada

Yes. You read that title right.

I’m aware that airlines are not known for their customer responsiveness. Ask anyone whose been trapped on a plane on the tarmac for 14 hours. You know you’ve really dropped the ball when Congress (which agrees on almost nothing) passes a customer bill of rights explicitly for your industry.

Air Canada, however, increasingly seems to be the exception to this rule. Their recent response to online customer feedback is instructive of why this is the case. For governments interested in engaging citizens online and improving services, Air Canada is an interesting case study.

The Background

Earlier this year, with great fanfare, Air Canada announced it was changing how it managed its frequent flyer reward system. Traditional, it had given out upgrade certificates which allowed customers who’d flown a certain number of flights the air-canada-logoability to upgrade themselves into business class for free. Obviously the people who use these certificates are some of Air Canada’s more loyal customers (to get certificates you have to be flying a fair amount). The big change was that rather than simple giving customers certificates after flying a certain number of miles, customers would earn “points” which they could allocate towards flights.

This was supposed to be a good news story because a) it meant that users had greater flexibility around how they upgraded themselves and b) the whole system was digitized so that travelers wouldn’t have to carry certificates around with them (this was the most demanded feature by users).

The Challenge

In addition to the regular emails and website announcement an Air Canada representative also posted details about the new changes on a popular air traveler forum called Flyertalk.com. (Note: Here is the first great lesson – don’t expect customers or citizens to come to you… go to where they hang out, especially your most hard core stakeholders).

flyertalk_logoVery quickly these important stakeholders (customers) began running the numbers and started discovering various flaws and problems. Some noticed that the top tier customers were getting a lesser deal than ordinary customers. Others began to sniff out how the new program essentially meant their benefits were being cut. In short, the very incentives the rewards program was supposed to create were being undermined. Indeed the conversation thread extended to over 113 pages. With roughly 15 comments per page, that meant around 1500 comments about the service.

This, of course, is what happens with customers, stakeholder and citizens in a digital world. They can get together. They can analyze your service. And they will notice any flaws or changes that do not seem above board or are worse than what previously existed.

So here, on Flyertalk, Air Canada has some of its most sophisticated and important customers – the people that will talk to everyone about Air travel rewards programs, starting to revolt against its new service which was supposed to be a big improvement. This was (more than) a little bit of a crisis.

The Best Practice

First, Air Canada was smart because it didn’t argue with anyone. It didn’t have communication people trying to explain to people how they were wrong.

Instead it was patient. It appeared silent. But in reality it was doing something more important, it was listening.

Remember many of these users know the benefits program better than most Air Canada employees. And it has real impact on their decisions, so they are going to analyze it up and down.

Second, When it finally did respond, Air Canada did several things right.

It responded in Flyertalk.com – again going to where the conversation was. (It subsequently sent around an email to all its members).

It noted that it had been listening and learning from its customers.

More than just listen, Air Canada had taken its customers feedback and used it to revise its air travel rewards program.

And, most importantly, the tone it took was serious, but engaging. Look at the first few sentences:

Thanks to everyone for the comments that have been posted here the last few days, and especially those who took the time to post some very valuable, constructive feedback. While it’s not our intent to address every issue raised on this forum on the changes to the 2011 Top Tier program, some very valid points were raised which we agree should be addressed to the best of our ability. These modifications are our attempt to do just that.

Governments, this is a textbook case on how to listen to citizens. They use your services. They know how they work. The single biggest take away here is, when they complain and construct logical arguments about why a service doesn’t make sense use that feedback to revise the service and make it better. People don’t want to hear why you can’t make it better – they want you to make it better. More importantly, these types of users are the ones who know your service the best and who talk to everyone about it. They are your alpha users – leverage them!

Again, to recap. What I saw Air Canada do that was positive was:

  • Engage their stakeholders where their stakeholders hang out (e.g. not on the Air Canada website)
  • Listen to what their stakeholders had to say
  • Use that feedback to improve the service
  • Communicate with customers in a direct and frank manner

Air Canada is doing more than just getting this type of engagement right. Their twitter account posts actual useful information, not just marketing glop and spin. I’m not sure who is doing social media for them, but definitely worth watching.

There’s a lot here for organizations to learn from. Moreover, for a company that used to be a crown corp I think that should mean there is hope for your government too – even if they presently ban access to facebook, twitter or say, my blog.

Big thank you to Mike B. for pointing out this cool case study to me.

London, UK, Open Government Data Camp Keynote – Nov 18, 2010

Here is my opening keynote to the Open Government Data Camp held in London earlier this year (2010) on Nov 18th. Pasted below is the day two keynote by the always excellent Tom Steinberg of mysociety.org.

Hope these are thought provoking for novice and veteran tech and open government types, as well as those just curious about how our world is changing.

My speech gets all very serious after the initial 12-year old boy moment. It was a great audience at the camp – really wonderful to have such an engaged group of people.

David Eaves from Open Knowledge Foundation on Vimeo.

Tim Berners-Lee, Prof. Nigel Shadbolt & Tom Steinberg from Open Knowledge Foundation on Vimeo.

An Open Letter on Open Government to the Access to Information, Privacy & Ethics Parliamentary Committee

The other week I received an invitation from the Canadian Standing Parliamentary Committee on Access to Information, Privacy & Ethics to come and testify about open government and open data on February 1st.

The Committee has talked a great deal about its efforts to engage in a study of open government and since February 1st is quite a bit away and I’d like to be helpful before my testimony, I thought I draft up some thoughts and suggestion for the committee’s strategy. I know these are unsolicited but I hope they are helpful and, if not, that they at least spark some helpful thoughts.

1. Establish a common understanding of the current state of affairs

First off, the biggest risk at the moment is that the Committee’s work might actually slow down efforts of the government to launch an open data strategy. The Committee’s work, and the drafting of its report, is bound to take several months, it would be a shame if the government were to hold back launching any initiatives in anticipation of this report.

Consequently, my hope is that the committee, at is earliest possible convenience, request to speak to the Chief Information Officer of the Government of Canada to get an update regarding the current status of any open government and open data initiatives, should they exist. This would a) create a common understanding regarding the current state of affairs for both committee members and witnesses; b) allow subsequent testimony and recommendations to take into consideration the work already done and c) allow the committee to structure its work so as to not slow down any current efforts that might be already underway.

2. Transform the Committee into a Government 2.0 Taskforce – similar to the Australian effort

Frankly, my favourite approach in this space has been the British. Two Government’s, one Labour, one Conservative have aggressive pursued an open data and open government strategy. This, would be my hope for Canada. However, it does not appear that is is presently the case. So, another model should be adopted. Fortunately, such a model exists.

Last year, under the leadership of Nicholas Gruen, the Australian government launched a Government 2.0 taskforce on which I had the pleasure of serving on the International Reference Group. The Australian Taskforce was non-partisan and was made up of policy and technical experts and entrepreneurs from government, business, academia, and cultural institutions. More importantly, the overwhelming majority of its recommendations were adopted.

To replicate its success in Canada I believe the Committee should copy the best parts of the Australian taskforce. The topic of Canadians access to their government is of central importance to all Canadians – to non-profits, to business interests, to public servants and, of course, to everyday citizens. Rather than non-partisan, I would suggest that a Canadian taskforce should be pan-partisan – which the Committee already is. However, like the Australian Taskforce it should include a number of policy and technical experts from outside government. This fill committee would this represent both a political cross-section and substantive knowledge in the emerging field of government 2.0. It could thus, as a whole, effectively and quickly draft recommendations to Parliament.

Best of all, because of step #1, this work could proceed in parallel to any projects (if any) already initiated by the government and possibly even inform such work by providing interim updates.

I concede such an approach may be too radical, but I hope it is at least a starting point for an interesting approach.

3. Lead by Example

There is one arena where politicians need not wait on the government to make plans: Parliament itself. Over the past year, while in conversations with the Parliamentary IT staff as well as the Speaker of the House, I have worked to have Parliament make more data about its own operations open. Starting in January, the Parliamentary website will begin releasing the Hansard in XML – this will make it much easier for software developers like the creators of Openparliament.ca as and howdtheyvote.ca to run their sites and for students, researchers and reporters to search and analyze our country’s most important public discussions. In short, by making the Hansard more accessible the Speaker and his IT staff are making parliament more accessible. But this is only the beginning of what parliamentarians could do to make for a truly Open Parliament. The House and Senate’s schedules and agendas, along with committee calendars should all be open. So to should both chambers seating arrangement. Member’s photos and bios should be shared with an unrestricted license as should the videos of parliament.

Leadership in this space would send a powerful message to both the government and the public service that Canada’s politicians are serious about making government more open and accessible to those who elect it. In addition, it could also influence provincial legislature’s and even municipal governments, prompting them to do the same and so enhance our democracy at every level.

4. Finally, understand your task: You are creating a Knowledge Government for a Knowledge Society

One reason I advise the Committee to take on external members is because, laudably, many admit this topic is new to them. But I also want the committee members to understand the gravity of their task. Open Government, Open Data and/or Government 2.0 are important first steps in a much larger project.

What you are really wrestling with here is what government is going to look like in an knowledge economy and a knowledge society. How is going to function with knowledge workers as employees? And, most importantly, how is it going to engage with knowledge citizens, many of whom can and want to make real contributions beyond the taxes they pay and don’t need government to self-organize?

In short, what is a knowledge based government going to look like?

At the centre of that question is how we manage and share information. The basic building block of a knowledge driven society.

Look around, and you can see how the digital world is transforming how we do everything. Few of us can imagine living today without access to the internet and the abundance of information it brings to us. Indeed, we have already become so used to the internet we forget how much it has radically changed whole swaths of our life and economy from the travel and music industry to the post to political fund-raising and to journalism.

If today our government still broadly looks and feels like an institution shaped by the printing press it is because, well it is. Deputy Ministers and Ministers still receive giant briefing binders filled with paper. This is a reflection of how we deal within information and knowledge in government, we move it around (for good reasons) in siloes, operating as though networks, advance search, and other innovations don’t exist (even though they already do).

How our government deals with information is at the heart of your task. I’m not saying you have to re-invent government or dismantle all the silos and ministries. Quite the contrary, I believe small changes can be made that will yield significant benefits, efficiencies and savings while enhancing our democracy. But you will be confronting decades, if not centuries of tradition, culture and process in an institution that is about to go through the biggest change since the invention of the printing press. You don’t have to do it all, but even some small first steps will not come easily. I share this because I want you going into the task with eyes wide open.

At the very least we aren’t going first, our cousins both across the Atlantic, the Pacific and our southern border have already taken the plunge. But this should add urgency to our task. We cannot afford to stand by while others renew their democratic institutions while simultaneously enhancing an emerging and critical pillar of a new knowledge economy and knowledge society.