Tag Archives: CPSR

Shared IT Services across the Canadian Government – three opportunities

Earlier this week the Canadian Federal Government announced it will be creating Shared Services Canada which will absorb the resources and functions associated with the delivery of email, data centres and network services from 44 departments.

These types of shared services projects are always fraught with danger. While they sometimes are successful, they are often disasters. Highly disruptive with little to show for results (and eventually get unwound). However, I suspect there is a significant amount of savings that can be made and I remain optimistic. With luck the analogy here is the work outgoing US CIO Vivek Kundra accomplished as he has sought to close down and consolidate 800 data centres across the US which is yielding some serious savings.

So here’s what I’m hoping Shared Services Canada will mean:

1) A bigger opportunity for Open Source

What I’m still more hopeful about – although not overly optimistic – is the role that open source solutions could play in the solutions Shared Services Canada will implement. Over on the Drupal site, one contributor claims government officials have been told to hold off buying web content management systems as the government prepares to buy a single solution for across all departments.

If the government is serious about lowering its costs it absolutely must rethink its procurement models so that open source solutions can at least be made a viable option. If not this whole exercise will mean the government may save money, but it will be the we move from 5 expensive solutions to one expensive solution variety.

On the upside some of that work has clearly taken place. Already there are several federal government websites running on Drupal such as this Ministry of Public Works website, the NRCAN and DND intranet. Moreover, there are real efforts in the open source community to accommodate government. In the United States OpenPublic has fostered a version of Drupal designed for government’s needs.

Open source solutions have the added bonus of allowing you the option of using more local talent, which, if stimulus is part of the goal, would be wise. Also, any open source solutions fostered by the federal government could be picked up by the provinces, creating further savings to tax payers. As a bonus, you can also fire incompetent implementors, something that needs to happen a little more often in government IT.

2) More accountability

Ministers Ambrose and Clement are laser focused on finding savings – pretty much every ministry needs to find 5 or 10% savings across the board. I also know both speak passionately about managing tax payers dollars: “Canadians work hard for their money and expect our Government to manage taxpayers dollars responsibly, Shared Services Canada will have a mandate to streamline IT, save money, and end waste and duplication.”

Great. I agree. So one of Shared Service Canada’s first act should be to follow in the footsteps of another Vivek Kundra initiative and recreate his incredibly successful IT Dashboard. Indeed it was by using the dashboard Vivek was able to “cut the time in half to deliver meaningful [IT system] functionality and critical services, and reduced total budgeted [Federal government IT] costs by over $3 billion.” Now that some serious savings. It’s a great example of how transparency can drive effective organizational change.

And here’s the kicker. The White House open sourced the IT Dashboard (the code can be downloaded here). So while it will require some work adapting it, the software is there and a lot of the heavy work has been done. Again, if we are serious about this, the path forward is straightforward.

3) More open data

Speaking of transparency… one place shared services could really come in handy is creating some data warehouses for hosting critical government data sets (ideally in the cloud). I suspect there are a number of important datasets that are used by public servants across ministries, and so getting them on a robust platform that is accessible would make a lot of sense. This of course, would also be an ideal opportunity to engage in a massive open data project. It might be easier to create policy for making the data managed by Shared Service Canada “open.” Indeed, this blog post covers some of the reasons why now is the time to think about that issue.

So congratulations on the big move everyone and I hope these suggestions are helpful. Certainly we’ll be watching with interest – we can’t have a 21st century government unless we have 21st century infrastructure, and you’re now the group responsible for it.

It's the icing, not the cake: key lesson on open data for governments

At the 2010 GTEC conference I did a panel with David Strigel, the Program Manager of the Citywide Data Warehouse (CityDW) at the District of Columbia Government. During the introductory remarks David recounted the history of Washington DC’s journey to open data.

Interestingly, that journey began not with open data, but with an internal problem. Back around 2003 the city had a hypothesis that towing away abandoned cars would reduce crime rates in the immediate vicinity, thereby saving more money in the long term than the cost of towing. In order to access the program’s effectiveness city staff needed to “mash-up” longitudinal crime data against service request data – specifically, requests to remove abandoned cars. Alas, the data sets were managed by different departments, so this was tricky task. As a result the city’s IT department negotiated bilateral agreements with both departments to host their datasets in a single location. Thus the DC Data Warehouse was born.

Happily, the data demonstrated the program was cost effective. Building on this success the IT department began negotiating more bilateral agreements with different departments to host their data centrally. In return for giving up stewardship of the data the departments retained governance rights but reduced their costs and the IT group provided them with additional, more advanced, analytics. Over time the city’s data warehouse became vast. As a result, when DC decided to open up its data it was, relatively speaking, easy to do. The data was centrally located, was already being shared and used as a platform internally. Extending this platform externally (while not trivial) was a natural step.

In short, the deep problem that needed to solved wasn’t open data. Its was an information management. Getting the information management and governance policies right was essential for DC to move quickly. Moreover, this problem strikes at the heart of what it means to be government. Knowing what data you have, where it is, and under a governance structure that allows it to be shared internally (as well as externally) is a problem every government is going to face if it wants to be efficient, relevant and innovative in the 21st century. In other words, information management is the cake. Open data – which I believe is essential – is however the sweet icing you smother on top of that dense cake you’ve put in place.

Okay, with that said two points that flow from this.

First: Sometime, governments that “do” open data start off by focusing on the icing. The emphasis in on getting data out there, and then after the fact, figuring out  governance model that will make sense. This is a viable strategy, but it does have real risks. When sharing data isn’t at the core function but rather a feature tacked on at the end, the policy and technical infrastructure may be pretty creaky. In addition, developers may not want to innovate on top of your data platform because they may (rightly) question the level of commitment. One reason DC’s data catalog works is because it has internal users. This gives the data stability and a sense of permanence. On the upside, the icing is politically sexier, so it may help marshal resources to help drive a broader rethink of data governance. Either way, at some point, you’ve got to tackle the cake, otherwise, things are going to get messy. Remember it took DC 7 years to develop its cake before it put icing on it. But that was making it from scratch. Today thanks to new services (armies of consultants on this), tools (eg. Socrata) and models (e.g. like Washington, DC) you can make that cake following a recipe and even use cake mix. As David Strigel pointed out, today, he could do it in a fraction of the time.

Second: More darkly, one lesson to draw from DC is that the capacity of a government to do open data may be a pretty good proxy for their ability to share information and coordinate across different departments. If your government can’t do open data in a relatively quick time period, it may mean they simply don’t have the infrastructure in place to share data internally all that effectively either. In a world where government productivity needs to rise in order to deal with budget deficits, that could be worrying.

A Response to a Ottawa Gov 2.0 Skeptic

So many, many months ago a Peter R. posted this comment (segments copied below) under a post I’d written titled: Prediction, The Digital Economy Strategy Will Fail if it Isn’t Drafted Collaboratively on GCPEDIA. At first blush Peter’s response felt aggressive. I flipped him an email to say hi and he responded in a very friendly manner. I’ve been meaning to respond for months but, life’s been busy.  However, over the break (and my quest to hit inbox 0) I finally carved out some time –  my fear is that this late response will sound like a counter attack – it isn’t intended as such but rather an effort to respond to a genuine question. I thought it would be valuable to post as many of the points may resonate with supporters and detractors of Gov 2.0 alike. Here’s my responses to the various charges:

The momentum, the energy and the excitement behind collaborative/networked/web 2.0/etc is only matched by the momentum, the energy and the excitement that was behind making money off of leveraged debt instruments in the US.

Agreed, there is a great deal of energy and excitement behind collaborative networks, although I don’t think this is sparked – as ensued, by something analogous to bogus debt instruments. People are excited because of the tangible results created by sharing and/or co-production networks like Wikipedia, Mozilla, Flickr Google search and Google Translate (your results improve based on users data) and Ushahidi inspire people because of the tremendous results they are able to achieve with a smaller footprint of resources. I think the question of what these types of networks and this type of collaboration means to government is an important question – that also means that as people experiment their will be failures – but to equate the entire concept of Gov 2.0 and the above cited organizations, tools and websites with financial instruments that repackaged subprime mortgages is, in my mind, fairly problematic.

David, the onus lies squarely with you to prove that policymakers across government are are incapable of finding good policy solutions WITHOUT letting everyone and his twitting brother chime in their two cents.

Actually the onus doesn’t lie squarely with me. This is a silly statement. In fact, for those of us who believe in collaborative technologies such as GCPEDIA or yammer this sentence is the single most revealing point in the Peter’s entire comment. I invite everyone and anyone to add to my rebuttal, or to Peter’s argument. Even those who argue against me would be proving my point – tools like blogs and GCPEDIA allow ideas and issues to be debated with a greater number of people and a wider set of perspectives. The whole notion that any thought or solution lies solely with one person is the type of thinking that leads to bad government (and pretty much bad anything). I personally believe that the best ideas emerge when they are debated and contested – honed by having flaws exposed and repaired. Moreover this has never been more important than today, when more and more issues cross ministerial divides. Indeed, the very fact that we are having this discussion on my blog, and that Peter deemed it worthy of comment, is a powerful counterpoint this statement.

Equally important, I never said policymakers across government are are incapable of finding good policy solutions. This is serious misinterpretation of what said. I did say that the Digital Economy Strategy would fail (and I’ll happily revise, and soften to say, will likely not be meaningful) unless written on GCPEDIA. I still believe this. I don’t believe you can have people writing policy about how to manage an economy who are outside of and/or don’t understand the tools of that economy. I actually think our public servants can find great policy solutions – if we let them. In fact, most public servants I know spend most of their time trying to track down public servants in other ministries or groups to consult them about the policy they are drafting. In short, they spend all their time trying to network, but using tools of the late 20th century (like email), mid 20th century (telephone), or mid 3rd century BC (the meeting) to do it. I just want to give them more efficient tools – digital tools, like those we use in a digital economy – so they can do what they are already doing.

For the last 8 years I’ve worked in government, I can tell you with absolute certainty (and credibility!) that good policy emerges from sound research and strategic instrument choice. Often (select) public consultations are required, but sometimes none at all. Take 3 simple and highly successful policy applications: seat belts laws, carbon tax, banking regulation. Small groups of policymakers have developed policies (or laws, regs, etc) to brilliant effect….sans web 2.0. So why do we need gcpedia now?

Because the world expects you to do more, faster and with less. I find this logic deeply concerning coming from a public servant. No doubts that government developed policies to brilliant effect before the wide adoption of the computer, or even the telephone. So should we get rid of them too? An increasing number of the world’s major corporations are, or are setting up an internal wiki/collaboration platform, a social networking site, even using microblogging services like Yammer to foster internal collaboration. Indeed, these things help us to do research and develop ideas faster, and I think, better. The question isn’t why do we need GCPEDIA now. The question is why aren’t we investing to make GCPEDIA a better platform? The rest of the world is.

I’ll put this another way: tons of excellent policy solutions are waiting in the shared drives of bureaucrats across all governments.

I agree. Let’s at least put it on a wiki where more people can read them, leverage them and, hopefully, implement them. You sitting on a great idea that three other people in the entire public service have read isn’t a recipe for getting it adopted. Nor is it a good use of Canadian tax dollars. Socializing it is. Hence, social media.

Politics — being what it is — doesn’t generate progressive out solutions for various ideological reasons (applies equally to ndp, lib, con). First, tell us what a “failed” digitial economy strategy (DES) looks like. Second, tell us what components need to be included in the DES for it be successful. Third, show us why gcpedia/wikis offer the only viable means to accumulate the necessary policy ingredients.

For the last part – see my earlier post and above. As for what a failed digital economy strategy looks like – it will be one that is irrelevant. It is one that will go ignored by the majority of people who actually work in the digital economy. Of course, an irrelevant policy will be better than a truly bad one which, which I suspect, is also a real risk based on the proceedings of Canada 3.0. (That conference seemed to be about “how do we save analog business that are about to be destroyed by the digital economy” – a link to one of my favourite posts). And of course I have other metrics that matter to me. That all said, after meeting the public servant in charge of the process at Canada 3.0, I was really, really encouraged – she is very smart and gets it.

She also found the idea of writing the policy on GCPEDIA intriguing. I have my doubts that that is how things are proceeding, but it gives me hope.

One Simple Step to Figure out if your Government "gets" Information Technology

Chris Moore has a good post up on his blog at the moment that asks “Will Canadian Cities ever be Strategic?” In it (and it is very much worth reading) he hits on a theme I’ve focused on in many of my talks to government but that I think is also relevant to citizens who care about how technologically sophisticated their government is (which is a metric of how relevant you think you government is going to be in a few years).

If you want to know how serious your government or ministry is about technology there is a first simple step you can take: look at the org chart.

Locate the Chief Information Officer (CIO) or Chief Technology Officer (CTO). Simple question: Is he or she part of the senior executive team?

If yes – there is hope that your government (or the ministry you are looking at) may have some strategic vision for IT.

If no – and this would put you in the bucket with about 80% of local governments, and provincial/federal ministries – then your government almost certainly does not have a strong vision, and it isn’t going to be getting one any time soon.

As Chris also notes and I’ve been banging away at (such as during my keynote to MISA Ontario), in most governments in Canada the CIO/CTO does not sit at the executive table. At the federal level they are frequently Director Generals, (or lower), not Assistant Deputy Minister level roles. At the local level, they often report into someone at the C-level.

This is insanity.

I’m trying to think of a Fortune 500 company – particularly one which operates in the knowledge economy – with this type of reporting structure. The business of government is about managing information… to better regulate, manage resources and/or deliver services. You can’t be talking about how to do that without having the CIO/CTO being part of that conversation.

But that’s what happens almost every single day in many govenrment orgs.

Sadly, it gets worse.

In most organizations the CIO/CTO reports into the Chief Financial Officer (CFO). This really tells you what the organization thinks of technology: It is a cost centre that needs to be contained.

We aren’t going to reinvent government when the person in charge of the infrastructure upon which most of the work is done, the services are delivered and pretty much everything else the org does depends on, isn’t even part of the management team or part of the organizations strategic conversations.

It’s a sad state of affairs and indicative of why our government’s are so slow in engaging in new technology.

Making StatsCan Data Free: Assessing the Cost

Regular readers of my blog will know that I’ve advocated that StatsCan’s data – and particularly its Census data – should be made open (e.g. free, unlicensed, and downloadable in multiple formats). Presently, despite the fact that Canadian tax dollars pay to collect (a sadly diminishing amount, and quality of,) data, it is not open.

The main defense I hear to why StatsCan’s data should not be free is because the department depends on the revenue the data generates.

So exactly how much revenue are we talking about? Thanks to the help of some former public servants I’ve been able to go over the publicly available numbers. The basic assessment – which I encourage people to verify and challenge – turns out not to be a huge a number.

The most interesting figure in StatsCan’s finances is the revenue it generates from its online database (e.g. data downloaded from its website). So how much revenue is it? Well in 2007/2008, it was $559,000.

That’s it. For $559,000 in lost government revenue Canadians could potentially have unlimited access to the Statscan census database their tax dollars paid to collect and organize. I suspect this is a tiny fraction of the value (and tax revenue) that might be generated by economic activity if this data were free.

Worse, the $559,000 is not profit. From what I can tell it is only revenue. Consequently, it doesn’t factor in collection costs StatsCan has to absorb to run and maintain a checkout system on its website, collect credit card info, bill people, etc… I’m willing to bet almost anything that the cost of these functions either exceed $559,000 a year, or come pretty close. So the net cost of making the data free could end up being a less.

StatsCan makes another $763,000 selling Statistics Canada publications (these are 243 data releases of the 29 major economic indicators StatsCan measures and the 5 census releases it does annually – in short these are non-customized reports). So for $1,422,000 Canadians could get access to both the online data statscan has and the reports the organization generates. This is such laughably (or depressingly) small number it begs the question – why are we debating this? (again this is revenue, not profit, so the cost could be much lower)

Of course, the figure that you’ll often hear cited is $100M in revenue. So what accounts for the roughly 100x difference between the above number and the alleged revenue? Well, in 2007/08 StatsCan did make $103,155,000 but this was from value added (e.g. customized) reports. This is very, very different product than the basic data that is available on its website. My sources tell me this is not related to downloaded data.

I think we should concede that if the entire StatsCan’s database were made open and free it would impact some of this revenue. But this would also be a good thing. Why is this? Let’s break it down:

  1. Increase Capacity and Data Literacy: By making a great deal of data open and free, StatsCan would make it easier for competitors to enter the market place. More companies and individuals could analyze the country’s census and other data, and so too could more “ordinary” Canadians than ever would be able to access the database (again, that their tax dollars paid to create). This might include groups like senior high school and university students, non-profits and everyday citizens who wanted to know more about their country. So yes, Statscan would have more competitors, but the country might also benefit from having a more data literate population (and thus potential consumers).
  2. Increase Accessibility of Canadian Data to Marginalized Groups: An increase in the country’s analysis capacity would drop the price for such work. This would make it cheaper and easier for more marginal groups to benefit from this data – charities, religious groups, NGO’s, community organizations, individuals, etc…
  3. Improve Competitiveness: It would also be good for Canadian competitiveness, companies would have to spend less to understand and sell into the Canadian market. This would lower the cost of doing business in Canada – helpful to consumers and the Canadian economy.
  4. StatsCan would not lose all or even most of its business: Those at StatsCan who fear the organization would be overwhelmed by a more open world should remember, not all the data can be shared. Some data – particularly economic data gathered from companies – is sensitive and confidential. As a result there will be some data that StatsCan retains exclusive access to, and thus a monopoly over analysis. More importantly, I suspect that were Statscan data made open the demand for data analysis would grow, so arguably new capacity might end up being devoted to new demand, not existing demand.
  5. It will Reduce the Cost of Government: Finally, the crazy thing about StatsCan is that it sells its data and services to other Ministries and layers of government. This means that governments are paying people to move tax payer money between government ministries and jurisdictions. This is a needless administrative costs that drives up everybody’s taxes and poorly allocates scarce government resources (especially at the local level). Assuming every town and city in Canada pays $50 – 1000 dollars to access statscan data may not seem like much, but in reality, we are really paying that, plus their and StatsCan’s staff time to manage all these transactions, enforce compliance, etc… all of which is probably, far, far more.

So in summary, the cost to Canada of releasing this data will likely be pretty marginal, while the benefits could be enormous.

At best, if costs half a million dollars in forgone revenue. Given the improved access and enormous benefits, this is a pittance to pay.

At worst, StatsCana would lose maybe 20-30 million – this is a real nightmare scenario that assumes much greater competition in the marketplace (again, a lot of assumptions in this scenario). Of course the improved access to data would lead to economic benefits that would far, far, surpass this lost revenue, so the net benefit for the country would be big, but the cost to StatsCan would be real. Obviously, it would be nice if this decline in revenue was offset by improved funding for StatsCan (something a government that was genuinely concerned about Canadian economic competitiveness would jump at doing). However, given the current struggles Statscan faces on the revenue front (cuts across the board) I could see how a worse case scenario would be nerve wracking to the department’s senior public servants, who are also still reeling from the Long Form Census debacle.

Ultimately, however, I think the worse case scenario is unlikely. Moreover, in either scenario the benefits are significant.

Bonus Material:

Possibly the most disconcerting part of the financial reports on StatsCan on Treasury Board’s website was the stakeholder consultation associated with access to statscan’s database. It claimed that:

Usability and client satisfaction survey were conducted with a sample of clients in early 2005. Declared level of satisfaction with service was very high.

This is stunning. I’ve never talked to anyone who has had a satisfactory experience on StatsCan’s website (in contrast to their phone support – which everyone loves). I refer to the statscan site where the place where what you want is always one click away.

I’m willing to bet a great deal that the consultations were with existing long term customers – the type of people that have experience using the website. My suspicion is that if a broader consultation was conducted with potential users (university students, community groups, people like me and you, etc…) the numbers would tank. I dare you to try to use their website. It is virtually unnavigable.

Indeed, had made its website and data more accessible I suspect it the department would engage Canadians and have more stakeholders. This would have been the single most powerful thing it could have done to protect itself from cuts and decisions like the Long Form fiasco.

I know this post may anger a number of people at Statscan. I’m genuinely sorry. I know the staff work hard, are dedicated and are exceedingly skilled and professional. This type of feedback is never flattering – particularly in public. It is because you are so important to the unity, economy and quality of life in our country that it is imperative we hold you to the highest possible bar – not just in the quality of that data your collect (there you already excel) but in the way you serve and engage Canadians. In this, I hope that you get the support you need and deserve.

Hello 2011! Coding for America, speaking, open data and licenses

Eaves.ca readers – happy new year! Here’s a little but of an overview into how we are kicking off 2011/

Thank you

First, if you are a regular reader…. thank you. Eaves.ca has just entered its 4th year and it all keeps getting more and more rewarding. I write to organize my thoughts and refine my writing skills but it is always most rewarding to see others benefit from/find it of value to read these posts.

Code for America

I’ve just landed San Francisco (literally! I’m sitting on the floor of the airport, catching the free SFO wifi) where I’ll be spending virtually all of January volunteering to help launch Code for America. Why? Because I think the organizations matters, its projects are important and the people are great. I’ve always enjoyed hanging out with technologists with a positive agenda for change (think Mozilla & OpenMRS) and Code for America takes all that fun and combines it with government – one of my other great passions. I hope to update you on the progress the organizations makes and what will be happening over the coming weeks. And yes, I am thinking about how Code for Canada might fit into all this.

Gov 2.0

I’ll also be in Ottawa twice in January. My first trip out is to present on a paper about how to write collaboratively using a wiki. With a little bit of work, I’ll be repositioning this paper to make it about how to draft public policy on a wiki within government (think GCPEDIA). With luck I’ll publish this in something like Policy Options or something similar (maybe US based?). I think this has the potential of being one of my most important pieces of the year and needless to say, I’m excited and will be grateful for feedback, both good and negative.

Open Data and Government

On my second trip to Ottawa I’ll be presenting on Open Data and Open Government to the Standing Committee on Access to Information, Privacy and Ethics. Obviously, I’m honored and thrilled they’ve asked me to come and talk and look forward to helping parliamentarians understand why this issue is so important and how they could make serious progress in short order if they put their minds to the task.

Licenses

So for the last two years we’ve been working hard to get cities to do open data with significant success. This year, may be the year that the Feds (more on that in a later post) and some provinces get in on the game, as well as a larger group of cities. The goal for them will be to build on, and take to the next level, the successes of the first movers like Vancouver, Edmonton, Toronto and Ottawa. This will mean one thing. Doing the licensing better. The Vancouver license, which has been widely copied was a good starting point for governments venturing into unknown territory (e.g. getting their toes wet). But the license has several problems –and there are several significantly better choices out there (I’m look over at you PDDL – which I notice the City of Surrey has adopted, nice work.). So, I think one big goal for 2011 will be to get governments to begin shifting to (ideally) the PDDL and (if necessary) something more equivalent. On this front I’m feeling optimistic as well and will blog on this in the near future.

Lots of other exciting things going on as well – I look forward to sharing them here in the blog soon.

All in all, its hard to to be excited about 2011 and I hope you are too. Thank you so much for being part of all of this.

The False choice: Bilingualism vs. Open Government (and accountability)

Last week a disturbing headline crossed my computer screen:

B.C. RCMP zaps old news releases from its website

2,500 releases deleted because they weren’t translated into French

1) The worst of all possible outcomes

This is a terrible outcome for accountability and open government. When we erase history we diminish accountability and erode our capacity to learn. As of today, Canadians have a poorer sense of what the RCMP has stood for, what it has claimed and what it has tried to do in British Columbia.

Consider this. The Vancouver Airport is a bilingual designated detachment. As of today, all press releases that were not translated were pulled down. This means that any press release related to the national scandal that erupted after Robert Dziekański – the polish immigrant who was tasered five times by the (RCMP) – is now no longer online. Given the shockingly poor performance the RCMP had in managing (and telling the truth about) this issue – this concerns me.

Indeed, I can’t think that anyone thinks this is a good idea.

The BC RCMP does not appear to think it is a good idea. Consider their press officer’s line: “We didn’t have a choice, we weren’t compliant.”

I don’t think there are any BC residents who believe they are better served by this policy.

Nor do I think my fellow francophone citizens believe they are better served by this decision. Now no one – either francophone or anglophone can find these press releases online. (More on this below)

I would be appalled if a similar outcome occurred in Quebec or a francophone community in Manitoba. If the RCMP pulled down all French press releases because they didn’t happen to have English translations, I’d be outraged – even if I didn’t speak French.

That’s because the one thing worse than not having the document in both official languages, is not having access to the document at all. (And having it hidden in some binder in a barracks that I have to call or visit doesn’t event hint of being accessible in the 21st century).

Indeed, I’m willing to bet almost anything that Graham Fraser, the Official Languages Commissioner – who is himself a former journalist – would be deeply troubled by this decision.

2) Guided by Yesterday, Not Preparing for Tomorrow

Of course, what should really anger the Official Languages Commissioner is an attempt to pit open and accountable government against bilingualism. This is a false choice.

I suspect that the current narrative in government is that translating these documents is too expensive. If one relies on government translators, this is probably true. The point is, we no longer have to.

My friend and colleague Luke C. pinged me after I tweeted this story saying “I’d help them automate translating those news releases into french using myGengo. Would be easy.”

Yes, mygengo would make it cheap at 5 cents a word (or 15 if you really want to overkill it). But even smarter would be to approach Google. Google translate – especially between French and English – has become shockingly good. Perfect… no. Of course, this is what the smart and practical people on the ground at the RCMP were doing until the higher ups got scared by a French CBC story that was critical of the practice. A practice that was ended even though it did not violate any policies.

The problem is there isn’t going to be more money to do translation – not in a world of multi-billion dollar deficits and in a province that boasts 63,000 french speakers. But Google translate? It is going to keep getting better and better. Indeed, the more it translates, the better it gets. If the RCMP (or Canadian government) started putting more documents through Google translate and correcting them it would become still more accurate. The best part is… it’s free. I’m willing to bet that if you ran all 2500 of the press releases through Google translate right now… 99% of them would come out legible and of a standard that would be good enough to share. (again, not perfect, but serviceable). Perhaps the CBC won’t be perfectly happy. But I’m not sure the current outcome makes them happy either. And at least we’ll be building a future in which they will be happy tomorrow.

The point here is that this decision reaffirms a false binary: one based on a 20th century assumptions where translations were expensive and laborious. It holds us back and makes our government less effective and more expensive. But worse, it ignores an option that embraces a world of possibilities – the reality of tomorrow. By continuing to automatically translate these documents today we’d continue to learn how to use and integrate this technology now, and push it to get better, faster. Such a choice would serve the interests of both open and accountable governments as well as bilingualism.

Sadly, no one at the head office of the RCMP – or in the federal government – appears to have that vision. So today we are a little more language, information and government poor.

Three asides:

1) I find it fascinating that the media can get mailed a press release that isn’t translated but the public is not allowed to access it on a website until it is – this is a really interesting form of discrimination, one that supports a specific business model and has zero grounding in the law, and indeed may even be illegal given that the media has no special status in Canadian law.

2) Still more fascinating is how the RCMP appears to be completely unresponsive to news stories about inappropriate behavior in its ranks, like say the illegal funding of false research to defend the war on drugs, but one story about language politics causes the organization to change practices that aren’t even in violation of its policies. It us sad to see more evidence still that the RCMP is one of the most broken agencies in the Federal government.

3) Thank you to Vancouver Sun Reporter Chad Skelton for updating me on the Google aspect of this story.

What Governments can Learn about Citizen Engagement from Air Canada

Yes. You read that title right.

I’m aware that airlines are not known for their customer responsiveness. Ask anyone whose been trapped on a plane on the tarmac for 14 hours. You know you’ve really dropped the ball when Congress (which agrees on almost nothing) passes a customer bill of rights explicitly for your industry.

Air Canada, however, increasingly seems to be the exception to this rule. Their recent response to online customer feedback is instructive of why this is the case. For governments interested in engaging citizens online and improving services, Air Canada is an interesting case study.

The Background

Earlier this year, with great fanfare, Air Canada announced it was changing how it managed its frequent flyer reward system. Traditional, it had given out upgrade certificates which allowed customers who’d flown a certain number of flights the air-canada-logoability to upgrade themselves into business class for free. Obviously the people who use these certificates are some of Air Canada’s more loyal customers (to get certificates you have to be flying a fair amount). The big change was that rather than simple giving customers certificates after flying a certain number of miles, customers would earn “points” which they could allocate towards flights.

This was supposed to be a good news story because a) it meant that users had greater flexibility around how they upgraded themselves and b) the whole system was digitized so that travelers wouldn’t have to carry certificates around with them (this was the most demanded feature by users).

The Challenge

In addition to the regular emails and website announcement an Air Canada representative also posted details about the new changes on a popular air traveler forum called Flyertalk.com. (Note: Here is the first great lesson – don’t expect customers or citizens to come to you… go to where they hang out, especially your most hard core stakeholders).

flyertalk_logoVery quickly these important stakeholders (customers) began running the numbers and started discovering various flaws and problems. Some noticed that the top tier customers were getting a lesser deal than ordinary customers. Others began to sniff out how the new program essentially meant their benefits were being cut. In short, the very incentives the rewards program was supposed to create were being undermined. Indeed the conversation thread extended to over 113 pages. With roughly 15 comments per page, that meant around 1500 comments about the service.

This, of course, is what happens with customers, stakeholder and citizens in a digital world. They can get together. They can analyze your service. And they will notice any flaws or changes that do not seem above board or are worse than what previously existed.

So here, on Flyertalk, Air Canada has some of its most sophisticated and important customers – the people that will talk to everyone about Air travel rewards programs, starting to revolt against its new service which was supposed to be a big improvement. This was (more than) a little bit of a crisis.

The Best Practice

First, Air Canada was smart because it didn’t argue with anyone. It didn’t have communication people trying to explain to people how they were wrong.

Instead it was patient. It appeared silent. But in reality it was doing something more important, it was listening.

Remember many of these users know the benefits program better than most Air Canada employees. And it has real impact on their decisions, so they are going to analyze it up and down.

Second, When it finally did respond, Air Canada did several things right.

It responded in Flyertalk.com – again going to where the conversation was. (It subsequently sent around an email to all its members).

It noted that it had been listening and learning from its customers.

More than just listen, Air Canada had taken its customers feedback and used it to revise its air travel rewards program.

And, most importantly, the tone it took was serious, but engaging. Look at the first few sentences:

Thanks to everyone for the comments that have been posted here the last few days, and especially those who took the time to post some very valuable, constructive feedback. While it’s not our intent to address every issue raised on this forum on the changes to the 2011 Top Tier program, some very valid points were raised which we agree should be addressed to the best of our ability. These modifications are our attempt to do just that.

Governments, this is a textbook case on how to listen to citizens. They use your services. They know how they work. The single biggest take away here is, when they complain and construct logical arguments about why a service doesn’t make sense use that feedback to revise the service and make it better. People don’t want to hear why you can’t make it better – they want you to make it better. More importantly, these types of users are the ones who know your service the best and who talk to everyone about it. They are your alpha users – leverage them!

Again, to recap. What I saw Air Canada do that was positive was:

  • Engage their stakeholders where their stakeholders hang out (e.g. not on the Air Canada website)
  • Listen to what their stakeholders had to say
  • Use that feedback to improve the service
  • Communicate with customers in a direct and frank manner

Air Canada is doing more than just getting this type of engagement right. Their twitter account posts actual useful information, not just marketing glop and spin. I’m not sure who is doing social media for them, but definitely worth watching.

There’s a lot here for organizations to learn from. Moreover, for a company that used to be a crown corp I think that should mean there is hope for your government too – even if they presently ban access to facebook, twitter or say, my blog.

Big thank you to Mike B. for pointing out this cool case study to me.

An Open Letter on Open Government to the Access to Information, Privacy & Ethics Parliamentary Committee

The other week I received an invitation from the Canadian Standing Parliamentary Committee on Access to Information, Privacy & Ethics to come and testify about open government and open data on February 1st.

The Committee has talked a great deal about its efforts to engage in a study of open government and since February 1st is quite a bit away and I’d like to be helpful before my testimony, I thought I draft up some thoughts and suggestion for the committee’s strategy. I know these are unsolicited but I hope they are helpful and, if not, that they at least spark some helpful thoughts.

1. Establish a common understanding of the current state of affairs

First off, the biggest risk at the moment is that the Committee’s work might actually slow down efforts of the government to launch an open data strategy. The Committee’s work, and the drafting of its report, is bound to take several months, it would be a shame if the government were to hold back launching any initiatives in anticipation of this report.

Consequently, my hope is that the committee, at is earliest possible convenience, request to speak to the Chief Information Officer of the Government of Canada to get an update regarding the current status of any open government and open data initiatives, should they exist. This would a) create a common understanding regarding the current state of affairs for both committee members and witnesses; b) allow subsequent testimony and recommendations to take into consideration the work already done and c) allow the committee to structure its work so as to not slow down any current efforts that might be already underway.

2. Transform the Committee into a Government 2.0 Taskforce – similar to the Australian effort

Frankly, my favourite approach in this space has been the British. Two Government’s, one Labour, one Conservative have aggressive pursued an open data and open government strategy. This, would be my hope for Canada. However, it does not appear that is is presently the case. So, another model should be adopted. Fortunately, such a model exists.

Last year, under the leadership of Nicholas Gruen, the Australian government launched a Government 2.0 taskforce on which I had the pleasure of serving on the International Reference Group. The Australian Taskforce was non-partisan and was made up of policy and technical experts and entrepreneurs from government, business, academia, and cultural institutions. More importantly, the overwhelming majority of its recommendations were adopted.

To replicate its success in Canada I believe the Committee should copy the best parts of the Australian taskforce. The topic of Canadians access to their government is of central importance to all Canadians – to non-profits, to business interests, to public servants and, of course, to everyday citizens. Rather than non-partisan, I would suggest that a Canadian taskforce should be pan-partisan – which the Committee already is. However, like the Australian Taskforce it should include a number of policy and technical experts from outside government. This fill committee would this represent both a political cross-section and substantive knowledge in the emerging field of government 2.0. It could thus, as a whole, effectively and quickly draft recommendations to Parliament.

Best of all, because of step #1, this work could proceed in parallel to any projects (if any) already initiated by the government and possibly even inform such work by providing interim updates.

I concede such an approach may be too radical, but I hope it is at least a starting point for an interesting approach.

3. Lead by Example

There is one arena where politicians need not wait on the government to make plans: Parliament itself. Over the past year, while in conversations with the Parliamentary IT staff as well as the Speaker of the House, I have worked to have Parliament make more data about its own operations open. Starting in January, the Parliamentary website will begin releasing the Hansard in XML – this will make it much easier for software developers like the creators of Openparliament.ca as and howdtheyvote.ca to run their sites and for students, researchers and reporters to search and analyze our country’s most important public discussions. In short, by making the Hansard more accessible the Speaker and his IT staff are making parliament more accessible. But this is only the beginning of what parliamentarians could do to make for a truly Open Parliament. The House and Senate’s schedules and agendas, along with committee calendars should all be open. So to should both chambers seating arrangement. Member’s photos and bios should be shared with an unrestricted license as should the videos of parliament.

Leadership in this space would send a powerful message to both the government and the public service that Canada’s politicians are serious about making government more open and accessible to those who elect it. In addition, it could also influence provincial legislature’s and even municipal governments, prompting them to do the same and so enhance our democracy at every level.

4. Finally, understand your task: You are creating a Knowledge Government for a Knowledge Society

One reason I advise the Committee to take on external members is because, laudably, many admit this topic is new to them. But I also want the committee members to understand the gravity of their task. Open Government, Open Data and/or Government 2.0 are important first steps in a much larger project.

What you are really wrestling with here is what government is going to look like in an knowledge economy and a knowledge society. How is going to function with knowledge workers as employees? And, most importantly, how is it going to engage with knowledge citizens, many of whom can and want to make real contributions beyond the taxes they pay and don’t need government to self-organize?

In short, what is a knowledge based government going to look like?

At the centre of that question is how we manage and share information. The basic building block of a knowledge driven society.

Look around, and you can see how the digital world is transforming how we do everything. Few of us can imagine living today without access to the internet and the abundance of information it brings to us. Indeed, we have already become so used to the internet we forget how much it has radically changed whole swaths of our life and economy from the travel and music industry to the post to political fund-raising and to journalism.

If today our government still broadly looks and feels like an institution shaped by the printing press it is because, well it is. Deputy Ministers and Ministers still receive giant briefing binders filled with paper. This is a reflection of how we deal within information and knowledge in government, we move it around (for good reasons) in siloes, operating as though networks, advance search, and other innovations don’t exist (even though they already do).

How our government deals with information is at the heart of your task. I’m not saying you have to re-invent government or dismantle all the silos and ministries. Quite the contrary, I believe small changes can be made that will yield significant benefits, efficiencies and savings while enhancing our democracy. But you will be confronting decades, if not centuries of tradition, culture and process in an institution that is about to go through the biggest change since the invention of the printing press. You don’t have to do it all, but even some small first steps will not come easily. I share this because I want you going into the task with eyes wide open.

At the very least we aren’t going first, our cousins both across the Atlantic, the Pacific and our southern border have already taken the plunge. But this should add urgency to our task. We cannot afford to stand by while others renew their democratic institutions while simultaneously enhancing an emerging and critical pillar of a new knowledge economy and knowledge society.

Launching datadotgc.ca 2.0 – bigger, better and in the clouds

Back in April of this year we launched datadotgc.ca – an unofficial open data portal for federal government data.

At a time when only a handful of cities had open data portals and the words “open data” were not being even talked about in Ottawa, we saw the site as a way to change the conversation and demonstrate the opportunity in front of us. Our goal was to:

  • Be an innovative platform that demonstrates how government should share data.
  • Create an incentive for government to share more data by showing ministers, public servants and the public which ministries are sharing data, and which are not.
  • Provide a useful service to citizens interested in open data by bringing it all the government data together into one place to both make it easier to find.

In every way we have achieved this goal. Today the conversation about open data in Ottawa is very different. I’ve demoed datadotgc.ca to the CIO’s of the federal government’s ministries and numerous other stakeholders and an increasing number of people understand that, in many important ways, the policy infrastructure for doing open data already exists since datadotgc.ca show the government is already doing open data. More importantly, a growing number of people recognize it is the right thing to do.

Today, I’m pleased to share that thanks to our friends at Microsoft & Raised Eyebrow Web Studio and some key volunteers, we are taking our project to the next level and launching Datadotgc.ca 2.0.

So what is new?

In short, rather than just pointing to the 300 or so data sets that exist on federal government websites members may now upload datasets to datadotg.ca where we can both host them and offer custom APIs. This is made possible since we have integrated Microsoft’s Azure cloud-based Open Government Data Initiative into the website.

So what does this mean? It means people can add government data sets, or even mash up government data sets with their own data to create interest visualization, apps or websites. Already some of our core users have started to experiment with this feature. London Ontario’s transit data can be found on Datadotgc.ca making it easier to build mobile apps, and a group of us have taken Environment Canada’s facility pollution data, uploaded it and are using the API to create an interesting app we’ll be launching shortly.

So we are excited. We still have work to do around documentation and tracking some more federal data sets we know are out there but, we’ve gone live since nothing helps us develop like having users and people telling us what is, and isn’t working.

But more importantly, we want to go live to show Canadians and our governments, what is possible. Again, our goal remains the same – to push the government’s thinking about what is possible around open data by modeling what should be done. I believe we’ve already shifted the conversation – with luck, datadotgc.ca v2 will help shift it further and faster.

Finally, I can never thank our partners and volunteers enough for helping make this happen.