Yearly Archives: 2010

Beautiful Wepages

For those not in the know the Webby Award nominations – probably the most important award for a website – were released a couple of weeks ago and voting, which does influence the outcome, will be closing soon.

If you haven’t voted yet I’d strongly urge you to head over to the website, check out the nominees, and vote. There are A LOT of categories so it can be a little overwhelming. That said, in the “Activism” section I am backing a horse and hope you’ll consider voting for them. My favourite – for both design and message – is the Stemcell Foundation’s “Renew The World” website by Manifest Communications.

Here’s a sample of the video off the website. In addition to voting, consider heading over to the site and sign the Stemcell Charter.

That said, if you like another website, vote for it! The Webby award webpage is filled with links to beautiful website and so poke around and get a sense of what is out there and what people think makes for a great site. Lots to learn.

Connectedness, Volleyball and Online Communities

I’m currently knee deep into Connected: The Surprising Power of Our Social Networks and How They Shape Our Lives by Christakis & Fowler and am thoroughly enjoying it.

One fascinating phenomenon the book explores is how emotions can spread from person to person. In other words, when you are happy you increase the odds your friends will be happy and, interestingly, that your friends’ friends will be happy. Indeed Christakis & Fowler’s research suggests that emotional states are, to a limited degree, contagious.

All this reminded me of playing competitive volleyball. I’ve always felt volleyball is one of the most psychologically difficult games to play. I’ve regularly seen fantastic teams collapse in front of vastly inferior opponents. I used to believe that the unique structure of volleyball accounted for this. The challenge is that the game pauses at the end of every point, allowing players to reflect on what happened and, more importantly, assign blame (which can often be allocated to a single individual on the team). This makes it easy for teams to start blaming a player, over-think a point, or get frustrated with one another.

As a result, even prior to reading Connected, I knew team cohesion was essential in volleyball (and, admittedly, any sport) . This is often why, between points, you’ll see volleyball teams come together and high-five or shake hands even if they lost the point. If emotions and confidence are contagious, I can now see why it is a team starts to lose a little confidence and consequently then plays a little worse causing them to lose still more confidence and then, suddenly they are in a negative rut and can’t escape.(Indeed, this peer reviewed paper showed that tactile touch among NBA players was a predictor of individual and team success)

Of course, I’ve also long believed the same is true of online (and, in particular, open source) communities. That poisonous and difficult people don’t just negatively impact the people they are in direct contact with, but also a broader group of people in the community. Moreover, because communication often takes place in comment threads the negative impact of poisonous people could potentially linger, dragging down the attitude and confidence of everyone else in the community. I’ve often thought that the consequence of negative behaviours in the online communities has been underestimated – Christakis and Fowler’s research suggests there are some more concrete ways to measure this negative impact, and to track it. Negative behaviour fosters (and possibly even attracts) still more negative behaviour, creating a downward loop and likely causing positive or constructive people to opt out, or even never join the community in the first place.

In the end, finding ways to identify, track and mitigating negative behaviour early on – before it becomes contagious – is probably highly important. This is just an issue of having people be positive, it is about creating a productive and effective space, be it in pursuit of an open source software product, or a vibrant and interesting discussion at the end of an online newspaper article.

Lesson in Misunderstanding Abundance and Scarcity: Quebec and Abitibibowater

One of the biggest problems old institutions have grasping the internet is how it changes notions of abundance and scarcity. We are used to a world of scarcity where, if you have something, I cannot also have it, so we need a way to figure out how to allocate it (market forces, government regulation, etc…). The two following examples are wonderful cases of acting like their is scarcity when there is abundance, and acting like there is abundance, where there is scarcity…

The Province of Quebec – Don’t help people vote!

The other day a friend of mine, a web developer who goes by the name Ducky, asked the Province of Quebec for image (KML) files for provincial ridings in Quebec. These files would allow her to show people, in say, Google Maps, which riding they live in, who their representative in the National Assembly might be, etc… The Government of Quebec was happy to share, but only on the condition that she never make money using the information:

From: Charlotte Perreault <Cperreault@dgeq.qc.ca>
Date: 2009/12/8
Subject: Shapefiles pour les circonscriptions électorales québécoises
To: ducky@*******.com
Cc: Mireille Loignon <Mloignon@dgeq.qc.ca>

Madame Duck Sherwood

Nous avons pris connaissance de votre demande concernant l’utilisation des cartes des circonscriptions électorales québécoises sur le site http://www.electionsquebec.qc.ca/fr/copyright.asp .  Après analyse, nous considérons que pour l’instant il n’y a pas d’utilisation commerciale de ces fichiers.  Par conséquent, vous êtes autorisée à les utiliser en mentionnant la source, soit la Commission de la représentation électorale du Québec.

Cependant, si vous souhaitez un jour commercialiser un produit fait à partir de ces cartes, vous devrez communiquer avec nous en nous transmettant plus de renseignements comme par, exemples, le prix de vente dudit produit, ou tout autre renseignement qui nous permettra d’évaluer correctement  votre demande.

Veuillez agréer, Madame, l’expression de nos meilleurs sentiments.

Charlotte Perreault, conseillère en communication
Direction des communications
Directeur général des élections du Québec
Édifice René-Lévesque
3460, rue de La Pérade
Québec (QC)
G1X 3Y5
So, if Ducky, or say a newspaper, wanted to create a website to help residents of Quebec identify what riding they live in… and were going to have advertising on the site, they couldn’t do that without permission from the Quebec government. Heaven forbid that someone offer citizens or voters a helpful or interesting service, especially using information that is in the public interest.
Why is this? I’m not sure. If Ducky created her site, it wouldn’t preclude someone else from using the exact same information to create their own site. No, the Quebec government is turning something that is infinitely abundant and that can be used over and over again (election data) and making it scarce by preventing anyone from using it in a way that could be useful and profitable. Maybe they think they could make some money from the data… but sadly, every academic anlsysis shows they won’t – usually money recovered for data like this is covers the cost of collecting the money, and nothing else. In the end, the losers in all this are… Quebec residents.
Abitibibowater – Print, Print and Print more!
On the flip side of this is Abitibibowater, the producer of “a wide range of newsprint, commercial printing papers, market pulp and wood products.” Apparently, they have become concerned about the number of people who include “do not print this e-mail” at the bottom of their emails. So they sent around this memo to their employees:

To all our employees:

At AbitibiBowater, we rely on the forest to make our products, and respect for the environment is a fundamental part of everything we do. Too often, however, we see the use of paper singled out as something ‘bad’ for the environment, which is the reason why some people include “do not print this e-mail” notes to the end of their electronic messages. There has been ongoing discussion concerning this matter in the media recently, given the proliferation of electronic readers. Here is a link to a March article that appeared on PBS’s website, which raises this issue.

We believe it is OK to print, especially if you use paper sourced from independently certified forests and recycle the paper after you are finished with it. Our Company works hard to operate in a way that is sustainable, and that reflects the values of our employees, our customers, and the communities in which we operate. We are continuously improving our manufacturing processes and reducing our environmental footprint.

We have developed a page on our website to address this issue (abitibibowater.com/print) that explains why you can feel comfortable knowing that using paper is an environmentally responsible choice. We have written our own e-mail footnote that makes this point and invites others to find out why:

It’s OK to print this e-mail. Paper is made from a renewable resource. Please choose paper sourced from independently certified sustainable forests and recycle. For more information visit abitibibowater.com/print.

Uh… actually it is still really, really, wasteful to print an email that you could just forward to someone. Even if that email is printed on certified paper. Paper may be a renewable resource, but that doesn’t mean that it isn’t scarce. Trees used for making paper aren’t being used for something else, like say, building homes, making tourists happy, or, say, converting CO2 into air.

Pretending like something is abundant when it is actually scarce is deeply irresponsible – and Abitibi’s email is prompting me to actually add the “do not print this email” signature to the bottom of my own emails.

The economics of abundance and scarcity matter – knowing which one you are dealing with matters. Make the wrong choice, and you could end up looking at best like a fool, at worst, deeply irresponsible.

Help with datadotgc.ca

For regular readers of my blog I promise not to talk too much about datadotgc.ca here at eaves.ca. I am going to today since I’ve received a number of requests from people asking if and how they could help so I wanted to lay out what is on my mind at the moment, and if people had time/capacity, how they could help.

The Context

Next wednesday I’ll be doing a small presentation to all the CIO’s of the federal public service. During that presentation I’d like to either go live to datadotgc.ca or at least show an up to date screen shot (if there is no internet). It would be great to have more data sets in the site at that time so I can impress upon this group a) how little machine readable data there is in Canada versus other countries (especially the UK and US) and b) show them what an effective open data portal should look like.

So what are the datadotgc.ca priorities at this moment?

1. Get more data sets listed in datadotgc.ca

There is a list of machine readable data sets known to exist in the federal government that has been posted here. For coders – the CKAN API is relatively straight forward to use. There is also an import script that can allow one to bulk import data lists into datadotgc.ca, as well as instructions posted here in the datadotgc.ca google group.

2. Better document how to bulk add data sets.

While the above documentation is good, I’d love to have some documentation and scripts that are specific to datadotgc.ca/ca.ckan.net. I’m hoping to recruit some help with this tonight at the Open Data hackathon, but if you are interested, please let me know.

3. Build better tools

One idea I had that I have shared with Steve T. is to develop a jet-pack add on for Firefox that, when you are on a government page scans for links to certain file types (SHAPE, XLS, etc…) and then let’s you know if they are already in datadotgc.ca. If not, it would provide a form to “liberate the dataset” without forcing the user to leave the government website. This would make it easier for non-developers to add datasets to datadotgc.ca.

4. Locate machine readable data sets

Of course, we can only add to datadotgc.ca data sets that we know about, so if you know about a machine readable datasets that could be liberated, please add it! If there are many and you don’t know how ping me, or add it directly to the list in the datadotgc.ca google group.

Open Government interview and panel on TVO's The Agenda with Steve Paikin

My interview on TVO’s The Agenda with Steve Paikin has been uploaded to Youtube (BTW, it is fantastic that The Agenda has a YouTube channel where it posts all its interviews. Kudos!). If you live outside Ontario, or were wrapped up in the Senators-Pens playoff game that was on at the same time (which obviously we destroyed in the ratings), I thought I’d throw it up here as a post in case it is of interest. The first clip is a one on one interview between myself and Paikin. The second clip is the discussion panel that occurred afterward with myself, senior editor of Reason magazine Katherine Mangu-Ward , American Prospect executive editor Mark Schmitt and the Sunlight Foundation’s Policy Director John Wonderlich.

Hope you enjoy!

One on one interview with Paikin:

Panel Discussion:

Why Old Media and Social Media Don't Get Along

Earlier today I did a brief drop in phone interview on CPAC’s Goldhawk Live. The topic was “Have social media and technology changed the way Canadians get news?” and Christoper Waddell, the Director of Carleton University’s School of Journalism and Chris Dornan, Director of Carleton University’s Arthur Kroeger School of Public Affairs were Goldhawk’s panel of experts.

Watching the program prior to being brought in I couldn’t help but feel I live on a different planet from many who talk about the media. Ultimately, the debate was characterized by a reactive, negative view on the part of the mainstream media supporters. To them, threats are everywhere. The future is bleak, and everything, especially democratic institutions and civilization itself teeter on the edge. Meanwhile social media advocates such as myself are characterized as delusional techno-utopians. Nothing, of course, could be further from the truth. Indeed, both sides share a lot in common. What distinguishes though, is that while traditionalists are doom and gloom, we are almost defined by the sense of the possible. New things, new ideas, new approaches are becoming available every day. Yes, there will be new problems, but there will also be new possibilities and, at least, we can invent and innovate.

I’m just soooooo tired of the doom and gloom. It really makes one want to give up on the main stream media (like many, many, many people under 30 have). But, we can’t. We’ve got to save these guys from themselves – the institutions and the brands matter (I think). So, in that pursuit, let’s tackle the beast head on, again.

Last, night the worse offender was Goldhawk, who tapped into every myth that surrounds this debate. Let’s review them one by one.

Myth 1: The average blog is not very good – so how can we rely on blogs for media?

For this myth, I’m going to first pull a little from Missing the Link, now about to be published as a chapter in a journalism textbook called “The New Journalist”:

The qualitative error made by print journalists is to assume that they are competing against the average quality of online content. There may be 1.5 million posts a day, but as anyone whose read a friend’s blog knows, even the average quality of this content is poor. But this has lulled the industry into a false sense of confidence. As Paul Graham describes: “In the old world of ‘channels’ (e.g. newspapers) it meant something to talk about average quality, because that’s what everyone was getting whether they liked it or not. But now you can read any writer you want. Consequently, print media isn’t competing against the average quality of online writing, they’re competing against the best writing online…Those in the print media who dismiss online writing because of its low average quality are missing an important point. No one reads the average blog.”

You know what though, I’m going to build on that. Goldhawk keeps talking about the average blog or average twitterer (which of course, no one follows, we all follow big names, like Clay Shirky and Tim O’Reilly). But you know what? They keep comparing the average blog to the best newspapers. The fact is, even the average newspaper sucks. The Globe represents the apex of the newspaper industry in Canada, not the average, so stop using it as an example. To get the average, go into any mid-sized town and grab a newspaper. It won’t be interesting. Especially to you – an outsider. It will have stories that will appeal to a narrow audience, and even then, many of these will not be particularly well written. More importantly still, there will little, and likely no, investigative journalism – that thing that allegedly separates blogs from newspapers. Indeed, even here in Vancouver, a large city, it is frightening how many times press releases get marginally touched up and then released as “a story.” This is the system that we are afraid of losing?

Myth 2: How will people sort good from low quality news?

I always love this myth. In short, it presumes that the one thing the internet has been fantastic at developing – filters – simple won’t evolve in a part of the media ecosystem (news) where people desperately want them. At best, this is naive. At worse, it is insulting. Filters will develop. They already have. Twitter is my favourite news filter – I probably get more news via it than any other source. Google is another. Nothing gets you to a post or article about a subject you are interested in like a good (old-fashioned?) google search. And yes, there is also going to be a market for branded content – people will look for that as short cut for figuring out what to read. But please people are smarter than you think at finding news sources.

Myth 3: People lack media savvy to know good from low quality news.

I love the elitist contempt the media industry sometimes has towards its readers. But, okay, let’s say this is true. Then the newspapers and mainstream media have only themselves to blame. If people don’t know what good news is, it is because they’ve never seen it (and by and large, they haven’t). The most devastating critique on this myth is actually delivered by one of my favourite newspaper men: Kenneth Whyte is his must listen-to Dalton Camp Lecture on journalism. In it Whyte talks about how, in the late 19th and early 20th century NYC had dozens and dozens of newspapers that fought for readership and people were media savvy, shifting from paper to paper depending on quality and perspective. That all changed with consolidation and a shift from paying for content to advertising for content. Advertisers want staid, plain, boring newspapers with big audiences. This means newspapers play to the lowest common denominator and are market oriented to be boring. It also leaves them beholden to corporate interests (when was the last time the Vancouver Sun really did a critical analysis of the housing industry – it’s biggest advertisement source?). If people are not media savvy it is, in part, because the media ecosystem demands so little of them. I suspect that social media can and will change this. Big newspapers may be what we know, but they may not be good for citizenship or democracy.

Myth 4: There will be no good (and certainly no investigative) journalism with mainstream media.

Possible. I think the investigative journalism concern is legitimate. That said, I’m also not convinced there is a ton of investigative journalism going on. There may also be more going on in the blogs than we might know. It could be that these stories a) don’t get prominence and b) even when they do, often newspapers don’t cite blogs, and so a story first broken by a blog may not be attributed. But investigative journalism comes in different shapes and sizes. As I wrote in one of my more viewed posts, The Death of Journalism:

I suspect the ideal of good journalism will shift from being what Gladwell calls puzzle solving to mystery solving. In the former you must find a critical piece of the puzzle – one that is hidden to you – in order to explain an event. This is the Woodward and Bernstein model of journalism – the current ideal. But in a transparent landscape where huge amounts of information about most organizations is being generated and shared the critical role of the journalist will be that of mystery solving – figuring out how to analyze, synthesize and discover the mystery within the vast quantity of information. As Gladwell recounts this was ironically the very type of journalism that brought down Enron (an organization that was open, albeit deeply  flawed). All of the pieces of that lead to the story that “exposed” Enron were freely, voluntarily and happily given to reports by Enron. It’s just a pity it didn’t happen much, much sooner.

I for one would celebrate the rise of this mystery focused style of “journalism.” It has been sorely needed over the past few years. Indeed, the housing crises that lead to the current financial crises is a perfect example of case where we needed mystery solving not puzzle solving, journalism. The fact that sub-prime mortgages were being sold and re-packaged was not a secret, what was lacking was enough people willing to analyze and write about this complex mystery and its dangerous implications.

And finally, Myth 5: People only read stories that confirm their biases.

Rather than Goldhawk it was Christopher Waddell who kept bringing this point up. This problem, sometimes referred to as “the echo chamber” effect is often cited as a reason why online media is “bad.” I’d love to know Waddell’s sources (I’m confident he has some – he is very sharp). I’ve just not seen any myself. Indeed, Andrew Potter recently sent me a link to “Ideological Segregation Online and Offline.” What is it? A peer reviewed study that found no evidence the Internet is becoming more ideologically segregated. And the comparison is itself deeply flawed. How many conservatives read the Globe? How many liberals read the National Post? I love the idea that somehow main stream media doesn’t ideologically segregate an audience. Hasn’t any looked at Fox or MSNBC recently?

Ultimately, it is hard to watch (or participate) in these shows without attributing all sorts of motivations to those involved. I keep feeling like people are defending the status quo and trying to justify their role in the news ecosystem. To be fair, it is a frightening time to be in media.

When someone demands to know how we are going to replace newspapers, they are really demanding to be told that we are not living through a revolution. They are demanding to be told that old systems won’t break before new systems are in place. They are demanding to be told that ancient social bargains aren’t in peril, that core institutions will be spared, that new methods of spreading information will improve previous practice rather than upending it. They are demanding to be lied to.

And I refuse to lie. It sucks to be a newscaster or a journalist or a columnist. Especially if you are older. Forget about the institutions (they’ve already been changing) but the culture of newsmedia, which many employed in the field cling strongly to, is evolving and changing. That is a painful process, especially to those who have dedicated their life to it. But that old world was far from perfect. Yes, the new world will have problems, but they will be new problems, and there may yet be solutions to them, what I do know is that there aren’t solutions to the old problems in the old system and frankly, I’m tired of those old problems. So let’s get on with it. Be critical, but please, stop spreading the myths and the fear mongering.

Datadotgc.ca Launched – the opportunity and challenge

Today I’m really pleased to announce that we’ve launched datadotgc.ca, a volunteer driven site I’m collaboratively creating with a small group of friends and, I hope, a growing community that, if you are interested, may include you.

As many of you already know I, and many other people, want our governments to open up and share their data, in useful, structured formats that people can actually use or analyze. Unlike our American and British peers, the Canadian Federal (and provincial…) government(s) currently have no official, coordinated effort to release government data.

I think that should change.

So rather than merely complain that we don’t have a data.gov or data.gov.uk in Canada, we decided to create one ourselves. We can model what we want our governments to do and even create limited versions of the service ourselves. So that is what we are doing with this site. A stab at showing our government, and Canada, what a federal open data portal could and should look like – one that I’m hoping people will want to help make a success.

Some two things to share.

First, what’s our goal for the site?

  • Be an innovative platform that demonstrates how government should share data.
  • Create an incentive for government to share more data by showing ministers, public servants and the public which ministries are sharing data, and which are not.
  • Provide a useful service to citizens interested in open data by bringing it all the government data together into one place to both make it easier to find.

Second, our big challenge.

As Luke C, one datadotgc.ca community member said to me – getting the site up is the easier part. The real challenge is building a community of people who will care for it and help make it a living, growing and evolving success. Here there is lots of work still to be done. But if you feel passionate about open government and are interested in joining our community, we’d love to have you. At the moment, especially as we still get infrastructure to support the community in place, we are convening at a google group here.

So what our some of the things I think are a priority in the short term?

  • Adding or bulk scraping in more data sets so the site more accurately displays what is available
  • Locating data sets that are open and ready to be “liberated”
  • Documenting how to add or scrape in a data set to allow people to help more easily
  • Implement a more formal bug and feature tracker
  • Plus lots of other functionality I know I at least (and I’m sure there are lots more ideas out there) would like to add (like “request a closed data set”)

As Clay Shirky once noted about any open source project, datadotgc.ca is powered by love. If people love the site and love what it is trying to accomplish, then we will have a community interested in helping make it a success. I know I love datadotgc.ca – and so my goal is to help you love it too, and to do everything I can to make it as easy as possible for you to make whatever contribution you’d like to make. Creating a great community is the hardest but best part of any project. We are off to a great start, and I hope to maybe see you on the google group.

Finally, just want to thank everyone who has helped so far, including the fine people at Raised Eyebrow Web Studio, Luke Closs, and a number of fantastic coders from the Open Knowledge Foundation. There are also some great people over at the Datadotgc.ca Google Group who have helped scrape data, tested for bugs and been supportive and helpful in so many ways.

Case Study: How Open data saved Canada $3.2 Billion

Note: I’ll be on TVO’s The Agenda with Steve Paikin tonight talking about Government 2.0.

Why does open data matter? Rather than talk in abstract terms, let me share a well documented but little known story about how open data helped expose one of the biggest tax frauds in Canada’s history.

It begins in early 2007 when a colleague was asked by a client to do an analysis of the charitable sector in Toronto. Considering it a simply consulting project, my colleague called the Canada Revenue Agency (CRA) and asked for all the 2005 T3010s – the Annual Information Returns where charities disclose to the CRA their charitable receipts and other information – in Toronto. After waiting several weeks and answering a few questions, the CRA passed along the requested information.

After spending time cleaning up the data my colleague eventually had a working excel spreadsheet and began to analyze the charitable sector in the Greater Toronto Area. One afternnon, on a lark, they decided to organize the charities by size of tax-receipted charitable donations.

At this point it is important to understand something about scale. The United Way of Greater Toronto is one of the biggest charities in North America, indeed its most recent annual charitable donation drive was the biggest on the continent. In 2008 – the year of the financial crisis started – the United Way of Greater Toronto raised $107.5 million.

So it was with some surprise that after sorting the charities by 2005 donation amounts my colleague discovered that the United Way was not first on the list. It wasn’t even second.

It was third.

This was an enormous surprise. Somewhere in Toronto, without anyone being aware of it, two charities had raised more money than the United Way (which in 2005 raised target of $96.1M). The larger one, the International Charity Association Network (ICAN) raised $248M in 2005. The other, the Choson Kallah Fund of Toronto had receipts of $120M (up from $6M in 2003).

Indeed, four out the top 15 charities on the list, including Millennium Charitable Foundation, Banyan Tree, were unknown to my colleague, someone who had been active in the Toronto charitable community for over a decade.

All told, my colleague estimated that these illegally operating charities alone sheltered roughly Half a billion dollars in 2005. Indeed, newspapers later confirmed that in 2007, fraudulent donations were closer to a billion dollars a year, with some some 3.2 billion dollars illegally sheltered, a sum that accounts for 12% of all charitable giving in Canada.

Think about this. One billion dollars. A year. That is almost .6% of the Federal Government’s annual budget.

My colleague was eager to make sure that CRA was taking action on these organizations, but it didn’t look that way. The tax frauds were still identified by CRA as qualified charities and were still soliciting donors with the endorsement of government. They knew that a call to CRA’s fraud tip line was unlikely to prompt swift action. The Toronto Star had been doing its own investigations into other instances of charity fraud and had been frustrated by CRA’s slow response.

My colleague took a different route. They gave the information to the leadership of the charitable sector and those organizations as a group took it to the leadership at CRA. From late 2007 right through 2009 the CRA charities division – now under new leadership – has systematically shut down charity tax shelters and are continuing to do so.  One by one, International Charity Association Network, Banyan Tree Foundation, Choson Kallah Fund, the Millennium Charitable Foundation and others identified by my colleague have lost their charitable status. A reported $3.2 billion in tax receipts claimed by 100,000 Canadian tax filers have so far been disallowed or are being questioned. A class action suit launched by thousands of donors against the organizers and law firm of Banyan Tree Foundation was recently certified. It’s a first. Perhaps the CRA was already investigating these cases. It must build its cases carefully as, if they end up in court and fail to successfully present their case, they could help legalize a tax loophole. It may just have been moving cautiously. But perhaps it did not know.

This means that, at best, government data – information that should be made more accessible and open in an unfettered and machine readable format – helped reveal one of the largest tax evasion scandals in the country’s history. But if the CRA was already investigating, scrutiny of this data by the public served a different purpose – helping to bring these issues out into the open, forcing CRA to take public action (suspending these organizations’ right to solicit more donations), sooner rather than later.  Essentially from before 2005-2007 dozens of charities were operating illegally. Had the data about their charitable receipts been available for the public’s routine review,  someone in the public might have taken notice and raised a fuss earlier. Perhaps even a website tracking donations might have been launched. This would have exposed those charities that had abnormally large donations with few programs to explain then. Moreover, it might have given some of the 100,000 Canadians now being audited a tool for evaluating the charities they were giving money to.

In the computer world there is something called Linus’ Law, which states: “given enough eyeballs, all bugs (problems) are shallow.” The same could be said about many public policy or corruption issues. For many data sets, citizens should not have to make a request. Nor should we have to answer questions about why we want the data. It should be downloadable in its entirety. Not trapped behind some unhelpful search engine. When data is made readily available in machine readable formats, more eyes can look at it. This means that someone on the ground, in the community (like, say, Toronto) who knows the sector, is more likely to spot something a public servant in another city might not see because they don’t have the right context or bandwidth. And if that public servant is not allowed to talk about the issue, then they can share this information with their fellow citizens.

This is the power of open data: The power to find problems in complicated environments, and possibly even to prevent them from emerging.

Opening Parliament and other big announcements

This is going to be an exciting week for online activists seeking to make government more open and engaged.

First off, openparliament.ca launched yesterday. This is a fantastic site with a lot going for it – go check it out (after reading my other updates!). And huge kudos to its creator Michael Mulley. Just another great example of how our democratic institutions can be hacked to better serve our needs – to make them more open, accessible and engaging. There is a ton of stuff that could be built on top of Michael’s and others – like Howdtheyvote, sites. I’ve written more about this in a piece on the Globe’s website titled If You Won’t Tell Us About Our MPs Well Do It For You.

Second, as follow on to the launch of openparliament.ca, I’ve been meaning to share for some time that I’ve been having conversations with the House of Parliament IT staff over the past couple of months. About a month ago parliament IT staff agreed to start sharing the Hansard, MP’s bios, committee calendars and a range of other information via XML (sorry for not sharing this sooner, things have been a little crazy). They informed me that they would start doing this before the year is over – so I suspect it won’t happen in the next couple of months, but will happen at some point in the next 6 months. This is a huge step forward for the house and hopefully not the last (also, there is no movement on the senate as of yet). There are still a ton more ways that information about the proceedings of Canada’s democracy could be made more easily available, but we have some important momentum with great sites like those listed above, and internal recognition to share more data. I’ll be having further conversations with some of the staff over the coming months so will try to update people on progress as I find out.

Finally, I am gearing up to launch datadotgc.ca. This is a project I’ve been working on for quite some time with a number of old and new allies. Sadly, the Canadian government does not have an open data policy and there is no political effort to create a data.gc.ca like that created by the Obama administration (http://www.data.gov/) or of the British Government (http://data.gov.uk/). So, I along with a few friends have decided to create one for them. I’ll have an official post on this tomorrow. Needless to same, I’m excited. We are still looking for people to help us populate the site with open government data sets – and have even located some that we need help scraping – so if you are interested in contributing feel free to join the datadotgc.ca google group and we can get you password access to the site.

Mozilla Drumbeat: Help keep the web open

Mozilla Drumbeat is the Mozilla Foundations new venture. An effort to reach out beyond those who have helped make the Firefox web browser to a broader community – those that care about keeping the internet open but who want to contribute in other ways.

Drumbeat will have three components:

1) Projects – many of which are already underway

2) Local events – like the upcoming on in Toronto on April 24th

3) A website – that ties it all together, a place where people can gather virtually to organize

So what can you do?

First, if you are interested in hosting a local Mozilla Drumbeat event, contact Nathaniel James at the Mozilla Foundation, there is going to be Facilitator Training event in Toronto on April 23-24th and in Berlin on May 7th-8th.

Second, consider participating in one of the Drumbeat projects listed on the Drumbeat website (still in beta).

I’m looking forward to see Drumbeat evolve and for it to become easier for people to participate in its various projects. Ultimately we need a way to protect the openness of the web – the thing that makes the web so fun and interesting for all us – and Drumbeat is a great starting point.