24 July 2009

Why the GNU GPL v3 Matters Even More

A little while back, I wrote a post called "Why the GNU GPL Still Matters". I was talking in general terms, and didn't really distinguish between the historical GNU GPL version 2 and the new version 3. That's in part because I didn't really have any figures on how the latter was doing. Now I do, because Matt Asay has just published some plausible estimates:

In July 2007, version 3 of the GNU General Public License barely accounted for 164 projects. A year later, the number had climbed past 2,000 total projects. Today, as announced by Google open-source programs office manager Chris DiBona, the number of open-source projects licensed under GPLv3 is at least 56,000.

And that's just counting the projects hosted at Google Code.

In a hallway conversation with DiBona at OSCON, he told me roughly half of all projects on Google Code use the GPL and, of those, roughly half have moved to GPLv3, or 25 percent of all Google Code projects.

With more than 225,000 projects currently hosted at Google Code, that's a lot of GPLv3.

If we make the reasonable assumption that other open-source project repositories Sourceforge.net and Codehaus have similar GPLv3 adoption rates, the numbers of GPLv3 projects get very big, very fast.

This is important not just because it shows that there's considerable vigour in the GNU GPL licence yet, but because version 3 addresses a particularly hot area at the moment: software patents. The increasing use of GPL v3, with its stronger, more developed response to that threat, is therefore very good news indeed.

Follow me @glynmoody on Twitter or identi.ca.

22 July 2009

Pat "Nutter" Brown Strikes Again

To change the world, it is not enough to have revolutionary ideas: you also have the inner force to be able to realise them in the face of near-universal opposition/indifference/derision. Great examples of this include Richard Stallman, who ploughed his lonely GNU furrow for years before anyone took much notice, and Michael Hart, who did the same for Project Gutenberg.

Another of these rare beings with both vision and tenacity is Pat Brown, a personal hero of mine. Not content with inventing one of the most important experimental tools in genomics - DNA microarrays - Brown decided he wanted to do something ambitious: open access publishing. This urge turned into the Public Library of Science (PLoS) - and even that is just the start:


PLoS is just part of a longer range plan. The idea is to completely change the way the whole system works for scientific communication.

At the start, I knew nothing about the scientific publishing business. I just decided this would be a fun and important thing to do. Mike Eisen, who was a post-doc in my lab, and I have been brain-storming a strategic plan, and PLoS was a large part of it. When I started working on this, almost everyone said, “You are completely out of your mind. You are obviously a complete idiot about how publishing works, and besides, this is a dilettante thing that you're doing.” Which I didn't feel at all.

I know I'm serious about it and I know it's doable and I know it's going to be easy. I could see the thermodynamics were in my favor, because the system is not in its lowest energy state. It's going to be much more economically efficient and serve the customers a lot better being open access. You just need a catalyst to GET it there. And part of the strategy to get it over the energy barrier is to apply heat—literally, I piss people off all the time.

In case you hadn't noticed, that little plan "to completely change the way the whole system works for scientific communication" is coming along quite nicely. So, perhaps buoyed up by this, Brown has decided to try something even more challenging:

Brown: ... I'm going to do my sabbatical on this: I am going to devote myself, for a year, to trying to the maximum extent possible to eliminate animal farming on the planet Earth.

Gitschier: [Pause. Sensation of jaw dropping.]

Brown: And you are thinking I'm out of my mind.

Gitschier: [Continued silence.]

Brown: I feel like I can go a long way toward doing it, and I love the project because it is purely strategy. And it involves learning about economics, agriculture, world trade, behavioral psychology, and even an interesting component of it is creative food science.

Animal farming is by far the most environmentally destructive identified practice on the planet. Do you believe that? More greenhouse production than all transportation combined. It is also the major single source of water pollution on the planet. It is incredibly destructive. The major reason reefs are dying off and dead zones exist in the ocean—from nutrient run-off. Overwhelmingly it is the largest driving force of deforestation. And the leading cause of biodiversity loss.

And if you think I'm bullshitting, the Food and Agricultural Organization of the UN, whose job is to promote agricultural development, published a study, not knowing what they were getting into, looking at the environmental impact of animal farming, and it is a beautiful study! And the bottom line is that it is the most destructive and fastest growing environmental problem.

Gitschier: So what is your plan?

Brown: The gist of my strategy is to rigorously calculate the costs of repairing and mitigating all the environmental damage and make the case that if we don't pay as we go for this, we are just dumping this huge burden on our children. Paying these costs will drive up the price of a Big Mac and consumption will go down a lot. The other thing is to come up with yummy, nutritious, affordable mass-marketable alternatives, so that people who are totally addicted to animal foods will find alternatives that are inherently attractive to eat, so much so that McDonald's will market them, too. I want to recruit the world's most creative chefs—here's a REAL creative challenge!

I've talked with a lot of smart people who are very keen on it actually. They say, “You have no chance of success, but I really hope you're successful.” That's just the kind of project I love.

Pat, the world desperately needs nutters like you. Let's just hope that the thermodynamics are in your favour once more.

Follow me @glynmoody on Twitter or identi.ca.

No Patents for Circuits? Since You Insist...

I love this argument:

Arguments against software patents have a fundamental flaw. As any electrical engineer knows, solutions to problems implemented in software can also be realized in hardware, i.e., electronic circuits. The main reason for choosing a software solution is the ease in implementing changes, the main reason for choosing a hardware solution is speed of processing. Therefore, a time critical solution is more likely to be implemented in hardware. While a solution that requires the ability to add features easily will be implemented in software. As a result, to be intellectually consistent those people against software patents also have to be against patents for electronic circuits.

People seem to think that this is an invincible argument *for* software patents; what the poor darlings fail to notice is that it's actually an invincible argument *against* patents for circuits.

Since software is just algorithms, which is just maths, which cannot be patented, and this clever chap points out that circuits are just software made out of hardware, it follows that we shouldn't allow patents for circuits (but they can still be protected by copyright, just as software can.)

So, thanks for the help in rolling back what is patentable...

21 July 2009

Has Google Forgotten Celera?

One of the reasons I wrote my book Digital Code of Life was that the battle between the public Human Genome Project and the privately-funded Celera mirrored so closely the battle between free software and Microsoft - with the difference that it was our genome that was at stake, not just a bunch of bits. The fact that Celera ultimately failed in its attempt to sequence and patent vast chunks of our DNA was the happiest of endings.

It seems someone else knows the story:

Celera was the company founded by Craig Venter, and funded by Perkin Elmer, which played a large part in sequencing the human genome and was hoping to make a massively profitable business out of selling subscriptions to genome databases. The business plan unravelled within a year or two of the publication of the first human genome. With hindsight, the opponents of Celera were right. Science is making and will make much greater progress with open data sets.

Here are some rea[s]ons for thinking that Google will be making the same sort of mistake as Celera if it pursues the business model outlined in its pending settlement with the AAP and the Author's Guild....

Thought provoking stuff, well worth a read.

Follow me @glynmoody on Twitter or identi.ca.

Building on Open Data

One of the great things about openness is that it lets people do incredible things by adding to it in a multiplicity of ways. The beatuy is that those releasing material don't need to try to anticipate future uses: it's enough that they make it as open as possible Indeed, the more open they make it, the more exciting the re-uses will be.

Here's an unusual example from the field of open data, specifically, the US government data held on Data.gov:


The purpose of Data.gov is to increase public access to high value, machine readable datasets generated by the Executive Branch of the Federal Government. Although the initial launch of Data.gov provides a limited portion of the rich variety of Federal datasets presently available, we invite you to actively participate in shaping the future of Data.gov by suggesting additional datasets and site enhancements to provide seamless access and use of your Federal data. Visit today with us, but come back often. With your help, Data.gov will continue to grow and change in the weeks, months, and years ahead.


Here's how someone intends to go even further:

Today I’m happy to announce Sunlight Labs is stealing an idea from our government. Data.gov is an incredible concept, and the implementation of it has been remarkable. We’re going to steal that idea and make it better. Because of politics and scale there’s only so much the government is going to be able to do. There are legal hurdles and boundaries the government can’t cross that we can. For instance: there’s no legislative or judicial branch data inside Data.gov and while Data.gov links off to state data catalogs, entries aren’t in the same place or format as the rest of the catalog. Community documentation and collaboration are virtual impossibilities because of the regulations that impact the way Government interacts with people on the web.

We think we can add value on top of things like Data.gov and the municipal data catalogs by autonomously bringing them into one system, manually curating and adding other data sources and providing features that, well, Government just can’t do. There’ll be community participation so that people can submit their own data sources, and we’ll also catalog non-commercial data that is derivative of government data like OpenSecrets. We’ll make it so that people can create their own documentation for much of the undocumented data that government puts out and link to external projects that work with the data being provided.

This the future.

Follow me @glynmoody on Twitter or identi.ca.

20 July 2009

British Library Turns Traitor

I knew the British Library was losing its way, but this is ridiculous:

The British Library Business & IP Centre at St Pancras, London can help you start, run and grow your business.

And how might it do that?

Intellectual property can help you protect your ideas and make money from them.

Our resources and workshops will guide you through the four types of intellectual property: patents, trade marks, registered designs and copyright.

This once-great institution used to be about opening up the world's knowledge for the benefit and enjoyment of all: today, it's about closing it down so that only those who can afford to pay get to see it.

What an utter disgrace.

Follow me @glynmoody on Twitter or identi.ca.

Patents *Are* Monopolies: It's Official

As long-suffering readers of this blog will know, I refer to patents and copyrights as intellectual monopolies because, well, that's what they are. But there are some who refuse to accept this, citing all kinds of specious reasons why it's not correct.

Well, here's someone else who agrees with me:


A further and more significant change may come from the President's nomination of David Kappos of IBM to be the next Director of the Patent Office. While in the past, IBM was a prolific filer of patent applications, many of them covering business methods and software, it has filed an amicus brief in Bilski opposing the patentability of business method patents. However, and perhaps not surprisingly, IBM defends approval of software patents.

Mr. Kappos announced his opposition to business method patents last year by stating that "[y]ou're creating a new 20-year monopoly for no good reason."

Yup: the next Director of the USPTO says patents are monopolies: it's official. (Via @schestowitz.)

Follow me @glynmoody on Twitter or identi.ca.

17 July 2009

Harvard University Press on Scribd

This sounds like a great move:


It’s a recession. Save the $200,000 you were going to spend on that Harvard education and check out some of the books Harvard University Press is selling on Scribd starting today.

It's so obviously right: saving trees, and making academic materials more readily available, thus boosting the recognition achieved by the authors - exactly what they want. Plus readers can buy the books much more cheaply, allowing many more people to access rigorous if rather specialist knowledge.

Or maybe not: a quick look through the titles shows prices ranging from mid-teens up to $45. Come on, people, these are *electrons*: they are cheap. The whole point is to use this fact to spread knowledge, reputation and joy more widely.

When will they (HUP) learn?

Follow me @glynmoody on Twitter or identi.ca.

Gadzooks - it's ZookZ from Antigua

I've been following the rather entertaining case of Antigua vs. US for a few years now. Basically, the US government has taken a "do as I say, not as I do" attitude to the WTO - refusing to follow the latter's rules while seeking to enforce them against others. The net result is that plucky little Antigua seems to have won some kind of permission to ignore US copyright - up to a certain point - although nobody really knows what this means in practice.

That's not stopping an equally cheeky Antigua-based company from trying to make money from this situation:

ZookZ provides a new way to get pure movie and music enjoyment. We deliver unlimited, high-quality movies and music through a safe, legal and secure platform for one low monthly subscription fee.

ZookZ makes it simple for any to enjoy digital entertainment. Our user-friendly interface provides access to all our digital assets. We offer unlimited downloads of all movies and music for one low monthly price. Files are delivered in MP3 and MP4 formats that are compatible with most mobile devices and players so you can enjoy your entertainment when, where and how you want. ZookZ is changing the way people use and enjoy digital entertainment. Unlike other companies, once you download the file, you can view or listen to it on any medium of your choice­ –without restrictions.

ZookZ is not a peer-to-peer file sharing system and prohibits that use of its product. Customers directly download safe content from our secure database, not from an unknown third party. ZookZ guarantees that all our digital media is free from viruses, adware and spyware. We are dedicated to providing high-quality, safe and secure digital files.

ZookZ operates under the parameters of the 2007 WTO ruling between Antigua and the United States, and is the only website that can legally offer members unlimited digital entertainment.


The FAQ has more details.

I doubt whether the US media industries will sit back and let ZookZ try to implement its plan, and I suspect that this could get rather interesting to watch.

16 July 2009

Why Most Newspapers are Dying

This is something that's struck me too:

as is oh-so-typical in these situations, Osnos does nothing at all to engage or respond to the comments that call out his mistakes. You want to know why newspapers are failing? It's not because of Google, it's because of this viewpoint that some journalists still hold that they're the masters of the truth, handing it out from on high, wanting nothing at all to do with the riff raff in the comments.

This is perhaps the biggest single clue that newspaper do not understand how the Internet has changed relationships between writers and readers. Indeed, one of my disappointments with the Guardian's Comment is Free site is that practically *never* do the writers deign to interact with their readers. Given that the Guardian is probably the most Web-savvy of the major newspapers, this does not augur well...

(Open) Learning from Open Source

As regular readers of this blog will know, I am intrigued by the way that ideas from free software are moving across to different disciplines. Of course, applying them is no simple matter: there may not be an obvious one-to-one mapping of the act of coding to activities in the new domain, or there may be significant cultural differences that place obstacles in the way of sharing.

Here's a fascinating post that explores some of the issues around the application of open source ideas to open educational resources (OER):


For all my fascination with all things open-source, I'm finding that the notion of open source software (OSS) is one that's used far too broadly, to cover more categories than it can rightfully manage. Specifically, the use of this term to describe collaborative open education resource (OER) projects seems problematic. The notion of OSS points to a series of characteristics and truths that do not apply, for better or worse, to the features of collaborative learning environments developed for opening up education.

While in general, open educational resources are developed to adhere to the letter of the OSS movement, what they miss is what we might call the spirit of OSS, which for my money encompasses the following:

* A reliance on people's willingness to donate labor--for love, and not for money.

* An embrace of the "failure for free" model identified by Clay Shirky in Here Comes Everybody.

* A loose collaboration across fields, disciplines, and interest levels.

Open educational resources are not, in general, developed by volunteers; they are more often the product of extensive funding mechanisms that include paying participants for their labor.

Unusually, the post does not simply lament this problem, but goes on to explore a possible solution:

an alternate term for OERs designed in keeping with the open source ideals: community source software (CSS)

Worth reading in its entirety, particularly for the light it sheds on things we take for granted in open source.

Follow me @glynmoody on Twitter or identi.ca.

Now You Too Can Contribute to Firefox...

...with your money:


This pilot allows developers to request an optional dollar amount for their Firefox Add-on. Along with requesting this amount, we’re helping developers tell their stories with our new “About the Developer” pages, which explain to prospective contributors the motivations for creating an add-on and its future road map. Since contributions are completely optional, users will have ample time to evaluate an add-on to determine whether or not they want to help a developer.

Some details:

How will payments work?

We are working with PayPal on this pilot to provide a secure and international solution for facilitating payments. Developers can optionally create a PayPal ID for each of their Firefox Add-ons. Users will be presented with a “Contribute” button that gives them the option of paying the suggested amount or a different amount.

This is a nice touch, too:

Why did you call this “Contributions” and not “Donations”?

At Mozilla, we use the word “Contributor” for community members who contribute time and energy to our mission of promoting choice and innovation on the Internet. Our goal is that users who contribute money to developers are supporting the future of a particular add-on, as opposed to donating for something already received.

Quite: this isn't just about getting some well-deserved dosh to the coders, but also about giving users a way to feel more engaged. Great move.

15 July 2009

Bill Gates Gets Sharing...Almost

Yesterday I wrote about Microsoft's attempt to persuade scientists to adopt its unloved Windows HPC platform by throwing in a few free (as in beer) programs. Here's another poisoned chalice that's being offered:

In between trying to eradicate polio, tame malaria, and fix the broken U.S. education system, Gates has managed to fulfill a dream of taking some classic physics lectures and making them available free over the Web. The lectures, done in 1964 by noted scientist (and Manhattan Project collaborator) Richard Feynman, take notions such as gravity and explains how they work and the broad implications they have in understanding the ways of the universe.

Gates first saw the series of lectures 20 years ago on vacation and dreamed of being able to make them broadly available. After spending years tracking down the rights--and spending some of his personal fortune--Gates has done just that. Tapping his colleagues in Redmond to create interactive software to accompany the videos, Gates is making the collection available free from the Microsoft Research Web site.

What a kind bloke - spending his *own* personal fortune of uncountable billions, just to make this stuff freely available.

But wait: what do we find when go to that "free" site:

Clicking will install Microsoft Silverlight.

So it seems that this particular free has its own non-free (as in freedom) payload: what a surprise.

That's a disappointment - but hardly unexpected; Microsoft's mantra is that you don't get something for nothing. But elsewhere in the interview with Gates, there's some rather interesting stuff:

Education, particularly if you've got motivated students, the idea of specializing in the brilliant lecture and text being done in a very high-quality way, and shared by everyone, and then the sort of lab and discussion piece that's a different thing that you pick people who are very good at that.

Technology brings more to the lecture availability, in terms of sharing best practices and letting somebody have more resources to do amazing lectures. So, you'd hope that some schools would be open minded to this fitting in, and making them more effective.

What's interesting is that his new-found work in the field of education is bringing him constantly face-to-face with the fact that sharing is actually rather a good thing, and that the more the sharing of knowledge can be facilitated, the more good results.

Of course, he's still trapped by the old Microsoft mindset, and constantly thinking how he can exploit that sharing, in this case by freighting it with all kinds of Microsoft gunk. But at least he's started on the journey, albeit unknowingly.

Follow me @glynmoody on Twitter or identi.ca.

14 July 2009

Hamburg Declaration = Humbug Declaration

You may have noticed that in the 10 years since Napster, the music industry has succeeded in almost completely ruining its biggest opportunity to make huge quantities of money, alienating just about anyone under 30 along the way (and a fair number of us old fogies, too).

Alas, it seems that some parts of the newspaper industry have been doing their job of reporting so badly that they missed that particular news item. For what does it want to do? Follow the music industry's lemming-like plunge off the cliff of "new intellectual property rights protection":

On the day that Commissioner Viviane Reding unveils her strategy for a Digital Europe during the Lisbon Council, and as the European Commission's consultation on the Content Online Report draws to a close this week, senior members of the publishing world are presenting to Information Society Commissioner Viviane Reding and Internal Market Commissioner Charlie McCreevy, a landmark declaration adopted on intellectual property rights in the digital world in a bid to ensure that opportunities for a diverse, free press and quality journalism thrive online into the future.

This is the first press communiqué on a significant meeting convened on 26th June in Berlin by news group Chief Executives from both the EPC and the World Association of Newspapers where the 'Hamburg Declaration' was signed, calling for online copyright to be respected, to allow innovation to thrive and consumers to be better served.

This comes from an extraordinary press release, combining arrogant self-satisfaction with total ignorance about how the Internet works:

A fundamental safeguard of democratic society is a free, diverse and independent press. Without control over our intellectual property rights, the future of quality journalism is at stake and with it our ability to provide our consumers with quality and varied information, education and entertainment on the many platforms they enjoy.

What a load of codswallop. What makes them think they are the sole guardians of that "free, diverse and independent press"? In case they hadn't noticed, the Internet is rather full of "quality and varied information, education and entertainment on the many platforms", most of it quite independent of anything so dull as a newspaper. As many others have pointed out, quality journalism is quite separate from old-style press empires, even if the latter have managed to produce the former from time to time.

Then there's this:

We continue to attract ever greater audiences for our content but, unlike in the print or TV business models, we are not the ones making the money out of our content. This is unsustainable.

Well, at least they got the last bit. But if they are attracting "ever greater audiences" for their content, but are not making money, does this not suggest that they are doing something fundamentally wrong? In a former incarnation, I too was a publisher. When things went badly, I did not immediately call for new laws: I tried again with something different. How about if newspaper publishers did the same?

This kind of self-pitying bleating would be extraordinary enough were it coming out of a vacuum; but given the decade of exemplary failure by the music industry taking *exactly* the same approach, it suggests a wilful refusal to look reality in the face that is quite extraordinary.

Speaking personally, the sooner all supporters of the Humbug Declaration are simply omitted from every search engine on Earth, the better: I'm sure we won't miss them, but they sure will miss the Internet...

Follow me @glynmoody on Twitter or identi.ca.

I Fear Microsoft Geeks Bearing Gifts...

Look, those nice people at Microsoft Research are saving science from its data deluge:

Addressing an audience of prominent academic researchers today at the 10th annual Microsoft Research Faculty Summit, Microsoft External Research Corporate Vice President Tony Hey announced that Microsoft Corp. has developed new software tools with the potential to transform the way much scientific research is done. Project Trident: A Scientific Workflow Workbench allows scientists to easily work with large volumes of data, and the specialized new programs Dryad and DryadLINQ facilitate the use of high-performance computing.

Created as part of the company’s ongoing efforts to advance the state of the art in science and help address world-scale challenges, the new tools are designed to make it easier for scientists to ingest and make sense of data, get answers to questions at a rate not previously possible, and ultimately accelerate the pace of achieving critical breakthrough discoveries. Scientists in data-intensive fields such as oceanography, astronomy, environmental science and medical research can now use these tools to manage, integrate and visualize volumes of information. The tools are available as no-cost downloads to academic researchers and scientists at http://research.microsoft.com/en-us/collaboration/tools.

Aw, shucks, isn't that just *so* kind? Doing all this out of the goodness of their hearts? Or maybe not:

Project Trident was developed by Microsoft Research’s External Research Division specifically to support the scientific community. Project Trident is implemented on top of Microsoft’s Windows Workflow Foundation, using the existing functionality of a commercial workflow engine based on Microsoft SQL Server and Windows HPC Server cluster technologies. DryadLINQ is a combination of the Dryad infrastructure for running parallel systems, developed in the Microsoft Research Silicon Valley lab, and the Language-Integrated Query (LINQ) extensions to the C# programming language.

So basically Project Trident is more Project Trojan Horse - an attempt to get Microsoft HPC Server cluster technologies into the scientific community without anyone noticing. And why might Microsoft be so keen to do that? Maybe something to do with the fact that Windows currently runs just 1% of the top 500 supercomputing sites, while GNU/Linux has over 88% share.

Microsoft's approach here can be summed up as: accept our free dog biscuit, and be lumbered with a dog.

Follow me @glynmoody on Twitter or identi.ca.

Batik-Makers Say "Tidak" to Copyright

Yesterday I was talking about how patents are used to propagate Western ideas and power; here's a complementary story about local artists understanding that copyright just ain't right for them:


Joko, speaking at this year’s Solo Batik Fashion Festival over the weekend, said that the ancient royal city was one of the principal batik cities in Indonesia, with no fewer than 500 unique motifs created here that are not found in any other region. The inventory process, however, was hampered by the reluctance of the batik makers to claim ownership over pieces.

The head of the Solo trade and industry office, Joko Pangarso, said copyright registration work had begun last year, but was constantly held up when it was found a particular batik only had a motif name because the creator declined to attach their own.

“So far only 10 motifs have been successfully included in the list,” he said. “The creators acknowledged their creations but asked for minimal exposure.

Interestingly, this is very close to the situation for software. The batik motifs correspond to sub-routines: both are part of the commons that everyone draws upon; copyrighting those patterns is as counter-productive as patenting subroutines, since it makes further creation almost impossible without "infringement". This reduces the overall creativity - precisely the opposite effect that intellectual monopolists claim. (Via Boing Boing.)

Follow me @glynmoody on Twitter or identi.ca.

13 July 2009

National Portrait Gallery: Nuts

This is so wrong:

Below is a letter I received from legal representatives of the National Portrait Gallery, London, on Friday, July 10, regarding images of public domain paintings in Category:National Portrait Gallery, London and threatening direct legal action under UK law. The letter is reproduced here to enable public discourse on the issue. For a list of sites discussing this event see User:Dcoetzee/NPG legal threat/Coverage. I am consulting legal representation and have not yet taken action.

Look, NPG, your job is to get people to look at your pix. Here's some news: unless they're in London, they can't do that. Put those pix online, and (a) that get to see the pix and (b) when they're in London, they're more likely to come and visit, no?

So you should be *encouraging* people to upload your pix to places like Wikipedia; you should be thanking them. The fact that you are threatening them with legal action shows that you don't have even an inkling of what you are employed to do.

Remind me not to pay the part of my UK taxes that goes towards your salary....

Are Patents Intellectual Monopolies? You Decide

Talking of intellectual monopolies, you may wonder why I use this term (well, if you've been reading this blog for long, you probably don't.) But in any case, here's an excellent exposition as to why, yes, patents are indeed monopolies:

On occasion you get some defender of patents who is upset when we use the m-word to describe these artificial state-granted monopoly rights. For example here one Dale Halling, a patent attorney (surprise!) posts about "The Myth that Patents are a Monopoly" and writes, " People who suggest a patent is a monopoly are not being intellectually honest and perpetuating a myth to advance a political agenda."

Well, let's see.

Indeed, do read the rest of yet another great post from the Against Monopoly site.

Follow me @glynmoody on Twitter or identi.ca.

What Are Intellectual Monopolies For?

If you still doubted that intellectual monopolies are in part a neo-colonialist plot to ensure the continuing dominance of Western nations, you could read this utterly extraordinary post, which begins:

The fourteenth session of the WIPO Intergovernmental Committee on Genetic Resources, Traditional Knowledge and Folklore (IGC), convened in Geneva from June 29, 2009 to July 3, 2009, collapsed at the 11th hour on Friday evening as the culmination of nine years of work over fourteen sessions resulted in the following language; “[t]he Committee did not reach a decision on this agenda item” on future work. The WIPO General Assembly (September 2009) will have to untangle the intractable Gordian knot regarding the future direction of the Committee.

At the heart of the discussion lay a proposal by the African Group which called for the IGC to submit a text to the 2011 General Assembly containing “a/(n) international legally binding instrument/instruments” to protect traditional cultural expressions (folklore), traditional knowledge and genetic resources. Inextricably linked to the legally binding instruments were the African Group’s demands for “text-based negotiations” with clear “timeframes” for the proposed program of work. This proposal garnered broad support among a group of developing countries including Malaysia, Thailand, Fiji, Bolivia, Brazil, Ecuador, Philippines, Sri Lanka, Cuba, Yemen India, Peru, Guatemala, China, Nepal and Azerbaijan. Indonesia, Iran and Pakistan co-sponsored the African Group proposal.

The European Union, South Korea and the United States could not accept the two principles of “text-based negotiations” and “internationally legally binding instruments”.

Australia, Canada and New Zealand accepted the idea of “text-based negotiations” but had reservations about “legally binding instruments” granting sui generis protection for genetic resources, traditional knowledge and folklore.

We can't possibly have dveloping countries protecting their traditional medicine and national lore - "genetic resources, traditional knowledge and folklore" - from being taken and patented by the Western world. After all, companies in the latter have an inalienable right to turn a profit by licensing that same traditional knowledge it back to the countries it was stolen from (this has already happened). That's what intellectual monopolies are for.

Follow me @glynmoody on Twitter or identi.ca.

10 July 2009

This Could Save Many Lives: Let's Patent It

Bill Gates is amazing; just look at this brilliant idea he's come up with:

using large fleets of vessels to suppress hurricanes through various methods of mixing warm water from the surface of the ocean with colder water at greater depths. The idea is to decrease the surface temperature, reducing or eliminating the heat-driven condensation that fuels the giant storms.

Against the background of climate change and increased heating of the ocean's surface in areas where hurricanes emerge, just imagine how many lives this could save - a real boon for mankind. Fantastic.

Just one problemette: he's decided to patent the idea, along with his clever old chum Nathan Myhrvold.

The filings were made by Searete LLC, an entity tied to Intellectual Ventures, the Bellevue-based patent and invention house run by Nathan Myhrvold, the former Microsoft chief technology officer. Myhrvold and several others are listed along with Gates as inventors.

After all, can't have people just going out there and saving thousands of lives without paying for the privilege, can we?

Follow me @glynmoody on Twitter or identi.ca.

Do We Need Open Access Journals?

One of the key forerunners of the open access idea was arxiv.org, set up by Paul Ginsparg. Here's what I wrote a few years back about that event:

At the beginning of the 1990s, Ginsparg wanted a quick and dirty solution to the problem of putting high-energy physics preprints (early versions of papers) online. As it turns out, he set up what became the arXiv.org preprint repository on 16 August, 1991 – nine days before Linus made his fateful “I'm doing a (free) operating system (just a hobby, won't be big and professional like gnu) for 386(486) AT clones” posting. But Ginsparg's links with the free software world go back much further.

Ginsparg was already familiar with the GNU manifesto in 1985, and, through his brother, an MIT undergraduate, even knew of Stallman in the 1970s. Although arXiv.org only switched to GNU/Linux in 1997, it has been using Perl since 1994, and Apache since it came into existence. One of Apache's founders, Rob Hartill, worked for Ginsparg at the Los Alamos National Laboratory, where arXiv.org was first set up (as an FTP/email server at xxx.lanl.org). Other open source programs crucial to arXiv.org include TeX, GhostScript and MySQL.

arxiv.org was and is a huge success, and that paved the way for what became the open access movement. But here's an interesting paper - hosted on arxiv.org:

Contemporary scholarly discourse follows many alternative routes in addition to the three-century old tradition of publication in peer-reviewed journals. The field of High- Energy Physics (HEP) has explored alternative communication strategies for decades, initially via the mass mailing of paper copies of preliminary manuscripts, then via the inception of the first online repositories and digital libraries.

This field is uniquely placed to answer recurrent questions raised by the current trends in scholarly communication: is there an advantage for scientists to make their work available through repositories, often in preliminary form? Is there an advantage to publishing in Open Access journals? Do scientists still read journals or do they use digital repositories?

The analysis of citation data demonstrates that free and immediate online dissemination of preprints creates an immense citation advantage in HEP, whereas publication in Open Access journals presents no discernible advantage. In addition, the analysis of clickstreams in the leading digital library of the field shows that HEP scientists seldom read journals, preferring preprints instead.

Here are the article's conclusions:

Scholarly communication is at a cross road of new technologies and publishing models. The analysis of almost two decades of use of preprints and repositories in the HEP community provides unique evidence to inform the Open Access debate, through four main findings:

1. Submission of articles to an Open Access subject repository, arXiv, yields a citation advantage of a factor five.

2. The citation advantage of articles appearing in a repository is connected to their dissemination prior to publication, 20% of citations of HEP articles over a two-year period occur before publication.

3. There is no discernable citation advantage added by publishing articles in “gold” Open Access journals.

4. HEP scientists are between four and eight times more likely to download an article in its preprint form from arXiv rather than its final published version on a journal web site.

On the one hand, it would be ironic if the very field that acted as a midwife to open access journals should also be the one that begins to undermine it through a move to repository-based open publishing of preprints. On the other, it doesn't really matter; what's important is open access to the papers. If these are in preprint form, or appear as fully-fledged articles in peer-reviewed open access journals is a detail, for the users at least; it's more of a challenge for publishers, of course... (Via @JuliuzBeezer.)

Follow me @glynmoody on Twitter or identi.ca.

08 July 2009

Not Kissing the Rod, Oh My Word, No

Becta today [6 July 2009] welcomes Microsoft's launch of the new Subscription Enrolment Schools Pilot (SESP) for UK schools, which provides greater flexibility and choice for schools who wish to use a Microsoft subscription agreement.

Great, and what might that mean, exactly?

The new licensing scheme removes the requirement that schools using subscription agreements pay Microsoft to licence systems that are using their competitor's technologies. So for the first time schools using Microsoft's subscription licensing agreements can decide for themselves how much of their ICT estate to licence.

So BECTA is celebrating that fact that schools - that is, we taxpayers - *no longer* have to "pay Microsoft to licence systems that are using their competitor's technologies"? They can now use GNU/Linux, for example, *without* having to pay Microsoft for the privilege?

O frabjous day! Callooh! Callay!

Follow me @glynmoody on Twitter or identi.ca.

Policing the Function Creep...

Remember how the poor darlings in the UK government absolutely *had to* allow interception of all our online activities so that those plucky PC Plods could maintain their current stunning success rate in their Whirr on Terruh and stuff like that? Well, it seems that things have changed somewhat:

Detectives will be required to consider accessing telephone and internet records during every investigation under new plans to increase police use of communications data.

The policy is likely to significantly increase the number of requests for data received by ISPs and telephone operators.

Just as every investigation currently has to include a strategy to make use of its subjects' financial records, soon CID officers will be trained to always draw up a plan to probe their communications.

The plans have been developed by senior officers in anticipation of the implementation of the Interception Modernisation Programme (IMP), the government's multibillion pound scheme to massively increase surveillance of the internet by storing details of who contacts whom online.

Er, come again? "CID officers will be trained to always draw up a plan to probe their communications"? How does that square with this being a special tool for those exceptional cases when those scary terrorists and real hard naughty criminals are using tricky high-tech stuff like email? Doesn't it imply that we are all terrorist suspects and hard 'uns now?

Police moves to prepare for the glut of newly accessible data were revealed today by Deputy Assistant Commissioner Janet Williams. She predicted always considering communications data will lead to a 20 per cent increase in the productivity of CID teams.

She told The Register IMP had "informed thinking" about use of communications data, but denied the plans gave the lie to the government line that massively increased data retention will "maintain capability" of law enforcement to investigate crime.

Well, Mandy Rice-Davies applies, m'lud...

Follow me @glynmoody on Twitter or identi.ca.

07 July 2009

Are Microsoft's Promises For Ever?

This sounds good:

I have some good news to announce: Microsoft will be applying the Community Promise to the ECMA 334 and ECMA 335 specs.

ECMA 334 specifies the form and establishes the interpretation of programs written in the C# programming language, while the ECMA 335 standard defines the Common Language Infrastructure (CLI) in which applications written in multiple high-level languages can be executed in different system environments without the need to rewrite those applications to take into consideration the unique characteristics of those environments.

"The Community Promise is an excellent vehicle and, in this situation, ensures the best balance of interoperability and flexibility for developers," Scott Guthrie, the Corporate Vice President for the .Net Developer Platform, told me July 6.

It is important to note that, under the Community Promise, anyone can freely implement these specifications with their technology, code, and solutions.

You do not need to sign a license agreement, or otherwise communicate to Microsoft how you will implement the specifications.

The Promise applies to developers, distributors, and users of Covered Implementations without regard to the development model that created the implementations, the type of copyright licenses under which it is distributed, or the associated business model.

Under the Community Promise, Microsoft provides assurance that it will not assert its Necessary Claims against anyone who makes, uses, sells, offers for sale, imports, or distributes any Covered Implementation under any type of development or distribution model, including open-source licensing models such as the LGPL or GPL.

But boring old sceptic that I am, I have memories of this:

The Software Freedom Law Center (SFLC), provider of pro-bono legal services to protect and advance free and open source software, today published a paper that considers the legal implications of Microsoft's Open Specification Promise (OSP) and explains why it should not be relied upon by developers concerned about patent risk.

SFLC published the paper in response to questions from its clients and the community about the OSP and its compatibility with the GNU General Public License (GPL). The paper says that the promise should not be relied upon because of Microsoft's ability to revoke the promise for future versions of specifications, the promise's limited scope, and its incompatibility with free software licenses, including the GPL.

That was then, of course, what about now? Well, here's what the FAQ says on the subject:

Q: Does this CP apply to all versions of the specification, including future revisions?

A: The Community Promise applies to all existing versions of the specifications designated on the public list posted at /interop/cp/, unless otherwise noted with respect to a particular specification.


Now, is it just me, or does Microsoft conspicuously fail to answer its own question? The question was: does it apply to all versions *including* future revision? And Microsoft's answer is about *existing* versions: so doesn't that mean it could simply not apply the promise to a future version? Isn't this the same problem as with the Open Specification Promise? Just asking.

03 July 2009

The Engine of Scientific Progress: Sharing

Here's a post saying pretty much what I've been saying, but in a rather different way:


Here we present a simple model of one of the most basic uses of results, namely as the engine of scientific progress. Research results are more than just accumulated knowledge. Research results make possible new questions, which in turn lead to even more knowledge. The resulting pattern of exponential growth in knowledge is called an issue tree. It shows how individual results can have a value far beyond themselves, because they are shared and lead to research by others.


Follow me @glynmoody on Twitter or identi.ca.

02 July 2009

Patents Don't Promote Innovation: Study

It's extraordinary how the myth that patents somehow promote innovation is still propagated and widely accepted; and yet there is practically *no* empirical evidence that it's true. All the studies that have looked at this area rigorously come to quite a different conclusion. Here's yet another nail in that coffin, using a very novel approach:

Patent systems are often justified by an assumption that innovation will be spurred by the prospect of patent protection, leading to the accrual of greater societal benefits than would be possible under non-patent systems. However, little empirical evidence exists to support this assumption. One way to test the hypothesis that a patent system promotes innovation is to simulate the behavior of inventors and competitors experimentally under conditions approximating patent and non-patent systems.

Employing a multi-user interactive simulation of patent and non-patent (commons and open source) systems ("PatentSim"), this study compares rates of innovation, productivity, and societal utility. PatentSim uses an abstracted and cumulative model of the invention process, a database of potential innovations, an interactive interface that allows users to invent, patent, or open source these innovations, and a network over which users may interact with one another to license, assign, buy, infringe, and enforce patents.

Data generated thus far using PatentSim suggest that a system combining patent and open source protection for inventions (that is, similar to modern patent systems) generates significantly lower rates of innovation (p<0.05), productivity (p<0.001), and societal utility (p<0.002) than does a commons system. These data also indicate that there is no statistical difference in innovation, productivity, or societal utility between a pure patent system and a system combining patent and open source protection.

The results of this study are inconsistent with the orthodox justification for patent systems. However, they do accord well with evidence from the increasingly important field of user and open innovation. Simulation games of the patent system could even provide a more effective means of fulfilling the Constitutional mandate ― "to promote the Progress of . . . useful Art" than does the orthodox assumption that technological innovation can be encouraged through the prospect of patent protection.

When will people get the message and start sharing for mutual benefit?

Follow me @glynmoody on Twitter or identi.ca.

01 July 2009

Help Me Go Mano a Mano with Microsoft

Next week, I'm taking part in a debate with a Microsoft representative about the passage of the OOXML file format through the ISO process last year. Since said Microsoftie can draw on the not inconsiderable resources of his organisation to provide him with a little back-up, I thought I'd try to even the odds by putting out a call for help to the unmatched resource that is the Linux Journal community. Here's the background to the meeting, and the kind of info I hope people might be able to provide....

On Linux Journal.

30 June 2009

Why Scientific Publishing Will Never be the Same

For those of us tracking open access and its wider import, it's pretty clear that scientific publishing has changed for ever. But for some within the industry, there remains the desperate hope that all this new-fangled open, collaborative stuff will just blow over.

Forget about it: as this magisterial analysis shows, not only is there no way for traditional publishing to get from there to here - it's a terrifying leap across a minimum to the next maximum - but the most exciting stuff is *already* happening in other forms of publishing:

What’s new today is the flourishing of an ecosystem of startups that are experimenting with new ways of communicating research, some radically different to conventional journals. Consider Chemspider, the excellent online database of more than 20 million molecules, recently acquired by the Royal Society of Chemistry. Consider Mendeley, a platform for managing, filtering and searching scientific papers, with backing from some of the people involved in Last.fm and Skype. Or consider startups like SciVee (YouTube for scientists), the Public Library of Science, the Journal of Visualized Experiments, vibrant community sites like OpenWetWare and the Alzheimer Research Forum, and dozens more. And then there are companies like Wordpress, Friendfeed, and Wikimedia, that weren’t started with science in mind, but which are increasingly helping scientists communicate their research. This flourishing ecosystem is not too dissimilar from the sudden flourishing of online news services we saw over the period 2000 to 2005.

It’s easy to miss the impact of blogs on research, because most science blogs focus on outreach. But more and more blogs contain high quality research content. Look at Terry Tao’s wonderful series of posts explaining one of the biggest breakthroughs in recent mathematical history, the proof of the Poincare conjecture. Or Tim Gowers recent experiment in “massively collaborative mathematics”, using open source principles to successfully attack a significant mathematical problem. Or Richard Lipton’s excellent series of posts exploring his ideas for solving a major problem in computer science, namely, finding a fast algorithm for factoring large numbers. Scientific publishers should be terrified that some of the world’s best scientists, people at or near their research peak, people whose time is at a premium, are spending hundreds of hours each year creating original research content for their blogs, content that in many cases would be difficult or impossible to publish in a conventional journal. What we’re seeing here is a spectacular expansion in the range of the blog medium. By comparison, the journals are standing still.

What's even better about this piece is that it's not content to point out why traditional publishing has big problems: it also offers some practical suggestions of what people *should* be looking at:

These opportunities can still be grasped by scientific publishers who are willing to let go and become technology-driven, even when that threatens to extinguish their old way of doing things. And, as we’ve seen, these opportunites are and will be grasped by bold entrepreneurs. Here’s a list of services I expect to see developed over the next few years. A few of these ideas are already under development, mostly by startups, but have yet to reach the quality level needed to become ubiquitous. The list could easily be continued ad nauseum - these are just a few of the more obvious things to do.

Fantastic stuff - do read it all if you have time.

Follow me @glynmoody on Twitter or identi.ca.

Squishing the Media Bugs

Belatedly, I'm writing about this very cool idea:


It’s called MediaBugs.org. And the idea is to create a web site, a web service, that people in a community, in this case the San Francisco Bay Area, can bring problems and errors that they find in media coverage and post them and try to get them fixed. [...]

The inspiration of the project is from what’s called a bug tracker in an open-source project. So if you’re developing open-source software, you have this project, and you put up a public website that anyone can bring these — file these bugs. If you’re using Firefox and something breaks, you go to their web site, and you tell them about it.

I fear the practice might be harder than the theory, but I certainly wish it well.

Follow me @glynmoody on Twitter or identi.ca.

Winning the Open Web

It seems an unfair fight. On the one hand, you have some of the biggest, most powerful multinationals, intent on defending their turf and extending their power and profits. On the other, you have a tiny number of ragtag idealists who believe that knowledge belongs to everyone, and that no one should have disproportionately long monopolies on its supply.

And yet: in the last few years a remarkable series of victories have been one by the latter against the former, to the extent that representatives of the big media industries have warned that they are losing the "battle".

Against that background of uneven forces - but not quite in the way the media companies mean it - sharing information about past successes so as to drive future ones is crucially important. And yet it is rarely done, probably because the practitioners are too busy fighting the battles to write about it.

Enter Becky Hogge, former Executive Director of the Open Rights Group, who happily has had some time on her hands to prepare a handy report entitled "Winning the Web":


Winning the Web is a 2009 report funded by the Open Society Institute and written by Becky Hogge, former Executive Director of the Open Rights Group. It examined 6 successful campaigns for intellectual property reform, in Brazil, Canada, the US, France, New Zealand, and the UK.

Lessons drawn from the study of the campaigns include the importance of coalition-forming, the best way to conduct online mobilisation campaigns, and the need for a more unified critique of current intellectual property regimes.

The introduction fleshes out the idea:

The global intellectual property regime is no longer fit for purpose. As the networked, digital age matures, it puts into the hands of millions of citizens the tools to access create and share “content”: text, pictures, music and video; data, news, analysis and art. Against this, the intellectual property regime falters. It presents citizens with a choice: stop using the technology – stop communicating, stop creating – or break the law.

Legal reform is presented with two separate challenges. The first is a small but vocal minority of entrenched corporate interests – the rightsholder lobby. Wedded to business models that pre-date the age of networked digital technology, they exploit their position as incumbents to influence legislators. Often representing the world’s biggest multinational corporations, they hijack a narrative that belongs to poor artists struggling in garrets and use the considerable profits they have made from exploiting these artists in the twentieth century to access the corridors of power and make their case.

That legislators listen is related to a second, geopolitical, challenge. Since the 1970s, the developed world has sought to use the global intellectual property regime to ensure its continued prosperity. Motivated by the ability of developing countries to undercut it on the global manufacturing market, it has sought to augment the financial privilege afforded to “knowledge workers”. The self-interest behind this practice is masked by a flawed orthodoxy that is rarely backed up by evidence – that more intellectual property provision is always good for economic growth.

Against this backdrop, a global IP reform movement (also called the access to knowledge movement) is emerging. Motivated by a range of concerns – from global justice, to the narrowing spectrum of permitted speech, to the broadening of surveillance power – these individuals and organisations approach their campaigning work with combined levels of ingenuity and intellectual rigour that make them stand out in the history of fledgling civil rights movements. Recently, these pockets of activism have taken IP reform issues to a wide audience, triggering sweeping civic action in the general population.

For me, the best bits are the detailed case studies of successful campaigns around the world. I knew the bare outlines, but the report really fleshes these out, and then uses them to provide concrete suggestions of what lessons can be learned for the future.

There's one other point is one that I've long thought absolutely crucial:

In the UK, citizens can engage with their elected representatives (including MEPs) using a one-click service called WriteToThem.com. Jim Killock is keen to stress that it is vital that such a tool be developed for all EU member states:

Writetothem.eu is absolutely critical if we want to run these campaigns in the next four years. It shows the contempt in which we seem to hold our European institutions and the irrelevance that they are felt to have across Europe.

This really must be a priority, or else all future campaigns in Europe will suffer as a result.

Follow me @glynmoody on Twitter or identi.ca.

29 June 2009

Watching the Watchers

I read with interest this morning the following:

“Snitchtown” is an essay by Cory Doctorow that first appeared in Forbes.com in June 2007. This SoFoBoMo project is an attempt to illustrate that essay with photographs of some of the 4.2 million CCTV cameras currently estimated to be active in Britain.

It got me thinking: how about setting up a database - a surveillance of surveillance database - that has pictures and locations of CCTVs in the UK? It could be crowd sourced, and anonymous, solving problems of scaling and legal issues. If nothing else, it would put the watchers on notice that they are being watched....

26 June 2009

The World Wins South Korea for Firefox

I've written before about the curious case of South Korea, where the use of Internet Explorer and ActiveX is almost mandatory. I rather despaired of anything changing this situation, since there didn't seem to be any way to get around it from outside. And yet, remarkably, change is coming:

Korea's multifaceted e-government services will be made available for those logging on from FireFox or Safari, web browsers that are gaining more popularity worldwide as an alternative to Internet Explorer.

According to the government Sunday, users of these "non-traditional" browsers will be able to file for year-end tax returns, sign up for a new passport or look for job openings and do much more at various service Web sites operated by the state.

The Ministry of Public Administration and Security, which is in charge of directing e-government initiatives, said that it will invest 11.5 billion won this year for technical projects to increase the browser compatibility of 49 e-government service Web sites.

Starting 2011, all of the 150 e-government Web sites are expected to be accessible from any browser.

That's remarkable, as is the reason for the change:

The development is expected to be useful for overseas Koreans or foreigners logging on to Web sites such as www.hikorea.go.kr from aboard through alternate browsers. Operated by the Ministry of Justice, the Web site is a comprehensive online repository of information for oversea Koreans, immigrants and foreign nationals.

Some civic groups have consistently raised the need to consider expanding e-government services to users of non-traditional browsers.

While the percentage of Koreans users of alternative browsers is still minimal, more netizens worldwide are increasingly surfing the net on browsers other than the Internet Explorer.

A ministry report showed that 21.7 percent of Web users worldwide are browsing on FireFox, and 8 percent on Safari, a browser developed by Apple. IE users make up 67.4 percent of the total Web population.

"We believe that enabling minor browsers to host our e-government services will help overseas Koreans to access the assistance they need and increase Korea's status as a leader in e-government initiatives," a ministry official said.

This shows that what's happening outside a country can still have considerable influence on the internal market - provided there is a big enough expatriate community that still "calls home". Given the increasingly globalised nature of computing, that offers hope for other parts of the world that may be lagging in their uptake of open source.

And if you're wondering why it matters anyway that the South Koreans should be able to use Firefox and other "non-standard" browsers - don't you just love that description? - it's because the country's users have some of the fastest broadband connections in the world; that means that new applications based on such connectivity may well emerge there first, so it's important that open source be available and viable for all kinds of uses. (Via Asa Dotzler.)

Follow me @glynmoody on Twitter or identi.ca.

Show Your Ardour for Ardour

Ardour is a fine open source music program; but like many fine open source programs, it has a problem: money - lack of it. In order to continue to improve the code, the Arbour team ideally needs dosh to pay for programmers and other such handy things; but it's not really happening:


As of now, June 25th, the financial side of things is not looking so good. Last month (May) didn't quite make the $4500 goal, and this month looks certain to fall short by quite a significant margin. There are currently 5-1/2 days left this month, and 28% of the target is still not met. There are no companies backing this project at this time, so its totally incumbent on those of you who use the program and have not yet helped pay to support it to step up and do the right thing. Thanks to everyone who has paid for their contributions and support.

Ardour will continue in some sense even if I find other work, and I believe that Carl, Dave, Hans and others will likely keep up some of their efforts anyway. Since the new download system started, there have been about 9000 OS X downloads and 6000 source code downloads. Less than 3% of the OS X downloads and only three source code downloads were associated with up-front payment, though it seems likely than many users donated after the fact. With a user-base like that, it seems to me that it should be possible to pay one full-time north american developer and to offer occasional payments to others for their outstanding work. What do you think?

Ardour is hardly the only project with these problems, which means that the open source world faces a larger issue: how to raise funds to pay for work that isn't being carried out mostly in bedrooms. It's not something many are thinking about (Matt Asay is an honourable exception), so it's not likely to get solved any time soon - which leaves Ardour in a bit of a pickle. Suggestions and contributions welcome.... (Via Leslie P. Polzer.)

Follow me @glynmoody on Twitter or identi.ca.

Eee, Look: A Useful E-petition Response

Even though I keep signing the wretched things, e-petitions have not generated much action from the UK government. Which makes the following case rather interesting.

The following e-petition:

“We the undersigned petition the Prime Minister to ask the Communities Secretary to require that all software produced by councils under the Timely Information to Citizens project be released under an open source licence.”

Produced this response:

The Government supports the principle that, where new software is being developed by the Timely Information to Citizens pilots, this should wherever possible be released under open source licence and available for use by other local authorities.

For many of the Timely Information to Citizens pilots, the focus is not on new software, but on how existing tools and techniques can be used to bring information together and present it in more useful and accessible ways. Several of the projects will utilise existing open source software to create new information sources and channels, and will share their experiences of doing so with other authorities.

Where the pilots will result in new software tools, ownership and intellectual property rights will usually remain with the individual local authorities. However, most of the authorities concerned have already made a commitment to make these tools available as open source software, or for use by their partner organisations, and we are working to secure the commitment of the remaining.

What impresses me is (a) the reasonableness of the response and (b) the fact that the release of government-developed software should be released as open source "in principle". I do believe we're getting there....

Next, Linux Revolutionises...Printers

Here's a new printer from HP:


Last June 22, HP announced its new all-in-one printer, the Photosmart Premium with TouchSmart Web. Aside from cramming a fax machine, copier, scanner, and a printer into one device, run of the mill technology by today's standards, this new printer can actually print straight from the Web using on-device applications fashioned specifically for this purpose.

As the headline to that story makes clear, that's a Linux-based printer: indeed, it's pretty much unthinkable that these innovative approaches could use anything else. Linux's small footprint, speed, customisability and low cost make it ideal - uniquely so. Where would we be without it?

25 June 2009

Authoring Beautiful HTML...

...ain't easy in the open source world, as David Ascher points out in this post:

However, for regular folks, life is not rosy yet in the Open Web world. Authoring beautiful HTML is, even with design and graphics talent, still way, way too hard. I’m writing this using Wordpress 2.8, which has probably some of the best user experience for simple HTML authoring. As Matt Mullenweg (the founder of Wordpress) says, it’s still not good enough. As far as I can tell, there are currently no truly modern, easy to use, open source HTML composition tools that we could use in Thunderbird for example to give people who want to design wholly original, designed email messages. That’s a minor problem in the world of email, which is primarily about function, not form, and I think we’ll be able to go pretty far with templates, but it’s a big problem for making design on the web more approachable.

There are some valiant efforts to clean up the old, crufty, scary composer codebase that Mozilla has relied on for years. There are simple blog-style editors like FCKEditor and its successor CKEditor. There are in-the-browser composition tools like Google Pages or Google Docs, but those are only for use by Google apps, and only work well when they limit the scope of the design space substantially (again, a rational choice). None of these can provide the flexibility that Ventura Publisher or PageMaker had in the dark ages; none of them can compete from a learnability point of view with the authoring tools that rely on closed stacks; none of them allow the essential polish that hand-crafted code can yield. That’s a gap, and an opportunity.

Let's hope people in the free software world seize it.

Follow me @glynmoody on Twitter or identi.ca.

Crowdsourcing Evil

This was inevitable:

a friend in Iran that I have been in touch with via Skype (which seems to work very well)” told him that a specific Web site, Gerdab.ir, is being used by the Iranian government to identify protesters by crowd-sourcing.

It's a tool: like all tools, it can be use for good...or not.... (Via @cshirky.)

Your Number is Up...

...and it's £2.2 trillion:

The gap between what the Government expects to spend and what it actually brings in has risen five-fold, from £120 billion to £608 billion in the space of six months.

At that rate, according to the Institute for Fiscal Studies, it will take 23 years to return government borrowing to anything like normal levels – Gordon Brown’s famous “golden rule”.

And of course, every year you borrow keeps adding to what you owe. Right now, the Government calculates that it owes a total of £2.2 trillion – about £144,000 per household. The figure has trebled since the bank bail-outs. Some traders are beginning to wonder if Britain can actually pay its debts. If they start pulling out, then we really are bust.

Tell me again why we can afford to spend £19 billion on ID cards and associated super-duper databases...?

24 June 2009

Pillars of Open Government

As you may have noticed, I'm writing more about open government these days, simply because there's more to write about - and that's great. Here's some clueful stuff from the other side of the globe. It's by Kate Lundy, a member of the Australian Senate, who really seems to get this openness thing:


For the Australian government, an opportunity to construct what I see as the three pillars of Open Government is presented. Each of these pillars assumes the basic principle of citizen engagement at every possible opportunity to both empower people, and to ensure the results are actually appropriate and useful.

The three pillars of open government.

* Citizen-centric services
* Open and transparent government
* Innovation facilitation

I particularly liked this one:

The second pillar is open and transparent government. This pillar builds on the principles that citizens have a right to the information they need to inform themselves about public and political affairs, and to participate in the democratic processes in an informed way. This second pillar is to ensure genuine means of engagement between citizens and the government in policy and decision-making. This is always harder than it sounds but it is essential to garner the wisdom of the crowd. It is vital that government engage with the broader community not just for a conversation, but in genuine partnership between political leaders and the people so we can as a society respond most effectively to the specific social and economic challenges communities confront. This localisation of policy solutions is essential to ensure relevance of government solutions to real situations, and essential to ensure a reasonable response time to new issues and emergencies. Open and transparent government will grow citizen trust and ultimately participation in policy development and government directions.

Great to see people all around the world working on this stuff. Pillars of open government, indeed.

Follow me @glynmoody on Twitter or identi.ca.

Sugar on a Stick v1 Strawberry is Out

Although I've been sceptical of the OLPC project, not least because of its ridiculous decision to offer a Windows XP version - putting them in thrall to a US monopolist is really good way to "help" the developing world, people - I'm a big fan of the GNU/Linux-based Sugar Learning platform. I'm also a big fan of using USB sticks as a way of providing complete software solutions based on GNU/Linux. So it should come as no surprise that I think this is fab:

Sugar Labs, nonprofit provider of the Sugar Learning Platform to over one-million children worldwide, announces the immediate availability of Sugar on a Stick v1 Strawberry. Available free for download at www.sugarlabs.org, Sugar on a Stick can be loaded onto an ordinary 1GB USB flash drive and used to reboot any PC or netbook directly into the award-winning Sugar environment. It runs on recent Macs with a helper CD and in Windows using virtualization. Sugar on a Stick is designed to work with a School Server that can provide content distribution, homework collection, backup services, Moodle integration, and filtered access to the Internet. Today’s Strawberry release is meant for classroom testing; feedback will be incorporated into the next version, available towards the end of 2009.

...

Learning Activities are at the heart of Sugar. Sugar on a Stick includes 40 Activities to interest young learners such as Read, Write, Paint, and Etoys. Hundreds more Activities are available free for download at the Sugar Activity Library. Most “Sugarized” Activities have student collaboration built-in; students and teachers work, play, and learn on the same Activities together. The Sugar Learning Platform is open, so by leveraging the work of other open source projects, existing software for children can be integrated; for example, the acclaimed GCompris suite of 100 Activities developed over the past five years by Bruno Coudoin was recently added to Sugar, including Activities such as Chess, Geography, and Sudoku. Teachers and parents interested in Sugar’s Activities and its modern interface for children can watch short videos on the recently opened Sugar Labs Dailymotion channel.

Note that this is a great way to (a) use old PCs (b) provide educational materials in a single package for free (c) to avoid security issues associated with Windows. What's not to like?

Follow me @glynmoody on Twitter or identi.ca.

23 June 2009

Why Open Source, Clouds and Crowds Rule

The Guardian's crowd-sourcing of the initial analysis of hundreds of thousands of PDFs of MPs' expenses is fast becoming mythic. If you want some more technical details, this post is a good place to start. I was particular struck by the following:

As well as the Guardian’s first Django joint, this was its first project with EC2, the Amazon contract-hosting service beloved by startups for its low capital costs.

Willison’s team knew they would get a huge burst of attention followed by a long, fading tail, so it wouldn’t make sense to prepare the Guardian’s own servers for the task. In any case, there wasn’t time.

“The Guardian has lead time of several weeks to get new hardware bought and so forth,” Willison said. “The project was only approved to go ahead less than a week before it launched.”

With EC2, the Guardian could order server time as needed, rapidly scaling it up for the launch date and down again afterward. Thanks to EC2, Willison guessed the Guardian’s full out-of-pocket cost for the whole project will be around £50.

As for the software, it was all open-source, freely available to the Guardian — and to anyone else who might want to imitate them. Willison hopes to organize his work in the next few weeks.

None of this happens without open source to allow zero-cost hacks; nothing happens without clouds, that allow immediate and low-cost scale-up. (Fifty quid? Blimey.) Bottom line: increasingly popular crowdsourcing efforts won't be happening without either.

Sarkozy Will Go "All the Way" for Monopolies

Curious stuff coming out of France:

“By defending copyright I do not just defend artistic creation, I also defend my idea of a free society where everyone’s freedom is based on respect for the rights of others. I am also defending the future of our culture. It is the future of creation.”

Er, sorry, mon brave, you seem to have forgotten that copyright is a monopoly: as such, it's antithetical to freedom. Indeed, it *takes away* the freedom from all those it is imposed upon, which is practically the entire population of the world. Artists create irrespective of copyright - they have to, because of an inner urge, not because copyright says they get a monopoly.

And as for the "future of our culture", you obviously don't understand that it is inextricably bound up with *past* culture. If it can't build on what everyone has created before, just as they did - and copyright makes this increasingly difficult - your culture won't have any future.

Follow me @glynmoody on Twitter or identi.ca.

Big Victory for FoI and UK Transparency

Kudos to Computer Weekly:


The information commissioner has ordered the opening of confidential files on a wide range of high-risk IT projects, including the ID cards scheme, joined up police intelligence systems and the NHS National Programme for IT (NPfIT).

It is the most far-reaching decision under the Freedom of Information Act for government IT.

It is also a victory for Computer Weekly’s campaign for the release of the results of Gateway reviews on the progress of major IT-based projects.

MPs have complained that the first they knew of problems on projects such as the IT to support tax credit and child support payments was when constituents contacted them.

Our campaign has been aimed at persuading government to release information about projects in time for MPs and others to ask informed questions, and possibly avert a failure.

I particularly liked the list of feeble excuses used for not giving out the information, especially the last one, which is extraordinary in its arrogance:

# It would prevent policy formulation or development taking place in the self-contained space needed to ensure it was done well.

# It would make policy development less effective because departments’ attention would be focused on obtaining a “green light”.

# It would cause reports to become bland and anodyne, defeating their purpose.

# It would make interviewees, senior responsible owners and the private sector less willing to participate in reviews or co-operate with interviewers.

# It would cause delays in the completion of reports as words and phrases would be argued over.

# It is unnecessary. The public interest is already met by the information about the programme in the public domain combined with parliamentary scrutiny.

and the list of responses from the Information Commissioner:

* It would allow the public a better understanding of the development of the programmes which are the subject of Gateway reviews.
* It would allow project risks and concerns to be identified.

* It would not damage the Gateway process in the way the OGC has suggested.

* The public scrutiny of projects by the National Audit Office and Public Accounts Committee involve largely historical and retrospective analyses. Gateway reviews “would provide a level of public scrutiny of current projects”.

* It would inform the debate as to the merits of the schemes, the practicalities involved and the feasibility.

* It would ensure that “schemes as complex as these are properly scrutinised and implemented”.

* It is unrealistic to imagine that civil servants will not participate if reviews are to be published. In accordance with the Civil Service Code, “civil servants must fulfil their duties and obligations responsibly.”

Those are crucially important points, because they apply to everything else, past, present and future.

Well done, Computer Weekly for waging and winning this battle: now let's all take it forward to make UK government even more transparent.

Follow me @glynmoody on Twitter or identi.ca.

GNU/Linux Tops TOP500 Supercomputers Again

The fact that GNU/Linux totally dominates the top 500 supercomputing list is hardly news, but the fact that it has managed to *increase* its market share yet further is.

Here are the results for June 2009:


GNU/Linux 443 (88.6%)
Windows 5 (1.0%)
Unix 22 (4.4%)

and here are the figures for six months ago:


GNU/Linux 439 (87.8%)
Windows 5 (1.0%)
Unix 23 (4.6%)

Notice that plucky little Windows, from that small and hopelessly out-gunned company up in Seattle has bravely managed to increase its share by precisely 0%: an impressive result considering the millions of dollars it has spent trying to break into this market.

Snarky? Moi?

Update: More details about the top 20, and GNU/Linux's dominance here.

Follow me @glynmoody on Twitter or identi.ca.

22 June 2009

Open Source Dendrochronology

How could I resist this story? Aside from the great headline, it's about old-style closed-source science being challenge by open science, with open data - the only kind, if you think about it:


Dendrochronology is the study of tree-rings to determine when and where a tree has grown. Everybody knows that trees produce one ring every year. But the rings also vary in width according to each year's local weather conditions. If you've got enough rings in a wood sample, then their widths form a unique "bar code". Collect enough samples of various ages from buildings and bog wood, and you can join the bar codes up to a reference curve covering thousands of years.

Dendrochronology has a serious organisational problem that impedes its development as a scientific discipline and tends to compromise its results. This is the problem of proprietary data. When a person or organisation has made a reference curve, then in many cases they will not publish it. They will keep it as an in-house trade secret and offer their paid services as dendrochronologists. This means that dendrochronology becomes a black box into which customers stick samples, and out of which dates come, but only the owner of the black box can evaluate the process going on inside. This is of course a deeply unscientific state of things. And regardless of the scientific issue, I am one of those who feel that if dendro reference curves are produced with public funding, then they should be published on-line as a public resource.

But there is a resistance movement: amateur dendrochronologists such as my buddies Torbjörn Axelsson and Åke Larsson. They practice open source data transparency on the net, which means that arguably amateur dendrochronology is at this time more scientific than the professional variety.

I think it's also interesting because the story shows that, even in specialised areas like dendrochronology, openness makes a big difference to how the science is conducted - and how reliable its results are. (Via @BoraZ.)

Follow me @glynmoody on Twitter or identi.ca.

MPs Plot Against Transparency - and Lose the Plot

They just don't get it, do they?


Parliament is planning to block the future release of expenses receipts after the humiliation endured by MPs this week, The Times has learnt.

Senior MPs have drawn up plans to replace the publication of every receipt with a spreadsheet detailing individual claims.

The changes would make less information available for public scrutiny, despite the anger caused this week by the way in which details were blacked out from the official files.

Look chaps, open means open, as in o-p-e-n: we're not going to settle for less. Get used to it, because we're going to keep coming back and coming back until we get audit trail clarity from our money in your pockets to every last expense.

Follow me @glynmoody on Twitter or identi.ca.