08 July 2009

Not Kissing the Rod, Oh My Word, No

Becta today [6 July 2009] welcomes Microsoft's launch of the new Subscription Enrolment Schools Pilot (SESP) for UK schools, which provides greater flexibility and choice for schools who wish to use a Microsoft subscription agreement.

Great, and what might that mean, exactly?

The new licensing scheme removes the requirement that schools using subscription agreements pay Microsoft to licence systems that are using their competitor's technologies. So for the first time schools using Microsoft's subscription licensing agreements can decide for themselves how much of their ICT estate to licence.

So BECTA is celebrating that fact that schools - that is, we taxpayers - *no longer* have to "pay Microsoft to licence systems that are using their competitor's technologies"? They can now use GNU/Linux, for example, *without* having to pay Microsoft for the privilege?

O frabjous day! Callooh! Callay!

Follow me @glynmoody on Twitter or identi.ca.

Policing the Function Creep...

Remember how the poor darlings in the UK government absolutely *had to* allow interception of all our online activities so that those plucky PC Plods could maintain their current stunning success rate in their Whirr on Terruh and stuff like that? Well, it seems that things have changed somewhat:

Detectives will be required to consider accessing telephone and internet records during every investigation under new plans to increase police use of communications data.

The policy is likely to significantly increase the number of requests for data received by ISPs and telephone operators.

Just as every investigation currently has to include a strategy to make use of its subjects' financial records, soon CID officers will be trained to always draw up a plan to probe their communications.

The plans have been developed by senior officers in anticipation of the implementation of the Interception Modernisation Programme (IMP), the government's multibillion pound scheme to massively increase surveillance of the internet by storing details of who contacts whom online.

Er, come again? "CID officers will be trained to always draw up a plan to probe their communications"? How does that square with this being a special tool for those exceptional cases when those scary terrorists and real hard naughty criminals are using tricky high-tech stuff like email? Doesn't it imply that we are all terrorist suspects and hard 'uns now?

Police moves to prepare for the glut of newly accessible data were revealed today by Deputy Assistant Commissioner Janet Williams. She predicted always considering communications data will lead to a 20 per cent increase in the productivity of CID teams.

She told The Register IMP had "informed thinking" about use of communications data, but denied the plans gave the lie to the government line that massively increased data retention will "maintain capability" of law enforcement to investigate crime.

Well, Mandy Rice-Davies applies, m'lud...

Follow me @glynmoody on Twitter or identi.ca.

07 July 2009

Are Microsoft's Promises For Ever?

This sounds good:

I have some good news to announce: Microsoft will be applying the Community Promise to the ECMA 334 and ECMA 335 specs.

ECMA 334 specifies the form and establishes the interpretation of programs written in the C# programming language, while the ECMA 335 standard defines the Common Language Infrastructure (CLI) in which applications written in multiple high-level languages can be executed in different system environments without the need to rewrite those applications to take into consideration the unique characteristics of those environments.

"The Community Promise is an excellent vehicle and, in this situation, ensures the best balance of interoperability and flexibility for developers," Scott Guthrie, the Corporate Vice President for the .Net Developer Platform, told me July 6.

It is important to note that, under the Community Promise, anyone can freely implement these specifications with their technology, code, and solutions.

You do not need to sign a license agreement, or otherwise communicate to Microsoft how you will implement the specifications.

The Promise applies to developers, distributors, and users of Covered Implementations without regard to the development model that created the implementations, the type of copyright licenses under which it is distributed, or the associated business model.

Under the Community Promise, Microsoft provides assurance that it will not assert its Necessary Claims against anyone who makes, uses, sells, offers for sale, imports, or distributes any Covered Implementation under any type of development or distribution model, including open-source licensing models such as the LGPL or GPL.

But boring old sceptic that I am, I have memories of this:

The Software Freedom Law Center (SFLC), provider of pro-bono legal services to protect and advance free and open source software, today published a paper that considers the legal implications of Microsoft's Open Specification Promise (OSP) and explains why it should not be relied upon by developers concerned about patent risk.

SFLC published the paper in response to questions from its clients and the community about the OSP and its compatibility with the GNU General Public License (GPL). The paper says that the promise should not be relied upon because of Microsoft's ability to revoke the promise for future versions of specifications, the promise's limited scope, and its incompatibility with free software licenses, including the GPL.

That was then, of course, what about now? Well, here's what the FAQ says on the subject:

Q: Does this CP apply to all versions of the specification, including future revisions?

A: The Community Promise applies to all existing versions of the specifications designated on the public list posted at /interop/cp/, unless otherwise noted with respect to a particular specification.


Now, is it just me, or does Microsoft conspicuously fail to answer its own question? The question was: does it apply to all versions *including* future revision? And Microsoft's answer is about *existing* versions: so doesn't that mean it could simply not apply the promise to a future version? Isn't this the same problem as with the Open Specification Promise? Just asking.

03 July 2009

The Engine of Scientific Progress: Sharing

Here's a post saying pretty much what I've been saying, but in a rather different way:


Here we present a simple model of one of the most basic uses of results, namely as the engine of scientific progress. Research results are more than just accumulated knowledge. Research results make possible new questions, which in turn lead to even more knowledge. The resulting pattern of exponential growth in knowledge is called an issue tree. It shows how individual results can have a value far beyond themselves, because they are shared and lead to research by others.


Follow me @glynmoody on Twitter or identi.ca.

02 July 2009

Patents Don't Promote Innovation: Study

It's extraordinary how the myth that patents somehow promote innovation is still propagated and widely accepted; and yet there is practically *no* empirical evidence that it's true. All the studies that have looked at this area rigorously come to quite a different conclusion. Here's yet another nail in that coffin, using a very novel approach:

Patent systems are often justified by an assumption that innovation will be spurred by the prospect of patent protection, leading to the accrual of greater societal benefits than would be possible under non-patent systems. However, little empirical evidence exists to support this assumption. One way to test the hypothesis that a patent system promotes innovation is to simulate the behavior of inventors and competitors experimentally under conditions approximating patent and non-patent systems.

Employing a multi-user interactive simulation of patent and non-patent (commons and open source) systems ("PatentSim"), this study compares rates of innovation, productivity, and societal utility. PatentSim uses an abstracted and cumulative model of the invention process, a database of potential innovations, an interactive interface that allows users to invent, patent, or open source these innovations, and a network over which users may interact with one another to license, assign, buy, infringe, and enforce patents.

Data generated thus far using PatentSim suggest that a system combining patent and open source protection for inventions (that is, similar to modern patent systems) generates significantly lower rates of innovation (p<0.05), productivity (p<0.001), and societal utility (p<0.002) than does a commons system. These data also indicate that there is no statistical difference in innovation, productivity, or societal utility between a pure patent system and a system combining patent and open source protection.

The results of this study are inconsistent with the orthodox justification for patent systems. However, they do accord well with evidence from the increasingly important field of user and open innovation. Simulation games of the patent system could even provide a more effective means of fulfilling the Constitutional mandate ― "to promote the Progress of . . . useful Art" than does the orthodox assumption that technological innovation can be encouraged through the prospect of patent protection.

When will people get the message and start sharing for mutual benefit?

Follow me @glynmoody on Twitter or identi.ca.

01 July 2009

Help Me Go Mano a Mano with Microsoft

Next week, I'm taking part in a debate with a Microsoft representative about the passage of the OOXML file format through the ISO process last year. Since said Microsoftie can draw on the not inconsiderable resources of his organisation to provide him with a little back-up, I thought I'd try to even the odds by putting out a call for help to the unmatched resource that is the Linux Journal community. Here's the background to the meeting, and the kind of info I hope people might be able to provide....

On Linux Journal.

30 June 2009

Why Scientific Publishing Will Never be the Same

For those of us tracking open access and its wider import, it's pretty clear that scientific publishing has changed for ever. But for some within the industry, there remains the desperate hope that all this new-fangled open, collaborative stuff will just blow over.

Forget about it: as this magisterial analysis shows, not only is there no way for traditional publishing to get from there to here - it's a terrifying leap across a minimum to the next maximum - but the most exciting stuff is *already* happening in other forms of publishing:

What’s new today is the flourishing of an ecosystem of startups that are experimenting with new ways of communicating research, some radically different to conventional journals. Consider Chemspider, the excellent online database of more than 20 million molecules, recently acquired by the Royal Society of Chemistry. Consider Mendeley, a platform for managing, filtering and searching scientific papers, with backing from some of the people involved in Last.fm and Skype. Or consider startups like SciVee (YouTube for scientists), the Public Library of Science, the Journal of Visualized Experiments, vibrant community sites like OpenWetWare and the Alzheimer Research Forum, and dozens more. And then there are companies like Wordpress, Friendfeed, and Wikimedia, that weren’t started with science in mind, but which are increasingly helping scientists communicate their research. This flourishing ecosystem is not too dissimilar from the sudden flourishing of online news services we saw over the period 2000 to 2005.

It’s easy to miss the impact of blogs on research, because most science blogs focus on outreach. But more and more blogs contain high quality research content. Look at Terry Tao’s wonderful series of posts explaining one of the biggest breakthroughs in recent mathematical history, the proof of the Poincare conjecture. Or Tim Gowers recent experiment in “massively collaborative mathematics”, using open source principles to successfully attack a significant mathematical problem. Or Richard Lipton’s excellent series of posts exploring his ideas for solving a major problem in computer science, namely, finding a fast algorithm for factoring large numbers. Scientific publishers should be terrified that some of the world’s best scientists, people at or near their research peak, people whose time is at a premium, are spending hundreds of hours each year creating original research content for their blogs, content that in many cases would be difficult or impossible to publish in a conventional journal. What we’re seeing here is a spectacular expansion in the range of the blog medium. By comparison, the journals are standing still.

What's even better about this piece is that it's not content to point out why traditional publishing has big problems: it also offers some practical suggestions of what people *should* be looking at:

These opportunities can still be grasped by scientific publishers who are willing to let go and become technology-driven, even when that threatens to extinguish their old way of doing things. And, as we’ve seen, these opportunites are and will be grasped by bold entrepreneurs. Here’s a list of services I expect to see developed over the next few years. A few of these ideas are already under development, mostly by startups, but have yet to reach the quality level needed to become ubiquitous. The list could easily be continued ad nauseum - these are just a few of the more obvious things to do.

Fantastic stuff - do read it all if you have time.

Follow me @glynmoody on Twitter or identi.ca.

Squishing the Media Bugs

Belatedly, I'm writing about this very cool idea:


It’s called MediaBugs.org. And the idea is to create a web site, a web service, that people in a community, in this case the San Francisco Bay Area, can bring problems and errors that they find in media coverage and post them and try to get them fixed. [...]

The inspiration of the project is from what’s called a bug tracker in an open-source project. So if you’re developing open-source software, you have this project, and you put up a public website that anyone can bring these — file these bugs. If you’re using Firefox and something breaks, you go to their web site, and you tell them about it.

I fear the practice might be harder than the theory, but I certainly wish it well.

Follow me @glynmoody on Twitter or identi.ca.

Winning the Open Web

It seems an unfair fight. On the one hand, you have some of the biggest, most powerful multinationals, intent on defending their turf and extending their power and profits. On the other, you have a tiny number of ragtag idealists who believe that knowledge belongs to everyone, and that no one should have disproportionately long monopolies on its supply.

And yet: in the last few years a remarkable series of victories have been one by the latter against the former, to the extent that representatives of the big media industries have warned that they are losing the "battle".

Against that background of uneven forces - but not quite in the way the media companies mean it - sharing information about past successes so as to drive future ones is crucially important. And yet it is rarely done, probably because the practitioners are too busy fighting the battles to write about it.

Enter Becky Hogge, former Executive Director of the Open Rights Group, who happily has had some time on her hands to prepare a handy report entitled "Winning the Web":


Winning the Web is a 2009 report funded by the Open Society Institute and written by Becky Hogge, former Executive Director of the Open Rights Group. It examined 6 successful campaigns for intellectual property reform, in Brazil, Canada, the US, France, New Zealand, and the UK.

Lessons drawn from the study of the campaigns include the importance of coalition-forming, the best way to conduct online mobilisation campaigns, and the need for a more unified critique of current intellectual property regimes.

The introduction fleshes out the idea:

The global intellectual property regime is no longer fit for purpose. As the networked, digital age matures, it puts into the hands of millions of citizens the tools to access create and share “content”: text, pictures, music and video; data, news, analysis and art. Against this, the intellectual property regime falters. It presents citizens with a choice: stop using the technology – stop communicating, stop creating – or break the law.

Legal reform is presented with two separate challenges. The first is a small but vocal minority of entrenched corporate interests – the rightsholder lobby. Wedded to business models that pre-date the age of networked digital technology, they exploit their position as incumbents to influence legislators. Often representing the world’s biggest multinational corporations, they hijack a narrative that belongs to poor artists struggling in garrets and use the considerable profits they have made from exploiting these artists in the twentieth century to access the corridors of power and make their case.

That legislators listen is related to a second, geopolitical, challenge. Since the 1970s, the developed world has sought to use the global intellectual property regime to ensure its continued prosperity. Motivated by the ability of developing countries to undercut it on the global manufacturing market, it has sought to augment the financial privilege afforded to “knowledge workers”. The self-interest behind this practice is masked by a flawed orthodoxy that is rarely backed up by evidence – that more intellectual property provision is always good for economic growth.

Against this backdrop, a global IP reform movement (also called the access to knowledge movement) is emerging. Motivated by a range of concerns – from global justice, to the narrowing spectrum of permitted speech, to the broadening of surveillance power – these individuals and organisations approach their campaigning work with combined levels of ingenuity and intellectual rigour that make them stand out in the history of fledgling civil rights movements. Recently, these pockets of activism have taken IP reform issues to a wide audience, triggering sweeping civic action in the general population.

For me, the best bits are the detailed case studies of successful campaigns around the world. I knew the bare outlines, but the report really fleshes these out, and then uses them to provide concrete suggestions of what lessons can be learned for the future.

There's one other point is one that I've long thought absolutely crucial:

In the UK, citizens can engage with their elected representatives (including MEPs) using a one-click service called WriteToThem.com. Jim Killock is keen to stress that it is vital that such a tool be developed for all EU member states:

Writetothem.eu is absolutely critical if we want to run these campaigns in the next four years. It shows the contempt in which we seem to hold our European institutions and the irrelevance that they are felt to have across Europe.

This really must be a priority, or else all future campaigns in Europe will suffer as a result.

Follow me @glynmoody on Twitter or identi.ca.

29 June 2009

Watching the Watchers

I read with interest this morning the following:

“Snitchtown” is an essay by Cory Doctorow that first appeared in Forbes.com in June 2007. This SoFoBoMo project is an attempt to illustrate that essay with photographs of some of the 4.2 million CCTV cameras currently estimated to be active in Britain.

It got me thinking: how about setting up a database - a surveillance of surveillance database - that has pictures and locations of CCTVs in the UK? It could be crowd sourced, and anonymous, solving problems of scaling and legal issues. If nothing else, it would put the watchers on notice that they are being watched....

26 June 2009

The World Wins South Korea for Firefox

I've written before about the curious case of South Korea, where the use of Internet Explorer and ActiveX is almost mandatory. I rather despaired of anything changing this situation, since there didn't seem to be any way to get around it from outside. And yet, remarkably, change is coming:

Korea's multifaceted e-government services will be made available for those logging on from FireFox or Safari, web browsers that are gaining more popularity worldwide as an alternative to Internet Explorer.

According to the government Sunday, users of these "non-traditional" browsers will be able to file for year-end tax returns, sign up for a new passport or look for job openings and do much more at various service Web sites operated by the state.

The Ministry of Public Administration and Security, which is in charge of directing e-government initiatives, said that it will invest 11.5 billion won this year for technical projects to increase the browser compatibility of 49 e-government service Web sites.

Starting 2011, all of the 150 e-government Web sites are expected to be accessible from any browser.

That's remarkable, as is the reason for the change:

The development is expected to be useful for overseas Koreans or foreigners logging on to Web sites such as www.hikorea.go.kr from aboard through alternate browsers. Operated by the Ministry of Justice, the Web site is a comprehensive online repository of information for oversea Koreans, immigrants and foreign nationals.

Some civic groups have consistently raised the need to consider expanding e-government services to users of non-traditional browsers.

While the percentage of Koreans users of alternative browsers is still minimal, more netizens worldwide are increasingly surfing the net on browsers other than the Internet Explorer.

A ministry report showed that 21.7 percent of Web users worldwide are browsing on FireFox, and 8 percent on Safari, a browser developed by Apple. IE users make up 67.4 percent of the total Web population.

"We believe that enabling minor browsers to host our e-government services will help overseas Koreans to access the assistance they need and increase Korea's status as a leader in e-government initiatives," a ministry official said.

This shows that what's happening outside a country can still have considerable influence on the internal market - provided there is a big enough expatriate community that still "calls home". Given the increasingly globalised nature of computing, that offers hope for other parts of the world that may be lagging in their uptake of open source.

And if you're wondering why it matters anyway that the South Koreans should be able to use Firefox and other "non-standard" browsers - don't you just love that description? - it's because the country's users have some of the fastest broadband connections in the world; that means that new applications based on such connectivity may well emerge there first, so it's important that open source be available and viable for all kinds of uses. (Via Asa Dotzler.)

Follow me @glynmoody on Twitter or identi.ca.

Show Your Ardour for Ardour

Ardour is a fine open source music program; but like many fine open source programs, it has a problem: money - lack of it. In order to continue to improve the code, the Arbour team ideally needs dosh to pay for programmers and other such handy things; but it's not really happening:


As of now, June 25th, the financial side of things is not looking so good. Last month (May) didn't quite make the $4500 goal, and this month looks certain to fall short by quite a significant margin. There are currently 5-1/2 days left this month, and 28% of the target is still not met. There are no companies backing this project at this time, so its totally incumbent on those of you who use the program and have not yet helped pay to support it to step up and do the right thing. Thanks to everyone who has paid for their contributions and support.

Ardour will continue in some sense even if I find other work, and I believe that Carl, Dave, Hans and others will likely keep up some of their efforts anyway. Since the new download system started, there have been about 9000 OS X downloads and 6000 source code downloads. Less than 3% of the OS X downloads and only three source code downloads were associated with up-front payment, though it seems likely than many users donated after the fact. With a user-base like that, it seems to me that it should be possible to pay one full-time north american developer and to offer occasional payments to others for their outstanding work. What do you think?

Ardour is hardly the only project with these problems, which means that the open source world faces a larger issue: how to raise funds to pay for work that isn't being carried out mostly in bedrooms. It's not something many are thinking about (Matt Asay is an honourable exception), so it's not likely to get solved any time soon - which leaves Ardour in a bit of a pickle. Suggestions and contributions welcome.... (Via Leslie P. Polzer.)

Follow me @glynmoody on Twitter or identi.ca.

Eee, Look: A Useful E-petition Response

Even though I keep signing the wretched things, e-petitions have not generated much action from the UK government. Which makes the following case rather interesting.

The following e-petition:

“We the undersigned petition the Prime Minister to ask the Communities Secretary to require that all software produced by councils under the Timely Information to Citizens project be released under an open source licence.”

Produced this response:

The Government supports the principle that, where new software is being developed by the Timely Information to Citizens pilots, this should wherever possible be released under open source licence and available for use by other local authorities.

For many of the Timely Information to Citizens pilots, the focus is not on new software, but on how existing tools and techniques can be used to bring information together and present it in more useful and accessible ways. Several of the projects will utilise existing open source software to create new information sources and channels, and will share their experiences of doing so with other authorities.

Where the pilots will result in new software tools, ownership and intellectual property rights will usually remain with the individual local authorities. However, most of the authorities concerned have already made a commitment to make these tools available as open source software, or for use by their partner organisations, and we are working to secure the commitment of the remaining.

What impresses me is (a) the reasonableness of the response and (b) the fact that the release of government-developed software should be released as open source "in principle". I do believe we're getting there....

Next, Linux Revolutionises...Printers

Here's a new printer from HP:


Last June 22, HP announced its new all-in-one printer, the Photosmart Premium with TouchSmart Web. Aside from cramming a fax machine, copier, scanner, and a printer into one device, run of the mill technology by today's standards, this new printer can actually print straight from the Web using on-device applications fashioned specifically for this purpose.

As the headline to that story makes clear, that's a Linux-based printer: indeed, it's pretty much unthinkable that these innovative approaches could use anything else. Linux's small footprint, speed, customisability and low cost make it ideal - uniquely so. Where would we be without it?

25 June 2009

Authoring Beautiful HTML...

...ain't easy in the open source world, as David Ascher points out in this post:

However, for regular folks, life is not rosy yet in the Open Web world. Authoring beautiful HTML is, even with design and graphics talent, still way, way too hard. I’m writing this using Wordpress 2.8, which has probably some of the best user experience for simple HTML authoring. As Matt Mullenweg (the founder of Wordpress) says, it’s still not good enough. As far as I can tell, there are currently no truly modern, easy to use, open source HTML composition tools that we could use in Thunderbird for example to give people who want to design wholly original, designed email messages. That’s a minor problem in the world of email, which is primarily about function, not form, and I think we’ll be able to go pretty far with templates, but it’s a big problem for making design on the web more approachable.

There are some valiant efforts to clean up the old, crufty, scary composer codebase that Mozilla has relied on for years. There are simple blog-style editors like FCKEditor and its successor CKEditor. There are in-the-browser composition tools like Google Pages or Google Docs, but those are only for use by Google apps, and only work well when they limit the scope of the design space substantially (again, a rational choice). None of these can provide the flexibility that Ventura Publisher or PageMaker had in the dark ages; none of them can compete from a learnability point of view with the authoring tools that rely on closed stacks; none of them allow the essential polish that hand-crafted code can yield. That’s a gap, and an opportunity.

Let's hope people in the free software world seize it.

Follow me @glynmoody on Twitter or identi.ca.

Crowdsourcing Evil

This was inevitable:

a friend in Iran that I have been in touch with via Skype (which seems to work very well)” told him that a specific Web site, Gerdab.ir, is being used by the Iranian government to identify protesters by crowd-sourcing.

It's a tool: like all tools, it can be use for good...or not.... (Via @cshirky.)

Your Number is Up...

...and it's £2.2 trillion:

The gap between what the Government expects to spend and what it actually brings in has risen five-fold, from £120 billion to £608 billion in the space of six months.

At that rate, according to the Institute for Fiscal Studies, it will take 23 years to return government borrowing to anything like normal levels – Gordon Brown’s famous “golden rule”.

And of course, every year you borrow keeps adding to what you owe. Right now, the Government calculates that it owes a total of £2.2 trillion – about £144,000 per household. The figure has trebled since the bank bail-outs. Some traders are beginning to wonder if Britain can actually pay its debts. If they start pulling out, then we really are bust.

Tell me again why we can afford to spend £19 billion on ID cards and associated super-duper databases...?

24 June 2009

Pillars of Open Government

As you may have noticed, I'm writing more about open government these days, simply because there's more to write about - and that's great. Here's some clueful stuff from the other side of the globe. It's by Kate Lundy, a member of the Australian Senate, who really seems to get this openness thing:


For the Australian government, an opportunity to construct what I see as the three pillars of Open Government is presented. Each of these pillars assumes the basic principle of citizen engagement at every possible opportunity to both empower people, and to ensure the results are actually appropriate and useful.

The three pillars of open government.

* Citizen-centric services
* Open and transparent government
* Innovation facilitation

I particularly liked this one:

The second pillar is open and transparent government. This pillar builds on the principles that citizens have a right to the information they need to inform themselves about public and political affairs, and to participate in the democratic processes in an informed way. This second pillar is to ensure genuine means of engagement between citizens and the government in policy and decision-making. This is always harder than it sounds but it is essential to garner the wisdom of the crowd. It is vital that government engage with the broader community not just for a conversation, but in genuine partnership between political leaders and the people so we can as a society respond most effectively to the specific social and economic challenges communities confront. This localisation of policy solutions is essential to ensure relevance of government solutions to real situations, and essential to ensure a reasonable response time to new issues and emergencies. Open and transparent government will grow citizen trust and ultimately participation in policy development and government directions.

Great to see people all around the world working on this stuff. Pillars of open government, indeed.

Follow me @glynmoody on Twitter or identi.ca.

Sugar on a Stick v1 Strawberry is Out

Although I've been sceptical of the OLPC project, not least because of its ridiculous decision to offer a Windows XP version - putting them in thrall to a US monopolist is really good way to "help" the developing world, people - I'm a big fan of the GNU/Linux-based Sugar Learning platform. I'm also a big fan of using USB sticks as a way of providing complete software solutions based on GNU/Linux. So it should come as no surprise that I think this is fab:

Sugar Labs, nonprofit provider of the Sugar Learning Platform to over one-million children worldwide, announces the immediate availability of Sugar on a Stick v1 Strawberry. Available free for download at www.sugarlabs.org, Sugar on a Stick can be loaded onto an ordinary 1GB USB flash drive and used to reboot any PC or netbook directly into the award-winning Sugar environment. It runs on recent Macs with a helper CD and in Windows using virtualization. Sugar on a Stick is designed to work with a School Server that can provide content distribution, homework collection, backup services, Moodle integration, and filtered access to the Internet. Today’s Strawberry release is meant for classroom testing; feedback will be incorporated into the next version, available towards the end of 2009.

...

Learning Activities are at the heart of Sugar. Sugar on a Stick includes 40 Activities to interest young learners such as Read, Write, Paint, and Etoys. Hundreds more Activities are available free for download at the Sugar Activity Library. Most “Sugarized” Activities have student collaboration built-in; students and teachers work, play, and learn on the same Activities together. The Sugar Learning Platform is open, so by leveraging the work of other open source projects, existing software for children can be integrated; for example, the acclaimed GCompris suite of 100 Activities developed over the past five years by Bruno Coudoin was recently added to Sugar, including Activities such as Chess, Geography, and Sudoku. Teachers and parents interested in Sugar’s Activities and its modern interface for children can watch short videos on the recently opened Sugar Labs Dailymotion channel.

Note that this is a great way to (a) use old PCs (b) provide educational materials in a single package for free (c) to avoid security issues associated with Windows. What's not to like?

Follow me @glynmoody on Twitter or identi.ca.

23 June 2009

Why Open Source, Clouds and Crowds Rule

The Guardian's crowd-sourcing of the initial analysis of hundreds of thousands of PDFs of MPs' expenses is fast becoming mythic. If you want some more technical details, this post is a good place to start. I was particular struck by the following:

As well as the Guardian’s first Django joint, this was its first project with EC2, the Amazon contract-hosting service beloved by startups for its low capital costs.

Willison’s team knew they would get a huge burst of attention followed by a long, fading tail, so it wouldn’t make sense to prepare the Guardian’s own servers for the task. In any case, there wasn’t time.

“The Guardian has lead time of several weeks to get new hardware bought and so forth,” Willison said. “The project was only approved to go ahead less than a week before it launched.”

With EC2, the Guardian could order server time as needed, rapidly scaling it up for the launch date and down again afterward. Thanks to EC2, Willison guessed the Guardian’s full out-of-pocket cost for the whole project will be around £50.

As for the software, it was all open-source, freely available to the Guardian — and to anyone else who might want to imitate them. Willison hopes to organize his work in the next few weeks.

None of this happens without open source to allow zero-cost hacks; nothing happens without clouds, that allow immediate and low-cost scale-up. (Fifty quid? Blimey.) Bottom line: increasingly popular crowdsourcing efforts won't be happening without either.

Sarkozy Will Go "All the Way" for Monopolies

Curious stuff coming out of France:

“By defending copyright I do not just defend artistic creation, I also defend my idea of a free society where everyone’s freedom is based on respect for the rights of others. I am also defending the future of our culture. It is the future of creation.”

Er, sorry, mon brave, you seem to have forgotten that copyright is a monopoly: as such, it's antithetical to freedom. Indeed, it *takes away* the freedom from all those it is imposed upon, which is practically the entire population of the world. Artists create irrespective of copyright - they have to, because of an inner urge, not because copyright says they get a monopoly.

And as for the "future of our culture", you obviously don't understand that it is inextricably bound up with *past* culture. If it can't build on what everyone has created before, just as they did - and copyright makes this increasingly difficult - your culture won't have any future.

Follow me @glynmoody on Twitter or identi.ca.

Big Victory for FoI and UK Transparency

Kudos to Computer Weekly:


The information commissioner has ordered the opening of confidential files on a wide range of high-risk IT projects, including the ID cards scheme, joined up police intelligence systems and the NHS National Programme for IT (NPfIT).

It is the most far-reaching decision under the Freedom of Information Act for government IT.

It is also a victory for Computer Weekly’s campaign for the release of the results of Gateway reviews on the progress of major IT-based projects.

MPs have complained that the first they knew of problems on projects such as the IT to support tax credit and child support payments was when constituents contacted them.

Our campaign has been aimed at persuading government to release information about projects in time for MPs and others to ask informed questions, and possibly avert a failure.

I particularly liked the list of feeble excuses used for not giving out the information, especially the last one, which is extraordinary in its arrogance:

# It would prevent policy formulation or development taking place in the self-contained space needed to ensure it was done well.

# It would make policy development less effective because departments’ attention would be focused on obtaining a “green light”.

# It would cause reports to become bland and anodyne, defeating their purpose.

# It would make interviewees, senior responsible owners and the private sector less willing to participate in reviews or co-operate with interviewers.

# It would cause delays in the completion of reports as words and phrases would be argued over.

# It is unnecessary. The public interest is already met by the information about the programme in the public domain combined with parliamentary scrutiny.

and the list of responses from the Information Commissioner:

* It would allow the public a better understanding of the development of the programmes which are the subject of Gateway reviews.
* It would allow project risks and concerns to be identified.

* It would not damage the Gateway process in the way the OGC has suggested.

* The public scrutiny of projects by the National Audit Office and Public Accounts Committee involve largely historical and retrospective analyses. Gateway reviews “would provide a level of public scrutiny of current projects”.

* It would inform the debate as to the merits of the schemes, the practicalities involved and the feasibility.

* It would ensure that “schemes as complex as these are properly scrutinised and implemented”.

* It is unrealistic to imagine that civil servants will not participate if reviews are to be published. In accordance with the Civil Service Code, “civil servants must fulfil their duties and obligations responsibly.”

Those are crucially important points, because they apply to everything else, past, present and future.

Well done, Computer Weekly for waging and winning this battle: now let's all take it forward to make UK government even more transparent.

Follow me @glynmoody on Twitter or identi.ca.

GNU/Linux Tops TOP500 Supercomputers Again

The fact that GNU/Linux totally dominates the top 500 supercomputing list is hardly news, but the fact that it has managed to *increase* its market share yet further is.

Here are the results for June 2009:


GNU/Linux 443 (88.6%)
Windows 5 (1.0%)
Unix 22 (4.4%)

and here are the figures for six months ago:


GNU/Linux 439 (87.8%)
Windows 5 (1.0%)
Unix 23 (4.6%)

Notice that plucky little Windows, from that small and hopelessly out-gunned company up in Seattle has bravely managed to increase its share by precisely 0%: an impressive result considering the millions of dollars it has spent trying to break into this market.

Snarky? Moi?

Update: More details about the top 20, and GNU/Linux's dominance here.

Follow me @glynmoody on Twitter or identi.ca.

22 June 2009

Open Source Dendrochronology

How could I resist this story? Aside from the great headline, it's about old-style closed-source science being challenge by open science, with open data - the only kind, if you think about it:


Dendrochronology is the study of tree-rings to determine when and where a tree has grown. Everybody knows that trees produce one ring every year. But the rings also vary in width according to each year's local weather conditions. If you've got enough rings in a wood sample, then their widths form a unique "bar code". Collect enough samples of various ages from buildings and bog wood, and you can join the bar codes up to a reference curve covering thousands of years.

Dendrochronology has a serious organisational problem that impedes its development as a scientific discipline and tends to compromise its results. This is the problem of proprietary data. When a person or organisation has made a reference curve, then in many cases they will not publish it. They will keep it as an in-house trade secret and offer their paid services as dendrochronologists. This means that dendrochronology becomes a black box into which customers stick samples, and out of which dates come, but only the owner of the black box can evaluate the process going on inside. This is of course a deeply unscientific state of things. And regardless of the scientific issue, I am one of those who feel that if dendro reference curves are produced with public funding, then they should be published on-line as a public resource.

But there is a resistance movement: amateur dendrochronologists such as my buddies Torbjörn Axelsson and Åke Larsson. They practice open source data transparency on the net, which means that arguably amateur dendrochronology is at this time more scientific than the professional variety.

I think it's also interesting because the story shows that, even in specialised areas like dendrochronology, openness makes a big difference to how the science is conducted - and how reliable its results are. (Via @BoraZ.)

Follow me @glynmoody on Twitter or identi.ca.

MPs Plot Against Transparency - and Lose the Plot

They just don't get it, do they?


Parliament is planning to block the future release of expenses receipts after the humiliation endured by MPs this week, The Times has learnt.

Senior MPs have drawn up plans to replace the publication of every receipt with a spreadsheet detailing individual claims.

The changes would make less information available for public scrutiny, despite the anger caused this week by the way in which details were blacked out from the official files.

Look chaps, open means open, as in o-p-e-n: we're not going to settle for less. Get used to it, because we're going to keep coming back and coming back until we get audit trail clarity from our money in your pockets to every last expense.

Follow me @glynmoody on Twitter or identi.ca.