11 November 2009

The Embedded Market Beds Down with Linux

Never a dull moment in the embedded Linux market. First Intel acquires Wind River, now the slightly less well-known Cavium acquires MontaVista:

Cavium Networks, a leading provider of highly integrated semiconductor products that enable intelligent processing for networking, wireless, storage and video applications, today announced that it is has signed a definitive agreement to acquire MontaVista Software for $50 million, comprised of approximately $16 million in cash and approximately $34 million in Cavium Networks common stock.

More interesting, perhaps, are the reason for the latest move:

Today embedded Linux is fast becoming the operating system of choice for hundreds of millions of devices ranging from very large carrier grade equipment to consumer electronics. Traditionally embedded devices used a proprietary OS or commercial real-time operating system. However, there is a major trend towards using embedded Linux. This rapid adoption of Linux in embedded networking, wireless, consumer electronics, mobile devices and storage is driving the demand for a high quality, commercial grade embedded Linux along with support for multi-core processors and embedded virtualization.

MontaVista Software is a leader in multi-core embedded Linux operating systems, virtualization, development tools and professional services with a broad array of Tier-1 customers. As the first commercial embedded Linux vendor, MontaVista provides the industry’s leading Carrier Grade Linux that has been widely adopted by industry leading companies that include Alcatel-Lucent, Cisco, Ericsson, Fujitsu, NEC, Nokia-Siemens, NTT, Motorola, Samsung and many other Tier-1 vendors. MontaVista is also the innovation leader in the embedded Linux market segment with deployments in Tier-1 Consumer Electronics manufactures such as Sony, Samsung and Philips; MID and Mobile vendors such as NEC and Garmin; Industrial Automation vendors such as HP, Kyocera-Mita and Fuji Xerox and leading Automotive infotainment suppliers. One of MontaVista’s signature, high profile deployments includes Dell’s latest innovative enterprise notebook the Dell Latitude ON that uses MontaVista's Montabello software platform.

Note that this very positive assessement comes not from some open source fanboy (like me), but from a semiconductor company that has just plunked down $50 million on its bet. Against that background, there seems little doubt that Linux will soon become the de facto standard in the world of embedded systems.

Follow me @glynmoody on Twitter or identi.ca.

Brazil to Allow Private Copying and Mashups

I always said Brazil was a civilised country:


O MinC proporá que a cópia privada de qualquer tipo de obra digital seja permitida sem a autorização expressa ou remuneração ao titular para uso privado e desde que seja apenas um exemplar, além de permitir o uso do conteúdo original em outra mídia que não aquela que o consumidor comprou originalmente.

Na prática, quem comprou um DVD ou um CD poderá criar uma cópia de backup para uso próprio ou então poderá repassar o filme ou as músicas para o computador, trocando o suporte físico (como o disco plástico do CD ou DVD) pelo digital (esteja o arquivo em MP3, WMA e OGG ou AVI e MPEG).

...

A reforma prevê ainda "a utilização (...) de pequenos trechos de obras preexistentes, sempre que a utilização em si não seja o objetivo principal da obra nova e que não prejudique a exploração normal da obra reproduzida", dando respaldo legal à criação de mashups musicais ou visuais.


[Via Google Translate: The MinC propose that private copying of any digital work is permitted without the express permission or compensation to the owner for private use and provided that only one copy, and allows the use of original content in other media than the one that consumer originally purchased.

In practice, those who bought a DVD or CD can create a backup copy for personal use or you can pass on the movie or the music to your computer, changing the physical medium (such as the hard plastic of the CD or DVD) to digital (whether the file to MP3, WMA and OGG or AVI and MPEG).

...

The reform also provides for "(...) the use of short extracts from existing works, where the use itself is not the main purpose of new work and does not prejudice the normal exploitation of the work reproduced, giving legal backing to the creation of mashups music or visual.]

(Via Remixtures.)

The Next Bill Gates, or the Next Tim B-L?

On Twitter, I have just been followed by @nextbillgates, which is associated with the eponymous Web site:

Consider yourself to be an IT wizard – just like Bill Gates?

If the answer is yes, why not enter our brand new, national search to find the UK version of the next Bill Gates?

This is rather sad. If we're searching for an "IT wizard" to hold up as an example, why not look closer to home, and choose Sir Tim Berners-Lee? He not only invented the Web, he *gave it away*, as one of the most glorious acts of altruism we've seen in recent years, kick-starting technological, economic and social changes that are probably unmatched since the Industrial Revolution.

Isn't that something worth celebrating, and encouraging others to emulate? Or are we so mired in greedy materialism that we must instead hold up a man who not only accumulated a truly obscene amount of money by overcharging people for defective software through smart marketing, but is also founder of a company that was found to be a monopoly?

What kind of future do we really want: one based on taking, or one based on giving?

Follow me @glynmoody on Twitter or identi.ca.

Why SAP is a Sap

There's some interesting turbulence in the blogosphere about the following call from Dr. Vishal Sikka, Chief Technology Officer (CTO) of SAP:

To ensure the continued role of Java in driving economic growth, we believe it is essential to transition the stewardship of the language and platform into an authentically open body that is not dominated by an individual corporation. Java should be free of any encumbrances to permit fair competition between compatible implementations for the benefit of customers. By preserving the integrity of Java, the IT industry can ensure a vibrant developer community and continued innovation for enterprise software customers. This ensures the continued global economic success brought about through open innovation.


Matt Asay rightly calls him out on this:

Irony, thy name is SAP.

SAP, after all, is hardly the most open-source or open-process friendly company on the planet. Despite early involvement in Eclipse, some interaction with MySQL (MaxDB), and a new commitment to the Apache Software Foundation, SAP remains a firmly proprietary company.

Even Microsoft, which arguably has the most to lose from open source, has consistently and continually experimented with greater open-source involvement.

SAP? Not so much. In large part, SAP hasn't been forced to embrace open source because it hasn't been threatened by it. ERP (enterprise resource planning) is such a complex beast that it has remained largely impervious to open source (with the exception of open-source start-ups like Compiere and Openbravo, to which I'm an adviser).

Now, Dirk Riehle is stepping into the fray:

I don’t think that this is a fair critique. SAP has always provided the source code of its main business applications suite to user-customers as part of a commercial license, and users have always customized SAP’s business suite to their heart’s content. In fact, it is the only way to make it work for their needs.

That may well be the case, but I think it's irrelevant.

The real reason SAP's call is hypocritical is this document [.pdf], essentially a love-letter to software patents, submitted as an amicus curiae brief to the European Patent Office. Software patents are simply incompatible with free software, because they are government-granted monopolies designed to *stop* people sharing stuff. They also prevent hackers from writing new code because they represent an ever-present digital sword of Damocles hanging over them.

SAP simply cannot claim to be a true friend of openness while it also supports software patents in any jurisdiction, in any form - the same applies to other companies, too, I should note. They can share as much code as they like, but until they repudiate software patents - for example, by placing their patent portfolios in the public domain - that's little more than window-dressing.

Follow me @glynmoody on Twitter or identi.ca.

06 November 2009

θαλασσα θαλασσα

Since I've been using the Web for over 15 years, it's not often that a site leaves me gob-smacked. But of all the sites I've seen recently, Marine Traffic is definitely one of the most amazing:

This web site is part of an academic, open, community-based project. It is dedicated in collecting and presenting data which are exploited in research areas, such as:

- Study of marine telecommunications in respect of efficiency and propagation parameters
- Simulation of vessel movements in order to contribute to the safety of navigation and to cope with critical incidents
- Interactive information systems design
- Design of databases providing real-time information
- Statistical processing of ports traffic with applications in operational research
- Design of models for the spotting of the origin of a pollution
- Design of efficient algorithms for sea path evaluation and for determining the estimated time of ship arrivals
- Correlation of the collected information with weather data
- Cooperation with Institutes dedicated in the protection of the environment

It provides free real-time information to the public, about ship movements and ports, mainly across the coast-lines of Europe and N.America. The project is currently hosted by the Department of Product and Systems Design Enginnering, University of the Aegean, Greece. The initial data collection is based on the Automatic Identification System (AIS). We are constantly looking for partners to take part in the community. They will have to install an AIS receiver and share the data of their area with us, in order to cover more areas and ports around the world.

Basically, then, it shows in *real-time* the positions of ships along major sea-lanes around the world. Hovering over their icons brings up information about that ship, including its speed and heading. Other pages have background data on ports and the ships. But the most amazing thing is just watching the shipping traffic gradually *move* in front of your eyes...

Free, open and gobsmacking: what more do you want? (Via Global Guerillas.)

Follow me @glynmoody on Twitter or identi.ca.

Microsoft's Biological Implants

Microsoft's up to its old tricks of offering pretty baubles to the innocent with The Microsoft Biology Foundation:

The bioinformatics community has developed a strong tradition of open development, code sharing, and cross-platform support, and a number of language-specific bioinformatics toolkits are now available. These toolkits serve as valuable nucleation points for the community, promoting the sharing of code and establishing de facto standards.

The Microsoft Biology Foundation (MBF) is a language-neutral bioinformatics toolkit built as an extension to the Microsoft .NET Framework. Currently it implements a range of parsers for common bioinformatics file formats; a range of algorithms for manipulating DNA, RNA, and protein sequences; and a set of connectors to biological Web services such as NCBI BLAST. MBF is available under an open source license, and executables, source code, demo applications, and documentation are freely downloadable from the link below.

Gotta love the segue from "strong tradition of open development, code sharing and cross-platform support" to "here, take these patent-encumbered .NET Framework toys to play with".

The point being, of course, that once you have dutifully installed the .NET framework, with all the patents that Microsoft claims on it, and become locked into it through use and habit, you are part of the Microsoft-controlled ecosystem. And there you are likely to stay, since Microsoft doesn't even pretend any of this stuff will be ported to other platforms.

For, under the misleading heading "Cross-platform and interoperability" it says:

MBF works well on the Windows operating system and with a range of Microsoft technologies.

Yeah? And what about non-Microsoft operating systems and technologies?

We plan to work with the developer community to take advantage of the extensibility of MBF and support an increasing range of Microsoft and non-Microsoft tools as the project develops.

Well, that's a complete irrelevance to being cross-platform: it just says it'll work with other stuff - big deal.

If I were a biologist I'd be insulted at this thinly-disguised attempt to implant such patent-encumbered software into the bioinformatics community, which has a long and glorious tradition of supporting free software that is truly free and truly cross-platform, and thus to enclose one of the most flourishing and vibrant software commons.

Follow me @glynmoody on Twitter or identi.ca.

03 November 2009

ACTA's All-out Assault on the Internet

Michael Geist has some deeply disturbing details about what may well be in the Internet section of ACTA:


1. Baseline obligations inspired by Article 41 of the TRIPs which focuses on the enforcement of intellectual property.

2. A requirement to establish third-party liability for copyright infringement.

3. Restrictions on limitations to 3rd party liability (ie. limited safe harbour rules for ISPs). For example, in order for ISPs to qualify for a safe harbour, they would be required establish policies to deter unauthorized storage and transmission of IP infringing content. Provisions are modeled under the U.S.-Korea Free Trade Agreement, namely Article 18.10.30. They include policies to terminate subscribers in appropriate circumstances. Notice-and-takedown, which is not currently the law in Canada nor a requirement under WIPO, would also be an ACTA requirement.

4. Anti-circumvention legislation that establishes a WIPO+ model by adopting both the WIPO Internet Treaties and the language currently found in U.S. free trade agreements that go beyond the WIPO treaty requirements. For example, the U.S.-South Korea free trade agreement specifies the permitted exceptions to anti-circumvention rules. These follow the DMCA model (reverse engineering, computer testing, privacy, etc.) and do not include a fair use/fair dealing exception. Moreover, the free trade agreement clauses also include a requirement to ban the distribution of circumvention devices. The current draft does not include any obligation to ensure interoperability of DRM.

5. Rights Management provisions, also modeled on U.S. free trade treaty language.

This is nothing less than the copyright cartel's last stand against the Internet - a desperate attempt to lock down everything. As Geist observes:

it provides firm confirmation that the treaty is not a counterfeiting trade, but a copyright treaty. These provisions involve copyright policy as no reasonable definition of counterfeiting would include these kinds of provisions.

That is, the Powers-that-Be *lied* to us, as usual. We must fight this, or we will be paying the consequences for years to come.

Follow me @glynmoody on Twitter or identi.ca.

WIPO Boss: ACTA Should be Open, Transparent

Wow:

On the secretive Anti-Counterfeiting Trade Agreement, Gurry said that WIPO too did not know a great deal about the talks.

“Naturally we prefer open, transparent international processes to arrive at conclusions that are of concern to the whole world,” he said, citing WIPO’s role as an international, United Nations agency. And, he added, “IP is of concern to the whole world.”

If even the head of WIPO is saying ACTA needs to be drawn up as part of an open, transparent process, isn't it time for the relevant governments to listen?

Follow me @glynmoody on Twitter or identi.ca.

28 October 2009

If a Sequoia Falls in the Forest...

...maybe somebody's listening:

Today, Sequoia Voting Systems officially introduced its latest revolutionary new offering – the Frontier Election System – the first transparent end-to-end election system including precinct and central count digital optical scan tabulators, a robust election management and ballot preparation system, and tally, tabulation, and reporting applications based on an open architecture with publicly disclosed source code developed specifically to meet current and future iterations of the federal Voting System Guidelines.


“Security through obfuscation and secrecy is not security,” said Eric D. Coomer, PhD, Vice President of Research and Product Development at Sequoia Voting Systems. “Fully disclosed source code is the path to true transparency and confidence in the voting process for all involved. Sequoia is proud to be the leader in providing the first publicly disclosed source code for a complete end-to-end election system from a leading supplier of voting systems and software. Sequoia’s Frontier Election System has been designed to comply with all the current Election Assistance Commission’s Voluntary Voting System Guidelines.”

As you may recall, Sequoia had previously been a big fan of precisely that "security through obfuscation and secrecy", so this is an astonishing turnaround. Kudos to them for finally seeing the light - its rivals will doubtless follow suit soon, because no one can be seen to be *against* transparency.

But even more kudos to all those people who fought the e-voting industry's attempts to shut them up and get on with murky business as usual. This will surely become a textbook example of how dogged and rigorous advocacy finally wins out.

Follow me @glynmoody on Twitter or identi.ca.

27 October 2009

Biophysical Economics: A Different View

One of the things that I have felt for a while is that mainstream economics isn't really the best way to look at free software, or any of the other intellectual commons or - even more importantly - the environmental commons, since economics is really about consumption. And now, it seems, some academics are beginning to call into question basic assumptions about that consumerist, consumptive viewpoint:

The financial crisis and subsequent global recession have led to much soul-searching among economists, the vast majority of whom never saw it coming. But were their assumptions and models wrong only because of minor errors or because today's dominant economic thinking violates the laws of physics?

A small but growing group of academics believe the latter is true, and they are out to prove it. These thinkers say that the neoclassical mantra of constant economic growth is ignoring the world's diminishing supply of energy at humanity's peril, failing to take account of the principle of net energy return on investment. They hope that a set of theories they call "biophysical economics" will improve upon neoclassical theory, or even replace it altogether.

Here's the heart of the problem:

Central to their argument is an understanding that the survival of all living creatures is limited by the concept of energy return on investment (EROI): that any living thing or living societies can survive only so long as they are capable of getting more net energy from any activity than they expend during the performance of that activity.

Great to see some new thinking in this area; I'm sure in time it will have knock-on consequences for the way we look at the commons, too.

Follow me @glynmoody on Twitter or identi.ca.

26 October 2009

How Proprietary JAWS Bites the Blind

Here's a heart-warming tale of those kind people who make proprietary software, specifically of the piquantly-named company Freedom Scientific, which produces a program called JAWS:

JAWS (an acronym for Job Access With Speech) is a screen reader, a software program for visually impaired users, produced by the Blind and Low Vision Group at Freedom Scientific of St. Petersburg, Florida, USA. Its purpose is to make personal computers using Microsoft Windows accessible to blind and visually impaired users. It accomplishes this by providing the user with access to the information displayed on the screen via text-to-speech or by means of a braille display and allows for comprehensive keyboard interaction with the computer.

Clearly, JAWS fulfils an important function for the visually impaired. One might presume it is a font of benevolence and altruism, doing its utmost to help a group of people who are already at a disadvantage. Maybe not, according to this petition:

Braille displays require a screen reader in order to work. Freedom Scientific has steadfastly refused to provide Braille display manufacturers with the driver development kit required to enable a particular Braille device to communicate with JAWS. Instead, the manufacturer must first pay an outrageous sum of money before support for the Braille device will be permitted. What's more, this charge to the Braille display manufacturer is not a one-time fee but is imposed annually.

Well, that doesn't sound very kind. So why on earth do people put up with this?

One might ask how Freedom Scientific can play the gatekeeper to its JAWS product where Braille driver support is concerned. The answer is simply and for no other reason because it can.

...

I for one am shocked, appalled, and amazed that Freedom Scientific would impose such limitations and restrictions not only upon its own customer base but also on those organizations which manufacture products that supplement the information that JAWS provides. This draconian and self-serving policy is not at all in keeping with the pro-Braille spirit exemplified by the Braille Readers are Leaders Initiative set into motion earlier this year by the National Federation of the Blind in honor of the Bicentennial celebration of Louis Braille. Instead of offering an additional opportunity to expand the usage of Braille, it stifles the ability of the blind consumer to choose the Braille display that will best meet his/her needs.

And the reason it can, of course, is because it is proprietary software, which means that nobody can route around the problem.

This episode shows once again why it is vital for such software to be open source so that there is no gatekeeper, and so that the community's needs come first, not the desire of a company to make as much money as possible regardless of the plight of the people it affects.

Follow me @glynmoody on Twitter or identi.ca.

23 October 2009

UK Government Blows it on Lobbying

If you wanted proof that the UK government is still an enemy of transparency, try this:

The Government is grateful to the Public Administration Select Committee for its examination of lobbying in the UK, which is the first Parliamentary inquiry on the subject since 1991.

It is right that the Government remains alert for signs of improper influence over any aspect of our public life and the Committee's Report provides a helpful opportunity to look again at arrangements and to ensure that it has the right framework in place to ensure confidence in the way outside interests interact with government.

In responding to the Committee's recommendations, it is first important to set out the context of this inquiry. While the Committee's Report focuses mainly on the relationship between the lobbying industry and Government, it must be remembered that lobbying goes much wider than this. Lobbying is essentially the activity of those in a democracy making representations to government on issues of concern. The Government is committed to protecting this right from improper use while at the same time seeking to avoid any unnecessary regulation or restriction. As well as being essential to the health of our democracy, its free and proper exercise is an important feature of good government.

What this conveniently glosses over is the difference between "making representations to government on issues of concern" - which is what you and I as citizens do all the time, mostly by sending emails to MPs and ministers - and *lobbying*, which is now an entire industry of people employed to use every trick in the book, from the most to least subtle, to get what their clients want.

The first - making representations - is just what it seems: someone expressing their view and/or asking for action. Lobbying, by contrast, is your typical iceberg, with most of its intent invisible below the surface. That is why a lobbyists' register is needed - so that others can work out the iceberg. The UK government's refusal to countenance this - and the pathetic excuse it offers for doing so - are yet another blot on this administration's record as far as openness is concerned.

And if you're wondering why it is so obstinate, maybe this has something to do with it:

The Government agrees that any system of regulation, whether it is voluntary self-regulation or statutory regulation, requires a register of lobbyists to ensure that lobbying activity is transparent. The Government agrees with most of the elements for such a register outlined by the Committee.

However, the Government does not agree that such a register should include the private interests of Ministers and civil servants. This should not be a matter for a register of lobbyists. Ultimately, major decisions are taken by Ministers. Information about Ministers' relevant private interests is now published as well as information in the Registers of Members' and Peers' Interests. In addition, relevant interests' of departmental board members are also available publicly. However, the Government believes that the proposal for a Register of the private interests of civil servants would be a disproportionate requirement that would place a significant burden on departments and agencies while adding very little to the regulation of lobbying. Both Ministers and civil servants are already subject to clear standards of conduct for dealing with lobbyists.

But hang on a minute: if the argument is that such information is *already* made available, then there would be no *extra* burden in providing it to both authorities. It would only be information not already declared that might require effort - and that is precisely what should be made available. Yet more pathetic - and transparently false - logic from the UK government, which is still trying to keep us from scrutinising the engines of political power.

Follow me @glynmoody on Twitter or identi.ca.

The Utter Moral Bankruptcy of the DNA Database

This is staggering:

Detections using the national DNA database have fallen over the past two years despite the number of profiles increasing by 1m and its running costs doubling to £4.2m a year.

A report on the database covering the years 2007-09, published today, shows that crimes cleared up as a result of a match on the DNA database fell from 41,148 to 31,915 over the period. At the same time the number of DNA profiles on the database – already the largest in the world – rose from 4.6m to 5.6m. Duplicates mean that the police database now holds details of 4.89 million individuals.

That is, despite increasing the size to getting on for 10% of the UK population, the number of crimes cleared *fell* by over 25%. How pathetic is that? Not as pathetic as this statement from the truly Orwellian "National Policing Improvement Agency":

Nevertheless, Peter Neyroud, the head of the National Policing Improvement Agency (NPIA), which hosts the DNA database, says in the report that it continues to provide the police with the most effective tool for the prevention and detection of crime since the development of fingerprint analysis more than a century ago.

Against the background that this "most effective tool for the prevention and detection of crime since the development of fingerprint analysis more than a century ago" is getting ever-less effective and more costly, and infringing on the rights of ever more people, this statement proves just one thing: that the British police are getting more and more incompetent, and have to rely on more and more Draconian laws and tools just to stop their already footling success rate dropping even more precipitously.

This is an utter scandal on so many levels, but above all because the UK government is continuing to foist this intrusive, disproportionate, racist and morally repugnant approach upon us when it's *own figures* demonstrate that it is failing more and more each year.

Follow me @glynmoody on Twitter or identi.ca.

22 October 2009

Of Open Source and Open Government

One of the key figures in the open government in Australia - and indeed globally, given the paucity of such people - is Kate Lundy. She's been speaking at the Free and Open Source Software for Geospatial Conference 2009. Understandably, her talk was mostly about geospatial data, but there was also this nice section:

FOSS is like a living blueprint – a map if you will – for trust, sustainability and interoperability in the implementation of Gov 2.0 principles. FOSS principles and methodologies are the best case studies you can find for tried and proven online collaboration with constructive outcomes. FOSS applications provide reference implementations for pretty much anything, which government can build upon and customise. We have a lot to learn from FOSS methods and practices, and I would ask all of you to assist us, in government, to understand.

Would that more politicians were so perspicacious.

Follow me @glynmoody on Twitter or identi.ca.

Artists to Fans to Artists: Positive Feedback

One of the sad things about the current mess in the music industry is that artists are too often pitted against fans, when in fact both want the same thing: good music in a convenient format at a fair price. Here's a welcome initiative that's trying to bridge that gulf of misunderstanding:

Artists need to be paid, and fans want to pay them.

Our goals at a2f2a are:

* Help each community better understand the other;

* Help find a practical and workable system which offers artists fair remuneration in exchange for access to material by fans; and

* Help set the agenda for discussions about the role P2P can play within the emergent digital record industry.

Together, we can do it – artist to fan to artist.

What I particularly like about this - aside from the dialogue - is that it starts from the premise that people *do* want to pay for stuff. I think that's absolutely central: most people realise that artists need to be supported, and that if everyone pays, the overall price will be lower. But the music industry likes to portray the public as split in two: those who don't want to pay anything, ever, and those who will meekly pay whatever exorbitant price the labels demand. It ain't that Manichean, and if a2f2a can help to dispel that myth, that's got to be good news.

Follow me @glynmoody on Twitter or identi.ca.

21 October 2009

No Patents on Seeds...or We're Really Stuffed

Good to see that I'm not a lone voice crying in the wilderness:


The continuing patenting of seeds, conventional plant varieties and animal species leads to far-reaching expropriations of farmers and breeders: farmers are deprived of their rights to save their seeds, and breeders are under strong limitations to use the patented seeds freely for further breeding. The patent holder controlls the sale of the seeds and the planting, decides about the use of herbicides and can even collect royalties at the harvest – up to the finished food product.

Our food security is increasingly dependent on a few transnational chemical and biotechnological companies.

The European Patent Office (EPO) has continuasly broadened the scope of patentability and undermined existing restrictions, in the interest of multinational companies.

Allthough plant varieties and animal species are by law exempt from patentability several hundret patents on genetically modified plants have been granted already. Basis for these decissions is the highly controversial EU Biotech Patents directive and a decission by the EPO's Enlarged Board of Appeal, which ruled in 1999 that in principle such patents could be granted.

Now the European Patent Office again has to deal with a basic question: Patents on conventional plants and animals!

The Enlarged Board of Appeal of the EPO will use a patent on broccoli (EP 1069819) for a fundamental ruling, on whether or not conventional plants are patentable. The broccoli in question was merely diagnosed using marker assisted breeding methods to identify its natural occuring genes. The genes were not modified. All other broccoli plants with similar genes are considered as "technical inventions“ by the patent. Thus even their use for breeding and the plants themselves are monopolised. Through this the provision which prohibits the patenting of "essentially biological proceses" is to be undermined. The EPO has already granted similar patents: e.g.: only recently the company Enza Zaden Beheer received a patents on pathogene resitant lettuce ( EP1179089B1)

Should the Enlarged Board of Appeal uphold the patent, then this decission (case T0083/05) will be binding for all other pending patent applications and even for animals and their offspring.

This exactly parallels the situation with software patents, where the EPO is using every trick in the book to approve them; except it's even worse.

Follow me @glynmoody on Twitter or identi.ca.

Why not Participatory Medicine?

If this participation thing is so great, why don't we apply it to something really important, like medicine? Why not, indeed?

Welcome to JoPM, a New Peer-Reviewed, Open Access Journal

Our mission is to transform the culture of medicine to be more participatory. This special introductory issue is a collection of essays that will serve as the 'launch pad' from which the journal will grow. We invite you to participate as we create a robust journal to empower and connect patients, caregivers, and health professionals.

More specifically:

Because the Journal of Participatory Medicine is a new journal publishing multidisciplinary articles on topics within a new, not yet defined field, we have established draft parameters that define the journal’s range of interest. We anticipate that these parameters will change somewhat as the field develops. In the meantime, the following characterize the field of participatory medicine.

I particularly liked the following section, with its emphasis on openness:

New Knowledge Creation stems from the collaboration of researchers and patients, as individuals and as groups.

1. Health professionals and patients sharing in the discussion of scientific methods, including open discussion about the level of evidence of the research

2. Open, transparent process that demonstrates collaboration and participation in research

3. Patients with significant interest in a topic joining together to create repositories for research, including (but not limited to) registries, tissue banks and genetic databases; demonstrating mutual respect for the contributions of the data owners and health research professionals with the tools to gain insight from those repositories. Interpretation of results and conclusions including involvement of all stakeholders.

Important stuff, worth a read.

Follow me @glynmoody on Twitter or identi.ca.

Won't Somebody Please Think of the Orphans?

This is droll: the European Commission is finally waking up to the copyright orphans problem - thanks to some healthy panic induced by Google's book digitisation plans [.pdf]:


Orphan works are works that are in copyright but whose right holders cannot be identified or located. Protected works can become orphaned if data on the author and/or other relevant right holders (such as publishers, photographers or film producers) is missing or outdated. A work can only be exploited only after obtaining prior permission from the right holders. In the case of orphan works, granting such authorisation is not possible. This leads to a situation where millions of works cannot be copied or otherwise used e.g. a photograph cannot be used to illustrate an article in the press, a book cannot be digitised or a film restored for public viewing. There is also a risk that a significant proportion of orphan works cannot be incorporated into mass-scale digitisation and heritage preservation efforts such as Europeana or similar projects.

Libraries, universities, archives, some commercial users and several Member States claim that the problem of existing instruments, such as the Commission Recommendation 2006/585/EC 7 or the 2008 Memorandum of Understanding on Orphan Works and the related diligent search guidelines, is that these are not legally binding acts and that the issue of mass digitisation has not been addressed. Since non-legislative initiatives neither provide sufficient legal certainty nor solve the fact that using orphan works constitutes a copyright infringement, they advocate a legislative approach at the European level to allow different uses of orphan works. It is also stressed that obstacles to intra-Community trade in orphan works may emerge if each Member State were to adopt its own set of rules to deal with the problem.

For publishers, collecting societies and other right holders, orphan works are a rights-clearance issue. They are sceptical about introducing a blanket exception to use orphan works. For them, the crucial issue is to ensure that a good faith due diligence search to identify and locate the right holders is carried out, using existing databases.

The utter cluelessness and fatuity of the publishers' response is breath-taking: the problem is that the rights-holders *cannot be found* - that's why they're called "orphans". Demanding "due diligence" just misses the point completely.

At least this proves that publishers simply have no credible arguments against introducing an "exception" to use orphan works, for example in digitisation projects. And against their non-existent downside, the upside is just immense. Let's hope the European Commission is beginning to understand this.

Follow me @glynmoody on Twitter or identi.ca.

20 October 2009

Racing to the Bottom of Openness

Here's some interesting news about Barnes & Noble's e-reader:

The reader, named the “Nook,” looks a lot like Amazon’s white plastic e-book, only instead of the chiclet-keyboard there is a color multi-touch screen, to be used as both a keyboard or to browse books, cover-flow style. The machine runs Google’s Android OS, will have wireless capability from an unspecified carrier and comes in at the same $260 as the now rather old-fashioned-looking Kindle.

Linux-based: no surprise there. But this is:

And over at the Wall Street Journal, somebody got a peek at an at ad set to run in the New York Times this coming Sunday. The ad features the line “Lend eBooks to friends”, and this has the potential to destroy the Kindle model. One of the biggest problems with e-books is that you can’t lend or re-sell them. If B&N is selling e-books cheaper than the paper versions, then the resale issue is moot. And lending, even if your friends need a Nook, too, takes away the other big advantage of paper.

In fact, this loaning function could be the viral feature that makes the device spread. Who would buy a walled-garden machine like the Kindle when the Nook has the same titles, cheaper, and you can borrow? The Nook is already starting to look like the real internet to the Kindle’s AOL.

It's a classic "race to the bottom", where the bottom is total openness: see you there, Amazon.

Follow me @glynmoody on Twitter or identi.ca.

19 October 2009

Monsanto: Making Microsoft Look Good

Following my recent post about Bill Gates helping to push genetically-modified and patented seeds towards needy African farmers, Roy Schestowitz kindly send me links to the follow-on story: Gates attacking anyone who dares to criticise that move:

Microsoft founder Bill Gates said on October 15 that environmentalists who are adamantly opposed to using genetically modified crops in Africa are hindering efforts to end hunger on that continent.

Gates was speaking at the annual World Food Prize forum, which honors those who make important contributions to improving agriculture and ending hunger. He noted that genetically modified crops, fertilizers, and chemicals could all help small African farms produce more food, but environmentalists who resist their use are standing in the way.

“This global effort to help small farmers is endangered by an ideological wedge that threatens to split the movement in two,” Gates told the forum. “Some people insist on an ideal vision of the environment. They have tried to restrict the spread of biotechnology into sub-Saharan Africa without regard to how much hunger and poverty might be reduced by it, or what the farmers themselves might want.”

This is, of course, a clever framing of the debate: if you're against patented GMOs it's because you're an "idealist" (now where have I heard that before?), with a hint of Luddite too. The same post - which writes from a very Gates-friendly viewpoint - quotes him as saying:

On one side is a technological approach that increases productivity.

On the other side is an environmental approach that promotes sustainability.

Productivity or sustainability — they say you have to choose.

It’s a false choice, and it’s dangerous for the field. It blocks important advances. It breeds hostility among people who need to work together. And it makes it hard to launch a comprehensive program to help poor farmers.

The fact is, we need both productivity and sustainability — and there is no reason we can’t have both.

Do genetically-modified seeds bring increased productivity? There seem doubts; but even assuming it's true, Gates sets up a false dichotomy: one reason GMO seeds aren't sustainable is because they are patented. That is, farmers *must* buy them year after year, and can't produce their own seeds. It's a situation that's relatively easy to solve: make GMOs patent-free; do not place restrictions on their use; let farmers do what farmers have done for millennia.

And look, there you have it, potentially: productivity and sustainability. But we won't get that, not because the idealistic environmentalist are blocking it, but because the seed industry wants farmers dependent on their technology, not liberated by it. It is sheer hypocrisy for a fan of patents to accuse environmentalists of being the obstacle to productivity and sustainability: that would be the industrial model of dependence, enforced by intellectual monopolies, and espoused by big companies like Monsanto, the Microsoft of plant software.

I wrote about the human price paid in India as a result of these patented seeds and the new slavery they engender a few months back. The key quotation:

Tara Lohan: Farmer suicides in India recently made the news when stories broke last month about 1,500 farmers taking their own lives, what do you attribute these deaths to?

Vandana Shiva:
Over the last decade, 200,000 farmers have committed suicide. The 1,500 figure is for the state of Chattisgarh. In Vidharbha, 4,000 are committing suicide annually. This is the region where 4 million acres of cotton have been grown with Monsanto's Bt cotton. The suicides are a direct result of a debt trap created by ever-increasing costs of seeds and chemicals and constantly falling prices of agricultural produce.

When Monsanto's Bt cotton was introduced, the seed costs jumped from 7 rupees per kilo to 17,000 rupees per kilo. Our survey shows a thirteenfold increase in pesticide use in cotton in Vidharbha. Meantime, the $4 billion subsidy given to U.S. agribusiness for cotton has led to dumping and depression of international prices.

Squeezed between high costs and negative incomes, farmers commit suicide when their land is being appropriated by the money lenders who are the agents of the agrichemical and seed corporations. The suicides are thus a direct result of industrial globalized agriculture and corporate monopoly on seeds.

Here's an excellent, in-depth feature from Vanity Fair on the tactics Monsanto uses in the US. A sample:

Some compare Monsanto’s hard-line approach to Microsoft’s zealous efforts to protect its software from pirates. At least with Microsoft the buyer of a program can use it over and over again. But farmers who buy Monsanto’s seeds can’t even do that.

...

Farmers who buy Monsanto’s patented Roundup Ready seeds are required to sign an agreement promising not to save the seed produced after each harvest for re-planting, or to sell the seed to other farmers. This means that farmers must buy new seed every year. Those increased sales, coupled with ballooning sales of its Roundup weed killer, have been a bonanza for Monsanto.

The feature is from last year, but I don't imagine the situation has got better since then. Indeed, the picture it paints of Monsanto is so bleak and depressing that I'm forced to admit that Microsoft in comparison comes off as almost benevolent. Given Monsanto's size, methods and evident ambitions, I fear I shall be writing rather more about this company in the future.

Follow me @glynmoody on Twitter or identi.ca.

18 October 2009

Opencourseware Comes Under Attack

It was bound to happen: opencourseware is under attack:

While seeking to make college more accessible, the Obama administration has launched a largely unnoticed assault upon the nation’s vibrant market in online learning. As part of an ambitious bill designed to tighten federal control over student lending, the House of Representatives included a scant few sentences green-lighting a White House plan to spend $500 million on an “Online Skills Laboratory,” in which the federal government would provide free online college courses in a variety of unspecified areas. The feds would make the courses “freely available” and encourage institutions of higher education to offer credit for them. The measure is now before the Senate.

Ah yes, "freely available": that communistic cancer again.

It is not clear what problem the administration is seeking to solve. The kinds of online courses that the administration is calling for already exist, and are offered by an array of publishers and public and private institutions. Online enrollment grew from 1.6 million students in 2002 to 3.9 million in 2007. Nearly 1,000 institutions of higher education provide distance learning.

More than half a dozen major textbook publishers, and hundreds of smaller providers, develop and distribute online educational content. To take one example, Pearson’s MyMathLab is a self-paced, customizable online course, which the University of Alabama uses to teach more than 10,000 students a year. Janet Poley, president of the American Distance Education Consortium, doesn’t see the need for federal dollars to be spent “reinventing courses that have already been invented.”

Since it's "not clear what problem the administration is seeking to solve", allow me to offer a little help.

The article suggests that the kinds of online courses that will be created are already on offer: well, no, because those produced by "major textbook publishers, and hundreds of smaller providers" are neither "free of charge" nor "free" in the other, more interesting sense that you can take them, rework them, reshape them, and then share them. And why might that be a good idea? Well, most importantly, because it means that you don't have to "reinvent courses that have already been invented."

Oh, but wait: isn't that what the article says the current situation avoids? Indeed: and that, of course, is where the article goes wrong. The existing courses, which are proprietary, and may not be copied or built on, cause *precisely* the kind of re-inventing of the wheel that opencourseware is accused of. That's because every publisher must start again, laboriously recreating the same materials, in order to avoid charges of copyright infringement.

That's an absurd waste of effort: the facts are the same, so once they are established it's clearly much more efficient to share them and then move on to create new content. The current system doesn't encourage that, which is why we need to change it.

Given the article gets things exactly the wrong way round, it's natural to ask how the gentleman who penned these words might have come to these erroneous conclusions. First, we might look at where he's coming from - literally:

Frederick M. Hess is director of education-policy studies at the American Enterprise Institute.

Here's what SourceWatch has to say on this organisation:

The American Enterprise Institute for Public Policy Research (AEI) is an extremely influential, pro-business, conservative think tank founded in 1943 by Lewis H. Brown. It promotes the advancement of free enterprise capitalism, and succeeds in placing its people in influential governmental positions. It is the center base for many neo-conservatives.

And if that doesn't quite explain why Mr Hess might be pushing a somewhat incorrect characterisation of the situation, try this:

In 1980, the American Enterprise Institute for the sum of $25,000 produced a study in support of the tobacco industry titled, Cost-Benefit Analysis of Regulation: Consumer Products. The study was designed to counteract "social cost" arguments against smoking by broadening the social cost issue to include other consumer products such as alcohol and saccharin. The social cost arguments against smoking hold that smoking burdens society with additional costs from on-the-job absenteeism, medical costs, cleaning costs and fires. The report was part of the global tobacco industry's 1980s Social Costs/Social Values Project, carried out to refute emerging social cost arguments against smoking.

So, someone coming from an organisation that has no qualms defending the tobacco industry is unlikely to have much problem denouncing initiatives that spread learning, participation, collaboration, creativity, generosity and general joy in favour of all their antitheses. And the fact that such a mighty machine of FUD should stoop to attack little old opencourseware shows that we are clearly winning.

17 October 2009

The Commons Meme Becomes More Common

One of the great knock-on benefits of Elinor Ostrom sharing the Nobel prize for Economics is that the concept of the commons is getting the best airing that it's ever had. Here's another useful meditation on the subject from someone who knows what he's talking about, since he's written a book on the subject:

Old fables die hard. That's surely been the history of the so-called "tragedy of the commons," one of the most durable myths of the past generation. In a famous 1968 essay, biologist Garrett Hardin alleged that it is nearly impossible for people to manage shared resources as a commons. Invariably someone will let his sheep over-graze a shared pasture, and the commons will collapse. Or so goes the fable.

In fact, as Professor Elinor Ostrom's pioneering scholarship over the past three decades has demonstrated, self-organized communities of "commoners" are quite capable of managing forests, fisheries and other finite resources without destroying them. On Monday, Ostrom won a Nobel Prize in Economics for explaining how real-life commons work, especially in managing natural resources.

As he notes:

Although Ostrom has not written extensively about the Internet and online commons, her work clearly speaks to the ways that people can self-organize themselves to take care of resources that they care about. The power of digital commons can be seen in the runaway success of Linux and other open-source software. It is evident, too, in the explosive growth of Wikipedia, Craigslist (classified ads), Flickr (photo-sharing), the Internet Archive (historical Web artifacts) and Public.Resource.org (government information). Each commons acts as a conscientious steward of its collective wealth.

And this is an acute observation:

A key reason that all these Internet commons flourish is because the commoners do not have to get permission from, or make payments to, a corporate middleman. They can build what they want directly, and manage their work as they wish. The cable and telephone companies that provide access to the Internet are not allowed to favor large corporate users with superior service while leaving the rest of us--including upstart competitors and non-market players--with slower, poorer-quality service.

In an earlier time, this principle was known as "common carriage"--the idea that everyone shall have roughly equivalent access and service, without discrimination. Today, in the Internet context, it is known as "net neutrality."

Neat: another reason we need to preserve Net neutrality is to preserve all the commons - past, present and future - it enables.

Follow me @glynmoody on Twitter or identi.ca.

15 October 2009

Open Sourcing America's Operating System

And how do you do that? By making all of the laws freely available - and, presumably, searchable and mashable:

Public.Resource.Org is very pleased to announce that we're going to be working with a distinguished group of colleagues from across the country to create a solid business plan, technical specs, and enabling legislation for the federal government to create Law.Gov. We envision Law.Gov as a distributed, open source, authenticated registry and repository of all primary legal materials in the United States.

This is great news, because Carl Malamud - the force behind this initiative - has been urging it for years: now it looks like it's beginning to take a more concrete form:

The process we're going through to create the case for Law.Gov is a series of workshops hosted by our co-conveners. At the end of the process, we're submitting a report to policy makers in Washington. The process will be an open one, so that in addition to the main report which I'll be authoring, anybody who wishes to submit their own materials may do so. There is no one answer as to how the raw materials of our democracy should be provided on the Internet, but we're hopeful we're going to be able to bring together a group from both the legal and the open source worlds to help crack this nut.

I particularly liked the following comment:

Law.Gov is a big challenge for the legal world, and some of the best thinkers in that world have joined us as co-conveners. But, this is also a challenge for the open source world. We'd like to submit such a convincing set of technical specs that there is no doubt in anybody's mind that it is possible to do this. There are some technical challenges and missing pieces as well, such as the pressing need for an open source redaction toolkit to sit on top of OCR packages such as Tesseract. There are challenges for librarians as well, such as compiling a full listing of all materials that should be in the repository.

What's interesting is that this recognises that open source is not just an inspiration, but a key part of the solution, because - like the open maths movement I wrote about below - it needs new kinds of tools, and free software is the best way to provide them.

Now, if only someone could do something similar in the UK....

Open Source Mathematics

This is incredibly important:

On 27 January 2009, one of us — Gowers — used his blog to announce an unusual experiment. The Polymath Project had a conventional scientific goal: to attack an unsolved problem in mathematics. But it also had the more ambitious goal of doing mathematical research in a new way. Inspired by open-source enterprises such as Linux and Wikipedia, it used blogs and a wiki to mediate a fully open collaboration. Anyone in the world could follow along and, if they wished, make a contribution. The blogs and wiki functioned as a collective short-term working memory, a conversational commons for the rapid-fire exchange and improvement of ideas.

The collaboration achieved far more than Gowers expected, and showcases what we think will be a powerful force in scientific discovery — the collaboration of many minds through the Internet.

You can read the details of what happened - and it's inspiring stuff - in the article. But as well as flagging up this important achievement, I wanted to point to some interesting points it makes:

The process raises questions about authorship: it is difficult to set a hard-and-fast bar for authorship without causing contention or discouraging participation. What credit should be given to contributors with just a single insightful contribution, or to a contributor who is prolific but not insightful? As a provisional solution, the project is signing papers with a group pseudonym, 'DHJ Polymath', and a link to the full working record. One advantage of Polymath-style collaborations is that because all contributions are out in the open, it is transparent what any given person contributed. If it is necessary to assess the achievements of a Polymath contributor, then this may be done primarily through letters of recommendation, as is done already in particle physics, where papers can have hundreds of authors.

The project also raises questions about preservation. The main working record of the Polymath Project is spread across two blogs and a wiki, leaving it vulnerable should any of those sites disappear. In 2007, the US Library of Congress implemented a programme to preserve blogs by people in the legal profession; a similar but broader programme is needed to preserve research blogs and wikis.

These two points are also relevant to free software and other open endeavours. So far, attribution hasn't really been a problem, since everyone who contributes is acknowledged - for example through the discussions around the code. Similarly, preservation is dealt with through the tools for source code management and the discussion lists. But there are crucial questions of long-term preservation - not least for historical purposes - which are not really being addressed, even by the longest-established open projects like GNU.

For example, when I wrote Rebel Code, I often found it hard to track down the original sources for early discussions. Some of them have probably gone for ever, which is tragic. Maybe more thought needs to be given - not least by central repositories and libraries - about how important intellectual moments that have been achieved collaboratively are preserved for posterity to look at and learn from.

Talking of which, the article quoted above has this to say on that subject:

The Polymath process could potentially be applied to even the biggest open problems, such as the million-dollar prize problems of the Clay Mathematics Institute in Cambridge, Massachusetts. Although the collaborative model might deter some people who hope to keep all the credit for themselves, others could see it as their best chance of being involved in the solution of a famous problem.

Outside mathematics, open-source approaches have only slowly been adopted by scientists. One area in which they are being used is synthetic biology. DNA for the design of living organisms is specified digitally and uploaded to an online repository such as the Massachusetts Institute of Technology Registry of Standard Biological Parts. Other groups may use those designs in their laboratories and, if they wish, contribute improved designs back to the registry. The registry contains more than 3,200 parts, deposited by more than 100 groups. Discoveries have led to many scientific papers, including a 2008 study showing that most parts are not primitive but rather build on simpler parts (J. Peccoud et al. PLoS ONE 3, e2671; 2008). Open-source biology and open-source mathematics thus both show how science can be done using a gradual aggregation of insights from people with diverse expertise.

Similar open-source techniques could be applied in fields such as theoretical physics and computer science, where the raw materials are informational and can be freely shared online. The application of open-source techniques to experimental work is more constrained, because control of experimental equipment is often difficult to share. But open sharing of experimental data does at least allow open data analysis. The widespread adoption of such open-source techniques will require significant cultural changes in science, as well as the development of new online tools. We believe that this will lead to the widespread use of mass collaboration in many fields of science, and that mass collaboration will extend the limits of human problem-solving ability.

What's exciting about this - aside from the prospect of openness spreading to all these other areas - is that there's a huge opportunity for the open source community to start, er, collaborating with the scientific one in producing these new kinds of tools that currently don't exist and are unlikely to be produced by conventional software houses (since spontaneously collaborative communities can't actually pay for anything). I can't wait.

Follow me @glynmoody on Twitter or identi.ca.

Gates Gives $300 million - but with a Catch

It's becoming increasingly evident that Bill Gates' philanthropy is not simple and disinterested, but has woven into it a complex agenda that has to do with his love of intellectual monopolies - and power. Here's the latest instalment:


The Bill and Melinda Gates Foundation, which is donating another $120 million to boosting agriculture in the developing world, will focus on self-help aid for poor farmers to sustain and grow production, a top adviser to the world's leading charitable foundation said.

Sounds good, no? Here are more details:

The Gates Foundation, with a $30 billion endowment to improve health and reduce poverty in developing countries, began investing in agricultural projects three years ago. The latest grants bring its farm sector awards to $1.4 billion.

One of its first investments was in African seeds through the Alliance for a Green Revolution in Africa (AGRA). The group is expected to introduce more than 1,000 new seed varieties of at least 10 crops to improve African production by 2016.

"Alliance for a Green Revolution in Africa" also sounds good; here's a little background on that organisation:

It has not gone unnoticed that AGRA falls under the direct supervision of the Global Development Program, whose senior programme officer is Dr. Robert Horsch, who worked for Monsanto for 25 years before he joined the Gates Foundation. Horsch was part of the scientific team in the company that developed Monsanto’s YieldGard, BollGard and RoundUp Ready technologies. Horsch’s task at the Gates Foundation is to apply biotechnology toward improving crop yields in regions including sub-Saharan Africa. Lutz Goedde another senior program officer of the Global Development Program, is also a recruit from the biotech industry as he used to head Alta Genetics, the world's largest privately owned cattle genetics improvement and artificial insemination Company, worth US$100 million.

That is, AGRA not only has close links with the Gates Foundation, but also with Monsanto - the Microsoft of the seed world.

If you read the rest of the document from which the above information was taken, you'll see that the AGRA programme is essentially promoting approaches using seeds that are genetically modified and patented. Here's the conclusion:

Sub-Saharan Africa represents an extremely lucrative market for seed companies. The development interventions by AGRA appear on the face of it, to benevolent. However, not only will AGRA facilitate the change to a market based agricultural sector in Africa replacing traditional agriculture, but it will also go a long way towards laying the groundwork for the entry of private fertilizer and agrochemical companies and seed companies, and more particularly, GM seed companies.

So Gates' donations are ultimately promoting an agriculture based on intellectual monopolies - just as Microsoft does in the software field. The latest $300 million doesn't sound quite so generous now, does it?

Follow me @glynmoody on Twitter or identi.ca.