Wednesday, March 07, 2007
Google Print: A Numbers Game
On February 6th, Google announced that the Princeton University library system agreed to participate in their Book Search Library Project. According to the announcement, Princeton and Google will identify one million works in the public domain for digitization. This follows the January 19th announcement that the University of Texas libraries, the fifth largest library in the U.S., also climbed on board the Library Project. Very quietly, the number of major research libraries participating in the project has more than doubled to twelve in the last two years. The seven new libraries will add millions of printed items to the tens of millions already held by the original five, and more fuel to the legal fire surrounding Google’s plan to scan library holdings and make the full texts searchable on the web.
The public discussion has been mostly one-sided, with Google supporters trying to hold the high moral ground. Their basic argument goes something like this: The universe of published works in the U.S. consists of some 32 million books. They argue that while 80 percent of these books were published after 1923, and, therefore, potentially protected by copyright, only 3 million of them are still in-print and available for sale. As a result, mountains of books have been unnecessarily consigned to obscurity.
No one has yet challenged the basic assumptions supporting this argument. Perhaps they’ve been scared off by Google’s reputation for creating clever algorithms that “organize the world’s information”. This one, though, doesn’t stand up to serious scrutiny.
The figures used by supporters of the Library Project come from a 2005 study undertaken by the Online Computer Library Center (OCLC), the largest consortium of libraries in the U.S. According to the OCLC study, its 20,000 member libraries hold 31,923,000 print books; the original five research libraries participating in the Google library scanning project hold over 18 million.
OCLC did not actually count physical books. They searched their massive database of one billion library holdings and isolated 55 million catalog records describing “language-based monographs”. This was further refined (eliminating duplicates) to 32 million “unique manifestations”, not including government publications, theses and dissertations. The reality of library classification, however, is such that “monographs” often include things like pamphlets, unbound documents, reports, manuals, and ephemera that we don’t usually think of as commercially published books.
The notion that 32 million U.S. published books languish on library shelves is absurd. Just do the math. That works out to more than 80,000 new books published every year since the first English settlement in Jamestown in 1607. Historical book production figures clearly show that the 80,000-threshold was not crossed until the 1980’s, after hovering around 10,000 for fifty years between 1910 to1958. The OCLC study showed, moreover, that member libraries added a staggering 17 million items (half of all print collections) since 1980. That averages out to 680,000 new print items acquired every year for 25 years, or more than the combined national outputs of the U.S., U.K., China, and Japan in 2004.
Not only will Google have to sift through printed collections to identify books, and then determine if they are in the public domain, but they will also have to separate out those published in the U.S. (assuming that their priority is scanning U.S.-based English-language books) from the sea of books published elsewhere. The OCLC study clearly showed that most printed materials held by U.S. libraries were not published in the U.S. The study counted more than 400 languages system-wide, and more than 3 million print materials published in French and German alone in the original Google Five. English-language print materials accounted for only 52% of holdings system-wide, and 49% in the Google Five. Since more than a few works were probably published in the United Kingdom, the total number of English-language books published in the U.S. will constitute less than half of all print collections, both system-wide and in Google libraries.
So how many U.S.-published books are there in our libraries? Annual book production figures show that some 4 million books have been published in the 125 years since figures were regularly compiled in 1880. If, very conservatively, we add an additional 1.5 million books to cover the pre-1880 years, and another 1.5 million to cover books published after 1880 that might have been missed, we get a much more realistic total of 7 million.
Using the lower baseline for published books tells a very different story than the dark one (that the universe of books consists of works that are out-of-print, in the public domain, or “orphaned” in copyright limbo) told by Google and their supporters. With some 3 million U.S. books in print, the inconvenient truth here is that 40% of all books ever published in the U.S. could still be protected by copyright. That would appear to jive with the OCLC finding that 75% of print items held by U.S. libraries were published after 1945, and 50% after 1974.
If we’re going to have a debate that may end up rewriting copyright law, let’s have one based on facts, not wishful thinking.
Andrew Grabois is a consultant to the publishing industry. He has compiled U.S. book production statistics since 1999. He can be reached at the following email address: agrabois@yahoo.com
Clarification update from Andrew: My post is not intended to be a criticism of the OCLC study ("Anatomy of Aggregate Collections: The Example of Google Print for Libraries") by Brian Lavoie et al, which is a valuable and timely look at print collections held by OCLC member libraries. What I am attempting to do here is point out how friends of the Google library project have misinterpreted the paper and cherry-picked findings and conclusions out of context to support their arguments.
Related articles:
Google Book Project (3/6/07)
Qualified Metadata (2/22/07)
Tuesday, March 06, 2007
Google Book Project
Publishers have elected to sue Google to protect their content rights and the content rights of their authors. At the same time, publishers have engaged with Google in participation in the Google Scholar program. Here publishers are equal partners and (I assume) negotiations for the acquisition of content by Google was negotiated in good faith and the results have been good to great for both parties. (Springer, Cambridge University). It is also no bad thing that Google’s content (digitization) programs have spurned other similar content initiatives particularly those of some of the larger trade and academic publishers.
The continued area of friction is the digitization project that Google initiated to scan all the books in as many libraries willing to participate. This is where publishers got upset. They were not consulted nor asked permission, they cannot approve the quality of the scanning, they will not participate in any revenue generated and they can not take for granted that the availability of the scanned book will not undercut any potential revenues they may generate on their own. The books in question are the majority of those published after 1925 or so (It's actually 1923: thanks to Shatzkin for noticing my error) and which are still likely to be under copyright protection of some sort.
Having said that, lets get one thing straight; having all books which exist in library stacks (or deep storage) available in electronic form so that they can be indexed, searched, reassembled, found at all and generally resourced in an easy way is a good thing and an important step forward and opportunity for libraries and library patrons. Ideally, it would lead to one platform (network) providing equal access to high quality, indexed e-book content which any library patron would be able to access via their local library. Sadly, while the vision is still viable the execution represented by the Google library program is not going to get us there.
Setting aside the copyright issue, the Google library program has been going on now for approximately 24mths and results and feedback is starting to show that the reality of the program is not living up to its promise. According to this post from Tim O’Reilly, the scans are not of high quality and importantly are not sufficient to support academic research. Assuming this is universally true (?), the program represents a fantastic opportunity lost for patrons, libraries and Google. BowerBird via O’Reilly states:
umichigan is putting up the o.c.r. from its google scans, for the public-domain books anyway, so the other search engines will be able to scrape that text with ease. what you will find, though, if you look at it (for even as little as a minute or two) is that the quality is so inferior it's almost worthless
Ironically, the lawsuit by the AAP could actually have a beneficial impact on the process of digitization. As some have noted, we may have underestimated the difficultly in finding relevant materials and resources once there is more content to search (this assuming full text is available for search). Initiatives are underway particularly by Library of Congress to address the bibliographic (metadata) requirements of a world with lots more content and perhaps the results of some of these bibliographic activities will result in a better approach to digitization of the more recent content (post 1923). Regrettably, some believe that since there may be only one opportunity to scan the materials in libraries that we may have lost the only opportunity to make these (older) materials accessible to users in an easy way.
Tomorrow, just what is the universe of titles in the post 1923 ‘bucket’? The supporters of the Google project speak about a universe of 30million books but deeper analysis suggests the number is wildly exaggerated.
Shaffer Announced as Chairman of Knovel
From the press release:
Chris Forbes, CEO of Knovel said: "David brings Knovel years ofexperience in managing companies that have successfully delivered highvalue information and productivity solutions to end users. We look forwardto his guidance as Knovel takes the next steps to dominate the engineeringinformation market." David Shaffer continued: "I am excited to join acompany that is a passionate leader in driving value for end users in thislarge and important market. This new role is a natural extension of mycareer at Thomson."
Monday, March 05, 2007
Bear Sterns Media Conference
Monday 3/5
Moodys - 10am. Webinar
Meridith - 11.00am. Webinar
CBS - 12.20pm. Webinar
Thomson - 2.40pm. Webinar
McGraw Hill - 3.20pm. Webinar
Tuesday 3/6
NYTimes - 9.25am. Webinar
Primedia - 10:05am Webinar
Dow Jones - 10:40am Webinar
Scripps - 2:40om Webinar
Reader's Digest Closes Sale
Sunday, March 04, 2007
Social BookMarking
Amazon visitors have not taken to tagging Amazon's books in significant numbers. With thousands of times the traffic, Amazon produced a tenth as many tags as LibraryThing. What's going on?
He goes on to talk about the power of numbers that inevitably feeds on itself; as more people tag more people find it useful which leads to a larger community. Passion also plays a part because the overwhelming number of active 'taggers' also have more books. Those that have less than 200 titles rarely tag books. (Since fees kick in over this number the users are also more financially committed to librarything).
Critical mass is important, even if we can't pinpoint the line. Ten tags are never enough; a thousand almost always is. Unfortunately, Amazon's low numbers translate into a broader failure to reach critical mass. With ten times as many tags overall, LibraryThing has fifteen times as many books with 100 tags, and 35 times as many with over 200 tags.
It is a very interesting article on the power of community and worth a read.
Lorcan Dempsey suggests these types of sites also represent examples of the 'Network effect':
Regular readers of this blog will not be surprised to hear me say that they also highlight a structural issue for libraries, how to provide services at the network level. These new services are network level services. They are aimed at the general user, not an audience circumscribed by region, or funding or institution. And, additionally, they provide an integrated service, moving the user quickly through whatever steps are needed to complete a task.
The other component of interest to me is that sites like librarything are becoming sources of viable bibliographic information and indeed the NY times article suggests that Librarything maybe selling or licensing some of its tagging data:
For example, he is in the process of selling some of his recommendations data, which is based in part on tagging statistics, to other sites that sell books and book information.
At Bowker one of our most important and costly editorial tasks was to assign subject classification to titles so that they could be found either in our own products or in the products of our customers. With well over 200,000 titles per year this is an expensive excercise and while it can be automated (and was) the process suffers from obvious limitations. Firstly, the subject classification methodology is quite rigid and is not always intuitive. Secondly, subjects change over time and books previously categorized could benefit from additional or changed subjects. Thirdly, the application of subjects is subjective and can pose a limitation on the subjects applied to the title. A subject expert wants to be as accurate as possible in applying the subject classification for relevancy and integrity. In BooksinPrint, we had an average of slightly more than two subjects per title over approximately five million records. Many had more and a few had more than 15 but the average was two. Many users of Librarything, myself included, have placed more than two tags against most titles.
The ability of a community to supply a range of subjects that reflect the work, what's important about the book and what becomes important in the wider world may provide a vast addition to traditional bibliographic database work. I believe the social network applications and the structured approaches will work in concert but increasingly it is the social aspect that needs to be actively solicited for inclusion into the traditional bibliographic databases.
Self-Serving List Of Best Blog
Friday, March 02, 2007
Check Prices and Get Recommendations by Phone
Rowling Suing EBAY
Neil Blair, Rowling’s legal adviser at the Christopher Little Literary Agency, said that she welcomed the court order. “Over the years eBay has appeared to be unwilling to control sellers on their site offering pirated or forged Harry Potter items for sale to innocent fans,” he said. “We have asked eBay on numerous occasions to assist by taking preventative steps to avoid these sales – steps that we are aware they can introduce. As these requests were not heeded we had no choice but to seek judicial intervention.Ebay has played the innocent in other similar cases - Tiffany in NYC and Dior in Paris - suggesting that it isn't partical for them to monitor every sale to ensure that the goods are legitmate. They just don't want to do the work.
Thursday, March 01, 2007
Harlequin (Torstar) Reports
On full year revenue:
Book Publishing revenue was $471.8 million in 2006, down $49.3 million from $521.1 million in 2005. Underlying revenue growth of $9.4 million was more than offset by a $30.0 million decline from the strengthening of the Canadian dollar during the year and $28.7 million from lower gains on U.S. dollar hedges year over year.
and on Operating Profit:
Book Publishing reported operating profit was $56.3 million in 2006, down $39.1 million from $95.4 million in 2005. Underlying operating profit was down only $2.5 million in the year while the strengthening Canadian dollar decreased profits by $7.9 million and lower gains on the U.S. dollar hedges decreased profits by $28.7 million year over year. Underlying results were up for North America Retail and Overseas but were more than offset by lower North America Direct-To-Consumer results.It is difficult to discern the true impact here as there may have been some currency impact in the operating profit in 2005. Nevertheless, taking the numbers at face value, underlying revenues were up $9.4mm versus 2005 but operating profit was down $2.5mm.
Here is the short version and the long version of the press releases.
Further detail on the publishing units performance is summarized as follows:
- North America Retail increased book sales in 2006 after stabilizing in 2005. Significant efficiency improvements were made to the series business in 2006 as fewer books were printed and distributed and more books were sold.
- The North America Direct-To-Consumer revenue decline in 2006 was due to both fewer shipments of a children’s direct-to-home continuity program and from shipping disruptions experienced early in the year associated with the bankruptcy of a key supplier. Improved sales through the Internet channel partially offset this decline.
- Overseas markets had mixed results in 2006 with improvements in the United Kingdom and the Nordic Group offset by lower results in Germany. Brazil, a joint venture launched in 2005, showed improvement in 2006 selling more books and making progress towards break-even.
- The outlook for 2007 is for stability. Harlequin has stabilized the total number of books sold over the past three years despite difficult trends in its direct-to-consumer operations.
- Investment will continue in innovation and new products including digital initiatives in 2007.
- Cost savings of approximately $3.0 million are expected from the restructuring undertaken in late 2006.
- Harlequin will continue to be subject to the impact of changes in the value of the Canadian dollar relative to the U.S. dollar and other currencies.
Wolters Kluwer Posts Full Year
Following is are highlights from the company press release:
- Revenues increased 9% to €3,693 million (2005: €3,374 million)
- Organic revenue growth of 3%, in line with full-year outlook
- Ordinary EBITA increased 16% to €618 million (2005: €533 million)
- Ordinary EBITA margin of 17% (2005: 16%)
- Investment in product development reached €272 million (an increase of 9% over 2005)
- Structural cost savings of €128 million (an increase of 28% over 2005)
- Strong free cash flow of €443 million (2005: €351 million)
- Ordinary diluted EPS increased 16% to €1.23 (an increase of 15% at constant currencies)
- Selective acquisitions to strengthen leading positions and enter high-growth adjacent markets
Fourth-Quarter 2006:
- Revenues increased 8% to €1,003 million (2005: €932 million)
- Organic revenue growth of 6% (2005: 3%) Ordinary EBITA increased 17% to €173 million (2005: €148 million)
- Ordinary EBITA margin of 17% (2005: 16%)
- Investment in product development reached €74 million (an increase of 7% over same period 2005)
- Structural cost savings of €37 million (an increase of 32% over same period 2005)
- Strong free cash flow of €204 million (2005: €208 million)
- Ordinary diluted EPS increased 9% to €0.34 (an increase of 15% at constant currencies)
A more detailed review is in the press release. Most of the business units are doing OK with Corporate and Financial Services showing the largest percentage revenue growth. Of Education they say the following:
Organic revenue growth of 2% with particularly strong performance in the United Kingdom following new product introductions. The review of strategic alternatives for Education announced in 2006 is expected to be completed in the first half of 2007.
The company has announced that this group may be sold and the statement seems to indicate that will be completed by the middle of the year.
The company is looking to build on the gains they have made over the past two years to migrate revenue to online, attack their cost structure and reinvigorate product development. Margin improvement was seem 2006 vs 2007 - of 1.0pt - but they are suggesting their EBITA margin will grow from 17% to closer to 20% which is significant progress.
Wednesday, February 28, 2007
Another Day another Deal: Springer Science
Tuesday, February 27, 2007
Simon & Schuster Strong Finish
The publishing unit revenues were $252.5 and $807mm for the quarter and year respectively and operating profit was $38.9 and $78.0mm. A full year operating margin of just short of 10% is good going in the consumer market and is virtually unchanged from last years margin performance. Penguin reported yesterday (with Pearson) and their operating marging was slightly lower than S&S at 7.8% on higher sales.
Here is more detail from the CBS press release:
For the quarter, Publishing revenues increased 7% to $252.5 million from $237.0
million, reflecting sales from top-selling fourth quarter 2006 titles, including YOU: On a Diet by Michael F. Roizen and Mehmet C. Oz, Lisey's Story by Stephen King and Joy of Cooking: 75th Anniversary Editionby Irma S. Rombauer, Marion Rombauer Becker and Ethan Becker. OIBDA increased 8% to $38.9 million from $36.1 million, and operating income increased 7% to $36.3 million from $33.9 million, reflecting the revenue increase partially offset by an increase in bad debt
expense.For the year, Publishing revenues increased 6% to $807.0 million from $763.6 million in 2005 due to sales of top-selling titles as well as higher distribution fee income. OIBDA increased 5% to $78.0 million and operating income increased 4% to $68.5 million, reflecting the revenue increase partially offset by higher expenses, primarily resulting from an increase in bad debt expense and higher production, employee-related and selling and marketing costs. Publishing results included stock-based compensation of$1.9 million and $.5 million for 2006 and 2005, respectively.
Sounds like the AMS situation had some impact but on the other hand it looks like the bonus' this year were relatively better than last year.
Rumor Mill and Thomson Learning
It is all speculation and you have to believe the betting is on pure Private Equity. Bertelsmann is a diversified media company but clearly not adverse to paper based publishing operations and in this case they may be especially interested in the electronic revenue potential that Thomson Learning could offer. Wait and see.
Monday, February 26, 2007
Pearson Post Record Results
In the company's release they pointed to the strong performance of all groups including the education group (both higher ed and k-12) which has been strong all year, FT group ad revenues and Penguin. The Penguin numbers may be at or slightly higher than other consumer publishing companies may be expecting. Here are the highlights from the Pearson release:
In their preview of 2007, it is basically more of the same with essentially the same outlook for education that they gave this time last year with growth in the 4-6% range with margin improvement in School and professional. Penguin margins are expected to continue to improve (and at 8% are already quite good for trade) and the FT group is expected to continue to benefit from the reorganization that they have implemented during the past 18mths or so. Margins are expected to improve there.Record results. Pearson reports its highest ever operating profits (adjusted operating profit up 15% to £592m), earnings (adjusted eps up 18% to 40.2p) and cash (free cash flow up £2m to £433m).
Sustained growth and market share gains. School sales up 6% and Higher Education sales up 4%, benefiting from leading position in content, assessment and technology; FT advertising revenues up 9%; Penguin sales up 3% despite tough consumer publishing market. Stronger margins and double-digit profit growth in all businesses.
Pearson margin up a percentage point to 13.4%. Education margin up to 14.1% and profits up 12%; FT Group margin up to 17.3% and profits up 18%; Penguin margin up to 7.8% and profits up 22%. Higher returns. Return on invested capital up to 8.0% (from 6.7% in 2005), above Pearson's weighted average cost of capital; dividend increased by 8.5% to 29.3p, the largest increase for a decade.
Naturally, there is no word on future acquisitions but growth from acquisitions has been critical to Pearson in growthing top line and margins recently - although underlying growth has been significant as well. Look for Pearson to continue to be active on this front particularly as Thomson Learning and Harcourt will be distracted by their respective divestiture processes.
Friday, February 23, 2007
Deals News: Harcourt, Reed, Riverdeep, Borders, B&N
In what could be construed as incredibly bad timing Riverdeep's accounting firm Ernst & Young resigned last week because they believed that "professional relationship between us and the company has irretrievably broken down." Too early to tell if this will be significant at all in Riverdeep's acquisition binge but Moody's is suggesting it will review their rating towards the end of March.
Martyn Daniels at the Booksellers Association reminds me of speculation (muted I would suggest) about a merger between B&N and Borders. Personally, nothing seems less likely. I think if B&N were to make any store related acquisition they would buy Chapters in Canada. That is if the government would allow and B&N decide going International would be a good thing.
Thursday, February 22, 2007
Ebay Media Marketplace
The media buying process is fairly rudimentary, inefficent and process bound. Staffed by under-paid and over worked ad reps/media buyers who cut their teeth in the business by executing media buys the marketplace has been ripe for change for many years. Curiously, it is not the media venues (networks) who are avidly pushing this concept rather it is some high profile and powerful advertisers who see an opportunity to increase their reach and effectiveness while at the same time paying a market driven price for the space.
MediaPost describes what the first version (beta) version of this product will do:
A copy of a document detailing the beta version that's being circulated for review reveals just how detailed and thorough the system that eBay has developed is. The exchange can go in either direction, where a marketer submits an RFP for a network to respond to, or a network posts inventory out for sale.
A buyer's RFP allows it to request that a network offer it a deal based on desired GRPs or budget it has to spend. The buyer can also request preferences, such as a primary and secondary demographic target; length of spots it wants to air; programming genres it is willing to run in; flight dates; and dayparts it would
air in. Requests can even extend to offerings such as product placement,
corollary Web placement or billboards.
It is a little jargony but esentially the system will automate what is currently done via phone, fax and email with several potential 'networks' all at the same time. The media rep has to gather and analyse what is available from the multitude of available space, pick the right mix and then get back on the phone and do the deal. This system has the potential to inventory all available media space, with associated demographic and Gross Rating Point info, potential ad conflicts, etc. and enable a media buyer to create an appropriate plan that meets their goals and budget. Once the 'RFP' has been created by the media buyer they offer it to the market (networks) for negotiation/bidding.
Selected networks could then respond with an offer indicating GRPs and CPMs it
would go with, as well as the number of ad units and other benefits (product
placement, Web presence, etc.) it is willing to offer.
The program is very much in its infancy and will not replace any complex or first run (upfront) media buying that generally require one-to-one negotiation. Nevertheless, a considerable amount of media spending is made day-to-day across a wide array of outlets (scatter) and this Ebay led marketplace has the potential to create significant efficiencies and better ad buys for advertisers.
Wednesday, February 21, 2007
Pat Sommers
The tip came from Laura Dawson who has more details.
WorldCat Registries
There are many other elements and functions which at this early stage will not be
populated consistently however, the possibilities exist to create a 'community' site for library information. Registered librarians are able to designate an editor or administrator who can validate information about specific libraries and the idea is to create valuable information that can be used by patrons, sales reps, publishers, and anyone interested in library specific information.What is great about this is that it is a web product that gets a running start because it is taking information that existed for other purposes and reasons recasts it and creates something new that has both relevance and scale. In other words, there is reason to come to the WorldCat Registry because from day one there is material amounts of data in the product.
More details are available via Lorcan Dempsey's Blog.
Here is a sample screen shot:

Access is currently free but registration is required and for anyone currently subscribing to American Library Directory, this is an interesting alternative to keep track of.
This is the ALD marketing blurb for those who are not familiar with the product:
"This acclaimed reference guide has provided librarians and library users with the most complete, current, and easily accessible information on libraries across North America for over half a century. In this new Web version, subscribers can find and view detailed profiles for more than 35,000 public, academic, special and government libraries, and library-related organizations in the United States, and Canada — including addresses, phone and fax numbers, and e-mail addresses, network participation, expenditures, holdings and special collections, key personnel, special services, and more — over 40 categories of library information in all.Free registered users can view physical address information only."
I think the free WorldCat registry product may already have more information than the free version available from ALD. In the case of the OCLC product the librarian is both customer/user and editor so they become engaged with the product in two ways which can be quite powerful. It will be interesting to see how the social aspects of this product develop and whether we will see a real community of interest(s) develop with the introduction of further technology.
Tuesday, February 20, 2007
Disney Wants the Honey Jar Too
Monday, February 19, 2007
Missing News: NBC
Monday Presidents Day
I mentioned Worldcat Identities last week and Tim O'Reilly had a nice plug for it as well. He also mentions the idea of some incorporation into wikipedea which has come up in conversation. Adding a resource like Worldcat to wikipedea would aid both authority and functionality. On any particular subject an editor or user has an ability to link to an appropriate and unambiguous data record - in this case a Worldcat Identity - via very simple linking functionality. They will add value to the item they are creating in wikipedea and can be confident in the accuracy and depth of the Worldcat Identity they have linked to.
This video from Librarybites. has been circulating the web for a few days and it is very well done. It comes from Kansas State University and attempts to show how rapidly our relationship with search, content and networking can change. It is about five minutes.
Google has an impressive list of employee benefits which includes inviting authors to speak to employees about their books. They also video the meetings/presentations and they are available to all here. I often wonder why B&N and Borders don't video their author readings....
There is a lot of hype about Second Life and just about everyone is setting up shop there or making some political announcement (that they could make just as easily on Speakers Corner), but is it all over-hyped?. Are there really millions of people populating the site or are there millions like me who tried it once to see what it was all about and have never been back? That and some other comments make this blog post from TeleRead thought provoking for any library(ian) or publisher that is thinking of setting up residence on the site. As he comments:
If libraries decide to invest in SL properties, they may suffer a virtual repeat of the Florida land boom of the 1920s. None other than Electronic Arts, the giant games company, is already bent on competing with SL.Now don't get me wrong, experimentation is a great thing as is becoming simpatico with the customers/patrons you are trying to service. Just don't loose perspective that's all. Also the following warning is also relevant given the silliness around a recent children's book ban:
The real mainstays of SL at this point are gambling and sex. Don’t libraries and schools have enough problems with American ayatollahs? No prudery here, just practicality. A virtual library should be able to integrate itself well with the rest of its cyberworld, but currently, youth-oriented library areas have to be isolated from SL as a whole.The European Commission is continuing a policy debate on the scientific publication system. Here is the text of their announcement. No word yet on any outcomes from the meeting last week.
In the context of the beginning of the Seventh Framework Programme, the research Directorate-General has launched a policy debate on the functioning and efficiency of the scientific publication system, understood as the practices, rules and mechanisms defining the process of scientific publication, as well as its exploitation. A first contribution to this debate is the EC-commissioned "Study on the economic and technical evolution of the scientific publication markets in Europe". This Study provides an economic analysis of European scientific publication markets and makes a series of policy recommendations. A public consultation was held from 31 March to 15 June 2006 on the basis of this Study.
An important milestone in the policy debate on the scientific information system is the joint Communication on "Scientific information in the digital age: access, dissemination and preservation" , presented by J. Potočnik, Commissioner for Science and Research, and V. Reding, Commissioner for Information society and media, and adopted in February 2007. This Communication offers an entry point for discussion within the Council of Ministers, at the Member State level, and within funding bodies and intergovernmental research organisations. Issues to be addressed include dissemination and access strategies (e.g. Open Access), publishing business models (e.g. reader-pay, author-pay), and the relation between scientific publication and research excellence. A further impulse was given by a conference hosted by the European Commission on 15-16 February 2007 in Brussels at the Charlemagne building.
No doubt the above is similar to a study the UK Parliament conduct two or three years ago. I wonder if the outcome will be any different.
Sunday, February 18, 2007
Sunday Funnies and Giggles
"pronunciation Eoin" - This person would no doubt have a problem with "Dun Loughrie" or "Ioan"
"Lindsay Lohan bra size" - which had me stumped I must say, but I subsequently found some spam on one of my old posts.
"What is the city of publication for King Lear" - I think this was my favorite.
"Suicide poems + reasons why" - Not getting this one at all.
The good people from latex.org paid a visit but it was fleeting.
Lastly, Mrs PND is brunette otherwise I would be in trouble for showing this video for a Mercedes Benz ad. You may have seen it.
http://www.youtube.com/watch?v=nAs8_N_tDoE
Friday, February 16, 2007
Deals: Harcourt, Reed, Riverdeep
The acquisition signified in a way that 'we were on our way' and that the company was forging ahead and fulfilling a strategic direction. Acquiring Harcourt seemed at the time to be something of a hedge; it was entirely unclear where the journals electronic revenues were going to end up. Elsevier Science was nascent and the journals business generally was getting a lot of flak from the market regarding pricing, archiving and access. The market was in total flux. On the advertising side, the business publications of Cahners (now Reed Business) were also starting to show declines in revenues and this became a major management issue. In this context acquiring Harcourt seemed a strategic attempt to spread revenue risk a little wider.
In news reports over the past day or so some analysts have suggested the acquisition may have been a mistake but this seems doubtful to me. If anything, selling the business is smart business because in today's Reed business it is easier to see where their core strengths are and they are not in education. There is a little confusion (maybe it is me) regarding the original purchase price since some assets were shared between Thomson and Reed. I am not certain if the original purchase of $4.5bill reflects the total or the portion of assets Reed got from the deal. I believe it represents the total since Reed Education has a current book value of $2.6billion. (Reuters) .
The FT reports that Riverdeep are prepared to pay £2.0billion which would make for a tidy profit on the original purchase. (The FT article says the assets purchased were worth $2.0Billion and Reuters says $4.0billion). Analysts suggest the ultimate purchase price will be between £1.8 and 2.2billion and management suggests a sale to be announced towards July.
Thursday, February 15, 2007
Reed Elsevier Selling Harcourt Education
During his analyst presentation Sir Crispin Davis said that Harcourt increasingly differs from the core RE business units in market growth expectations and revenue model and that the opportunities for electronic and online revenues are "less clear". This also comes in the context of a 20% drop in operating profit for the unit.
For the year RE Group revenues were up 6% with 5% from organic growth and operating profit (despite the 20% drop for Harcourt) was up 9% for the 12mths. Here is the analyst presentation. Here is their summary press release.
Davis went on to say that the divestiture would probably occur in the second half of 2007 and that the proceeds would be divested to shareholders. The unit would be expected to sell for over £1.0billion however, the market for large educational assets is a little glutted right now.
Importantly, Harcourt is focused on k-12 so revenues are focused and RE expects some key adoptions during 2007 - for which they have invested pre-pub expense - and if those are won then the company will be more attractive (obviously). They may have some advantage but we will see. Private equity has to be the prime candidate for acquisition.
Here is a Times article.
Here is a local Oxford Mirror article
Interestingly, I can't help thinking that because Thomson and Elsevier compete aggressively in certain areas that RE do not want to loose out if Thomson take their Education Divesture dollars and go after something big that places them either in greater competition with RE or takes off the market a business that RE could also logically acquire. Their stated strategic goals are similar - grow electronic and recurring revenues and provide integration tools for users of their products. Both see the last part of that strategy as vitial because achieving it can raise switching costs for customers and effectively embed the publishers products into the work flows. Once one of these players gets a leg up the market could become hard to crack for the "looser." Despite suggesting they will return the money to shareholders, RE has no reason to do this and watch as a prime asset goes (cheaply) to one of their primary competitors. It may be fun to watch.
Wednesday, February 14, 2007
WorldCat Identities: Making Data Work Harder

Rolling Stones Rock n Roll Circus
There is more on the DVD than was broadcast on PBS including a longer set from Taj Mahal. He still comes to NYC regularly and I hope to check out a concert the next time he is here. Watching the show inspired me to start reading According to The Rolling Stones published by Chronicle Books. It has been sitting on the shelf for two years gathering dust. So far it is incredibly interesting to learn how they all came together in Ealing and Edith Grove. Curiously, there is virtually nothing in this book about the Circus. (So, here is wikipedia). Maybe there is something to the notion that they thought they were out-done by The Who. I thought The Stones sounded great but I did notice that there seemed to be no interaction at all between the members of the band. Poor Brian was dead 7 mths later.
Tuesday, February 13, 2007
Tuesday Take-up
Since it is coming into Academy Award season here is a list of popular books that haven't made it onto the screen. The list includes The Catcher in the Rye, Atlas Shrugged and Mein Kampf (Selznick produces and Hitch directs).
The posts on Hitwise are always interesting and here is a recent one on search terms.
From Public Lending Right in the UK is an annual listing of the most popular titles checked out of UK libraries. The winner is a children's writer Jacqueline Wilson.
Titles by the immensely popular children’s writer were lent almost two million
times between July 2005 and June 2006. Wilson’s nearest rivals are the adult
novelists, James Patterson, Josephine Cox, Danielle Steel and Ian Rankin who
each need to clock up at least another half million loans annually before they
can pose a threat to the "queen of library lending".
The article also notes the strength in lending of Dan Brown and J.K. Rowling which is interesting to me given the huge numbers of copies sold through retail, the abundance of second hand copies, continued strength of paper and mass market versions and yet they are still strong through the library. What does that mean?
When Martha Stewart published her first cookbook there were rumors the recipies hadn't been fully tested. I don't know if this was true or malicious however Mrs PND has never cooked a bad meal out of one of the MS books. Along the same lines is a cautionary tale from India. Do not include meat in your recipies.
Saturday, February 10, 2007
New Shoes, New Music
http://www.youtube.com/watch?v=EiXD2eAFxE0
As I understand it, Paulo's mum and dad live over a chip shop. Boy's definitely on his way.
Thursday, February 08, 2007
Thomson Learning: Let the Auction Begin
The Learning segment has been classified under discontinued operations and the overview comments are as follows:
Thomson Learning accounted for the majority of results in Discontinued
operations, with revenues of $672 million in the fourth quarter and $2.4 billion
for the full year of 2006. Thomson Learning’s operating profit was $161
million in the fourth quarter and $383 million for the full year. The
majority of Learning’s results were generated by the higher education, careers
and library reference businesses as well as NETg and Prometric, which were
announced for sale in the fourth quarter of 2006.
In their executive presentation the detailed results for Learning show decent results as compared with industry leader Pearson. Top line gained 5% and Operating Profit of $359MM was 6% better than prior with a very slightly higher operating margin. Pearson which will report at the end of February has consistently grown a 1-2pts better than Thomson over the past five years and their operating margin has steadily improved over the same period. It will be interesting to see the comparison at the end of the month.
It would be a big surprise if Thomson Learning does not go to a private equity buyer; there has been a lot of interest thus far.
Space Oddity
Wednesday, February 07, 2007
Blogging In Baghdad
Written in a flat, unemotional style, the entries relate the bombings,
blockades, shootings, threats, shortages and petty frustrations that make up
everyday life for the cadre of civil servants working at Iraq’s main cultural
and literary storehouse. A complaint that heating fuel prices are 40 times
higher than in the fall is followed by a report on the assassination of one of
the library’s bright young Web designers and the need to ask the government to
keep the electricity on.
Tuesday, February 06, 2007
Tuesday Trivia
Unless you live in the Arctic with a million penguins (and New York feels that way at the moment) you will have heard of the wikinovel idea launched last week. They are jumping on the wiki band wagon and you have to give them credit for the experiement but it is hard to see how anything coherent is going to come of this. The construct of the novel - some lone guy sitting naked in a room for months writing - doesn't travel far and certainly not to the collaborative mess that a wiki novel could create. That doesn't mean that creative expression that results in some type of novel-like work can't be done collaboratively just not the way we know it as yet. To me, it may end up like some weird combination of Penny Lane and Second Life.
With all my filters set to capture publishing related news and stuff the announcement from Library of Congress regarding a grant they recieved to digitize some text material seems to have slipped through. Thanks to the millionsblog for finding it. Millions also has a list with covers of all the Booker Prize long list books. (I've only read one of them).
In a similar vein, here is an article from BBC News looking at how governments could build libraries that preserve both digital and print versions of content. The article is a case study of the approach Canada has taken to begin preserving digital content.
Instead, the rules focus on online material that is considered to be in "publication" form......As part of the deposit process, publishers can choose between open access, which allows the public to view and download the publication through the internet, or restricted access, which limits public access to selected computer terminals at the LAC's main building in Ottawa. The LAC encourages publishers to select open access whenever possible.I am not sure if this means anything, but in 1980 a hardback cost $14.00 and Levis 501's about $20. This year the Harry Potter book will be $35 and 501's cost $45. What's the problem? Here is a perspective from SeekingAlpha. (Oh and I still wear the same size).
From mediabistro here is a longer review of the recent NY Public Google fest. Nothing hugely new here:
On went the catchphrases at the one-day event, carefully designed to tell attending publishing industry types that if you're not moving with the digital times, you're just not a 21st Century publisher. And as the day wore on, it seemed as if the audience had been completely lulled into submission by the overly positive mood and the plentiful supply of food.It was by many accounts quite the love-in. Was is Devo that said Are We Not Ludites?
And then, the moderator cut off questioning and the conference returned to its shiny, happy, artificial equilibrium. But for that single moment, the tension between Google and the publishing world was on display -– as well as the larger sense that there's no holding back the tide of change to an industry still struggling to understand what the digital age is all about. "The publishing industry has become increasingly irrelevant," said author and UnBound speaker Seth Godin in a recent telephone interview. "They need to stop thinking about selling paper, when the last big changes to that model took place over sixty years ago with popularization of the paperback."Here is a link to Seth's new book. Not too good for him are we?
And here is a longer review of the new Amazon product wiki from O'Reilly.
Amapedia is very tag focused. They can be used for searching, navigating, and comparing. There are two different types of tags fact tags (very similar to Flickr's new machine tags - Radar Post) and category tags )such as Real-Time Strategy Games). The fact tags are name:value pairs such as Number of Factions. Most articles when only have a couple of category tags; they are supposed to answer the question "What is this?".
That's it for a while.
Gather Looking Like Prodigy?
The announcement states that Publishers' Weekly will:
In the books category, Publishers Weekly will offer a behind-the-scenes look at the publishing world, while getting feedback from avid readers. "On Gather.com, our experts will not only share what they discover, but learn about what book lovers think and want to know," said David Nudo, publisher of Publishers Weekly, in a statement. "This information is invaluable to us."
And,
Columbia, meanwhile, is developing a site for Gather.com's music channel to promote the label's new releases and artists like Beyonce, Bob Dylan and John
Mayer. The label's recording stars will join in online sessions with users as part of promotional efforts around new records.
Currently the interface to books looks like this:
but I suspect they are going to have to change it to accomodate the new partners and give them brand space.
Perhaps more interesting to me will be the possible impact this alliance will have on Publishers' Weekly which has stuggled somewhat with their internet forays over the years and haven't effectively created a consumer profile for the brand. Their reviews program is the BIG exception to this. Meanwhile the significant B2B segment of their business has been diminishing which has place significant pressure on the company. If they are able to channel consumer interest, information, content, etc and extend their brand with consumers but also importantly creating new value for their B2B (Publishers, Booksellers) then they will really have accomplished something.
Friday, February 02, 2007
It's not Mr.Dewey's Search Engine Anymore
Read/WriteWeb is a popular weblog that provides Web Technology news, reviews and analysis. It began publishing on April 20, 2003 and is now one of the most widely read and respected Web 2.0 blogs. Read/WriteWeb is ranked among Technorati’s Top 100 blogs in the world. The site is edited by Richard MacManus, a recognized thought leader in the Internet industry.
I say check it out. And specifically look at this recent article on search engines - that is the other 100. It is facinating to look at some of the examples he links to and I recommend MsDewey (who is significantly more attractive than you might imagine) and LivePlasma. If you get to MsDewey search for book or books several times in a row and see what happens. It is facinating. (Note: I just went back to MsDewey is seemed to get stumped; regardless, a benefit is to remain transfixed on MsDewey).
Thursday, February 01, 2007
Feedburner
Here is the new RSS url:
http://feeds.feedburner.com/Personanondata
If you are using MS Explorer you may also be able to see two RSS choices clicking on the RSS icon on your nav bar. Please switch over from the .atom feed to Feedburner
Thanks,
Wednesday, January 31, 2007
Jimmy Wales Discusses Wikia Search
The meeting at NYU was both a official class meeting and a hosted meeting of the FreeCulture society. Purely by chance, I found out about it by glancing at a copy of AM New York this morning as I came back from a breakfast meeting. This is the great thing about New York that these types of things go on all the time like no other place on earth. (I will have another post tomorrow about another meeting I attended earlier in the week).
Wales suggested that he was taken aback by the attention given by the media world on this initiative and he claims that he accidentally dropped the hint about it at the end of last year. I have some doubts about the story. As he explained the search ‘tool’ will become a legitimate competitor to the commercial providers particularly Google and Yahoo. He even suggested that some second tier search providers have approached him to offer assistance and support and he reasons that these companies recognize that a legitimate competitor to Google et al is good for all the non-major players. He didn’t directly state that an objective is to make basic search a commodity but this does seem to be the central objective of the initiative. Value-added services then would ride along or on top of basic search thereby providing unique business offerings.
With respect to the three core criteria he views as essential to the initiative all algorithms will be published, testable and researchable which supports his transparency goal. Establishing a participatory environment will be dependent on the relevancy and usefulness of the engine. As one student suggested, if the tool sucks then no-one will participate to which Wales noted that he is in the process of hiring the best researchers in search technology and is well aware that the first release has to be impressive. He also went on to say that they want to include the best elements of wikipedia participation coupled with the trusted network of key participants. Within wikipedia there is a core group of 1000-2000 contributors who are unlike gatekeepers and more like collaborators. Lastly, the search tool will be free which he defined by reciting the four freedoms of software. These are the ability to copy, modify, redistribute and redistribute with modifications.
Other than the fact that I was in a room full of under-graduates and feeling very old this was a very interesting discussion. Questions towards the end reflected concern over privacy issues and why Google and the other services are not ‘free’. I was curious about why the wikipedia model hasn’t yet transferred well to the world of educational publishing and journal publishing because so far those initiatives appear indifferent but I didn’t get the chance to ask.
Tuesday, January 30, 2007
Google TV: Is it for real? UPDATED
Update:
I think the following YouTube video from the same source as the GoogleTV episode proves GoogleTV is a hoax. This video is about recharging alkaline batteries and to quote "it is called electrical tape because it conducts electricity."
Well it was fun to think about....
In the run up to the law suit by US publishers against Google, an article was circulated (and I forget where it was published) to all of us on the AAP board which described a meeting between Google execs and NBC in Los Angeles. Apparently, Google had been storing the NBC feed for months and presented their proud new ideas for TV over the internet. NBC were not similarly impressed and the idea was buried.
Now comes word of GoogleTV which if true could be the resolution of this earlier idea to deliver on-demand television to your computer. How cool will that be. Just think in six months you can use your browser on your new iPhone to watch the program of your choice when and where ever you want. Extending that your iPhone becomes the distribution device for your own tv, radio/music, movie channel. The YOU channel.
Again it could all be a well constructed hoax. It could also be a hoax but very close to the truth.
Here are links:
http://www.youtube.com/watch?v=wNjlGm-YIKg
http://www.youtube.com/watch?v=7MulSMSJV-U
Thursday, January 25, 2007
Joining the Network
Ebay and Amazon are the most obvious network level players but others are becoming increasingly prevalent and increasingly have at their core a set of integrated web service applications. Google, for example is known mostly as a search platform; however, they are investing in many types of applications from calendar functions, to spreadsheets, to blog software that in effect creates a potential network level desktop. It is entirely possible in the Google environment to become completely untethered from your traditional 'physical' desktop and increasingly this will be the way people work. As a consequence of the developing Google network we will benefit from more integration of communication - calendars, email, blogging, - between users. Who knows the level of integration that may result once browser type, hardware, productivity application don't matter.
The beauty of the Network level application is that it can function as a component of work flow or as the work flow itself. The applications are built to standard specifications and are interchangeable, upgradeable and reusable. Web services applications are the most common facility by which modular component software applications are brought together to produce a work flow application. By definition, these web service applications are not tied to any specific operating system or programming language.
Amazon.com has aggressively promoted (and was an early adopter) of web services. Most online booksellers would face much higher monthly operating expenses and would also not have access to other seller tools (comparative pricing, availability information, etc.) were it not for Amazons web services. Simple cover art is available to all online retailers via the Amazon web services widget. It would be difficult and time consuming for a small book retailer to scan and upload cover art and the fact that this and many other functions have been removed from the local store level to the network is an example of a network effect. Anyone who has sold on Ebay over the past ten years will recognize how the process gets easier and easier as Ebay has added content, applications and services that the average garage sale seller could never develop by themselves. As Amazon has done, they to have developed web services applications that others can use on their own auction or retail sales sites.
The Network effect is coming to the library, publisher and bookseller market. (There is still some opportunity in book selling despite Amazon). It is interesting to think further about the Network effect on publishing but I think that we are seeing the first stages of a radical change in publishing with the development of self-publishing houses (AuthorHouse, Lulu, iUniverse), publishing applications (Blurb, Picaboo) and blogging tools (wetpaint, Wordpress, Blogger) all of which represent a very different way of publishing. I think it is the beginning of the death of publishing as we know it but by no means the death of publishing.
Similarly, in the library community software vendors sell expensive software implementations for local catalogs (OPAC) which are proprietary and often islands of information with only minimal integration with the outside world. More often than not the applications themselves are filled with features and gee wiz stuff that no one needs or uses. Rise up to the network level all of this functionality and the libraries is able select the applications they want as components parts and assemble them as they please. Due to the increasing strength and decreasing cost of communications and bandwidth the library can run the critical tools they need via a set of network level applications. Importantly they do this without a large expensive investment in hardware or software and they get continual access to software development improvements.
It is an interesting time and I think I will think more about the network-level impact on book selling and libraries.
Wednesday, January 24, 2007
Spooky Cry for Help? All the books Art Garfunkel has read
Interestingly, 52 people have tagged this web page in delicious. I think most are in awe.
A new take on personal cataloging of books etc is gurulib which I have yet to test out but it has a similar purpose to Librarything. We all need a little competition. This site is free for the moment although Librarything was hardly expensive.
Tuesday, January 23, 2007
Google News and Other News
Pearson were cagey all year about their results suggesting that the second half of the year would be much harder. Well they are on target for record earnings. Their stock declined.
Here are the Edgar Award nominees.
The French apparently love James Bond.
There was a Google love in at the New York Public Library last week. Predictably publishers are wrong and Google is right.
If you see the local Barnes & Noble windows filled with Chocolate in a few weeks you will know why. In these competitions they never say where the people who live in New York get to go if they win. I guess the presumption is we get to go to Omaha. It's just not fair.
Monday, January 22, 2007
Archiving Special Collections.
As more and more libraries start to digitize their 'special' collections (and not via Google either) it will be interesting to see how they 'display' these collections in the on-line context. I believe small thriving businesses will be developed that help libraries create on-line or electronic shows. The online version of the material that typically shows up in the glass display cases in the library lobby.
It is great to have all this 'special' collection material available for research and access but for the casual library patron some filtering and explanation/analysis is important for enjoyment and value. This is why I think you will find digital archivists selling their services to libraries to create these representative packages or shows of the material. These digital archivists will present the best parts of the collection and 'design' or curate these online shows. Viewers will be able to access the material via the web but perhaps they could also view the material in the library via web enabled kiosks.
There is so much of this material coming on-line. (OCLC has three different programs designed to collect this material). I think only a small percentage of stuff has thus far been digitised and we run the risk of having a glut of material that if not organized and presented to the typical patron may never be seen. This would be akin to the local public library collection that sits undisturbed in the special viewing room that isn't open to the public.
The need to create logical presentations of this valuable material does draw attention to the importance of the local librarian in selecting or editing this material. The process also creates new opportunities to add to the material in ways that perhaps was discouraged or difficult. This is especially the case with oral histories. Creating a visual respresenation of the material in a special collection and then encouraging some long lived local patrons to add their vocal history to the presentation would add significantly to the relevance and importance of the collection.
As I said, I think this is coming because I think there is an obvious need for services to support this type of activity.
Friday, January 19, 2007
Private Equity Deals
With Thomson Education on the market, Houghton Mifflin sold, Wiley purchasing Blackwell Publishing, Wolters Kluwer Education for sale, there is a lot on offer and this week the rumors were flying around Pearson. Pearson has made no public comments regarding selling any of the company yet the PE bankers and the equity analysts think something is a foot. The Pearson stock closed at a four year high today (CNN). No doubt this is very pleasing to all the management options holders.
Last week, Reed Elsevier was also touted as a non-too-obvious PE candidate and it would seem you are not worth your salt as an equity analyst in the media industry if you don't declare one of these conglomerates as an ideal candidate for Private Equity players. Reed have a complicated ownership structure which, while simpler than when Reed and Elsevier were put together, seems to be a concern. When (or if) it comes down to it I doubt this will be an issue. As with Pearson, I suspect that the managers with options will be very happy with the stock price escalation. Last week the stock was at 600p now it is at 611p. When I was at Reed in 2000, the stock was over 700p and this was before they purchased Harcourt and the significant slump in advertising revenues that hit their trade publications business. Point is, I wonder if a valuation will push the share price up a lot higher?
In recent years Pearson has been chastised for under-managing a collection of valuable publishing assets but this died down a little over the past 12 months. Their financial results and forecasts have been on target and results in the Education group have been particularly strong. This is especially true when Pearson education is compared to Thomson in terms of acquisitions made, growth rate and operating margin. In Thomson's defence, I am sure that it has been difficult operating in within the Thomson corporate environment when reservations may have been expressed about Education for a while. Thomson relative to Pearson has probably lost momentum which a sale will quickly fix and we will see a renewed Thomson Education as a result.
Currently, the PE bankers see an opportunity to buy a large 'holding company' put some lipstick on it and sell parts off in short order. In selling the parts - The Financial Times, Pearson Education - they would expect to make a killing. (Pearson Education could be split into College and School). Management is clearly not happy with the current debate but if an offer is made then they obviously need to consider it. What may result is some appeasement resulting in the sale of one of the tastier chunks. As I predicted for 2007: Murdoch will buy The Financial Times.
Thursday, January 18, 2007
News From Australia
When the judge then suggested that he live with his grandparents, the boy cried out that they also beat him. After considering the remainder of the immediate family and learning that domestic violence was apparently a way of life among them, the judge took the unprecedented
step of allowing the boy to propose who should have custody of him.
After two recesses to check legal references and confer with child welfare officials, the judge granted temporary custody to the English Cricket Team, whom the boy firmly believes are not capable of beating anyone.
Mr. Charkin had a recent post along the same lines. Dispair.
Spare a thought for my parents who have taken to spending the English winter in Australia but this year they got torential rains on Christmas Day, the Cricket team sucked and this week brush fires caused a power outage in Melbourne. Oh well...
Tuesday, January 16, 2007
Worst Cover Award & Australia Week

And I think that is prettty ordinary. I think they are selling linens not excellent Aussie food. And using the light green lettering means you can barely make out his name against the grey background.
Incidentally, it is Australia week here in New York; who knew? Many events but no book related ones and doesn't that figure. With a little concentration you could just about see that table top display of Aussie authored books at the Union Square Barnes & Noble. I guess not...
More thoughts on book covers generally visit http://nytimesbooks.blogspot.com/index.html.