You can tell the sort of industry we are becoming by the language we use to describe what is happening. Does an industry which refers to data “mining”, or entity “extraction”, seem to you to want to align itself to the softer values of literary publishing? Our senior management teams are now replete with data or content “architects” working alongside data or process “engineers” to ensure that we handle data as content in the right way for today, while staying “agile” in terms of new product development. We are “solutions” orientated because now, for the first time in history, we really can tell how our content is being used, what problems users commonly encounter, and how we can ease their processes, help their learning, improve their workflow or deepen their insight by adjusting, or helping them to self-adjust. The way in which data-as-content is recorded in our systems creates new dataflows which are all about those reactions. We used to throw that data away, some of us, because we could not “read” it. Now the “exhaust data” blown out of the back of our machines when they are running full tilt, could be just the place to pan for gold.

And just as diesel is apparently more noxious than petrol, and heavy vehicle than modest family car, there are clearly many different varieties of exhaust. I have always worried greatly about the use to which events organizers have used the rich data derived from registrations, exhibitor profiles, attendee tracking and preference listings. Given privacy constraints there is clearly scope here to add third party data from venues and elsewhere and go beyond the needs of an individual show and into service development for the target group more generally. I have been told in the past that there is too much data to handle or too little to give significant results – all excuses which become increasingly pale in the age of data. And the same opportunities exist in the creation of usage data in online services universally.

But much of the exhaust data potential is less obvious. Jose Ferreira, founder and CEO at Knewton, notes in his latest blog (http://www.knewton.com/blog/ceo-jose-ferreira/):

“OER (Open Education Resources) represents a tectonic shift in education materials. Try typing “mitosis” into Google. Almost every search result on the first few pages is for OER exploring the process of cell division. The same is true for nearly any other concept you type in: “subject-verb agreement,” “supply and demand,” “Pythagorean theorem” — you name it. And what you can find today on the Internet is probably less than one tenth of one percent of the OER out there. Most is trapped on teachers’ PCs.”

And I bet he is right. Services already exploit this exhaust from the teaching processes of individual teachers (TES Connect, www.teacherspayteachers.com). But Jose’s argument goes further. If you are able to employ the OER (what I think I used to call the “learning object” then you are able to see who stumbled over it, what the exhaust data of assessment shows about understanding and accomplishment of learning objectives, and then you should be able to move towards a genuinely adaptive learning that understands learning difficulty and recognises speed of learning acquisition.

Another form of feedback loop came to light this week in a note from f1000Research, the Open Access service in STM which is clearly bent on adding fresh layers of meaning to the expression “on the fly”. Using studies of Drosophila Melanogaster (fly – geddit ?) in his paper on genetic variations in different populations (http://f1000research.com/articles/3-176/v1.) the Professor of Neurogenetics at Regensburg and f1000 release for the first time an article in which not only are the professor’s data changeable as fresh evidence emerges, but other labs are invited to add their own data to one part of the data to get a comparative view. This article is then a “living” entity, showing fresh results – “on the fly” – every time it is opened. It also, of course, allows every lab to make comparative studies of its results to the Regensburg results, introducing a fresh instance of the “repeatability” principle to peer review. And the interactions of other labs with the article produces a fresh stream of exhaust data, some of which may itself be citable in this instance.

Like “robotic milking”, the new craze in farming, this should be seen as a great gift to publishers. A robotic cash cow that milks itself! But I fear it will be very specialised in its applications, since looking a gift cow in the mouth, or swatting a data fly, are more traditional pastimes for those-once-called-publishers than searching for gold in the exhaust.

The past week demonstrates triumphantly that human ingenuity knows no bounds in finding fresh ways to invest in content and software marketplaces, with appetites apparently undaunted by the many examples available which show how hard it is to get growth in traditional media and how hard it is to get margins in the newer variety. Pearson’s results, despite positive underlying trends and signs of more normal market conditions in the US, failed to set investors alight. The mean machine at Amazon, trying to cope with the margin constriction of online retail while desperately seeking other sources of value add profits in logistics and AWS got a chilly response from markets. And yet, during the week, a number of unrelated deals demonstrated a continuing hunger for media and information marketplace assets that belies the difficulties and provides new exemplars of market reconstruction and consolidation at work.

We are all used of course to the cellular division process employed by McGraw Hill and News Corp to create a Good Bank/Bad Bank division of assets that enables the bit with prospects to be revalued and become a new growth point. This week we could call this the Murdoch Gambit, as Twentieth Century Fox, aka Good Assets, went after Time Warneur while the latter was in process of casting its Bad element, aka Time Inc, overboard. The acres of screenspace devoted to discussing this rather obscured remarkable goings-on in the wholly less glamourous but once far more profitable field of building and construction industry information. A few weeks ago McGraw announced that it was selling its properties in this area, clearly not relishing the build to a workflow-based BIM marketplace, with market players only slowly migrating to towards a new world of data handling. Last week Reed Elsevier went one further, by selling its RS Means building costs division to Warburg Pincus (who own the competitor, Gordian Group) and including 49% of Reed Construction in the deal. This demonstrates two interesting possibilities: Reed really are a portfolio player now, with a clear strategy on the rules of investment engagement and a determination to let others share the risk when retooling and re-investment becomes necessary; and private equity is becoming recognised again as a good place to go for those re-investment activities. Warburg Pincus in particular can point to their years of patient market and service development work at GlobalSpec, now a key element in the IHS positioning at the front of the Engineering information market.

And this was not the only interesting news from Reed Elsevier. For a start, it’s revenues are now 82% “digital”, a figure that only financial analysts seem to care about, long after the rest of us had assumed the figure was 100%! And for a moment midweek we could have been forgiven for thinking we were returning to the “Happy Families” consolidations of the 1990s as Reed (Lexis) sold its Polish law assets to Wolters Kluwer, who with equal solemnity sold their Canadian assets to Lexis. It all made perfect sense. Neither Thomson Reuters or Lexis ever made Germany work, yet WK did. Poland was the same, only smaller. Canada was much more comfortable for Lexis, which had considerable assets there already. One can only wonder why rationalisation sometimes takes so long. Whatever the answer, looking at the assets as a portfolio investment manager and not as a committed investor in certain markets and geographies certainly aids the thought process and clarifies the rules. One of Reed’s mantras in recent years has been reducing reliance on unstable advertising marketplaces. This week’s results indicated that advertising is now down to 2% of gross revenues. Mission accomplished then, since Reed are clearly not interested in the marketing services environments which will succeed old-style advertising, and which created what for me was Deal of the Week: the sale of Bizo to LinkedIn. When we look back for benchmarks of the recognition of marketing services online as a wholly new service concept, then Russell Glass’s company, itself a breakout from ZoomInfo, will be the measure.

So should we expect more weeks like this as the industry vertically restructures and consolidates? Will Wolters Kluwer seek a revaluation of its wonderful health portfolio by floating it separately from the less vibrant business, law and tax divisions. Informa, who recently announced a very logical and much more service-centric structure, could take a similar view, since the relationships between, for example, their academic research and their trade exhibitions businesses are pretty tenuous. My guess however is that the real control here will not be the investment savvy of the suits at head office, but market tolerance and utility. In markets where data availability inside workflow driven models becomes the expectation, and each offering must be content complete in order to compete, there will seldom be more than two competitors. The portfolio investor decision is the oldest on record: stick, or …twist.

« go backkeep looking »