Recommind and LexisNexis Team to Generate More Revenue
May 9, 2011
Recommind has moved from eDiscovery to enterprise search and back again. The latest tactic in the firm’s growth strategy is a tie up with LexisNexis. This unit of Reed Elsevier has emerged as one of the leading non US owned firms delivering legal information in America and elsewhere. LexisNexis has been working overtime to cope with changing buying patterns among consumers of high end commercial online content. LexisNexis has branched into new markets, including data analytics and various legal back office services.
Recommind announced in “Recommind Forms Strategic Alliance With LexisNexis for Hosted eDiscovery Service” a new deal with LexisNexis. The idea is to apply the well known online dream of 1+1=2, maybe 3 or more. The news announcement said the tie up was “A strategic hosting and sales alliance” the two companies promises “rapid deployment of [Recommind’s] Axcelerate On-Demand” and LexisNexis’ Hosted Litigation Solutions group.
The goal we learned is to:
“provide more options and greater flexibility in discovery. . . dramatically reduce the costs and timelines associated with document review and analysis as part of litigation and regulatory investigations.”
In addition, the alliance offers “top-tier infrastructure capabilities, globally diverse IP network,” as well as security against disastrous loss events.
The business alliance will answer 2010 customer demand “for Axcelerate On-Demand with Predictive Coding.” It is designed to offer corporations and law firms to meet their review information needs, budgetary demands, and critical timelines for all of their cases, no matter how complex, changing “’the way corporations and law firms manage litigation in 2011 and into the future.’”
Sounds very good. Now we have to wait to see if there is an impact on other competitors in the over-crowded legal sector and if the river of revenues pulls a swollen Mississippi or maintains the current flow.
Jane Livingston, May 9, 2011
Freebie unlike commercial online legal and news information or special purpose search solutions
New Spin for OmniFind: Content Analytics
May 2, 2011
IBM has dominated my thinking with its bold claims for Watson. In the blaze of game show publicity, I lost track of the Lucene-based search system OmniFind 9.x. My Overflight system alerted me to “Content Analytics Starter Pack.” According to the April 2011 announcement:
The Starter Pack offers an advanced content analytics platform with Content Analytics and industry-leading, knowledge-driven enterprise search with OmniFind Enterprise Edition in a combined package. IBM Content Analytics with Enterprise Search empowers organizations to search, assess, and analyze large volumes of content in order to explore and surface relevant insight quickly to gain the most value from their information repositories inside and outside the firewall.
The product allows IBM licensees to:
- Find relevant enterprise content more quickly
- Turn raw text into rapid insight from content sources internal and external to your enterprise
- Customize rapid insight to industry and customer specific needs
- Enable deeper insights through integration to other systems and solutions.
At first glance, I thought IBM Content Analytics V2.2 was one program. I noticed that the OmniFind Enterprise Edition 9.1 has one set of hardware requirements at http://goo.gl/Wie0X and another set of hardware requirements for the analytics component at http://goo.gl/5J1ox. In addition, there are specific software requirements for each product.
The “new” product includes “improved support for content assessment, Cognos® Business Intelligence, and Advanced Case Management.”
Is IBM’s bundling of analytics and search a signal that the era of traditional search and retrieval has officially ended? Base image source: www.awesomefunnyclever.com
When you navigate to http://goo.gl/he3NR, you can see the different configurations available for this combo product.
What’s the pricing? According to IBM, “The charges are unchanged by this announcement.” The pricing seems to be based on processor value units or PVUs. Without a link, I am a bit at sea with regards to pricing. IBM does point out:
For clarification, note that if for any reason you are dissatisfied with the program and you are the original licensee, you may obtain a refund of the amount you paid for it, if within 30 days of your invoice date you return the program and its PoE to the party from whom you obtained it. If you downloaded the program, you may contact the party from whom you acquired it for instructions on how to obtain the refund. For clarification, note that for programs acquired under the IBM International Passport Advantage Agreement, this term applies only to your first acquisition of the program.
Watson and Its Methods
April 30, 2011
In the Fast Company article by Ariel Schwartz, IBM is partnering with Caltrans and the University of California at Berkeley to create a “personalized commuter forecast” for individuals living in large cities and high traffic areas. This p.c.f. will be dubbed “Watson” because everyone needs a trusty sidekick.
Schwartz’s article “IBM Will Go All Watson On Your Commute, Keep You Out Of Traffic” explains that the program, which is still in its prototypes stage, will use the GPS on your phone to analyze traffic on your daily route to traffic and suggest the route that will get you to work the fastest. (According to IBM you’re still S.O.L. if you live in an area with no or few alternate routes, go figure.)
Instead of slogging through the traffic, your phone recommends that you drive halfway to work, park in the BART parking lot, and take the subway system the rest of the way. If you leave now, you’ll make your way through traffic just in time to catch the next train to work.
I feel like that’s a little too good to be true. Though IBM’s willingness to utilize already in place technologies such as the road sensors used by Berkeley and Caltrans is admirable (no wonder they’re partnered.)
Let’s face it. It all boils down to money, IBM is hoping to generate cash based upon sales to different transportation entities, merchants who would build along newly used transit systems and sales from advertisements–in exchange for IBM knowing your every location. IBM is not the government. Some people may take the position, “What’s one more conglomerate tracking user behavior?”
Leslie Radcliff, April 30, 2011
Freebie
Libraries Like the Snow Leopard May Be Endangered
April 29, 2011
We should have known this day would come. At silicon.com, Peter Cochrane blogs the question: is it “Time Libraries Were Shelved?” He asserts:
“Does it matter anyway? The debate goes on but I must admit that I cannot remember the last time I visited a physical library. I give away far more books than I read.”
Humph.
His questions were prompted by cuts to public libraries in the U.K. That story is already in progress here in the U.S. Are we about to become an illiterate society?
Budget woes pushed the trend, of course, but perhaps it was inevitable. Many feel that books are simply an outdated technology. I see their point but, at the risk of sounding outdated myself, there’s just no substitute for a real book in my real hands.
Sure, I can curl up in my comfy chair with an eReader, but it’s just not the same. I enjoy the different weights of different books, the feel of turning a real page, even the smell of ink and paper. And those sensations are part of what enticed me to become a reader in the first place! I can’t be the only one.
Besides, without libraries, how will folks get free access to knowledge? Ben Franklin would be very disappointed. Online is useful, but it does not answer * every * question a research may have.
Cynthia Murrell April 29, 2011
Freebie
Amazon: Insight into Search, Engineering, and Cloud Computing
April 28, 2011
In order to locate data, one must be able to search for it. If search does not work, data are lost. Seems obvious but one of the consequences of the Amazon cloud outage was that I had to think about the online big box store again. Amazon is, to me, a convenient way to get books and buy a gift or a replacement BlackBerry battery. Even when the A9 service was a priority, Amazon’s ability to make information findable was hit and miss.
Even today, I have a tough time thinking of Amazon as giant, reliable, low cost information utility. I have difficulty finding lists of books “about” a subject. Sometimes I stumble upon this user created content; other times, I have no idea how to find this useful information. When I want a book, I don’t know how to NOT out books that are available from those that will be published in the future. I cannot find information about the credits I “earn” when I buy Kindle books or products using my Amazon credit card. The snail mail coupons I used to get have disappeared, and I don’t have a clue about “finding” this information.
Several years ago, we did a close look at how Amazon handled glitches. The information was not that different from other companies we had examined. However, one approach was interesting. When an outage took place, a small team was assembled to figure out what happened and to fix it. This approach has its upside such as speed and fluid problem solving. The downside, in my opinion, was that solutions could be ad hoc. In my view, the next time a problem cropped up, the Amazon approach I probed three years ago meant that the next problem solving team had to figure out what the previous team did. No big deal until the problem of figuring out everything consumed lots of time.
We are not using Amazon Web services. Call me old fashioned but I prefer to have data storied on local devices with appropriate backups on media in an off site location.
For another, unrelated project we ran a series of tests in 2010 on the take up of the phrase “cloud computing.” What we learned was that the actual traffic generated by the phrase “cloud computing” was far less than our client anticipated.
After a six month text, we concluded:
- There was a large amount of information about cloud computing from a bewildering range of vendors big and small
- The interest in cloud computing was less than in some other words and bound phrases we tested
- The information about cloud computing was a cloud of semantic fuzziness; that is, it was difficult to pin down specifics within the documents written about cloud computing.
What happens when you combine a retail store with a cloud computing service? You get an anchor point. Amazon becomes associated with certain words and phrases, but these may not have much meaning. Examples range from acronyms from S3 to EC2.
What happens when a company which has associated itself with this difficult to define subject has an outage? The problems of Amazon immediately diffuse across other products and services available in the cloud.
You can see an example of this semantic drift in “Amazon: Some Data Won’t Be Recovered after Cloud Outage.” The article points out that the Amazon “outage” has resulted in data that “won’t be recovered.” The problem is no one that Amazon and its customers must resolve.
Amazon’s close association with cloud computing has made the Amazon incident the defining case for the risks of cloud computing. Even worse, unrecoverable data cannot be found. Search and retrieval does little good if the data no longer exist. Services which depend on their customers locating information are effectively stranded. Those affected include “Quora, Sencha, Reddit, and FourSquare.”
So what?
This problem at Amazon provides some insight into the firm’s engineering approach. In a larger arena, the close association of Amazon with cloud computing has had a somewhat negative impact on the concept of cloud computing. To sum up:
- You can’t find information if it is not “there”
- Amazon’s engineering methods are interesting and may give some companies some additional analysis to perform
- The impact of the outage has created some pushback for other cloud computing vendors.
Will this be a defining moment for Amazon? Probably not, but it is an interesting moment. Non-recoverable is a disturbing notion to those who have to find a fact, entity, or a concept. Amazon has figured out some aspects of eCommerce. Other areas warrant additional investment which may be why Amazon’s costs are skyrocketing.
Stephen E Arnold, April 28, 2011
Freebie
Google Tries Social Inputs for News
April 27, 2011
In the midst of farming and Panda bashing, pundits and poobahs did not pound the drum for Google’s social ranking tests. Nieman Journalism Lab did examine one of Google’s projects in “The Layered Look: How Google News Is Integrating the Social Web.” Google keeps plugging away with social features, of course, but this one is a little different. It’s a layer of social networking- based rankings integrated into the Google News page. Writer Simon Owens explains:
I spoke with Jeannie Hornung, a spokesperson for Google News, about the various ways Twitter and other social media platforms are used by the aggregator. She pointed first to the ‘Most Shared’ section, found on the right sidebar near the bottom of the main page. At the time of this writing, it displays headlines from a mixture of blogs and more traditional news outlets and allows you to sort the most popular stories by day, week, and month.
Owens was unable to get straight answers as to how exactly some of the ranking works. Twitter feeds factor prominently, but it looks like the share feature also plays a part.
The component may seem superfluous to some. But for those of us looking for the most important (or at least most shared) news stories in a limited browsing time, it can be a helpful tool.
One thing is certain, some Googlers will respond to the money carrot that Google is dangling in front of its engineers. Will those crunchy treats result in real progress in the social media sector? Facebook, according to some, is on a roll. To date, Microsoft, not Google, has a stake in Facebook. More significantly, when making news, a Googler in Egypt mentioned Facebook as one of the conduits for the social movement. Will that type of nuance be hooked into the new social news algorithm?
Cynthia Murrell April 27, 2011
Freebie
Vertical Blog: A New Angle for Online
April 27, 2011
Our Overflight intelligence system tracks certain types of information. There are some basic Overflight services available from the ArnoldIT.com Web log. We have other systems running as well. One of these identified a new blog called Backnotch. Published by Jean Glaceau appears to cover one narrow segment of online information; namely, transactions related to Angola. What’s interesting about the publication is that the content appears to be summaries of publicly-accessible information. The Backnotch service is similar to a traditional abstracting service. The principal difference is that the contributors are offering some broad editorial comments. These comments, plus the collection of articles, comprise a useful resource for anyone looking at what types of open source information cover certain activities associated with Angola and related topics.
According to the About page of the blog:
In my first week of work, I decided to narrow my focus to a handful of issues which are covered in the open source literature. The information I located struck me as similar in some ways to a fictional story or a Hollywood film. Going forward, I want to continue to explore how the open source information follows a particular story and what entities surface in those stories.
The publisher is Jean Glaceau. When we did a couple of queries for him, we found a number of individuals in the hit list. We were not able to determine which Glaceau was running the research project behind the information service. We wrote the email address for the blog, but we had not received an answer as we queued this story for publication.
We checked out the search engine for the service, and it appears to have a backfile of about 60 articles. If Mr. Glaceau keeps up his current pace of content production, the service will generate about 50 to 60 stories each month. Our view is that online has moved from vertical search to vertical “finding” services.
We will check back with Backnotch in a couple of months. Worth a look.
Stephen E Arnold, April 27, 2011
Freebie
Ducks and Alphas: Wolfram Alpha and DuckDuckGo Unite
April 25, 2011
“Wolfram|Alpha and DuckDuckGo Partner on API binding and Search Integration,” touts Wolfram Alpha’s own blog. Both organizations have brought something unique to the Search universe, so we’re interested to see what comes of this. Will it be more agile than a Google and Godzilla would? (Googzilla?)
Wolfram|Alpha’s Computational Knowledge Engine not only retrieves data but crunches it for you—very useful, if you phrase your query well. Play with that here.
DuckDuckGo’s claim to fame is that they don’t track us; privacy champions like that. A lot. The site provides brief info, say from a dictionary or Wikipedia, as well as related topics at the top of the results page. It’s also blissfully free of advertising clutter. Check that out here.
According to the Wolfram Alpha blog, they are combining the Wolfram|Alpha functionality with the DuckDuckGo search:
So what does this new partnership mean for you? If you are a DuckDuckGo user, you’ll start to notice expanded Wolfram|Alpha integration. DuckDuckGo will start adding more Wolfram|Alpha functionality and datasets based on users’ suggestions. If there’s a specific topic area you’d like to see integrated into DuckDuckGo, your suggestions are welcome.
And for developers, DuckDuckGo will maintain the free Wolfram Alpha API Perl binding. With that, you can integrate Wolfram|Alpha into your application. Keep in mind that InQuira and Attensity are “products” of similar tie ups.
We’ll enjoy watching the progress of this hybrid beast.
Cynthia Murrell April 25, 2011
Freebie
Autonomy Financials via a Mid Tier Consultant
April 23, 2011
In my email this morning, was a short item that pointed me to Autonomy’s 2011, first quarter financial results. I took a quick look at the top line revenues, multiplied by four and concluded:
- Autonomy has a better than even chance of breaking $1 billion in revenue before the end of its current fiscal year
- Autonomy was growing and rolling out new products and services, including an interesting medical and health product, other vendors of search were floundering (Google), giving away search as part of bundles and other deals (Microsoft, Oracle), struggling to be findable by potential customers (Thunderstone, a search vendor whose name is now used by a band and a game), or repositioning themselves to be something other than a vendor of enterprise search (Brainware for scanning, Coveo for customer support).
- Autonomy was reporting growth in its various of lines of business at a decent rate; 28 percent organic growth if I read the report correctly.
The story was ignored by most of the financial wizards who monitor search for the bottom tier and mid tier consulting firms. I read one “analysis” from an outfit called Gerson Lehrman Group which was written by a single individual but presented with a royal “we”. What struck me was that individuals seem happy pontificating about search, financials, and a darned complex technology using sentences that remind me of the rhetoric for the royal wedding. Wedding coverage has more substance than analyses of enterprise search I think.
In my new landscape of search study for Pandia.com, I analyze Autonomy, finding enough bone and gristle to fill 13 pages with technical goodies, comments, and critical evaluation of a company that blew past Convera, Delphes, Endeca, Entopia, Fast Search & Transfer, Powerset, Radar Networks, and a bunch of others.
If you want a free run down on what Autonomy has been doing in the last two years, just do the query “Autonomy” in the search box on the splash page of this blog or click this link. We changed our search results display to make it easier for users to get a sense of search vendor activities. For the more timely information, click this link for my free Overflight “what’s happening” report.
Stephen E Arnold, April 23, 2011
Freebie unlike low and mid tier consulting services
Google Campaigns Against Web Censorship
April 22, 2011
The worldwide web could possibly undergo some drastic and even somewhat controversy changes in the future. According to the Ars Technica article “Google: Don’t Give Private Trolls Web Censorship Power” the US government is currently working on legislation in support of Web site blocking at the domain level. If this bill comes to pass it would mean that online ad networks as well as credit card companies would be required to cease working with any site that falls on the blocked list. Eventually, “private companies get the right to bring a censorship action in court without waiting for government to act.” The web giant Google has wasted little time speaking out against this potential bill. Appearing at today’s “Legitimate Sites v. Parasites” hearing before the House Judiciary Committee, Google’s Kent Walker was clear: a private right of action to bring a COICA claim would give rights holders tremendous leverage over Google.
Walker went so far as to warn of “shakedowns” from private companies wanting to force changes in Google’s behavior. Google has been under noted scrutiny due to controversial search results. Google prefers to take a more neutral role and states that it “doesn’t want to be the “judge, jury, and executioner” for Web sites. Finally, the company simply doesn’t know what sites are “authorized” outlets for music and movies, nor when some use is fair. Legislators want to pass a law that allows them to go after anyone who they feel has inappropriate content on their Web site, from child porn to counterfeit products. Though evidence clearly shows that there are questionable sites online, if search results are allowed to be censored, changed or filtered by the wrong people according to a particular guideline, when it comes to search engine results it will become difficult to determine truth versus fiction.
We don’t want to be sticks in the mud. But: If certain content is excluded because it lacks “value”, is that not a form of censorship? Nah, probably not.
April Holmes, April 22, 2011
Freebie