January 22, 2014
Did you know that there was an open source version of ClearForest called Calais? Neither did we, until we read about it in the article posted on OpenCalais called, “Calais: Connect. Everything.” Along with a short instructional video, is a text explanation about how the software works. OpenCalais Web Service automatically creates rich semantic metadata using natural language processing, machine learning, and other methods to analyze for submitted content. A list of tags are generated and returned to the user for review and then the user can paste them onto other documents.
The metadata can be used in a variety of ways for improvement:
“The metadata gives you the ability to build maps (or graphs or networks) linking documents to people to companies to places to products to events to geographies to… whatever. You can use those maps to improve site navigation, provide contextual syndication, tag and organize your content, create structured folksonomies, filter and de-duplicate news feeds, or analyze content to see if it contains what you care about.”
The OpenCalais Web Service relies on a dedicated community to keep making progress and pushing the application forward. Calais takes the same approach as other open source projects, except this one is powered by Thomson Reuters.
January 15, 2014
The article on the Thetus blog titled Thetus South is Now in Business celebrates the opening of a Thetus office in Nashville, Tennessee. With headquarters in Portland, Oregon, Thetus was incorporated in 2003 and has since developed semantic modeling technologies. Savanna 3.3 is analytic software that offers search and discovery, among other tools. The article on the blog mainly examines the building Thetus will now call home.
The article states:
“Located in the heart of Nashville’s historic arts district, the newest addition to the Thetus family embodies the classic Thetus feel while creating its own identity in unique and historic Cummins Station. Originally built in 1907 , Cummins Station is famous for once being the warehouse for Maxwell House Coffee. Over a century later, it’s making a new mark on Nashville as a fabulous and creative LEED Gold mixed-use space…. a perfect match for Thetus!”
The company offered some pictures of the gorgeous windows in the space, and gave some details about enhancements made (such as removing the paint from the brick and removing the carpet. Cummins Station is home to over 140 businesses ranging from IT to restaurant to a gym. The blog also encouraged job-seekers to apply. They are also embracing Nashville’s culture, sponsoring the 2013 Hack Nashville, a coding event.
Chelsea Kerwin, January 15, 2014
January 3, 2014
I am sure that you have heard that Big Data is capable of working with all data types. According to Information Week’s article, “Variety’s The Spice Of Life-And Bane Of Big Data” the newest trend has problem and the biggest is handling the variety of data. The article argues that it is impractical to funnel ad hoc data source into a central schema. The better alternative is to use optional schema also known as RDF standards or semantic web technologies. How can trading one schema for another make a difference?
The article states:
“When data is accessible using the simple RDF triples model, you can mix data from different sources and use the SPARQL query language to find connections and patterns with no need to predefine schema. Leveraging RDF doesn’t require data migration, but can take advantage of middleware tools that dynamically make relational databases, spreadsheets, and other data sources available as triples. Schema metadata can be dynamically pulled from data. It is stored and queried the same way as data.”
RDF acts like a sieve. It allows the unnecessary data information to run through the small holes, leaving the relevant stuff behind for quicker access. It sounds like a perfect alternative, except it comes with its own set of challenges. The good news is that these problems can easily be resolved with a little training and practice.
Whitney Grace, January 03, 2014
December 28, 2013
The article titled How Semantic Search is Killing the Keyword on iMedia heralds the end of keyword-driven search in place of semantic search, or the user’s intention. Based mainly on Google’s work on the Knowledge Graph, a web of information that attempts to connect related data and provide a user with answers to questions they might not have known to ask. The article goes so far as to call keyword-centered content a thing of the past.
The article explains:
“Current trends in search engine marketing and optimization have been underscored by Google’s Knowledge Graph and its Hummingbird update. Both emphasize the importance of not only having high-quality content but also adding semantic markup… to your content, which makes it possible for machines to detect meaning or intent. Marking up your site’s content is crucial to its success in search engines… It not only improves the display of search results, but it makes it much easier for users to find desired web pages.”
The article goes on to describe what exactly would replace boring text-based search results. With semantic markup, the search results can consist of more visuals, ratings, reviews, anything to keep the user interested (the article considers semantic markup search as having more “personality.” Who needs words anyway?
Chelsea Kerwin, December 28, 2013
December 20, 2013
Partnerships develop when companies each possess a strength and then combine forces to build a beneficial relationship. The CogBlog, Cognition’s Semantic NLP Blog, announced a new relationship in the post, “Cognition To Power Grabbit’s Online Recommendation Engine.” Cognition is a leading name in semantic analysis and language process and Grabbit is the developer of a cloud-hosted suite of Web services. Together they have formed a strategic partnership that will combine Cognition’s natural language processing technology with Grabbit’s patent-pending system for making online recommendations of products, content, and people. The idea behind pairing the two technologies is that the semantic software would analyze social media content and then Grabbit’s software would then make product recommendations based on the data.
The article states:
“Cognition provides a powerful set of semantic tools to power Grabbit’s new web services. The scope of Cognition’s Semantic Map is more than double the size of any other computational linguistic dictionary for English, and includes more than ten million semantic connections that are comprised of semantic contexts, meaning representations, taxonomy and word meaning distinctions. The Map encompasses over 540,000 word senses (word and phrase meanings); 75,000 concept classes (or synonym classes of word meanings); 8,000 nodes in the technology’s ontology or classification scheme; and 510,000 word stems (roots of words) for the English language. Cognition’s lexical resources encode a wealth of semantic, morphological and syntactic information about the words contained within documents and their relationships to each other. These resources were created, codified and reviewed by lexicographers and linguists over a span of more than 25 years.”
Why do I get the feeling that online shopping is going to get even more complicated? Personal qualms aside, Cognition and Grabbit are not the first companies that come to mind when it comes to social media analytics and e-commerce. This partnership is not the first endeavor to cash in on Internet sales.
Whitney Grace, December 20, 2013
December 18, 2013
The article titled The Stealthy Rise of Semantic Search on Search Engine Journal relates the outmoding of SEOs with natural language search. The article explains that Boolean expressions leaped search forward with the abilities of query modifiers. Many people use semantic search without realizing it, but it works nonetheless to determine intent instead of just matching search terms.
The article explains:
“Vertical search engines such as Hakia, Lexxe, and VSW were at the forefront of semantic technology long before Bing, Facebook, and Google. These engine providers built their business around that potential and are using it to create new distribution and business models to deliver options for content makers well beyond keyword search and SEO.
Advertisers will reap huge benefits from semantic search because it increases the relevance of all forms of advertising.”
Semantic search will also decrease the likelihood of ads being paired with unsuitable content. Soon, users will be presented with information they “didn’t even know they wanted.” This may even lead to interconnected devices within the home. The article offers the possibility of the refrigerator notifying the homeowner that they are running low on milk. (Smart House, anyone?) Whether scary or exciting, the article would have us believe that these developments are inevitable.
Chelsea Kerwin, December 18, 2013
December 13, 2013
EasyAsk Ranked Among Top Four Providers of E-Commerce Technology, an article on Virtual-Strategy Magazine, recognizes the achievements of EasyAsk, the natural language search company. EasyAsk was recently named one of the top 4 vendors (out of 1,000) in driving e-commerce sales by the E-Retailers Guide. Craig Bassin, CEO of EasyAsk, expressed no surprise at this, since reports show that 43% of visitors to a given website will head straight for the search box.
Bassin expanded on his company’s progress:
“”EasyAsk is poised to capture a significant share of the growing spend on e-commerce technology, said Bassin. “EasyAsk eCommerce Edition delivers amazing value to our clients. EasyAsk is embedded within Infor Storefront and has out-of-the-box integrations with the leading e-commerce platforms, such as IBM Websphere Commerce, Magento, Hybris and Netsuite. Our customers consistently tell us we help them turn shoppers into buyers.” Gartner Inc. estimates that retailers spent about $3 billion on e-commerce technology in 2012. “
Semantic search has become unavoidably important, with Google and Microsoft adopting their own offerings since in the last two years. But EasyAsk stands out as offering “natural language search for e-commerce enterprise, on-premise and cloud platforms.” Their work in raising online revenue by allowing users to search in plain English and receive specific and relevant results has made them a leader in the field.
Chelsea Kerwin, December 13, 2013
December 5, 2013
Good news for TEMIS, everyone! According to The Sacramento Bee in the article, “OECD Chooses TEMIS To Semantically Structure Its Knowledge And Information Management Process,” TEMIS has a new and very big client. The OECD stands for the Organization for Economic Cooperation and Development and they have selected TEMIS’s semantic content enrichment solution Luxid. OECD has started a new project called the Knowledge and Information Management (KIM) Program to create framework for managing and delivering information as well as improving its accessibility and presentation. The OECD collects and analyzes data for over thirty-four member governments and over one hundred countries to help them sustain economic growth, boost employment, and raise the standard of living. The KIM Program will be a portal for the organization’s information and will hopefully increase findability and search.
What will TEMIS do? the article explains:
“In this context, the OECD has chosen TEMIS’s flagship Luxid® Content Enrichment Platform to address all Semantic Enrichment stages of the KIM framework. Luxid® will help OECD to consistently enrich document metadata in alignment with its taxonomies and ontologies, providing a genuinely semantic integration layer across heterogeneous document storage and content management components. This semantic layer will both enable new search and browsing methods and improved relevance and accuracy of search results, as well as progressively build an integrated map of OECD knowledge.”
Glad to see that enriched search and findability are not dead yet. Metadata still has its place, folks. How else will the big data people be able to find their new insights if metadata is not used?
Whitney Grace, December 05, 2013
November 12, 2013
If you are a fan of semantic methods, you may find the Siderean Software profile a useful case study. You can find the write up, among others, at this location. The chatter at conferences about semantic methods is finally burning out. Nevertheless, semantic methods bubble beneath the surface of many modern search systems. The Siderean case is an example of what types of content processing operations are required to perform “deep indexing” or “rich metadata extraction.” The first step, as you will learn, is to have content tagged. That means SGML or XML.
The question becomes, “How do I get my content into these formats?” The answer, for many budgets, is a deal breaker. One the content is processable, then a number of manipulations are possible. Think of Siderean’s system as delivering the type of flip and flop of data that Excel provides in its pivot table. Now ask yourself, “How often do I use a pivot table?” Exactly.
Remember. I am posting pre-publication drafts of analyses that may have been used, recycled, or just ripped off by various “real” publishers over the years. If there are errors in these drafts, you can “correct” them by adding a comment to this post in Beyond Search. The archive of case studies or profiles will not be updated.
I am providing these for personal use. If a frisky soul wants to use them for commercial purposes, I will take some type of action. If you were in my lecture at the enterprise search conference in New York last week, you will know that I called attention to one of the most slippery of the azure chip consulting firms. I showed a slide that listed the same “expert” twice on a $3,500 report. Not bad, since the outfit’s expert did not create the information in the report.
Stephen E Arnold, November 12, 2013
September 15, 2013
In the article titled Is Semantic Search the Solution to Information Retrieval in Sharepoint? on CMS Wire, a survey taken by a UK firm is under discussion. The survey reveals that up to half of Sharepoint users have trouble with finding internal information or getting the full support for the business in which Sharepoint is being deployed. The article does stress that the survey does not specifically ask what version of Sharepoint the respondents use. The article explains,
“The research points out that SharePoint is primarily an enterprise collaboration platform and not a search tool… It also shows that many information professionals believe they are getting less out of SharePoint enterprise search functions than their peers. Even worse, a quarter of SharePoint users believe that the SharePoint search function is offering them access to less than half the information contained in the enterprise. It also shows that most users think that information workers in other enterprises have better search capabilities.”
With nearly 100 million users globally, and 78% of Fortune 500 companies using Sharepoint, the research suggests that the search function was the most problematic.
Chelsea Kerwin, September 15, 2013