On his personal website, Frederick Giasson reports, “We just released a new UMBEL web service endpoint and online tool: the Concept Tagger Plain. This plain tagger uses UMBEL reference concepts to tag an input text. The OBIE (Ontology-Based Information Extraction) method is used, driven by the UMBEL reference concept ontology. By plain we mean that the words (tokens) of the input text are matched to either the preferred labels or alternative labels of the reference concepts. The simple tagger is merely making string matches to the possible UMBEL reference concepts.” Read more
Posts Tagged ‘Ontology’
Cognitum’s year got off to a good start, with an investment from the Giza Polish Ventures Fund, and it plans to apply some of that funding to building its sales and development teams, demonstrating the approaches to and benefits of semantic knowledge engineering, and focusing on big implementations for recognizable customers. The company’s products include Fluent Editor 2 for editing and manipulating complex ontologies via controlled natural language (CNL) tools, and its NLP-fronted Ontorion Distributed Knowledge Management System for managing large ontologies in a distributed fashion (both systems are discussed in more detail in our story here). “The idea here is to open up semantic technologies more widely,” says CEO Pawel Zarzycki.
To whom? Zarzycki says the company currently has pilot projects underway in the banking sector, which see opportunities to leverage ontologies and semantic management frameworks that provide a more natural way for sharing and reusing knowledge and expressing business rules for purposes such as lead generation and market intelligence. In the telco sector, another pilot project is underway to support asset management and impact assessment efforts, and in the legal arena, the Poland-based company is working with the Polish branch of international legal company Eversheds on applying semantics to legal self-assessment issues. Having a semantic knowledge base can make it possible to automate the tasks behind assessing a legal issue, he says, and so it opens the door to outsourcing this job directly to the entity pursuing the case, with the lawyer stepping in mostly at the review stage. That saves a lot of time and money.
Ontologies are getting a thumbs up to serve as the basis for the Office of Financial Research’s Instruments database. Last week, the Data & Technology Subcommittee of the OFR Financial Research Advisory Committee (FRAC) recommended that the OFR “adopt the goal of developing and validating a comprehensive ontology for financial instruments as part of its overall effort to meet its statutory requirement to ‘prepare and publish’ a financial instrument reference database.”
The Instruments database will define the official meaning of financial instruments for the financial system — derivatives, securities, and so on. The recommendation by the subcommittee is that the OFR conduct its own evaluation of private sector initiatives in this area, including the Financial Industry Business Ontology (FIBO), to assess whether and how ontology can support transparency and financial stability analysis.
FIBO, which The Semantic Web Blog discussed in detail most recently here, is designed to improve visibility to the financial industry and the regulatory community by standardizing the language used to precisely define the terms, conditions, and characteristics of financial instruments; the legal and relationship structure of business entities; the content and time dimensions of market data; and more. The effort is spearheaded by the Object Management Group and the Enterprise Data Management (EDM) Council.
Everyone knows The Clapper for turning electric equipment on and off, right? Sing along: “Clap-on, clap-off….The Clapper.”
Things have come a long way since then, with security, energy management, and more coming along to help turn the average house into a smarter home. Now comes a chance to take things to another level, with semantic-based resource discovery and orchestration in home and building automation. Research led by Michele Ruta, assistant professor at Technical University of Bari, takes on the challenge of bringing together the worlds of semantic web and automation in order to improve what Ruta says are very poor user interaction scenarios.
How? In home automation solutions today, he says, the user is limited to very basic scenarios and very static interaction that requires pre-programming the capabilities the home can assume. “It should be possible to have dynamic interaction, more intelligent interaction with the user, and decisions should be done according to user interest, to a user’s profile,” Ruta says. Semantic technology can be called upon to annotate users’ profiles, interests, and needs against home automation profile options, and make the match between them, he says.
Look, up in the sky! It’s a bird, it’s a plane, no – it’s an Amazon drone!
Admittedly, Amazon Prime Air’s unmanned aerial vehicles in commercial use are still a little ways off. But such technology – along with other recent innovations, such as the use of unmanned aircraft in crop-dusting or even Department of Homeland Security border applications, or future capabilities to extend the notion of auto-piloting in passenger airplanes using autonomous machine logic to control airspace and spacing between planes –needs to be accounted for in terms of its impact on the air space. The Next-Generation Air Transportation System is taking on the change in the management and operation of the national air transportation system.
And semantic technology, natural language processing, and machine learning, too, will have a hand in helping out, by fostering collaboration among the agencies that will be working together to develop the system, including the Federal Aviation Administration, the U.S. Air Force, U.S. Navy, and the National Aeronautics and Space Administration, under the coordination of the Joint Planning and Development Office. These agencies will need to leverage each other’s knowledge and research, as well as ensure – as necessary – data privacy.
Industry leaders in sectors including banking and financial services look to have high hopes for semantic technology. They’re thinking about FIBO (Financial Industry Business Ontology) and leveraging semantic technology for more traditional types of data integration and analytics projects. At Cognizant, Thomas Kelly, a director in its Enterprise Information Management practice – and the author of this white paper on How Semantic Technology Drives Agile Business – sees the positive development that clients in the Fortune 500 space like these “are maturing in their use of semantic technology, from a project focus to more enterprise initiatives.”
The interest in FIBO, he says, is representative of an overall interest across in industries in leveraging industry ontologies as mechanisms to help companies better standardize, align and learn from the output of industry-wide efforts. The attention that industry analysts, including Gartner, have put on the semantic web in the last year – not to mention regulators beginning to consider its use in sharing information on a regulatory basis – have helped increase interest by commercial organizations, Kelly notes. That’s also evident in the life sciences sector, as another example, with the efforts of the FDA/PhUSE Semantic Technology Working Group Project to include a draft set of existing CDISC standards in RDF.
The pickup in attention to many things semantic ties to the different perspectives that organizations need to manage about their data, which include “how they currently think of their data, how it is currently perceived in managing business operations; and where they are looking to go in the future that makes it more inclusive of what’s going on in the world outside their walls – that is, how the rest of the industry looks at this data and uses it to support their business processes,” he says.
Picking up from where we left off yesterday, we continue exploring where 2014 may take us in the world of semantics, Linked and Smart Data, content analytics, and so much more.
Marco Neumann, CEO and co-founder, KONA and director, Lotico: On the technology side I am personally looking forward to make use of the new RDF1.1 implementations and the new SPARQL end-point deployment solutions in 2014 The Semantic Web idea is here to stay, though you might call it by a different name (again) in 2014.
Bill Roberts, CEO, Swirrl: Looking forward to 2014, I see a growing use of Linked Data in open data ‘production’ systems, as opposed to proofs of concept, pilots and test systems. I expect good progress on taking Linked Data out of the hands of specialists to be used by a broader group of data users.
Yesterday we said a fond farewell to 2013. Today, we look ahead to the New Year, with the help, once again, of our panel of experts:
Phil Archer, Data Activity Lead, W3C:
For me the new Working Groups (WG) are the focus. I think the CSV on the Web WG is going to be an important step in making more data interoperable with Sem Web.
I’d also like to draw attention to the upcoming Linking Geospatial Data workshop in London in March. There have been lots of attempts to use Geospatial data with Linked Data, notably GeoSPARQL of course. But it’s not always easy. We need to make it easier to publish and use data that includes geocoding in some fashion along with the power and functionality of Geospatial Information systems. The workshop brings together W3C, OGC, the UK government [Linked Data Working Group], Ordnance Survey and the geospatial department at Google. It’s going to be big!
[And about] JSON-LD: It’s JSON so Web developers love it, and it’s RDF. I am hopeful that more and more JSON will actually be JSON-LD. Then everyone should be happy.
News came the other week that Senzari had announced the MusicGraph knowledge engine for music. The Semantic Web Blog had a chance to learn a little bit more about it what’s underway thanks to a chat with Senzari’s COO Demian Bellumio.
MusicGraph used to go by the geekier name of Adaptable Music Parallel Processing Platform, or AMP3 for short, for helping users control their Internet radio. “We wanted to put more knowledge into our graph. The idea was we have really cool and interesting data that is ontologically connected in ways never done before,” says Bellumio. “We wanted to put it out in the world and let the world leverage it, and MusicGraph is a production of that vision.”
Since its announcement earlier this month about launching the consumer version on the Firefox OS platform that lets users make complex queries about music and learn and then listen to results, Senzari has submitted its technology to be offered for the iOS, Android, and Windows Mobile platforms. “You can ask anything you can think of in the music realm. We connect about 1 billion different points to respond to these queries,” he says. Its data covers more than twenty million songs, connected to millions of individual albums and artists across all genres, with extracted information on everything from keys to concept extractions derived from lyrics.
The health care industry – and the American citizenry at large – has been focused of late on the problems surrounding the implementation of the Affordable Care Act, the federal website’s issues foremost among them. But believe it or not, there are other things the healthcare industry needs to prepare for, among them the October 1, 2014 date for replacing the World Health Organization’s International Statistical Classification of Diseases and Related Health Problems ICD-9 code sets used to report medical diagnoses and inpatient procedures by ICD-10 code sets. ICD-9 uses 14,000 diagnosis codes which will increase to 68,000 in ICD-10, which is a HIPAA (Health Insurance Portability and Accountability Act) code set requirement.
Natural language processing has had the primary role in many solutions aimed at transforming large volumes of unstructured clinical data into information that healthcare IT application vendors and their hospital customers can leverage. But there’s an argument being made that understanding unstructured text of clinical notes that contain a huge stash of information and then mapping them to fine-grained ICD-10 coding schemes requires a combination of NLP, advanced linguistics, machine learning and semantic web technologies, and Amit Sheth, professor of computer science and engineering at Wright State University and director of the Kno.e.sis Center is making them. (See our story yesterday for a look at how the NLP market is evolving overall, including in healthcare.)
“ICD-10 has thousands of codes with millions of possible permutations and combinations. A rule-based approach is not effective to cover the huge number of ICD-10 codes.” Sheth says. Extracting the correct concepts, identifying the relationship between these concepts and mapping them to the correct code is a major challenge, with codes often formed by information from various sections of a clinical document that itself is subject to individual physicians’ style of recording information, among other factors.