Posts Tagged ‘linked open data’

Art Lovers Will See There’s More To Love With Linked Data

The team behind the data integration tool Karma this week presented at LODLAM (Linked Open Data in Libraries, Archives & Museums), illustrating how to map museum data to the Europeana Data Model (EDM) or CIDOC CRM (Conceptual Reference Model). This came on the heels of its earning the best-in-use paper award at ESWC2013 for its publication about connecting Smithsonian American Art Museum (SAAM) data to the LOD cloud.

The work of Craig KnoblockPedro SzekelyJose Luis AmbiteShubham GuptaMaria MusleaMohsen Taheriyan, and Bo Wu at the Information Sciences InstituteUniversity of Southern California, Karma lets users integrate data from a variety of data sources (hierarchical and dynamic ones too) — databases, spreadsheets, delimited text files, XML, JSON, KML and Web APIs — by modeling it according to an ontology of their choice. A graphical user interface automates much of the process. Once the model is complete, users can publish the integrated data as RDF or store it in a database.

The Smithsonian project builds on the group’s work on Karma for mapping structured sources to RDF. For the Smithsonian project (whose announcement we covered here), Karma converted more than 40,000 of the museum’s holdings, stored in more than 100 tables in a SQL Server Database, to LOD, leveraging EDM, the metamodel used in the Europeana project to represent data from Europe’s cultural heritage institutions.

Read more

Self Medicating? Stay Safe With Semantic Tech’s Help

It’s pretty common these days for people to hit the web in search of medication advice to deal with symptoms they’re experiencing.  The trouble is, most people don’t approach the process in a truly safe manner.

Semantic technology can help rectify the situation. In fact, it’s already doing so in France, where Olivier Curé, an associate professor in computer science at the University of Paris-Est in France, created a web application based on the book by pharmacology expert and educator Jean-Paul Giroud Médicaments sans ordonnance: Les bons et les mauvais!, of which he is a co-author with Catherine Cupillard. The app is made available to their consumers via three big insurance companies there, in order to help the companies save costs on reimbursing them for buying drugs that won’t actually help their condition, or direct them to the appropriate drugs at pharmacies with which the insurers may have relationships to supply them at lower costs. An iPhone version of the app was just released to accompany the web version.

Read more

Addressing Price-Performance And Curation Issues For Big Data Work In The Cloud

The cloud’s role in processing big semantic data sets was recently highlighted in early April when DERI and Fujitsu Laboratories announced a new data storage technology for storing and querying Linked Open Data that resides on a cloud-based platform (see our story here).

The cloud conversation, with storage as one key discussion point, will continue to be an active one in Big Data circles, whether users are working with massive, connected Linked Data sets or trying to run NLP across the Twitter firehose. CloudSigma, for example, recently publicly disclosed that it is using an all solid-state drive (SSD) solution for its public cloud offering that lets users purchase CPU, RAM, storage and bandwidth independently. The use of SSD, says CEO Robert Jenkins, avoids the problem that spinning disks have with the randomized, multi-tenant access of a public cloud that leads to storage bottlenecks and curbs performance.

That, combined with the company’s approach of letting customers size virtual machine resources as they like, as well as leverage exposed advanced hypervisor settings to optimize for their particular applications, he says, brings the use of the public cloud infrastructure closer to what companies can get out of private cloud environments, and at a price-performance win.

Read more

Fujitsu Labs And DERI To Offer Free, Cloud-Based Platform To Store And Query Linked Open Data

The Semantic Web Blog reported last year about a relationship formed between the Digital Enterprise Research Institute (DERI) and Fujitsu Laboratories Ltd. in Japan, focused on a project to build a large-scale RDF store in the cloud capable of processing hundreds of billions of triples. At the time, Dr. Michael Hausenblas, who was then a DERI research fellow, discussed Fujitsu Lab’s research efforts related to the cloud, its huge cloud infrastructure, and its identification of Big Data as an important trend, noting that “Linked Data is involved with answering at least two of the three Big Data questions” – that is, how to deal with volume and variety (velocity is the third).

This week, the DERI and Fujitsu Lab partners have announced a new data storage technology that stores and queries interconnected Linked Open Data, to be available this year, free of charge, on a cloud-based platform. According to a press release about the announcement, the data store technology collects and stores Linked Open Data that is published across the globe, and facilitates search processing through the development of a caching structure that is specifically adapted to LOD.

Read more

Opening Up Publically Funded Research in Europe

Anna Leach of the Wall Street Journal reports, “New scientific research must be published for free online, the vice-president of the European Commission said, in a move designed to increase the knowledge pool open to small business and lead to more innovative products. All scientists receiving European Union funding will have to publish their results in an open-access format, Neelie Kroes, the commissioner responsible for Europe’s digital agenda, said Monday in Stockholm.  Ms. Kroes also  launched the global Research Data Alliance — a group committed to pooling and co-ordinating scientific data so it can be shared better.”

Leach continues, “Opening up scientific research is good for small business, said Victor Henning, CEO of British startup Mendeley, which aims to make academic research more connected. He has noticed the demand for access to academic research from small businesses. Read more

List of Thousands of Public Data Sources

A website called BigML (for Big Machine Learning) has compiled a great list of freely available public data sources. The article begins: “We love data, big and small and we are always on the lookout for interesting datasets. Over the last two years, the BigML team has compiled a long list of sources of data that anyone can use. It’s a great list for browsing, importing into our platform, creating new models and just exploring what can be done with different sets of data. In this post, we are sharing this list with you. Why? Well, searching for great datasets can be a time consuming task. We hope this list will support you in that search and help you to find some inspiring datasets. ” Read more

Announcing the Launch of the GBPN Knowledge Platform

Martin Kaltenbock of the Semantic Web Company reports, “The brand new web based GBPN Knowledge Platform has been launched on 21 February 2013. It helps the building sector effectively reduce its impact on climate change! It has been designed as a participative knowledge hub and data hub harvesting, sharing and curating best practice policies in building energy performance globally. Available in English and soon in Mandarin, this new web-based tool of the Global Buildings Performance Network (GBPN) aims to stimulate collective research and analysis from experts worldwide to promote better decision-making and help the building sector effectively reduce its impact on climate change. Read more

Semantic Tech Outlook: 2013

Photo Courtesy: Flickr/Lars Plougmann

In recent blogs we’ve discussed where semantic technologies have gone in 2012, and a bit about where they will go this year (see here, here and here).

Here are some final thoughts from our panel of semantic web experts on what to expect to see as the New Year rings in:

John Breslin,lecturer at NUI Galway, researcher and unit leader at DERI, creator of SIOC, and co-founder of Technology Voice and StreamGlider

Broader deployment of the schema.org terms is likely. In the study by Muehlisen and Bizer in July this year, we saw Open Graph Protocol, DC, FOAF, RSS, SIOC and Creative Commons still topping the ranks of top semantic vocabularies being used. In 2013 and beyond, I expect to see schema.org jump to the top of that list.

Christine Connors, Chief Ontologist, Knowledgent:

I think we will see an uptick in the job market for semantic technologists in the enterprise; primarily in the Fortune 2000. I expect to see some M&A activity as well from systems providers and integrators who recognize the desire to have a semantic component in their product suite. (No, I have no direct knowledge; it is my hunch!)

We will see increased competition from data analytics vendors who try to add RDF, OWL or graphstores to their existing platforms. I anticipate saying, at the end of 2013, that many of these immature deployments will leave some project teams disappointed. The mature vendors will need to put resources into sales and business development, with the right partners for consulting and systems integration, to be ready to respond to calls for proposals and assistance.

Read more

Semantic Tech: It’s Moving Mainstream, Playing To The Data-Is-An-Asset Crowd, And Living Life Out Loud

At the recent SemTech conference in NYC, The Semantic Web Blog had an opportunity to ask some leaders in the field about where semantic technology has been, and where it’s going.

David Wood, CTO, 3RoundStones:

The short take: Hiring has been on in a big way at semantic tech players as enterprises are moving in greater numbers to buy semantic software, recognizing their traditional vendors won’t solve their interoperability issues. Sem tech vendors should have a happy 2013 as semantics continues going mainstream.

The full take:

Read more

Good-Bye to 2012: Continuing Our Look Back At The Year In Semantic Tech

Courtesy: Flickr/LadyDragonflyCC <3

Yesterday we began our look back at the year in semantic technology here. Today we continue with more expert commentary on the year in review:

Ivan Herman, W3C Semantic Web Activity Lead:

I would mention two things (among many, of course).

  •  Schema.org had an important effect on semantic technologies. Of course, it is controversial (role of one major vocabulary and its relations to others, the community discussions on the syntax, etc.), but I would rather concentrate on the positive aspects. A few years ago the topic of discussion was whether having ‘structured data’, as it is referred to (I would simply say having RDF in some syntax or other), as part of a Web page makes sense or not. There were fairly passionate discussions about this and many were convinced that doing that would not make any sense, there is no use case for it, authors would not use it and could not deal with it, etc. Well, this discussion is over. Structured data in Web sites is here to stay, it is important, and has become part of the Web landscape. Schema.org’s contribution in this respect is very important; the discussions and disagreements I referred to are minor and transient compared to the success. And 2012 was the year when this issue was finally closed.
  •  On a very different aspect (and motivated by my own personal interest) I see exciting moves in the library and the digital publishing world. Many libraries recognize the power of linked data as adopted by libraries, of the value of standard cataloging techniques well adapted to linked data, of the role of metadata, in the form of linked data, adopted by journals and soon by electronic books… All these will have a profound influence bringing a huge amount of very valuable data onto the Web of Data, linking to sources of accumulated human knowledge. I have witnessed different aspects of this evolution coming to the fore in 2012, and I think this will become very important in the years to come.

Read more

<< PREVIOUS PAGENEXT PAGE >>