Christopher Tozzi of The VAR Guy reports, “PredictionIO, the open source machine learning platform, has received a big boost with the announcement of $2.5 million in seed funding, which it plans to use to make its automated data interpretation and prediction platform widely available to open source developers. PredictionIO’s goal is to make it easy for developers and companies of all sizes to integrate machine learning —i.e., software that can interpret data intelligently to make automated decisions and predictions—into their products. ‘PredictionIO aims to be the Machine Learning server behind every application,’ according to the company. ‘Building Machine Learning in software will be as common as search soon with PredictionIO’.” Read more
Posts Tagged ‘open source’
Penny Wolf of IT News in Australia reports, “Yahoo7 is rolling out a semantic publishing platform across the web sites that promote Channel 7’s local free to air programs, building its new content systems atop an open source graph database. Craig Penfold, CTO Yahoo7 was inspired by a similar project undertaken by the BBC, which initially launched semantic web publishing for the channel’s 2010 World Cup website to increase viewer engagement. After conducting research and discussions with BBC and Yahoo, the Yahoo7 team chose to change the datastore from a standard SQL database to a graph database.” Read more
New Startup Skymind Offers Support for Open Source Deep Learning
Derrick Harris of GigaOM reports, “A San Francisco-based startup called Skymind launched on Monday to offer support and services for deeplearning4j, an open source deep learning project it has created. It’s early to tell how much traction deep learning will gain among mainstream companies or even web companies, but the technology does hold a lot of promise. The existence of open source libraries backed by professional services could certainly help spur adoption – especially for a field of data analysis previously relegated to top universities and research labs at companies such as Google, Microsoft, Facebook and Baidu.” Read more
Octo Consulting, a technology solutions and management services company for both the intelligence and healthcare sectors, recently published an infographic exploring the intersection among the Semantic Web, Linked Data and Health IT as it relates to accessing and interacting with data from an array of sources in the healthcare chain. “Our point of view is that in healthcare there are multiple data sources and so much data – especially when it comes to clinical trials, pharmaceuticals research and scientific data,” says CTO Ashok Nare. “It’s very possible that each of those data elements is represented in a different format, so how to take them all and connect them to ask questions you aren’t able to ask otherwise. That’s where semantic technologies are extremely useful.”
One health-care sector project in which Octo is putting semantic technologies to use these days is an effort it has underway with the U.S.’s medical research agency, the National Institutes of Health, whose mission includes providing grants to the scientific community to engage in research and experiments “to enhance health, lengthen life, and reduce illness and disability,” as its web site explains. Now, not only does the NIH want to understand what it’s funding and how those grants are progressing, but also “what opportunities it may be missing out on,” Nare explains.
That means continually assessing not only what’s in its portfolio but also what research gaps there are, which requires conducting analysis on more and more data sources and investigating more queries: That could mean more development and expense, without the help of semantic web technologies.
According to a new announcement from the company 3 Round Stones, Callimachus Open Source v1.2 has been released. Callimachus is a Linked Data application server used by the Federal Government for publishing open data on the Web and Fortune 1000 for consuming and visualizing a combination of enterprise and open content. 3 Round Stones is working to make Callimachus the choice of Web developers who want to rapidly write and host data-driven Web applications. Callimachus Open Source 1.2 provides a number of improvements across the board – some which improve use experience for building visualizations with the Chart Wizard and some back end changes aimed at improving scaling and performance. As always, for full details about the project and downloads check out http://callimachusproject.org. Here are some highlights.” Read more
09-24-2013 06:12 PM CET — The EU’s “digital.me” project brings Fraunhofer IAO together with seven research and industry partners to develop a system for user-controlled social networks and services that can serve as a central hub for managing a user’s various digital identities. The project has now released the source code from its software development work as an open-source project.
The use of personal information for private and business life is a trend in our increasingly information-driven society. With the rise of social media, individuals are revealing more personal data online than ever before. This data disclosure provides value to users, such as enhancing social contacts or obtaining personalized services and products. However, the existing social internet makes it difficult for using personal information in a controlled way while retaining privacy where required. Read more
Simon Phipps of InfoWorld recently wrote, “At OSCON in Portland, Ore., last month, I had the chance to meet Jeff Hawkins, the inventor of the Palm Pilot and arguably the father of the smartphone. I learned that he is now pioneering the analysis of huge streams of real-time data using insights gained as a neuroscientist. His company offers a product that can learn the characteristics of data streams, predict their future actions, and identifiy anomalies. He has just recently taken the core of that product and released it as a GPLv3-licensed open source project on GitHub so that anyone can build machine intelligence into their systems. Below is a video of our discussion, followed by an edited version of the interview.” Read more
Justin Kern of Information Management reports, “The U.S. government is outlining its new program for an open source repository to foster collaboration on getting more information to citizens in a faster manner. Federal CTO Todd Park formally introduced Project Open Data on Thursday in a blog post, and gave an update on its first days of activity. In the first 24 hours after Project Open Data was published, more than two dozen contributors submitted to its GitHub platform, including fixes to broken Web links and policy input. Other, meatier contributions, or ‘pull requests,’ included a tool that converts spreadsheets and databases into APIs for ease of use by developers, and code that translates geographic data from locked formats into open, available formats, according to Park.” Read more
Shahid Qadri has written an article for Med City News about how to use WebID to create single sign-on access for health care systems. He writes, “The Simple Sign-on challenge sponsored by the ONC through the Health 2.0 challenge was an exciting opportunity for us to learn about a sophisticated technology protocol and then being able to hack several open source system to implement a single sign on solution based on the protocol. This was a challenge that was truly a ‘challenge’ for me, but an exciting and rewarding one (our solution was the second place winner!).” Read more
Search is a fundamental, a system building block, and something that should be a critical part of enterprise architectures. That’s what Grant Ingersoll, co-founder and CTO at search, discovery and analytics vendor LucidWorks – which leverages the Apache Lucene/Solr open source search project – told an audience at last week’s GigaOM Structure Data event.
The company late last year launched LucidWorks Big Data for developing Big Data applications, which builds on top of its heritage developing the LucidWorks Search solution. “It’s a platform for organizations and developers to build out next-generation data applications,” Ingersoll said in a conversation with the Semantic Web Blog. Its focus is on tight integration of key Apache open source projects and layering with a REST API, to provide developers single-source access to the stack’s richness for creating applications that provide comprehensive search, discovery and analysis of an organization’s vast content and user interactions.
LucidWorks Big Data is made up of Apache Hadoop; the Apache Mahout machine-learning library; Hive, a data warehouse system for Hadoop that facilitates data summarization, ad-hoc queries, and the analysis of large datasets stored in Hadoop-compatible file systems; and Apache OpenNLP, a machine-learning based toolkit for the processing of natural language text that supports common NLP tasks.
NEXT PAGE >>