Machine Learning

Baidu Research Achieves Speech Recognition Breakthrough With “Deep Speech”

Baidu logoSUNNYVALE, CA, Dec 18, 2014 (Marketwired via COMTEX) — Baidu Research, a division of Baidu, Inc. today announced initial results from its Deep Speech speech recognition system.

Deep Speech is a new system for speech, built with the goal of improving accuracy in noisy environments (for example, restaurants, cars and public transportation), as well as other challenging environments (highly reverberant and far-field situations).

Key to the Deep Speech approach is a well-optimized recurrent neural net (RNN) training system that uses multiple GPUs, as well as a set of novel data synthesis techniques that allowed Baidu researchers to efficiently obtain a large amount of varied data for training. Read more

4 Open Source Machine Learning Projects

deepdiveSerdar Yegulalp of InfoWorld recently wrote, “Over the last year, as part of the new enterprise services that IBM has been pushing on its reinvention, Watson has become less of a “Jeopardy”-winning gimmick and more of a tool. It also remains IBM’s proprietary creation. What are the chances, then, of creating a natural-language machine learning system on the order of Watson, albeit with open source components? To some degree, this has already happened — in part because Watson itself was built in top of existing open source work, and others have been developing similar systems in parallel to Watson. Here’s a look at four such projects.” Read more

How AppZen Uses Natural Language Process to Automate Expense Reports

appzenConner Forrest of Tech Republic reports, “AppZen, a Sunnyvale, California-based startup, is leveraging natural language processing and machine learning to automate the expense report process. The company bills itself as an ambient expense manager, and it works in the background by tracking your expenses as they happen and creates a report for you. The process begins with the mobile app that runs on Android or iOS, which is able to track an employee’s calendars, itineraries, and credit card charges to build out an expense report as it happens.” Read more

Yandex Takes Its Machine Learning Smarts Into New Role as Business Services Provider

yandex-logo (1)Yandex is going beyond web search and into the enterprise. This week it announced a new venture, the Yandex Data Factory, which will apply its machine-learning products and algorithms – which power more than 70 percent of its own products and applications – to business’ Big Data issues.

Using a client’s pre-existing data, the press release notes, Yandex creates an algorithmic model, which it then applies to the client’s new data to predict what will happen next in various scenarios. “This is exactly what is happening every second on Yandex’s services when we personalize search suggestions, recommend music, recognize speech or images, or target ads,” the release notes.

The model cases for Yandex Data Factory include: churn prediction by running segmentation and micro-segmentation algorithms on the data to find patterns in customer behavior that indicate they’re heading for the exit or that possible fraud activity is underway; personalizing cross-sell and up-sell recommendations based on user profiles built upon the searches they made, links or ads they clicked, videos they watched, and other activities; using its speech-to-text technology to analyze call center or other support call speech streams and detect anomalies in interactions to drive employee interaction quality and improve skills.

It also uses history-based prediction technology and its own computer vision and image recognition technologies to enables businesses to analyze large volumes of images and videos to spot anomalies, find recurring objects or events, and other things that will help them assess conditions and assure productivity.

Read more

How Quoc Le is Teaching Machines to Understand Emotions Through Deep Learning

Google LogoDaniela Hernandez of Wired recently wrote that Quoc Le “works on the Google Brain, the search giant’s foray into ‘deep learning,’ a form of artificial intelligence that processes data in ways that mimic the human brain—at least in some ways. Le was one of the main coders behind the widely publicized first-incaration of the Google Brain, a system that taught itself to recognize cats on YouTube images, and since then, the 32-year-old Vietnam-native has been instrumental in helping to build Google systems that recognize your spoken words on Android phones and automatically tag your photos on the web, both of which are powered by deep-learning technology.” Read more

Technology Brief: Machine Learning on Natural Language Text and Log Data

Skytree - the machine learning company - LogoSkytree, “The Machine Learning Company,” has published a technology brief entitled, “Machine Learning on Natural Language Text and Log Data.”

The author of the brief is Nick Pendar, PhD, who serves as NLP Data Scientist for the company.

Pendar states, “Critical business information is often in the form of unstructured and semi-structured data that can be hard or impossible to interpret with legacy systems. In this brief, discover how you can use machine learning to analyze both unstructured text data and semi-structured log data, providing you with the insights needed to achieve your business goals.”

Download your free copy today from the DATAVERSITY website:

Download the Skytree technology brief today.

Google’s Process for Translating Pictures into Words

Google LogoThe MIT Technology Review recently wrote, “Translating one language into another has always been a difficult task. But in recent years, Google has transformed this process by developing machine translation algorithms that change the nature of cross cultural communications through Google Translate. Now that company is using the same machine learning technique to translate pictures into words. The result is a system that automatically generates picture captions that accurately describe the content of images. That’s something that will be useful for search engines, for automated publishing and for helping the visually impaired navigate the web and, indeed, the wider world.” Read more

Carnegie Mellon Researchers Identify Brain Regions That Encode Words, Grammar, Story

cmuSome people say that reading “Harry Potter and the Sorcerer’s Stone” taught them the importance of friends, or that easy decisions are seldom right. Carnegie Mellon University scientists used a chapter of that book to learn a different lesson: identifying what different regions of the brain are doing when people read. Read more

Google Researchers Use End-to-End Neural Network To Caption Pictures

pizzaGoogle researchers have announced the development of a machine-learning system that can automatically produce captions to accurately describe images in properly formed sentences the first time it sees them.

“This kind of system could eventually help visually impaired people understand pictures, provide alternate text for images in parts of the world where mobile connections are slow, and make it easier for everyone to search on Google for images,” report research Scientists Oriol Vinyals, Alexander Toshev, Samy Bengio, and Dumitru Erhan in a blog about how they’re building a neural image caption generator.

Getting there, the researchers say, involved merging recent computer vision and language models into a single jointly trained system that can directly produce a human readable sequence of words to describe a given image. The task is no easy one, they point out, explaining that unlike image classification or object recognition on its own, their work has to account not only for the objects contained in the image, but also for expressing how these objects relate to each other, as well as their attributes and the activities they are involved in.

The approach leverages an end- to-end neural network that can automatically view an image and generate a plan English description of it.

Read more

Cornell Announces New Connected Experiences Laboratory

cornell_big_red_logoThe Cornell Chronicle reports that “AOL, the Jacobs Technion-Cornell Institute at Cornell Tech and New York City officials have announced the creation of the Connected Experiences Laboratory, or ConnX, aimed at exploring new technologies at the forefront of the digital age. Announced Nov. 12, an AOL gift will fund a four-year partnership between New York City-based AOL and Cornell Tech. The collaboration will extend to the Technion – Israel Institute of Technology, Cornell’s partner in the Jacobs Institute, with part of the gift supporting activities at the Technion in coordination with AOL Israel.” Read more

NEXT PAGE >>