Improving speech recognition for contact centers

Contact centers are critical to many businesses, and the right technologies play an important role in helping them provide outstanding customer care. Last July, we announced Contact Center AI to help businesses apply artificial intelligence to greatly improve the contact center experience. Today, we’re announcing a number of updates to the technologies that underpin the Contact Center AI solution—specifically Dialogflow and Cloud Speech-to-Text—that improve speech recognition accuracy by over 40% in some cases to better support customers and the agents that help them.These updates include:Auto Speech Adaptation in Dialogflow (beta)Speech recognition baseline model improvements for IVRs and phone-based virtual agents in Cloud Speech-to-TextRicher manual Speech Adaptation in Dialogflow and Cloud Speech-to-Text (beta)Endless streaming in Cloud Speech-to-Text (beta)MP3 file format support in Cloud Speech-to-TextImproving speech recognition in virtual agentsVirtual agents are a powerful tool for contact centers, providing a better user experience around the clock while reducing wait times. However, the automated speech recognition (ASR) that virtual agents require is much harder to do on noisy phone lines than in the lab. And even at high recognition-accuracy rates (~90%), ASR can sometimes still result in a frustrating customer experience, as you can see below.To help virtual agents quickly understand what customers need, and respond accurately, we’re introducing an exciting new feature in Dialogflow.Auto Speech Adaptation in Dialogflow BetaJust like knowing the context in a conversation makes it easier for people to understand one another, ASR improves when the underlying AI understands the context behind what a speaker is saying. We use the term speech adaptation to describe this learning process.In Dialogflow—our development suite for creating automated conversational experiences—knowing context can help virtual agents respond more accurately. Using the example in the animation above, if the Dialogflow agent knew the context was “ordering a burger” and that “cheese” is a common burger ingredient, it would probably understand that the user meant “cheese” and not “these”. Similarly, if the virtual agent knew that the term “mail” is a common term in the context of a product return, it wouldn’t confuse it with the words “male” or “nail”. To meet that goal, the new Auto Speech Adaptation feature in Dialogflow helps the virtual agent automatically understand context by taking all training phrases, entities, and other agent-specific information into account. In some cases, this feature can result in a 40% or more increase in accuracy on a relative basis.It’s easy to activate Auto Speech Adaptation: just click the “on” switch in the Dialogflow console (off by default), and you’re all set!Cloud Speech-to-Text baseline model improvements for IVRs and phone-based virtual agentsIn April 2018, we introduced pre-built models for improved transcription accuracy from phone calls and video. We followed that up last February by announcing the availability of those models to all customers, not just those who had opted in to our data logging program. Today, we’ve further optimized our phone model for the short utterances that are typical of interactions with phone-based virtual agents. The new model is now 15% more accurate for U.S. English on a relative basis beyond the improvements we previously announced. Applying speech adaptation can also provide additional improvements on top of that gain. We’re constantly adding more quality improvements to the roadmap—an automatic benefit to any IVR or phone-based virtual agent, without any code changes needed–and will share more about these updates in future blog posts.Improving transcription to better support human agentsAccurate transcriptions of customer conversations can help human agents better respond to customer requests, resulting in better customer care. These updates improve the quality of transcription accuracy to support human agents.Richer manual speech adaptation tuning in Cloud Speech-to-TextWhen using Cloud Speech-to-Text, developers use what are called SpeechContext parameters to provide additional contextual information that can make transcription more accurate. This tuning process can help improve recognition of phrases that are common in the specific use case involved. For example, a company’s customer service support line might want to better recognize the company’s product names. Today, we are announcing three updates, all currently in beta, that make SpeechContext even more helpful for manually tuning ASR to improve transcription accuracy. These new updates are available in both the Cloud Speech-to-Text and Dialogflow APIs.SpeechContext classes BetaClasses are pre-built entities reflecting popular/common concepts, which give Cloud Speech-to-Text the context it needs to more accurately recognize and transcribe speech input. Using classes lets developers tune ASR for a whole list of words at once, instead of adding them one by one.For example, let’s say there is an utterance that would normally result in the transcription, “It’s twelve fifty one”. Based on your use case, you could use a SpeechContext class to refine the transcription in a few different ways:A number of other classes are available to similarly provide context around digit sequences, addresses, numbers, and money denominations—you can see the full list here. SpeechContext boost BetaTuning speech recognition with tools like SpeechContext increases the likelihood of certain phrases getting captured—which will both reduce the number of false negatives (when a phrase was mentioned, but does not appear in the transcript), but can also potentially increase the number of false positives (when a phrase wasn’t mentioned, but appears in transcript). The new “boost” feature lets developers use the best speech adaptation strength for their use case.Example:SpeechContext expanded phrase limit BetaAs part of the tuning process, developers use “phrase hints” to increase the probability that commonly used words or phrases related to their business or vertical will be captured by ASR. The maximum number of phrase hints per API request has now been raised by 10x, from 500 to 5,000, which means that a company can now optimize transcription for thousands of jargon words (such as product names) that are uncommon in everyday language.  In addition to these new adaptation-related features, we’re announcing a couple of other highly requested enhancements that improve the product experience for everyone. Endless streaming Beta in Cloud Speech-to-TextSince we introduced Cloud Speech-to-Text nearly three years ago, long-running streaming has been one of our top user requests. Until now, Cloud Speech-to-Text only supported streaming audio in one-minute increments, which was problematic for long-running transcription use cases like meetings, live video, and phone calls. Today, the session time limit has been raised to 5 minutes. Additionally, the API now allows developers to start a new streaming session from where the previous one left off—effectively making live automatic transcription infinite in length, and unlocking a number of new use cases involving long-running audio.MP3 file format support Beta in Cloud Speech-to-TextCloud Speech-to-Text has supported seven file formats up until now (list here). Up until now, processing MP3 files required first expanding them into the LINEAR16 format, which requires maintaining additional infrastructure. Cloud Speech-to-Text now natively supports MP3 so there are no additional conversions needed. Woolworths’ use of conversational AI to improve the contact center experienceWoolworths is the largest retailer in Australia with over 100,000 employees, and has been serving customers since 1924. “In partnership with Google, we’ve been building a new virtual agent solution based on Dialogflow and Google Cloud AI. We’ve seen market-leading performance right from the start,” says Nick Eshkenazi, Chief Digital Technology Officer for Woolworths. “We were especially impressed with accuracy of long sentences, recognition of brand names, and even understanding of the format of complex entities, such as ‘150g’ for 150 grams. “Auto Speech Adaptation provided a significant improvement on top of that and allowed us to properly answer even more customer queries,” says Eshkenazi. “In the past, it used to take us months to create a high quality IVR experience. Now we can build very powerful experiences in weeks and make adjustments within minutes.”“For example, we recently wanted to inform customers about a network outage impacting our customer hub and were able to add messaging to our virtual agent quickly. The new solution provides our customers with instant responses to questions with zero wait time and helps them connect instantly with the right people when speaking to a live agent is needed.”Looking forwardWe’re excited to see how these improvements to speech recognition improve the customer experience for contact centers of all shapes and sizes—whether you’re working with one of our partners to deploy the Contact Center AI solution, or taking a DIY approach using our conversational AI suite. Learn more about both approaches via these links:Contact Center AI solutionsDialogflowCloud Speech-to-TextCloud Text-to-Speech
Quelle: Google Cloud Platform

IoT sensors and wearables revolutionize patient care

When was the last time you or a loved one went to the doctor or hospital? Things have changed dramatically over the last few years, with kiosks to register, portals to track your health history, and texts reminding you about upcoming appointments.

These changes have made a difference in how we interact with our healthcare providers. But there are more changes, not on the horizon, but here today. It is estimated as many as 50 billion medical devices will connect to clinicians, health systems, patients, and to each other.

Cardiac patient monitoring improvements

Imagine that you or a family member have periodic symptoms of irregular heartbeats, an all too common medical disorder known as an arrhythmia. If persistent, an arrhythmia can cause blood to clot in the heart, significantly increasing the risk of a heart attack or stroke. If caught early, a clot or blockage can be contained or cleared away and stent can be put in to keep blood flowing normally. In the US, more than 1.8 million stents are implanted annually, along with countless other preventative cardiac procedures to treat the 28.2 million US adults with diagnosed heart disease. Following any cardiac related procedure, a patient is typically counseled about the importance of exercise and nutrition then sent home. Post procedure, it’s common to be concerned about another cardiac related event in the immediate days following discharge, but what if the patient could be proactive, as well as reactive, to cardiac disease? 

Peerbridge Health, a New York-based remote patient monitoring company has developed the Peerbridge Cor™ (Cor), an award-winning, multi-channel wearable electrocardiogram (ECG) to better assist physicians and their patients detect and treat irregular heart activity, expeditiously. Prescribed by a cardiologist, the Cor is an elegant wearable worn 24 hours a day for up to 7 days with the ability to record every single heartbeat. In the event of abnormal cardiac activity, the patient can transmit select ECG activity to the prescribing physician’s care team for analysis at the press of a button. This continuous recording with transmitted events provides an unparalleled “window” into the patient’s heart activity as they go about their daily activities. Finally, an ECG monitoring solution providing critical data transmission patient’s expect with modern medicine.

Frustrated by all the wires in the hospital, Peerbridge Health was founded by Dr. Angelo Acquista when he was caring for his father in the ECU in 2006. Instead of getting up and walking around, his father was covered in wires and sensors unable to move, like most cardiac patients. Shortly after this experience, Dr. Acquista started Peerbridge Health, determined to change how this chronic disease is managed. Today, Peerbridge is a leading-edge manufacturer of the Peerbridge Cor (pictured above), the smallest and lightest FDA-cleared, multi-channel, wireless ECG.

The Peerbridge team selected Microsoft’s Azure IoT platform for their cardiac monitor because they, “saw the Microsoft IoT platform as being a foundational ingredient to help them grow and scale.” Peerbridge CEO, Adrian Gilmore states, “Azure IoT not only provides the secure data stream that is needed to monitor patients, it also offers cloud tools enabling us to present data in formats physicians expect, making the entire system a real revolution in cardiac care.” He continued, “Our engagement at the Microsoft AI and IoT Insider Lab, was the perfect opportunity for us to sharpen our team’s digital strategy, ensuring we optimize the company’s cloud architecture, and take full advantage of the variety of data services Microsoft offers.”

Avoiding diabetic amputations

Another company, Sensoria Health, a Seattle-based company, has taken on the problem of diabetes-related amputations. Why diabetes? Well, the statistics from the American Podiatric Medical Association are staggering:

More than 400 million people have diabetes worldwide
32 million people in the US have diabetes, costing more than 327 billion dollars
In the world today, a lower limb is lost to diabetes every 20 seconds
Cost in the US is estimated to be about 20 billion dollars

The typical progression to the amputation of a toe, foot, or more, always begins with a foot ulcer. The team at Sensoria asked themselves, “What can be done to expedite the healing of foot wounds to avoid amputations?” In response, Sensoria joined forces Optima to develop the Motus Smart powered by Sensoria®. It combines Sensoria® Core technologies, together with the clinically-tested Optima Molliter Offloading System, to take the pressure off the area of ulceration to improve blood circulation which is a critical factor to improve chance of healing.

Originally unveiled at the Consumer Electronics Show in January 2018, where it won Innovation Honoree Award, the Motus Smart, leverages Sensoria® Core to monitor activity and compliance, and is a clinically-proven and viable alternative to a total contact cast, and non-removable cam boot. The Sensoria® sensors work with a real-time app and alert system and an Azure based dashboard to inform patients, caregivers, and clinicians of non-compliant patients, allowing for easy and immediate intervention. The expensive and uncomfortable cast finally has an IoT, viable, and clinically-proven alternative with Motus Smart.

Why did Sensoria choose Microsoft’s Azure IoT platform for their patient monitoring devices? Davide Vigano, co-founder and CEO of Sensoria, shares in this video the three reasons why they selected Azure:

The richness of the development tools and already knowing how to use them
The openness of the platform and ability to use open source
Microsoft’s understanding and command of the enterprise market segment

Furthermore, Sensoria is using the Microsoft cloud and the Azure IoT platform to build a connected medical device platform, as they continue to develop new patient monitoring devices, like their smart sock v2.0 and Sensoria® Core, that drive improved outcomes for a variety of conditions. 

Learn more

Want to learn more about Microsoft and our work in healthcare? Check out our healthcare microsite, detailing our approach to the cloud and security, as well as compelling customer stories from Ochsner Health, BD, and others.
Quelle: Azure

An OpenShift Administrator’s Guide to Onboarding Applications

Infrastructure teams managing Red Hat OpenShift often ask me how to effectively onboard applications in production. OpenShift embeds many functionalities in a single product and it is fair to imagine an OpenShift administrator struggling to figure out what sort of conversations his team must have with an application team before successfully running an application on […]
The post An OpenShift Administrator’s Guide to Onboarding Applications appeared first on Red Hat OpenShift Blog.
Quelle: OpenShift