PLS - what is it, what “gap” does it fill?

What PLS did is this:

  • First, it provided a single, standard XML-based language for describing pronunciations, both for speech recognizers and for speech synthesizers.
  • Second, it requires support for IPA, the International Phonetic Alphabet. This Alphabet is a standard symbol set for representing pronunciations of all the languages of the world.

With PLS it is now possible to write one lexicon document that can be used by any speech recognizer and/or any speech synthesizer that supports it. One document for all of your pronunciations, independent of your voice technology vendor.  [click heading for more]

W3C Standard Simplifies Creation of Speech-Enabled Web Applications


W3C published today a standard that will simplify the development of Web applications that speak and listen to users. The Pronunciation Lexicon Specification (PLS) 1.0 is the newest piece of W3C's Speech Interface Framework for creating Web applications driven by voice and speech. PLS can reduce the cost of developing these applications by allowing people to share and reuse pronunciation dictionaries. In addition, PLS can make it easier to localize applications by separating pronunciation concerns from other parts of the application. [click heading for more]

W3C examines the next generation of speech technology

[nik's note:]

The W3C on Tuesday said the next generation of VoiceXML will include specifications for speaker verification.

"Speaker verification and identification is not only the best biometric for securing telephone transactions and communications, it can work seamlessly with speech recognition and speech synthesis in VoiceXML deployments," Ken Rehor, newly elected chairman of the VoiceXML Forum, said in a statement.
The W3C has now completed its desired requirements for VoiceXML 3.0 and expects to have a working draft of the specifications by the end of the first quarter, said James Larson, co-chair of the W3C Voice Browser Working Group.
In addition to the speaker identification requirements for VoiceXML 3.0, the W3C addressed the issue of extending its Speech Synthesis Markup Language (SSML) functionality to certain languages including Mandarin, Japanese and Korean.

[click heading for more]

Vicorp strengthens its relationship with the US market

UK based Vicorp is delighted to announce its partnership with Newfound Communications, a US based speech solutions provider. Not only does the partnership increase Vicorp presence in the US market, but also broadens their portfolio.

Vicorp and Newfound Communications are established in their own countries, but this partnership presents both companies with exciting opportunities to extend their reach into the US and Europe.

"Our companies share the same desire to help organisations and developers deliver superior features, which enhance over-the-phone contact centre experiences for the end caller," says Craig Ashmole, Vicorp VP of Alliances and New Business.

"Newfound Communications have enabled us to extend our portfolio to offer assisted outbound IVR, voice recording and enriched self-service branding. This capability takes us ever closer to our mission to provide 100% call resolution for our customers. We are excited by the potential of this partnership."

Both Vicorp's and Newfound Communications' technology is XML based and complements one another for a richer integrated offering. Vicorp's service creation and execution environment, xMP, in conjunction with Newfound Communications' products including call recording, media mixer and outbound dialler all contribute to a stronger portfolio for both companies. In using Vicorp service creation and execution environment, xMP, Newfound Communications will now be able to pass onto their customers benefits such as editable and re-usable component technology. [click heading for more]

Voice enabling XML, Part 1: Develop a voice-enabled RSS reader

RSS is a hot topic these days, as it provides an easy way to stream data online. This article, the first of a four-part series on developing VoiceXML applications, shows you how to develop a voice-enabled RSS reader. The input to the application is RSS data, and the output is VoiceXML that can be read and spoken by your favorite compatible voice application.