Distributed Speech Processing in MiPad’sMultimodal User Interface
-
Upload
madhav-krishna -
Category
Documents
-
view
225 -
download
0
Transcript of Distributed Speech Processing in MiPad’sMultimodal User Interface
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
1/22
Distributed SpeechProcessing in MiPads
Multimodal User Interface
Presented by-Madhav Krishna
C0-2
1120111
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
2/22
1. Introduction-
1.1 GUI vs Multimodal Interface-
GUI relies heavily on a graphical display, keyboard andpointing devices that are not alays available!
Mobile co"p#ters have constraints on physical si$e and
battery poer, or present li"itations d#e to hands-b#syeyes-b#sy scenarios hich "ake traditional GUI achallenge!
%poken lang#age enabled "#lti"odal inter&aces are idelybelieved to be capable o& dra"atically enhancing the
#sability o& co"p#ters beca#se GUI and speech haveco"ple"entary strengths! 'hile spoken lang#age has thepotential to provide a nat#ral interaction "odel, thedi(c#lty in resolving the a"big#ity o& spoken lang#age andthe high co"p#tational re)#ire"ents o& speech technologyhave so &ar prevented it &ro" beco"ing "ainstrea" in aco"p#ter*s #ser inter&ace! MiPad, M#lti"odal Interactive
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
3/22
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
4/22
1. More about MiPad!"or#ing$-MiPad intends to alleviate a prevailing proble" o& pecking ithtiny styl#ses or typing on "in#sc#le keyboards in today*sP+s by adding speech capability thro#gh a b#ilt-in "icrophone!
MiPad is designed to s#pport a variety o& tasks s#ch as /-"ail,voice-"ail, calendar, contact list, notes, eb brosing, "obilephone, and doc#"ent reading and annotation! his collection o&nctions #nies the vario#s "obile devices into a single,co"prehensive co""#nication and prod#ctivity tool!'hile the entire nctionality o& MiPad can be accessed by penalone, it as &o#nd that a better #ser eperience can beachieved by co"bining pen and speech inp#ts! 3ther pointingdevices, s#ch as a roller on the side o& the device, device &ornavigating a"ong the inp#t elds, can also be e"ployed to
enable one handed operation!.he speech inp#t "ethod, called Tap & Talk, not only indicateswhere the recognized tet sho#ld go b#t also serves as a p#sh totalk b#tton! Tap & Talk narrows down the number of possibleutterances for the spoken lang#age processing "od#le! 4or
ea"ple, selecting the 5To: eld on an e-mail applicationdisplay indicates that the #ser is abo#t to enter a na"e! his
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
5/22
3ne key &eat#re o& MiPad is a general p#rpose 5Co""and6 eld to hicha #ser can iss#e nat#rally spoken co""ands s#ch as 5%ched#le a "eeting
ith redd to"orro at to o*clock!64ro" the #ser*s perspective, MiPad not only recogni$es b#t understandsthe command by Miad e!ecuting the necessary actions conveyed in thespoken co""ands!In response to the above co""and, MiPad ill display a 5"eetingarrange"ent6screen ith related elds s#ch as date, ti"e, attendees, etc!. lledappropriately based on the #ser*s #tterance! MiPad lly i"ple"entsPersonal In&or"ation Manage"ent PIM. nctions incl#ding e"ail,calendar, notes, task, and contact list!+ll MiPad applications are cong#red in a client7server architect#re asshon-
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
6/22
he client on the le&t side o& 4ig! 2 is MiPad poered by Microso&t'indos C/ operating syste" that s#pports 1. so#nd capt#re, 2.
&ront-end aco#stic processing incl#ding noise red#ction, channelnor"ali$ation, &eat#re co"pression, and error protection, 8. GUIprocessing, and 9. a &a#lt-tolerant co""#nication layer thatallos the syste" to recover gracelly &ro" netork connection&ail#res!+ ireless local area netork ':+;. connects MiPad to a host
"achine server. here the contin#o#s speech recognition C% o& r#nti"e heap, and "erely cons#"es
approi"ately 8=? o& CPU load ith iP+@*s 20A MB$ %trong+
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
7/22
1.% &ationale 'ehind MiPads (rchitecture-+ltho#gh c#sto"i$ed syste" so&tare and hardare have been
reported to bring etra benets and eibility in tailoringapplications to "obile environ"ents, the MiPad proDect #tili$esonly oE-the-shel& hardare and so&tare! Given the rapidi"prove"ents in the hardare and syste" so&tare capabilities,it is believed s#ch an approach is a reasonable one!
%econd, altho#gh speaker independent speech recognition has"ade signicant strides d#ring the past to decades, e havedeliberately positioned MiPad as apersonal de#ice where the #serprole can be #tili$ed to enrich applications and co"ple"enttechnological shortco"ings! 4or speech, this "eans it "ay #sespeaker dependent recognition, thereby avoiding the challenges&aced by other approaches! In addition to enabling higherrecognition acc#racy, #ser specic in&or"ation can also be storedlocally and speaker specic processing can be carried o#t on theclient device itsel&! his architect#re allos #s to create #serc#sto"i$ed applications #sing generic
servers, thereby i"proving overall scalability!
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
8/22
. &obustness to (coustic)nvironments-I""#nity to noise and channel distortion is one o& the "ost
i"portant design considerations &or MiPad! 'ith theconvenience o& #sing the b#ilt-in "icrophone, noise rob#stnessbeco"es a key challenge to "aintaining desirable speechrecognition and #nderstanding per&or"ance!his section ill present "ost recent res#lts in the &ra"eorko& distrib#ted speech recognition %
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
9/22
.1 'asic version of SP*I+)-
%P:IC/ is a &ra"e-based, bias re"oval algorith" &or cepstr#"enhance"ent #nder additive noise, channel distortion or aco"bination o& the to!%P:IC/ ass#"es no eplicit noise "odel, and the noisecharacteristics are e"bedded in the pieceise linear "appingbeteen the 5stereo6 clean and distorted speech cepstralvectors! he pieceise linearity is intended to approi"ate thetr#e nonlinear relationship beteen the to!%P:IC/ is potentially able to handle a ide range o& distortions,incl#ding non-stationary distortion, Doint additive and
convol#tional distortion, and nonlinear distortion in ti"e-do"ain. beca#se the stereo data provides acc#rate esti"ates o&the bias or correction vectors itho#t the need &or an eplicitnoise "odel!3ne key re)#ire"ent &or the s#ccess o& the basic version o&
%P:IC/ described here is that the distortion conditions #nderhich the correction vectors are
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
10/22
. )nhancing SP*I+) b, emporal Smoothing-
In this enhanced version o& %P:IC/, e not only "ini"i$e thestatic deviation &ro" the clean to noisy cepstral vectors as inthe basic version o& %P:IC/., b#t also seek to "ini"i$e thedyna"ic deviation!he basic %P:IC/ opti"ally processes each &ra"e o& noisyspeech independently! +n obvio#s etension is to Dointly processa seg"ent o& &ra"es! In this ay, altho#gh the deviation &ro"the clean to noisy speech cepstra &or an individ#al &ra"e co#ldbe #ndesirably greater than that achieved by the basic, static%P:IC/, the overall deviation that takes into acco#nt the holese)#ence o& &ra"es and the "is"atch o& slopes ill be red#ced
co"pared ith the basic %P:IC/!
.% )nhancing SP*I+) b, oise estimation and oiseormali/ation-In this enhance"ent o& %P:IC/, diEerent noise conditions
beteen the %P:IC/ training set and test set are nor"ali$ed!he research shoed that the eEectiveness o& the above noise-
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
11/22
.%.1 $on-stationary $oise %stimation by terati#e 'tochastic
(ppro!imation-+ novel algorith" is proposed, i"ple"ented, and eval#ated &orrec#rsive esti"ation o& para"eters in a nonlinear "odel involvinginco"plete data! he algorith" is applied specically to ti"e-varying deter"inistic para"eters o& additive noise in a "ildlynonlinear "odel that acco#nts &or the generation o& the cepstraldata o& noisy speech &ro" the cepstral data o& the noise andclean speech!
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
12/22
%. 0eature +ompression and )rrorProtection-
his is intended to address the three key re)#ire"ents &ors#ccessl deploy"ent o& distrib#ted speech recognitionassociated ith the client7server approachF 1. co"pression o&cepstral &eat#res via )#anti$ation. "#st not degrade speechrecognition per&or"ance 2. the algorith" &or so#rce andchannel coding "#st be rob#st to packet losses, b#rsty orotherise and8. the total ti"e delay d#e to the coding, hich res#lts &ro" aco"bined )#anti$ation delay, error-correction coding delay, andtrans"ission delay, "#st be kept ithin an acceptable level!
%.1 0eature +ompression-+ ne so#rce coding algorith" has been developed thatconsists o& to se)#ential stages! +&ter the standard Mel-cepstra are etracted, each speech &ra"e is rst classied to aphonetic category e!g!, phone"e. and then is vector )#anti$ed
H@. #sing the split-H@ approach! he "otivation behind thisne so#rce coder is that the speech signal can be co"posed o&
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
13/22
%. )rror Protection-+ novel channel coder has also been developed to protectMiPad*s Mel-cepstral &eat#res based on the client7serverarchitect#re! he channel coder assigns #ne)#al a"o#nts o&red#ndancy a"ong the diEerent so#rce bits, giving a greatera"o#nt o& protection to the "ost i"portant bits here thei"portance
is "eas#red by the contrib#tions o& these bits to the ord errorrate in speech recognition!+ )#antiable proced#re to assess the i"portance o& each bit isdeveloped, and the channel coder eploits this #tility nction &orthe opti"al &orard error correction 4/C. assign"ent!
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
14/22
. +ontinuous Speech recognitionand Understanding-
'hile the co"pressed and error-protected Mel-cepstral&eat#res are co"p#ted in the MiPad client, "aDor co"p#tation&or contin#o#s speech recognition decoding. resides in theserver! he entire set o& the lang#age "odel, hidden Markov"odels BMMs., and leicon that are #sed &or speech decodingall reside in the server, hich processes the Mel-cepstral&eat#res trans"itted &ro" the client!MiPad is designed to be a personal device! +s a res#lt, speechrecognition #ses speaker-adaptive aco#stic "odels BMMs. anda #ser-adapted leicon to i"prove recognition acc#racy!he speech recognition engine in MiPad #ses the #nied
lang#age "odel that takes advantage o& both r#le-based anddata-driven approaches! Consider to training sentencesF)Meeting at three with *ohn 'mith" #ersus )Meeting at four Mwith +erek"'ithin a p#re -gra" &ra"eork, e need to esti"ate ohn
three ith. and erek PM ith. individ#ally! his "akes itvery di(c#lt to capt#re the obvio#sly needed long-span
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
15/22
4or the ea"ple listed here, e "ay have C4Gs &or ;+M/ and IM/
respectively, hich can be derived &ro" the &actoid gra""ars o& s"allersi$es!he training sentences no look likeF)Meeting at three:TM% with *ohn 'mith:$(M% , and )Meeting at four M:TM% with +erek: $(M% "'ith parsed training data, e can no esti"ate the n-gra" probabilities
as #s#al! 4or ea"ple, the replace"ent o& ohn three ith.J;+M/I JIM/ ith . "akes s#ch 5n -gra"6 representation "ore"eaningl and "ore acc#rate!Inside each C4G, hoever, e can still derive 5ohn %"ith6L J;+M/.and 5&o#r PM6 JIM/. , &ro" the eisting n-gra" n-gra" probabilityinheritance. so that they are appropriately nor"ali$ed!
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
16/22
2. MiPad User Interface Designand )valuation-
+s "entioned previo#sly, MiPad does not e"ploy speechsynthesis as an o#tp#t "ethod! his design decision is"otivated "ainly by the &olloing to reasons! 4irst, despitethe signicant progress in synthesis technologies, especially inthe area o& concatenated ave&or"s, the )#ality o& synthesi$edspeech has re"ained #nsatis&actory &or large scaledeploy"ents! he "ost critical draback o& speech o#tp#t isith the non-persistent or #olatile nature of speechpresentation!he h#"an #ser "#st process the speech "essage and"e"ori$e the contents o& the "essage in real ti"e! here is no
knon #ser inter&ace design that can elegantly assist theh#"an #ser &or the cases here the speech ave&or" cannotbe easily heard and #nderstood, or there is si"ply too "#chin&or"ation to be absorbed! In contrast, a graphical display canrender a large a"o#nt o& in&or"ation persistently &or leis#re
cons#"ption, avoiding the a&ore"entioned proble"s!
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
17/22
MiPad takes advantage o& the graphical display in UI design!he graphical display si"plies dra"atically the dialog"anage"ent! 4or instance, MiPad is able to considerablystrea"line the conr"ation and error repair strategy as all thein&erred #ser intentions are conr"ed implicitly on the screen".hene#er an error occ#rs, the #ser can correct it thro#gh theGUI or speech "odalities that are appropriate and appear "orenat#ral to the #ser! hanks to the display persistency, #sers arenot obligated to correct errors i""ediately a&ter they occ#r!
he display also allos MiPad to conr" and ask the #ser "any)#estions in a single t#rn!
2.1 ap 3 al# Interface->eca#se o& MiPad*s s"all &or"-&actor, the present pen-based"ethods &or getting tet into a P +are potential barriers tobroad "arket acceptance! %peech is generally not as precise as
a "o#se or a pen to per&or" position-related operations!%peech interaction can also be adversely aEected by
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
18/22
espite these disadvantages, speech co""#nication is not
only nat#ral b#t also provides a poerl co"ple"entary"odality to enhance the pen-based inter&ace i& the strengthso& #sing speech can be appropriately leveraged and thetechnology li"itations be overco"e!In able II, e elaborate several cases hich sho that penand speech can be co"ple"entary and #sed eEectively &orhandheld devices! he advantage o& pen is typically theeakness o& speech and vice versa!
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
19/22
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
20/22
2. 4isual 0eedbac#s for Speech Inputs-
Processing latency is a ell recogni$ed iss#e in #ser inter&acedesign! his is even "ore so &or MiPad in hich distrib#tedspeech recognition is e"ployed! In addition to the recognitionprocess itsel&, the ireless netork rther introd#ces "orelatency that so"eti"es is not easily controllable! Conventionalisdo" &or UI design dictates that lling the ti"e ith vis#al
&eedback not only signicantly i"proves the #sability, b#t alsoprevents #sers &ro" adversely intervening an ongoing processthat cannot be easily recoverable! 4or these reasons, MiPadadopts a vis#al &eedback &or speech inp#ts!
2.% User Stud, &esults-
he goal is to "ake MiPad prod#ce real val#e to #sers! It isnecessary to have a rigoro#s eval#ation to "eas#re the #sabilityo& the prototype! 3#r "aDor concerns areF)s the task completion time much better/ and )s it easier toget the 0ob done/
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
21/22
52 s t %asier to 6et The *ob +one/: 4i&teen o#t o& the 1Aparticipants in the eval#ation stated that they pre&erred #sing
the Tap & Talk inter&ace &or creating ne appoint"ents and all1A said they pre&erred it &or riting longer e"ails!
-
7/25/2019 Distributed Speech Processing in MiPads Multimodal User Interface
22/22
5. &eferences-