16:01:07 RRSAgent has joined #htmlspeech 16:01:07 logging to http://www.w3.org/2011/06/16-htmlspeech-irc 16:01:17 trackbot, start telcon 16:01:19 RRSAgent, make logs public 16:01:21 Zakim, this will be 16:01:21 I don't understand 'this will be', trackbot 16:01:22 Meeting: HTML Speech Incubator Group Teleconference 16:01:22 Date: 16 June 2011 16:01:26 zakim, aacc is Dan_Druta 16:01:26 +Dan_Druta; got it 16:01:32 + +1.425.830.aadd 16:01:43 zakim, nick DanD is Dan_Druta 16:01:43 ok, burn, I now associate DanD with Dan_Druta 16:01:52 zakim, aadd is Charles_Hemphill 16:01:52 +Charles_Hemphill; got it 16:02:09 zakim, nick Charles is Charles_Hemphill 16:02:09 ok, burn, I now associate Charles with Charles_Hemphill 16:02:14 zakim, who's here? 16:02:14 On the phone I see +1.818.237.aaaa, Milan_Young, [Microsoft], Michael_Johnston, Dan_Burnett, Michael_Bodell, Olli_Pettay, Dan_Druta, Charles_Hemphill 16:02:16 On IRC I see RRSAgent, satish, DanD, mbodell, Charles, Zakim, smaug, burn, Milan, Robert, trackbot 16:02:35 zakim, aaaa is Patrick_Ehlen 16:02:35 +Patrick_Ehlen; got it 16:02:49 zakim, [Microsoft] is Robert_Brown 16:02:49 +Robert_Brown; got it 16:03:01 MichaelJ has joined #htmlspeech 16:03:44 -Olli_Pettay 16:04:20 Scribe: Patrick_Ehlen 16:05:37 ehlen has joined #htmlspeech 16:05:38 satish, will you be joining us today? or anyone from Google? 16:05:47 zakim, nick ehlen is Patrick_Ehlen 16:05:47 ok, burn, I now associate ehlen with Patrick_Ehlen 16:05:52 ScribeNick: ehlen 16:06:11 zakim, who's here? 16:06:11 On the phone I see Patrick_Ehlen, Milan_Young, Robert_Brown, Michael_Johnston, Dan_Burnett, Michael_Bodell, Dan_Druta, Charles_Hemphill 16:06:14 On IRC I see ehlen, MichaelJ, RRSAgent, satish, DanD, mbodell, Charles, Zakim, burn, Milan, Robert, trackbot 16:06:14 smaug has joined #htmlspeech 16:06:27 Chair: Dan_Burnett 16:06:39 Agenda: http://lists.w3.org/Archives/Public/public-xg-htmlspeech/2011Jun/0033.html 16:06:45 burn: I can't join the conf call as I'm on a train, hence only in IRC 16:06:56 burn: Bjorn is still on paternity leave, not sure about Glen Shires 16:07:24 +??P21 16:07:29 satish, thanks. Hopefully Glen will join. We will be making decisions today about other discussion topics 16:07:33 Zakim, ??21 is Olli_Pettay 16:07:33 sorry, smaug, I do not recognize a party named '??21' 16:07:51 Zakim, nick smaug is Olli_Pettay 16:07:51 sorry, smaug, I do not see a party named 'Olli_Pettay' 16:08:01 Topic: New design decisions? 16:08:18 Zakim, ??P21 is Olli_Pettay 16:08:20 +Olli_Pettay; got it 16:08:34 Zakim, nick smaug is Olli_Pettay 16:08:34 ok, smaug, I now associate you with Olli_Pettay 16:08:37 robert: is audio recording without recognition be supported? 16:09:00 are there important scenarios for supporting recording without recognition 16:09:51 satish, any update on markup binding? 16:10:12 topic: markup binding 16:11:33 burn: None, Bjorn was collecting input from the chrome team and since he has gone on leave I have no contact on what the status was. 16:12:21 satish, can you please check? we are not waiting on the answer, but it would be nice to have the input 16:12:21 robert: google issue on whether there should be a button to press 16:12:43 burn: yes, I can take an action to get a definitive answer in the next few days. 16:13:07 burn: satish will take this on w/ the chrome team 16:13:12 topic: discussion time 16:13:36 topic: do we need to support audio recording with recognition? 16:14:38 burn: an advantage could be endpointing. 16:14:47 ... is that an important criteria in this case as well? 16:15:15 charles: another question is how real-time is the reco response? 16:15:23 ... a recording may result in reco later 16:16:22 ... an identifier might later associate the recording with a reco transcription 16:16:50 burn: brings up question of whether we support reco on recorded audio 16:17:54 robert: garbage models could be used to make recording in edge cases 16:18:01 ... "overloading" recognition 16:18:15 ... or will recording be a more common task 16:18:48 ... Do we think recording with endpointing is important? 16:20:18 milan: channel adaptation, sharing headers in same structure, parameters could be reused; sharing the same network paths -- convenient to use same 16:21:09 Charles: Also, the on-line vs. off-line cases 16:22:20 milan: would most recording be associated with an attempt to understand the text in the recording? 16:23:00 burn: Most significant feature is the endpointing 16:23:26 milan: in that case, why not just use dict model, do reco, and save the waveform as backup? 16:24:06 ... and how common would that be. If not so common, could use a garbage model (even a "first-class" one) 16:24:41 burn: seems strange to call recording a weird special case of reco 16:26:04 burn: in favor of using the recording resource as described in mrcp 16:27:16 robert: though endpointing may be valuable, would we support a "record" object in the API? how would this go all the way to the developer? 16:27:44 burn: does not seem to be in our scope 16:28:00 olli: there are other proposals that would handle recording 16:28:54 charles: channel adaptation 16:30:39 burn: channel normalization is not a valid reason for recording support 16:31:32 charles: should probably also include built-in record grammar 16:32:04 (milan above) 16:35:03 milan: use case: may want to to do dictation in parallel with c&c 16:35:21 ... e.g., provide a c&c followed immediately by dictation 16:36:18 burn: but does that really belong as a built-in type in a grammar? 16:36:33 ... sounds like there is not real consensus today vis-a-vis supporting a recording capability 16:37:06 robert: have not heard a compelling reason to support recording 16:38:08 burn: consensus not to do it now 16:38:23 milan: would like a standard way to do it, should the need arise 16:38:55 burn: we could state that we reserve this for the future 16:39:14 milan: there should be some consistent and portable way to do this across engines 16:39:59 robert: could be done as a proprietary extension 16:40:37 milan: at least provide a consistent hack, like builtin:record 16:40:45 robert: that's what the garbage model recording would be 16:41:27 milan: that's fine, as long as all engines support this type of garbage model 16:42:30 burn: to summarize, can't agree on specific recording scenarios 16:42:46 (robert above) 16:43:06 ... should agree on supporting garbage-recording scenario 16:43:26 burn: as a group, agree not to define an explicit recording capability at this time. 16:43:47 ... can be supported using a garbage model, or capabilities defined outside this group 16:45:05 topic: what are the built-ins, and what does that mean? 16:45:55 milan: existing builtins: dictation, search, address, numbers 16:46:07 robert: already agreed there should be a certain set of predefined grammars 16:46:14 ... so how do we refer to those? 16:46:53 burn: 2 things make builtins interesting: (1) parameterization; (2) no language is required 16:50:05 milan: markup already has certain defined types, parameters, etc, as native to HTML5. Would make sense to pay attention to that here 16:50:34 burn: an unconstrained text box should naturally bind to a dictation model 16:50:54 milan: should we remap the names of the builtins? 16:51:58 burn: argue strongly for using html as a starting point 16:52:33 robert: These should be builtins, not re-used vxml grammars 16:53:19 could someone paste a link to voicexml's builtin grammars ? 16:53:27 charles: they've become a de facto standard; not supporting them is awkward 16:54:16 these are the HTML input types: http://www.w3.org/TR/html5/the-input-element.html#attr-input-type 16:54:24 burn: if someone wants to support legacy builtins in a way that doesn't break existing builtins, that's not a problem 16:54:28 perhaps have builtins that match these 16:57:00 charles: there needs to be some way to include these 16:57:23 (milan above) 16:57:51 ... is there something about this that can't be represented by a query string? 16:59:06 michael: do you want to reference, for example, an html number type, or some arbitrary number? 17:00:37 milan: easier to use old builtins & augment them 17:00:52 charles: need to look at greater good of using html vs vxml 17:01:21 Widely implemented? See http://en.wikipedia.org/wiki/URI_scheme 17:04:51 burn: michael, how would you reference grammars that are assoc. with html input types? 17:06:00 michael: an html ruleref, with various attributes; or don't specify URI and ref them by markeup AP... 17:06:16 ... most important is associating grammars with individual input elements 17:08:37 ... not a strong use case to have URIs for these things, or ability for user to write their own that reference these 17:10:05 burn: when people want to hack something up quickly, common input types should lend themselves to being included as part of a larger utterance 17:10:20 michael: may be other ways to specify input for that type of scenarios 17:11:33 burn: maybe reference not the grammar but the input type itself 17:12:58 charles: similar input types not always require the same grammar 17:14:02 burn: but the app author may want a way to link these different types of builtin grammars together 17:14:50 milan: perhaps just do the proposal 17:15:43 burn: who on the call is interested in builtin models? 17:16:16 charles: interested in it; this group seems focused on web search and dictation, as opposed to broader html cases 17:16:25 17:17:28 michael: there will probably be a standard set of grammar libraries, though perhaps the market will provide those 17:17:57 johnston: can't see us requiring something like a "zip code" lib, for internationalization reasons 17:18:10 burn: HTML has already handled a lot of these issues 17:18:21 (milan above) 17:18:43 (michael, above, actually) 17:18:51 s/burn/michael/ 17:20:04 milan: should there be an html binding? 17:20:42 michael: would be better if you could speech enable certain input types with little work 17:22:09 robert: if no builtins were specified, what are the consequences? 17:22:49 burn: if you want broad adoptability and usage, it needs to be as easy to create simple apps as vxml 17:23:15 robert: we need it to do the html binding. 17:23:49 ... so how much do we need the html binding part? 17:24:06 milan: definitely need the capability to specify search, dictation, etc. 17:24:31 robert: that's different from looking at html input types, etc. that's a complex problem 17:24:55 milan: would like to have a notion of how to solve binding problem before we do dictation 17:25:58 robert: does anyone have a proposal to volunteer? 17:26:11 milan: perhaps can do it after I get the dictation stuff out 17:26:56 micheal: there is a topic in the API about markup bindings. 17:27:10 burn: true that it's a binding issue 17:28:13 burn: without a proposal, it doesn't happen. 17:28:23 ... so it will be up to someone to write a proposal 17:28:36 milan: perhaps sending a message to google on this 17:28:42 robert: or to satish 17:29:11 burn: action item for milan to talk with satish and ask for help on structuring a proposal 17:29:34 burn: reminder: no call next week 17:29:49 robert: but there will be a protocol meeting 17:30:52 -Michael_Bodell 17:30:53 -Milan_Young 17:30:56 -Robert_Brown 17:31:07 -Olli_Pettay 17:31:08 -Patrick_Ehlen 17:31:09 -Michael_Johnston 17:31:23 -Dan_Burnett 17:31:24 zakim, bye 17:31:24 Zakim has left #htmlspeech 17:31:25 leaving. As of this point the attendees were +1.818.237.aaaa, Milan_Young, +1.973.452.aabb, Michael_Johnston, Dan_Burnett, Michael_Bodell, Olli_Pettay, +1.425.580.aacc, Dan_Druta, 17:31:27 ... +1.425.830.aadd, Charles_Hemphill, Patrick_Ehlen, Robert_Brown 17:31:33 rrsagent, make log public 17:31:38 rrsagent, draft minutes 17:31:38 I have made the request to generate http://www.w3.org/2011/06/16-htmlspeech-minutes.html burn 17:32:34 Regrets: Raj_Tumuluri, Bjorn_Bringert 17:35:28 s/+1.818.237.aaaa, // 17:35:41 s/+1.973.452.aabb, // 17:35:57 s/+1.425.580.aacc, // 17:36:12 s/+1.425.830.aadd, // 17:36:16 rrsagent, draft minutes 17:36:16 I have made the request to generate http://www.w3.org/2011/06/16-htmlspeech-minutes.html burn 17:37:18 rrsagent, bye 17:37:18 I see no action items