16:01:07 RRSAgent has joined #htmlspeech
16:01:07 logging to http://www.w3.org/2011/06/16-htmlspeech-irc
16:01:17 trackbot, start telcon
16:01:19 RRSAgent, make logs public
16:01:21 Zakim, this will be
16:01:21 I don't understand 'this will be', trackbot
16:01:22 Meeting: HTML Speech Incubator Group Teleconference
16:01:22 Date: 16 June 2011
16:01:26 zakim, aacc is Dan_Druta
16:01:26 +Dan_Druta; got it
16:01:32 + +1.425.830.aadd
16:01:43 zakim, nick DanD is Dan_Druta
16:01:43 ok, burn, I now associate DanD with Dan_Druta
16:01:52 zakim, aadd is Charles_Hemphill
16:01:52 +Charles_Hemphill; got it
16:02:09 zakim, nick Charles is Charles_Hemphill
16:02:09 ok, burn, I now associate Charles with Charles_Hemphill
16:02:14 zakim, who's here?
16:02:14 On the phone I see +1.818.237.aaaa, Milan_Young, [Microsoft], Michael_Johnston, Dan_Burnett, Michael_Bodell, Olli_Pettay, Dan_Druta, Charles_Hemphill
16:02:16 On IRC I see RRSAgent, satish, DanD, mbodell, Charles, Zakim, smaug, burn, Milan, Robert, trackbot
16:02:35 zakim, aaaa is Patrick_Ehlen
16:02:35 +Patrick_Ehlen; got it
16:02:49 zakim, [Microsoft] is Robert_Brown
16:02:49 +Robert_Brown; got it
16:03:01 MichaelJ has joined #htmlspeech
16:03:44 -Olli_Pettay
16:04:20 Scribe: Patrick_Ehlen
16:05:37 ehlen has joined #htmlspeech
16:05:38 satish, will you be joining us today? or anyone from Google?
16:05:47 zakim, nick ehlen is Patrick_Ehlen
16:05:47 ok, burn, I now associate ehlen with Patrick_Ehlen
16:05:52 ScribeNick: ehlen
16:06:11 zakim, who's here?
16:06:11 On the phone I see Patrick_Ehlen, Milan_Young, Robert_Brown, Michael_Johnston, Dan_Burnett, Michael_Bodell, Dan_Druta, Charles_Hemphill
16:06:14 On IRC I see ehlen, MichaelJ, RRSAgent, satish, DanD, mbodell, Charles, Zakim, burn, Milan, Robert, trackbot
16:06:14 smaug has joined #htmlspeech
16:06:27 Chair: Dan_Burnett
16:06:39 Agenda: http://lists.w3.org/Archives/Public/public-xg-htmlspeech/2011Jun/0033.html
16:06:45 burn: I can't join the conf call as I'm on a train, hence only in IRC
16:06:56 burn: Bjorn is still on paternity leave, not sure about Glen Shires
16:07:24 +??P21
16:07:29 satish, thanks. Hopefully Glen will join. We will be making decisions today about other discussion topics
16:07:33 Zakim, ??21 is Olli_Pettay
16:07:33 sorry, smaug, I do not recognize a party named '??21'
16:07:51 Zakim, nick smaug is Olli_Pettay
16:07:51 sorry, smaug, I do not see a party named 'Olli_Pettay'
16:08:01 Topic: New design decisions?
16:08:18 Zakim, ??P21 is Olli_Pettay
16:08:20 +Olli_Pettay; got it
16:08:34 Zakim, nick smaug is Olli_Pettay
16:08:34 ok, smaug, I now associate you with Olli_Pettay
16:08:37 robert: is audio recording without recognition be supported?
16:09:00 are there important scenarios for supporting recording without recognition
16:09:51 satish, any update on markup binding?
16:10:12 topic: markup binding
16:11:33 burn: None, Bjorn was collecting input from the chrome team and since he has gone on leave I have no contact on what the status was.
16:12:21 satish, can you please check? we are not waiting on the answer, but it would be nice to have the input
16:12:21 robert: google issue on whether there should be a button to press
16:12:43 burn: yes, I can take an action to get a definitive answer in the next few days.
16:13:07 burn: satish will take this on w/ the chrome team
16:13:12 topic: discussion time
16:13:36 topic: do we need to support audio recording with recognition?
16:14:38 burn: an advantage could be endpointing.
16:14:47 ... is that an important criteria in this case as well?
16:15:15 charles: another question is how real-time is the reco response?
16:15:23 ... a recording may result in reco later
16:16:22 ... an identifier might later associate the recording with a reco transcription
16:16:50 burn: brings up question of whether we support reco on recorded audio
16:17:54 robert: garbage models could be used to make recording in edge cases
16:18:01 ... "overloading" recognition
16:18:15 ... or will recording be a more common task
16:18:48 ... Do we think recording with endpointing is important?
16:20:18 milan: channel adaptation, sharing headers in same structure, parameters could be reused; sharing the same network paths -- convenient to use same
16:21:09 Charles: Also, the on-line vs. off-line cases
16:22:20 milan: would most recording be associated with an attempt to understand the text in the recording?
16:23:00 burn: Most significant feature is the endpointing
16:23:26 milan: in that case, why not just use dict model, do reco, and save the waveform as backup?
16:24:06 ... and how common would that be. If not so common, could use a garbage model (even a "first-class" one)
16:24:41 burn: seems strange to call recording a weird special case of reco
16:26:04 burn: in favor of using the recording resource as described in mrcp
16:27:16 robert: though endpointing may be valuable, would we support a "record" object in the API? how would this go all the way to the developer?
16:27:44 burn: does not seem to be in our scope
16:28:00 olli: there are other proposals that would handle recording
16:28:54 charles: channel adaptation
16:30:39 burn: channel normalization is not a valid reason for recording support
16:31:32 charles: should probably also include built-in record grammar
16:32:04 (milan above)
16:35:03 milan: use case: may want to to do dictation in parallel with c&c
16:35:21 ... e.g., provide a c&c followed immediately by dictation
16:36:18 burn: but does that really belong as a built-in type in a grammar?
16:36:33 ... sounds like there is not real consensus today vis-a-vis supporting a recording capability
16:37:06 robert: have not heard a compelling reason to support recording
16:38:08 burn: consensus not to do it now
16:38:23 milan: would like a standard way to do it, should the need arise
16:38:55 burn: we could state that we reserve this for the future
16:39:14 milan: there should be some consistent and portable way to do this across engines
16:39:59 robert: could be done as a proprietary extension
16:40:37 milan: at least provide a consistent hack, like builtin:record
16:40:45 robert: that's what the garbage model recording would be
16:41:27 milan: that's fine, as long as all engines support this type of garbage model
16:42:30 burn: to summarize, can't agree on specific recording scenarios
16:42:46 (robert above)
16:43:06 ... should agree on supporting garbage-recording scenario
16:43:26 burn: as a group, agree not to define an explicit recording capability at this time.
16:43:47 ... can be supported using a garbage model, or capabilities defined outside this group
16:45:05 topic: what are the built-ins, and what does that mean?
16:45:55 milan: existing builtins: dictation, search, address, numbers
16:46:07 robert: already agreed there should be a certain set of predefined grammars
16:46:14 ... so how do we refer to those?
16:46:53 burn: 2 things make builtins interesting: (1) parameterization; (2) no language is required
16:50:05 milan: markup already has certain defined types, parameters, etc, as native to HTML5. Would make sense to pay attention to that here
16:50:34 burn: an unconstrained text box should naturally bind to a dictation model
16:50:54 milan: should we remap the names of the builtins?
16:51:58 burn: argue strongly for using html as a starting point
16:52:33 robert: These should be builtins, not re-used vxml grammars
16:53:19 could someone paste a link to voicexml's builtin grammars ?
16:53:27 charles: they've become a de facto standard; not supporting them is awkward
16:54:16 these are the HTML input types: http://www.w3.org/TR/html5/the-input-element.html#attr-input-type
16:54:24 burn: if someone wants to support legacy builtins in a way that doesn't break existing builtins, that's not a problem
16:54:28 perhaps have builtins that match these
16:57:00 charles: there needs to be some way to include these
16:57:23 (milan above)
16:57:51 ... is there something about this that can't be represented by a query string?
16:59:06 michael: do you want to reference, for example, an html number type, or some arbitrary number?
17:00:37 milan: easier to use old builtins & augment them
17:00:52 charles: need to look at greater good of using html vs vxml
17:01:21 Widely implemented? See http://en.wikipedia.org/wiki/URI_scheme
17:04:51 burn: michael, how would you reference grammars that are assoc. with html input types?
17:06:00 michael: an html ruleref, with various attributes; or don't specify URI and ref them by markeup AP...
17:06:16 ... most important is associating grammars with individual input elements
17:08:37 ... not a strong use case to have URIs for these things, or ability for user to write their own that reference these
17:10:05 burn: when people want to hack something up quickly, common input types should lend themselves to being included as part of a larger utterance
17:10:20 michael: may be other ways to specify input for that type of scenarios
17:11:33 burn: maybe reference not the grammar but the input type itself
17:12:58 charles: similar input types not always require the same grammar
17:14:02 burn: but the app author may want a way to link these different types of builtin grammars together
17:14:50 milan: perhaps just do the proposal
17:15:43 burn: who on the call is interested in builtin models?
17:16:16 charles: interested in it; this group seems focused on web search and dictation, as opposed to broader html cases
17:16:25
17:17:28 michael: there will probably be a standard set of grammar libraries, though perhaps the market will provide those
17:17:57 johnston: can't see us requiring something like a "zip code" lib, for internationalization reasons
17:18:10 burn: HTML has already handled a lot of these issues
17:18:21 (milan above)
17:18:43 (michael, above, actually)
17:18:51 s/burn/michael/
17:20:04 milan: should there be an html binding?
17:20:42 michael: would be better if you could speech enable certain input types with little work
17:22:09 robert: if no builtins were specified, what are the consequences?
17:22:49 burn: if you want broad adoptability and usage, it needs to be as easy to create simple apps as vxml
17:23:15 robert: we need it to do the html binding.
17:23:49 ... so how much do we need the html binding part?
17:24:06 milan: definitely need the capability to specify search, dictation, etc.
17:24:31 robert: that's different from looking at html input types, etc. that's a complex problem
17:24:55 milan: would like to have a notion of how to solve binding problem before we do dictation
17:25:58 robert: does anyone have a proposal to volunteer?
17:26:11 milan: perhaps can do it after I get the dictation stuff out
17:26:56 micheal: there is a topic in the API about markup bindings.
17:27:10 burn: true that it's a binding issue
17:28:13 burn: without a proposal, it doesn't happen.
17:28:23 ... so it will be up to someone to write a proposal
17:28:36 milan: perhaps sending a message to google on this
17:28:42 robert: or to satish
17:29:11 burn: action item for milan to talk with satish and ask for help on structuring a proposal
17:29:34 burn: reminder: no call next week
17:29:49 robert: but there will be a protocol meeting
17:30:52 -Michael_Bodell
17:30:53 -Milan_Young
17:30:56 -Robert_Brown
17:31:07 -Olli_Pettay
17:31:08 -Patrick_Ehlen
17:31:09 -Michael_Johnston
17:31:23 -Dan_Burnett
17:31:24 zakim, bye
17:31:24 Zakim has left #htmlspeech
17:31:25 leaving. As of this point the attendees were +1.818.237.aaaa, Milan_Young, +1.973.452.aabb, Michael_Johnston, Dan_Burnett, Michael_Bodell, Olli_Pettay, +1.425.580.aacc, Dan_Druta,
17:31:27 ... +1.425.830.aadd, Charles_Hemphill, Patrick_Ehlen, Robert_Brown
17:31:33 rrsagent, make log public
17:31:38 rrsagent, draft minutes
17:31:38 I have made the request to generate http://www.w3.org/2011/06/16-htmlspeech-minutes.html burn
17:32:34 Regrets: Raj_Tumuluri, Bjorn_Bringert
17:35:28 s/+1.818.237.aaaa, //
17:35:41 s/+1.973.452.aabb, //
17:35:57 s/+1.425.580.aacc, //
17:36:12 s/+1.425.830.aadd, //
17:36:16 rrsagent, draft minutes
17:36:16 I have made the request to generate http://www.w3.org/2011/06/16-htmlspeech-minutes.html burn
17:37:18 rrsagent, bye
17:37:18 I see no action items