09:44:18 RRSAgent has joined #immersive-web 09:44:22 logging to https://www.w3.org/2023/09/15-immersive-web-irc 09:44:22 tidoust has joined #immersive-web 09:46:37 topic: https://github.com/immersive-web/spatial-favicons/issues/7 09:46:37 present+ 09:46:37 q+ 09:46:37 Zakim has joined #immersive-web 09:47:08 present+ 09:47:09 present+ 09:47:09 rrsagent, make log public 09:47:10 q+ 09:47:10 q- 09:47:17 present+ 09:47:17 meeting: Immersive Web WG/CG TPAC 2023 Day3 09:47:17 present+ 09:47:17 present+ 09:47:17 scribenick: vicki 09:47:22 gombos has joined #immersive-web 09:47:30 Present+ Laszlo_Gombos 09:47:41 ada: this might be a long one. we have a repo in the CG for spatial favicons, not a lot has happened. 09:47:44 agenda: https://github.com/immersive-web/administrivia/tree/main/TPAC-2023#agenda 09:47:48 rrsagent, publish minutes 09:47:49 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html atsushi 09:47:56 shiling has joined #immersive-web 09:48:04 ... maybe the favicon for the web app manifest, maybe for the page 09:48:06 TPAC 2023 Day2 minutes -> https://www.w3.org/2023/09/12-immersive-web-minutes.html 09:48:40 topic: Agenda bashing 09:48:40 ... might not totally be within our charter but let's talk about tit 09:48:40 s/tit/it 09:48:57 q+ 09:49:00 ada: seems like a very cool thing. could user agents punch out to an SVG to a blobby thing to make a 3d favicon look good? 09:49:05 rrsagent, publish minutes 09:49:06 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html atsushi 09:49:12 ack cabanier 09:49:15 ack cabanier 09:49:21 ... any thoughts? or should we archive this? 09:49:51 s/topic: Agenda bashing// 09:50:02 cabanier: magic leap implement this a few years ago, made orig proposal 09:50:12 s|TPAC 2023 Day2 minutes -> https://www.w3.org/2023/09/12-immersive-web-minutes.html|| 09:50:14 ... that the time there were a lot of questions 09:50:22 ...icons are underspecified 09:50:30 they are not fetched by the page, by the browser 09:50:41 ... I'm not sure that matters 09:50:49 i|topic: https://github.com/immersive-web/spatial-favicons/issues/7|TPAC 2023 Day2 minutes -> https://www.w3.org/2023/09/12-immersive-web-minutes.html 09:50:50 rrsagent, publish minutes 09:50:52 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html atsushi 09:50:59 as long as we don't say, let's support more mime types 09:51:11 q? 09:51:12 I think there's a nice fallback mechanism that would be nice to see 09:51:50 ada: I think it makes sense. I want to just drop in my Cistine chapel.gltf 09:52:06 cabanier: it would be great if there were an organization or association to define that 09:52:44 ada: maybe the Alliance for Open USD 09:52:44 q+ 09:52:44 ... something with a very limited feature set, very small 09:52:47 ... I'd love to see what people do with a restricted medium like this 09:52:47 ack Brandel 09:53:14 Brandel: AOUSD initial agreement is to go through current material and turn it into a standard 09:53:28 ... and next, to find out where other relevant standards are necessary 09:53:30 ... web is next 09:55:02 ... one benefit of NURBS is that they can be used to represent low-res things 09:55:08 q? 09:55:41 ada: it would be fun to see the standards evolve for small model formats 09:56:08 cabanier: would it make sense to reach out to the working group to see 09:56:18 ada: that would be a good idea, this kind of thing takes a long time 09:56:43 ... work out the WHATWG for page favicons, or Web App WG for web app manifest 09:57:16 ... wasn't there something like, "its not up to us to decide what formats to do". wasn't there a list somewhere? 09:57:26 .. this was pre-model 09:58:07 cabanier: if image supports this, maybe support for model too 09:58:21 ada: browsers that don't support it could just ignore it 09:58:35 ... if someone were to implement it... 09:58:44 cabanier: ideally there would be a document somewhere to specify 09:58:57 ada: esp now that SVG favicons are supported 09:59:22 ada: do people still feel positive? we'd probably be the lynchpin 10:00:07 ... should we do a CG repor to establish what it means to do a small one? 10:00:07 ...we can do it in the spatial favicons repo 10:00:13 ... establish some base guidelines for what a small model should be 10:00:54 ... guidelines for both gltf and usd 10:01:09 ... this is approx what we think it should do. mention NURBS, mention low-poly, etc 10:01:20 proposal for size constraints from magic leap: https://ml1-developer.magicleap.com/en-us/learn/guides/portal-icon-guidelines 10:01:40 ...I don't think the spatial favicons effort will ever leave the CG, more like "this is a things we think" vs a standard 10:01:51 ... +1 if you think this is a good idea 10:01:53 0 10:01:57 +1 for "small models" 10:01:58 +1 10:01:59 +0.25 10:02:01 +1 10:02:05 +1 10:02:15 ... 0 for neutral, -1 if against 10:02:55 cabanier: I'm a little worried bc I don't know who is making the right decision, there could be arguing g 10:03:10 ...maybe we'll end up just doing our own thing 10:03:38 ada: the history of favicon is pretty loose, I think someone from Netscape was just like "here's a thing I added" one evening 10:03:51 ...this might be all of the discussion we need around this 10:04:18 ...let's pick up another 15 min issue 10:05:08 https://github.com/immersive-web/administrivia/issues/201 10:05:08 topic: https://github.com/immersive-web/administrivia/issues/201 10:05:41 yonet: is there anything else we should add/ask from our community? there's a Help Wanted tag on this 10:05:41 ...one is types 10:05:41 cabanier: I think editors have not really gone through this, some issues are very old 10:05:45 ...maybe we should start by going through this 10:06:36 https://github.com/immersive-web/webxr/issues?q=is%3Aopen+is%3Aissue+label%3A%22help+wanted%22 10:07:00 ada: it would be nice if /help added the Help Wanted label 10:07:50 ...should I remove the HW thing from email? do people actually read it? 10:08:19 cabanier: maybe assign it to someone and they get an email every day 10:08:46 yonet: if we have anything to ask we could promote it a little 10:08:56 ada: I guess now we are getting closer to Rec 10:09:24 ...maybe it could be useful for like, "Hey, we want feedback for this" if someone wants to leave comments, it would be welcome on these issues 10:09:52 ...action: add a new label for Needs Feedback, put that in the email instead, and add a new command for people to request feedback 10:10:14 ...I don't think anyone reads the mailing list. or the emails I send out! :p 10:10:24 ...I guess people do turn up for the F2F 10:10:38 ...I think this will be a good action for us to do (yonet agrees) 10:11:32 ... BTW if anyone in this room wants navigation to work, feel free to drop in and work on it! 10:14:39 break time! \o/ 10:17:02 CharlesL has joined #immersive-web 10:17:12 present+ 10:29:10 tidoust has joined #immersive-web 10:30:46 etienne has joined #immersive-web 10:31:01 scribenick: cabanier 10:31:01 present+ 10:31:08 ada: thank you for joining us 10:31:21 ... we'd like to expose more a11y through webxr 10:31:22 jamesn has joined #immersive-web 10:31:30 present+ 10:31:40 present+ 10:31:45 present+ 10:31:49 present+ 10:32:26 ... for new to webxr, it's an API that sits on top of xr 10:32:46 ... some devices are attached to a computer but most frequently, the display is the computer 10:32:53 ... such as the vision pro and quest 10:32:56 present+ 10:33:03 ... or augmented reality on smart phone 10:33:29 ... the issue we have with adding a11y to WebXr is that we are on top of WebGL which has no a11y features 10:33:41 ... we have some api's that can increase semantics 10:33:51 ... the pixels on the screen are rendered by the user 10:33:58 ... we need help with some of these issues 10:34:21 ... the other item we are working on, unrelated to WebXR is a new API called the element 10:34:31 ... it's a new html element that can display 3D model 10:34:34 Peter has joined #immersive-web 10:34:41 .. they are no immersive but can be integrated in the page 10:34:57 jamesn: is this what we discussed at the last TPAC? 10:35:31 ada: yes 10:35:31 Accessibility considerations: https://github.com/immersive-web/webxr/blob/main/accessibility-considerations-explainer.md 10:36:11 ada: it would be a good idea for the developer to expose what is on the screen at the time 10:36:25 ... so we can report back to the user what we are looking at 10:36:38 jocelyntran has joined #immersive-web 10:36:52 ... the naive thing would be to label everything on the screen but it doesn't give you the context 10:37:09 ... the developer can confer that information on what they're rendering on the screen 10:37:24 ... if there was a way to properly get that information out 10:37:48 q+ 10:37:59 ack vicki 10:38:05 ... this api should use the proper tools for the job 10:38:31 q+ 10:38:50 ack jamesn 10:38:57 vicki: are there other elements where you're drawing with webgl, what do the other a11y tools look like 10:39:16 jamesn: canvas had an experiment with html overlays 10:39:36 ... so rather than trying every line being available, you could do potentially do something 10:39:44 q+ 10:39:51 ... so there's no single description for an entire screen 10:40:09 https://www.bbc.co.uk/accessibility/forproducts/xr/presentations/ 10:40:18 bkardell_: could someone break down if there is research or experiments in this area? 10:40:35 ... there are numerous other things that have been explored 10:40:44 q+ 10:40:51 ada: I wish that the people from xr access where here 10:40:59 q+ 10:41:12 yonet: one thing that they did, it was to say what the objects are around you 10:41:14 q+ 10:41:25 q+ 10:41:30 ack bkardell_ 10:41:30 ... that is one of the challenges but they didn't have solutions for that 10:41:33 ack yonet 10:41:36 ack CharlesL 10:41:49 CharlesL: xr access has been doing a lot of research at cornell 10:41:54 https://xraccess.org/workstreams/prototype-for-the-people/ 10:42:03 ... they have prototype 10:42:15 ... they want people to experiment 10:42:32 ... and need help. They broke it down for the different types of disability 10:42:50 ... the developer knows the scene such as a meeting room 10:43:17 ... the user could have preferences on how the room is laid out 10:43:23 q+ 10:43:33 ... depending on the level they are querying, the metadata can be there 10:43:33 ack cabanier 10:43:37 scribenick: ada 10:43:45 ack cabanier 10:44:47 q+ 10:44:49 cabanier: I know that at TPAC Japan an example from 2D canvas that was only implemented in firefox for an a11y tree that when you draw pixels you also render a semntic tree. 10:45:40 cabanier: screenreaders would need to be existed with it to work and shouldn't be too bad for developers. It wouldn't be hard for developers to render it. But it would fall to the developers. 10:45:49 scribenick: cabanier 10:45:53 ack tidoust 10:46:21 tidoust: I want to find solutions to expose the a11y information 10:46:22 q+ 10:46:36 ... so we need to teach developers on how to expose this 10:46:43 ... if there's information in USD 10:47:12 ... how can this format be extended to give accessible information 10:47:23 .. that can easily be targeted at developers 10:47:41 ack ada 10:47:43 yonet: I'm concerned that if you have a model, there are still a lot of other things around you 10:47:52 ada: we can do both 10:48:07 ... webxr has a long time problem that needs to be solved 10:48:32 ... for both webxr and model 10:48:43 ... adding a11y information to the models is important 10:49:13 ada: as cabanier said, having a render buffer where you render colors referencing a tree 10:49:32 q+ 10:49:41 ... this could work really well. You can probably do it within a single render 10:49:45 q+ 10:49:58 ack jamesn 10:50:01 ... it could be a good approach to see if it makes sense 10:50:11 jamesn: I want to jump back to canvas 10:50:35 ... if we do something with a dom renderer, having something that is overlaid 10:50:45 ... rather than something separately 10:51:06 ada: unfortunately there are issues here because you can't show html in webxr 10:51:22 ... we've been trying to fix this over the year but extremely hard 10:51:42 ... dom content is typically rectangular which doesn't match webgl 10:52:07 ... you can have more advantage if you have a pixel map 10:52:15 ack bkardell_ 10:52:29 bkardell_: I mentioned the research 10:52:37 https://abilitynet.org.uk/news-blogs/apple-vision-pro-has-vision-disabled-inclusion 10:52:47 ... apple launched the vision pro with a lot of a11y 10:53:19 ... it would be great if people could share your a11y research 10:53:34 ... we need to gather more information and do studies 10:53:47 ... hard to come up with answers in this room 10:54:12 ack CharlesL 10:54:17 yonet: the meta research team wanted to present but didn't get permission on time 10:54:41 CharlesL: can we have different rendering on top would be nice for color blind people 10:55:09 ... the mention of eye tracking, my eyes shake so I'm worried about this technology 10:55:59 yonet: for hololens we weren't reactive to the eye movements. The system itself would also train on you 10:56:02 ack cabanier 10:56:11 CharlesL: I'm using different headsets to try this out 10:56:11 scribenick:ada 10:57:59 q+ 10:58:22 ack bkardell 10:58:25 cabanier: The tree itself would be written out into the DOM, but the colours would map to the elements in this tree. 10:58:31 bkardell_: I assume that this works on simple canvases 10:58:47 yonet: it would be good if the same view was used for everyone 10:58:48 ... does this work on more complex content 10:59:23 cabanier: the colour index buffer is just used for mapping pixels to elements rather than being visible 10:59:28 q+ 11:00:25 the canvas dom overlay technique, it works for a certain size but does it actually work for things are infinitely big e.g. google maps? Or an open world game with a massive amount of data. 11:00:28 ack ada 11:01:07 cabanier: as you walk through the universe you would update the DOM 11:01:17 bkardell_: does this work today 11:01:27 cabanier: it was removed, it was liked but it was removed 11:01:36 q- 11:01:56 ada: you would only apply colors to the leaf nodes? 11:02:42 q+ 11:03:09 q+ 11:04:28 ack yonet 11:04:35 ack CharlesL 11:04:53 CharlesL: I was trying to figure out, is there an aria type solution? 11:05:40 ada: is there a way to output arbitrary information for a screen reader 11:05:40 q+ 11:05:44 ... today, it has to be a dom tree and the user has to traverse it 11:05:45 q+ 11:05:58 CharlesL: my thinking we only have screen readers for headsets 11:06:12 ... do we need screen readers for headset? 11:06:29 ada: the headset can be a dumb screen that is attached to a computer 11:06:44 ... it's up to the browser itself 11:07:16 bkardell_: are you sure? This is needed for all the apps 11:07:16 ... for browsers, you have an OS level a11y tree 11:07:52 ... with the exception of vision pro, they're all android based so could use virtual touch 11:08:06 yonet: it would be great what was done for canvas 11:08:20 ... and would like to know the challenges in the XR space 11:08:30 ack yonet 11:08:32 ... how can we work with the canvas group 11:08:50 jamesn: the aria group is not necessarily the right group 11:08:58 ... we know a11y api very well 11:09:30 jcraig has joined #immersive-web 11:09:50 ... it's unclear if they have accessibility APIs 11:09:57 ... apart from vision pro 11:10:06 q? 11:10:09 ... I'd love to be involved in this conversation 11:10:14 ack jamesn 11:10:27 rrsagent, make minutes 11:10:28 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html jcraig 11:11:44 https://www.w3.org/TR/xaur/ 11:11:54 topic: semantic labels 11:11:58 scribenick: ada 11:12:56 s/ the aria group is not necessarily the right group/ the APA group is not necessarily the right group/ 11:14:07 cabanier: we added support for semantic labelling in webxr with gives you information about the room you are in, when the user sets up their device they can annotate their environment floors, tables, walls, paintings, this information is giving to WebXR and used to prevent users from running into objects. It seems useful even at a system level. This is useful for everybody. There are 11:14:08 only a limited set of labels right now. E.g. No ottoman label. 11:14:16 q+ 11:14:23 ack CharlesL 11:14:46 s/it's unclear if they have accessibility APIs/it's unclear if the people in the APA group have experience with accessibility APIs, those with the relevant API experience are more likely to be in ARIA/ 11:14:51 rrsagent, make minutes 11:14:53 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html jcraig 11:15:30 q+ 11:15:30 Semantic labels: https://github.com/immersive-web/semantic-labels 11:15:30 CharlesL: is this done through a W3C registry? 11:15:40 cabanier: right now it's in a repo people can contribute to 11:16:14 q? 11:16:15 CharlesL: intnl, the hardware can do it 11:16:26 ack jcraig 11:16:32 cabanier: the hardware can do it 11:16:49 https://www.aswf.io/blog/community-support-needed-for-new-accessibility-initiative/ 11:17:02 myles has joined #immersive-web 11:17:58 james craig: Apple have been working with the ??? foundation to improve a11y. They are going to be working with the USD format to... [connection lost] 11:18:01 rrsagent, make minutes 11:18:02 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html CharlesL 11:18:22 https://github.com/immersive-web/webxr-accessibility 11:18:47 i/ada: you would only apply colors to the leaf/scribe+ cabanier/ 11:18:56 rrsagent, publish minutes 11:18:58 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html atsushi 11:19:11 scribe- cabanier 11:19:16 https://wiki.aswf.io/plugins/servlet/mobile?contentId=74940495#content/view/74940495 11:21:57 Quoting from last linkā€¦ 11:22:47 USD Accessibility needs : 11:22:47 - Method to label an object (container or leaf node) 11:22:48 - Method to label a time range (possibly use a caption format? like VTT metadata) 11:22:48 - If this format can include video and/or audio, how is loc/lang handled? We may need something similar same for captions and audio descriptions. 11:22:54 rrsagent, make minutes 11:22:55 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html jcraig 11:24:11 s/Apple have been /Apple has been / 11:25:12 s/??? foundation/Academy Software Foundation (ASWF)/ 11:25:34 rrsagent, make minutes 11:25:36 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html jcraig 11:52:33 myles has joined #immersive-web 12:15:26 myles has joined #immersive-web 12:23:46 tidoust has joined #immersive-web 12:25:28 myles has joined #immersive-web 12:28:55 atsushi has joined #immersive-web 12:35:39 bialpio has joined #immersive-web 12:37:54 present+ 12:37:58 rrsagent, publish minutes 12:37:59 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html atsushi 12:38:05 scribenick: ada 12:38:05 present+ 12:38:05 present+ 12:38:35 yonet has joined #immersive-web 12:38:39 topic: https://github.com/immersive-web/webxr-ar-module/issues/90 12:39:49 cabanier: this issue is something we have noticed that start of in VR that then progress into AR, so for a big portion of AR you are doing all the AR things, cameras, depth, reprojection which is a lot of wasted power. 12:39:51 q+ 12:39:54 q+ 12:40:17 cabanier: I Don't think it is something we the browser can figure out 12:40:26 scribenick:yonet 12:40:51 ada: It might be more useful to change session time within the session 12:41:50 ...if you can pre ask the features. It might be request a change for XR session. I think it would be a more useful way , like going from inline to iimmersive modes 12:42:02 q+ 12:42:09 ack ada 12:42:32 rik: you are proposing while in session, a new request for a different session? 12:43:46 ada: or devs can request for more than one immersive sesion and we give the new xr session. 12:43:46 ...I don't think it would break any current usage 12:44:09 ...it would be something like xrsession.transition or something 12:44:23 rik: I think it would be awkward 12:44:53 ada: if you put it as an optionall feature 12:45:02 rik: you can't make it an optional feature 12:45:57 ada: if we request as a feature, it would be like showing another room when you want to switch the mode. 12:46:11 Rik: I would prefer if it was in the render state. 12:46:42 ada: when you are in immersive-vr you are not going to test for hit test or anything 12:47:19 rik: if you know what the room looks like, you could do hit testing on the room. You want to keep the environment if you want to keep the AR mode 12:47:36 ...creating a new session will have discontinuity 12:48:30 ada: I think you are right, especially on handheld, in that situation what you want is still be in the handheld mode but switch from ar to vr 12:48:48 rik: do you think a flag? 12:49:20 ada: I don't think every platform can do it instantanously, so it should be a promise 12:50:02 ...if it is system app start taking over doing ar stuff and you couldn't get access again for immersive ar session 12:50:27 rik: yes. I think I can see it both ways. 12:51:01 ada: for some things like change the mode from transparency 12:51:20 rik: Hololens and a device like that would fail the promise. 12:52:03 ada: you might also want to fire it as an optional feature, that way the device can get ready for the session change 12:52:07 q? 12:52:16 ack bialpio 12:53:22 bialpio: vr to ar is tricky. I don't know if we want to allow that. How do you display permission priompt 12:53:48 ada: you can still have ar session. You call the session immediately 12:54:19 bialpo: we don't need that feature now right. 12:55:20 ada: For the escape room case, for the situation where you fail the camera. More work need to be done. You say this is reserved 12:55:20 vicki has joined #immersive-web 12:55:55 bialpo: for the vr, the device can render over everything. After the ar mode, it will render everything over again. I think this is over optimization 12:56:08 ada: you are not going to get the camera feed instantly 12:56:18 ...we will need a transition state 12:57:25 bialpo: I agree if update render state might not know if the camera is rendered or not, for those cases we might need a promise 12:58:18 ...I think we will know if we are on additive devices or not. This case is not enough for the feature as a promise. 12:59:08 ...for hit test we don't require ar session. you could expose hit test. the only thing is that we don't have a way of hit testing in vr in virtual objects 12:59:17 q+ 12:59:18 q? 12:59:25 ack alcooper 12:59:27 bialpio has joined #immersive-web 12:59:37 scribenick: ada 12:59:39 Leonard has joined #immersive-web 12:59:53 present+ 13:00:12 q+ 13:00:14 alcooper: is there anythign in OpenXR that has an example for how they have done this and whether they actually can stop doing work even if we ask it not to. 13:01:08 alcooper: I think as long as the session is going the camera will still be going but that could just be for our implementation, 13:01:31 alcooper: I don't see going from immersive-ar to immersive-vr beign something feasible since they are essentially two seperate runtimes. 13:01:36 ack cabanier 13:02:08 q+ 13:03:19 cabanier: on OpenXR there is a system ready call and do passthrough, and a pass through layer which can be removed from the lift of layers. 13:03:53 cabanier: to bialpio since it could be a hint for additive devices they can just fulfill the promise immediately. 13:04:11 ack bialpio 13:04:31 q+ 13:04:42 bialpio: I think this makes sense, I wanted to add some details about ARCore in which case i think we can immediately switch at least in one direction or maybe both for the camera availability. 13:05:20 alcooper: Yeah we should be able to do that but we may at least save a little optimisation although it may do it anyway. 13:05:39 bialpio: yeah it might be a very small performance improvement for android 13:06:58 q+ 13:07:01 bialpio: there might be some features that the site may want access to which we may be able to disable but I don't think we would be able to disable these on the fly in ARCore they need to be set on session creation which add latency to the camera such as image tracking which add a couple of ms of delay. Which is noticable, we probaby wouldn't be able to turn these off. 13:07:15 ack cabanier 13:07:25 ack alcooper 13:07:37 cabanier: i agree that depth sensing or image tracking are turned on then it won't do anything 13:08:43 alcooper: one thought i had regarding passthrough, would it make sense to have pass through be a request in immersive-vr too so pass through is just a feature that can be turned on or off 13:09:19 scribenick: yonet 13:09:39 q+ 13:10:01 cabanier: I don't think the developer thinks of their app as vr. 13:11:15 q+ 13:11:33 scribenick: ada 13:11:37 alcooper: to rik's point, maybe it is weird to have these ar vs vr with the pass through 13:11:47 ack bialpio 13:13:18 bialpio: i think this is starting to blur the lines between AR and VR I don't think many of the feature sassociated with AR are blocked in VR sessions and it's worth while considering AR sessions as more sensitive in terms of privacy so perhaps some features should be restricted to AR only. 13:13:18 q? 13:13:18 q+ 13:13:18 q+ 13:13:19 ack cabanier 13:14:52 cabanier: one more reason they should start from AR then disable, is preerving the mode type in android since we don't want people starting in cardboard mode then going AR 13:14:52 bialpio: also it would need to solve getting the permission prompt 13:15:25 ack alcooper 13:15:26 cabanier: it would also cause the session to grid to a halt for a few seconds 13:15:32 [AFK, BRB] 13:15:55 q+ 13:16:10 alcooper: if it is a hint we need to have someway to show it can fail e.g. if you are doing raw camera and ask us to disable your camera that should probably fail 13:16:11 scribenick: yonet 13:16:15 q- 13:16:28 ack bialpio 13:17:16 bialpio: re raw camera access we would need to start handing off camera frames whilst the user cannot see what is being visible by the camera. 13:17:19 q+ 13:17:37 ack ada 13:18:14 ada: if you request a raw camera access, what is stopping the dev to render a big square to hide the environment 13:18:53 alcooper: for what it's worth you could do something with getUserMedia. 13:19:06 s/something/the same thing 13:19:22 topic: https://github.com/immersive-web/webxr/issues/1345 13:20:07 ada: I've chat with hdr experts. I've been trying to work out what we need to expose to the user 13:20:25 eeeps has joined #immersive-web 13:20:41 ...there are some xr devices, like vision pro, meta quest pro, have a display that is hdr 13:20:53 ...currently all xr sessions are srgb 13:21:28 ...what was it that you explained about quest pro? do you just stretch out the color space? 13:21:30 q+ 13:22:21 rik: most websites do it in srgb but spec says it is linear 13:23:23 ...I think this is how webgl works too. 13:23:42 ...we were creating linear colorspace. some browsers implementation colors are wrong 13:24:00 ada: what Rik told me gl layers can be floats 13:24:14 rik: spec does not allow floats but can be updated 13:24:41 ...webgl can render to floating point rgb buffer for hdr, that is totally possible 13:24:56 ...if not, we need to request the color spaces 13:25:30 ada: I was thinking having a prefered color spaces. list of supported ones 13:26:21 ...there needs to be a way to say what color space you are using, the color space you are rendering needs to be supported by the frameworks. You need to be able to tell which one you are doing. 13:26:46 q 13:26:54 ...for headsets like vision pro, head room might change 13:27:05 q+ 13:28:11 ...we provide the user list of color space, and the developer can tell us which color space they are rendering. There might be some reason they might do headspace 13:28:23 QUEUE? 13:28:26 ...I think we will have to draft something 13:28:49 ack Leonard 13:30:25 Leonard: I think this may be resolved, but is this about colourspaces or HDR? 13:30:25 Brandel: This is more sRGB vs P# which is also HDR 13:31:32 Leonard: These calculations are done in Linear colourspace and then would need to go through tonemapping to go the displays. 13:31:33 gombos has joined #immersive-web 13:32:16 ack cabanier 13:32:17 Leonard: You Emmet Liash from Google is an expert in this field. 13:34:20 References: P3 - https://en.wikipedia.org/wiki/DCI-P3; sRGB - https://en.wikipedia.org/wiki/SRGB 13:34:20 The HDR on the Web breakout session (https://www.w3.org/events/meetings/009a5b81-0459-4ae4-9b33-f88dd9a9d89f/) was good and a few of the issues discussed there seem relevant 13:35:38 Different color spaces have a major impact on ensuring consistency of display results between different devices 13:36:34 cabanier: @ada re the headroom we need to expose, WebXR is based on a canvas the developer instantiates before the sessio nstarts 13:37:11 ada: the canvas context needs to know the details of the devie rather the computer so the information eeds to be present on the XR session 13:38:30 cabanier: the canvas already needs to have a way to retarget itself to what ever the display is what ever they do with that we can treat the canvas context the same on the headset. 13:38:53 q? 13:39:18 Related, 2d canvas has a way to specify color space: https://chromestatus.com/feature/5807007661555712 13:40:18 need to get the list of supported depth formats for the system compositor so that the developer can choose the correct. 13:40:46 rik: I am not sure if we support floating points 13:42:19 ada: it seems like we are going to be blocked. on the hdr meeting they showed us all required. 3 thinks were blockers 13:42:45 q+ 13:42:51 ack Leonard 13:43:28 Leonard: are we going to discuss this on our regular meetings? I want to make sure the experts will join 13:43:50 ada: I will bring the topic back, otherwise the issue is going to get worse with more devices 13:44:20 Leonard: the color space is the issue, different than hdr 13:44:45 ada: I guess only hdr is dependent on buffers 13:45:02 rik: I don't think we have a problem currently about the color space 13:45:14 ada: we should at least expose prefer color spaces 13:45:27 rik: you mean, prefered texture formats 13:45:30 q+ 13:45:54 ada: color space 13:46:24 ack bialpio 13:46:28 Leonard: is this the final output or the colour space physically based rendering is done in 13:46:35 Brandel: it's the tonemapping, the final output 13:47:07 ada: I was pretty sure threejs is using linear these days and then convert to srgb at the end 13:48:35 bialpio: opengl assumes srgb. You all the math in linear space, and then turn it into srgb. We might be confused about some of this. 13:49:03 ...webgl and opengl will always assume linear 13:50:18 ada: does correct colors won't work in gl 13:50:18 rik: yes 13:50:52 q? 13:50:52 rik: we have to do texture formats 13:51:24 ada: down the road, we will have to display everything to the user. 13:51:57 Do you really want to commit to WebGL as it is [slowly] dropping out in favour of WebGPU? 13:51:57 ...if you have 3 headsets connected to a computer they each will have a different profile 13:53:30 scribenick: cabanier 13:53:30 TY all - great meeting. I need to leave now. 13:54:05 Thanks Leonard 13:54:05 topic: Automated testing of WebXR sessions with WebDriver 13:54:10 ada: there's a w3c group wrt the webdriver 13:54:25 ... it would be great if webxr could use webdriver 13:54:25 q+ 13:54:42 ... I think it would be useful for developer building complext website 13:54:50 ... but also in depth testing for webxr 13:55:02 ... I was wondering if people think our group should do 13:55:15 ... and then submit it to the group that manager webdriver 13:55:20 ack alcooper 13:55:30 vicki has joined #immersive-web 13:55:36 alcooper: we have the webxr test api 13:55:44 ... which chrome uses to drive tests 13:55:54 ... I believe Apple contributed to that as well 13:55:56 WebXR Test API: https://immersive-web.github.io/webxr-test-api/ 13:56:06 ... maybe more could be done there 13:56:11 ada: that's great. 13:56:20 ... maybe that could be the basis for this type of work 13:56:32 ... that sounds great 13:56:39 q+ 13:56:46 ... should the group be involved in this 13:56:51 ack bialpio 13:56:56 bialpio: do we know what is expected here? 13:57:06 ... we already have the web platform tests 13:57:12 ... it mostly mocks some data 13:57:22 ... depending on what is needed, it's sufficient 13:57:39 ... in chromium, you can't take advantage of this outside of web platform tests 13:57:54 ... we don't initialize the components all the time. It only exists in our tests 13:57:59 q+ 13:58:24 ada: what I'd like to see happen, is that I can put my hmd in the computer and then run my tests 13:58:42 ... and then dump it to a file, generate screenshots, diff them , etc 13:59:01 ... so if I make a change, everything still works as expected 13:59:16 ... it's not just mocking, I want to actually drive the hardware 13:59:28 bialpio: you can send triangles through the mock api 13:59:42 ... we have an implementation that drives this 13:59:56 ... you can't control ref space, input sources 14:00:10 ... so you could write a script that moves you around 14:00:32 ... related to plugging in the headset, ARCore allows us to record an AR session 14:00:43 ... this is how we do testing internally 14:00:51 ... check how the api is working 14:01:13 ... If you go to session recording, it's very backend specific 14:01:37 ... we don't something special, just feed the precorded session to the application 14:01:45 ... the test API might be sufficient 14:01:54 ada: is this API defined? 14:02:05 bialpio: yes, Alex and Manish are the editors 14:02:24 ... for all the APIs we ship, we've been updating the data 14:02:42 ack alcooper 14:02:42 ... we have a javascript implementation that implements the runtime 14:02:55 alcooper: the goal of the test API is to serve as a fake device 14:03:09 ... if you want to inspect bits of state 14:03:18 ... it drives and controls the state 14:03:39 ... you need a special build of chrome. We only do it as a polyfill 14:04:33 q+ 14:04:33 ... some of our tests, do rely on generic webdriver functionality 14:04:42 ada: it sounds that we want to build on top of this 14:05:07 ... we don't want to have users wear the headset while the tests run 14:05:13 ack bialpio 14:05:15 alcooper: we don't ship any of that 14:05:28 bialpio: just to clarify, it's not really a polyfill 14:05:41 ... because the native code still runs 14:06:00 ... we talk to a javascript backend. Just not arcore 14:06:30 ... I'm looking at the spec and it looks like we need to relax it if we want it to run on shipped builds 14:07:03 ada: if we go down the selenium group, we want to say that it references this api 14:07:08 ... but that it can ship in browser 14:07:24 alcooper: I believe Manish implemented this in servo 14:07:43 ... I think it's fair to discuss a separate interface 14:07:56 ... that can control a generic device 14:08:41 ada: a developer could use the test api withouth selenium without us needing to make changes there 14:09:12 ... I'm glad we had this conversation 14:12:46 myles has joined #immersive-web 14:12:59 rrsagent, publish minutes 14:13:01 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html atsushi 14:14:11 i/Leonard: I think this may be resolved, but is/scribe+ ada/ 14:14:19 rrsagent, publish minutes 14:14:20 I have made the request to generate https://www.w3.org/2023/09/15-immersive-web-minutes.html atsushi 15:05:07 eeeps has joined #immersive-web 15:34:09 dino has joined #immersive-web 16:33:06 Zakim has left #immersive-web 17:34:06 dino has joined #immersive-web