14:33 | <nicolo-ribaudo> | Why in the HTML PR preview diff the keyboard shortcut for "next" is J and for "prev" is K? The "next" button is on the right of the "prev" button, but J is on the ledt of K |
14:47 | <evilpie> | Why in the HTML PR preview diff the keyboard shortcut for "next" is J and for "prev" is K? The "next" button is on the right of the "prev" button, but J is on the ledt of K |
16:47 | <Domenic> | annevk: think we can merge https://github.com/whatwg/html/pull/10188 ? |
16:53 | <annevk> | Domenic: sounds good to me. |
16:53 | annevk | wonders if Domenic is traveling |
17:06 | <krosylight> | re: sensitive=true, I think the conversation fully went to the direction about server-side browser AI feature, which makes sense, but the original GH issue I think is more focused on AI crawlers which users don't have control of. Maybe do some enum attribute that selectively signal those crawlers for initial simplicity, like donotcollectby=crawler or something |
17:07 | <krosylight> | Context: today's whatnot meeting on https://github.com/whatwg/html/issues/10519 |
17:57 | <annevk> | Why would a crawler see end user sensitive data? That would be a website bug, no? |
19:28 | <Timo Tijhof> | Why would a crawler see end user sensitive data? That would be a website bug, no? I suppose it's possible for cookieless authenticated URLs to be leaked or indirectly discovered by a crawler. Akin to "invite/share by link" private URLs, or internal mechanisms where emails point to APIs, static file servers, etc that are private where the URL is the secret. However this is an area where I'd expect norobots to be used on each page, and/or for these to be in a sub tree excluded wholesale by robots.txt |
19:31 | <Timo Tijhof> | Why would a crawler see end user sensitive data? That would be a website bug, no? |