stream: IETF area: Internet status: informational author: Leon van Kammen date: 2023-04-12T00:00:00Z workgroup: Internet Engineering Task Force value: draft-XRFRAGMENTS-leonvankammen-00
This draft offers a specification for 4D URLs & navigation, to link 3D scenes and text together with- or without a network-connection. The specification promotes spatial addressibility, sharing, navigation, query-ing and interactive text across for (XR) Browsers. XR Fragments allows us to enrich existing dataformats, by recursive use of existing technologies like URI Fragments & visual-meta.
How can we add more features to existing text & 3D scenes, without introducing new dataformats? Historically, there’s many attempts to create the ultimate markuplanguage or 3D fileformat. However, thru the lens of authoring their lowest common denominator is still: plain text. XR Fragments allows us to enrich existing dataformats, by recursive use of existing technologies:
#pos=0,0,0&t=1,100
e.g.)#q=cube
){::boilerplate bcp14-tagged}
Here’s an ascii representation of a 3D scene-graph which contains 3D objects (◻
) and their metadata:
+--------------------------------------------------------+
| |
| index.gltf |
| │ |
| ├── ◻ buttonA |
| │ └ href: #pos=1,0,1&t=100,200 |
| │ |
| └── ◻ buttonB |
| └ href: other.fbx |
| |
+--------------------------------------------------------+
An XR Fragment-compatible browser viewing this scene, allows the end-user to interact with the buttonA
and buttonB
.
In case of buttonA
the end-user will be teleported to another location and time in the current loaded scene, but buttonB
will
replace the current scene with a new one (other.fbx
).
Text in XR has to be unobtrusive, for readers as well as authors. We think and speak in simple text, and given the new paradigm of XR interfaces, logically (spoken) text must be enriched afterwards (lazy metadata). Therefore, XR Fragment-compliant text will just be plain text, and not yet-another-markuplanguage. In contrast to markup languages, this means humans need to be always served first, and machines later.
Basically, a direct feedbackloop between unobtrusive text and human eye.
Reality has shown that outsourcing rich textmanipulation to commercial formats or mono-markup browsers (HTML) have there usecases, but also introduce barriers to thought-translation (which uses simple words). As Marshall MCluhan said: we have become irrevocably involved with, and responsible for, each other.
In order enjoy hasslefree batteries-included programmable text (glossaries, flexible views, drag-drop e.g.), XR Fragment supports visual-meta(data).
The XR Fragment specification bumps the traditional default browser-mimetype
text/plain;charset=US-ASCII
into:
text/plain;charset=utf-8;visual-meta=1
This means that visual-meta(data) can be appended to plain text without being displayed.
+--------------------------------------------------------------+ +------------------------+
| | | author.com/article.txt |
| index.gltf | +------------------------+
| │ | | |
| ├── ◻ article_canvas | | Hello friends. |
| │ └ src: ://author.com/article.txt | | |
| │ | | @{visual-meta-start} |
| └── ◻ note_canvas | | ... |
| └ src:`data:welcome human @{visual-meta-start}...` | +------------------------+
| |
| |
+--------------------------------------------------------------+
The difference is that text (+visual-meta data) in Data URI is saved into the scene, which also promotes rich copy-paste. In both cases will the text get rendered immediately (onto a plane geometry, hence the name ‘_canvas’). The enduser can access visual-meta(data)-fields only after interacting with the object.
NOTE: this is not to say that XR Browsers should not load HTML/PDF/etc-URLs thru
src
-metadata, it is just thattext/plain;charset=utf-8;visual-meta=1
is the minimum requirement.
+---------------------------------------------------------------+
| |
| index.gltf |
| │ |
| ├── ◻ todo |
| │ └ src:`data:learn about ARC @{visual-meta-start}...`|
| │ |
| └── ◻ ARC |
| └── ◻ plane |
| └ src: `data:ARC was revolutionary |
| @{visual-meta-start} |
| @{glossary-start} |
| @entry{ |
| name = {ARC}, |
| description = {Engelbart Concept: |
| Augmentation Research Center, |
| The name of Doug's lab at SRI. |
| }, |
| }` |
| |
+---------------------------------------------------------------+
Here we can see an 3D object of ARC, to which the enduser added a textnote (basically a plane geometry with src
).
The enduser can view/edit visual-meta(data)-fields only after interacting with the object.
This allows the 3D scene to perform omnidirectional features for free, by omni-connecting the word ‘ARC’:
The previous example, offers something exciting compared to simple textual copy-paste. , XR Fragment offers 4D- and HYPER- copy/paste: time, space and text interlinked. Therefore, the enduser in an XR Fragment-compatible browser can copy/paste/share data in these ways:
In contrast to markuplanguage, the (dictated/written) text needs no parsing, stays intact, by postponing metadata to the appendix.
This allows for a very economic XR way to:
NOTE: visualmeta never breaks the original intended text (in contrast to forgetting a html closing-tag e.g.)
Here’s an ascii representation of a 3D scene-graph with 3D objects (◻
) which embeds remote & local 3D objects (◻
) (without) using queries:
+--------------------------------------------------------+ +-------------------------+
| | | |
| index.gltf | | ocean.com/aquarium.fbx |
| │ | | │ |
| ├── ◻ canvas | | └── ◻ fishbowl |
| │ └ src: painting.png | | ├─ ◻ bass |
| │ | | └─ ◻ tuna |
| ├── ◻ aquariumcube | | |
| │ └ src: ://rescue.com/fish.gltf#q=bass%20tuna | +-------------------------+
| │ |
| ├── ◻ bedroom |
| │ └ src: #q=canvas |
| │ |
| └── ◻ livingroom |
| └ src: #q=canvas |
| |
+--------------------------------------------------------+
An XR Fragment-compatible browser viewing this scene, lazy-loads and projects painting.png
onto the (plane) object called canvas
(which is copy-instanced in the bed and livingroom).
Also, after lazy-loading ocean.com/aquarium.gltf
, only the queried objects bass
and tuna
will be instanced inside aquariumcube
.
Resizing will be happen accordingly to its placeholder object (aquariumcube
), see chapter Scaling.
TODO Security
This document has no IANA actions.
TODO acknowledge.