XR Fragments


stream:    IETF
area:      Internet
status:    informational
author:    Leon van Kammen
date:      2023-04-12T00:00:00Z
workgroup: Internet Engineering Task Force
value:     draft-XRFRAGMENTS-leonvankammen-00

Abstract

This draft offers a specification for 4D URLs & navigation, to link 3D scenes and text together with- or without a network-connection. The specification promotes spatial addressibility, sharing, navigation, query-ing and interactive text across for (XR) Browsers. XR Fragments allows us to enrich existing dataformats, by recursive use of existing technologies like URI Fragments & visual-meta.

Introduction

How can we add more features to existing text & 3D scenes, without introducing new dataformats? Historically, there’s many attempts to create the ultimate markuplanguage or 3D fileformat. However, thru the lens of authoring their lowest common denominator is still: plain text. XR Fragments allows us to enrich existing dataformats, by recursive use of existing technologies:

Conventions and Definitions

{::boilerplate bcp14-tagged}

Navigating 3D

Here’s an ascii representation of a 3D scene-graph which contains 3D objects () and their metadata:

  +--------------------------------------------------------+ 
  |                                                        |
  |  index.gltf                                            |
  |    │                                                   |
  |    ├── ◻ buttonA                                       |
  |    │      └ href: #pos=1,0,1&t=100,200                 |
  |    │                                                   |
  |    └── ◻ buttonB                                       |
  |           └ href: other.fbx                            |
  |                                                        |
  +--------------------------------------------------------+

An XR Fragment-compatible browser viewing this scene, allows the end-user to interact with the buttonA and buttonB. In case of buttonA the end-user will be teleported to another location and time in the current loaded scene, but buttonB will replace the current scene with a new one (other.fbx).

Navigating text

Text in XR has to be unobtrusive, for readers as well as authors. We think and speak in simple text, and given the new paradigm of XR interfaces, logically (spoken) text must be enriched afterwards (lazy metadata). Therefore, XR Fragment-compliant text will just be plain text, and not yet-another-markuplanguage. In contrast to markup languages, this means humans need to be always served first, and machines later.

Basically, a direct feedbackloop between unobtrusive text and human eye.

Reality has shown that outsourcing rich textmanipulation to commercial formats or mono-markup browsers (HTML) have there usecases, but also introduce barriers to thought-translation (which uses simple words). As Marshall MCluhan said: we have become irrevocably involved with, and responsible for, each other.

In order enjoy hasslefree batteries-included programmable text (glossaries, flexible views, drag-drop e.g.), XR Fragment supports visual-meta(data).

Default Data URI mimetype

The XR Fragment specification bumps the traditional default browser-mimetype

text/plain;charset=US-ASCII

into:

text/plain;charset=utf-8;visual-meta=1

This means that visual-meta(data) can be appended to plain text without being displayed.

URL and Data URI

  +--------------------------------------------------------------+  +------------------------+
  |                                                              |  | author.com/article.txt |
  |  index.gltf                                                  |  +------------------------+
  |    │                                                         |  |                        |
  |    ├── ◻ article_canvas                                      |  | Hello friends.         |
  |    │    └ src: ://author.com/article.txt                     |  |                        |
  |    │                                                         |  | @{visual-meta-start}   |
  |    └── ◻ note_canvas                                         |  | ...                    |
  |           └ src:`data:welcome human @{visual-meta-start}...` |  +------------------------+ 
  |                                                              | 
  |                                                              |
  +--------------------------------------------------------------+

The difference is that text (+visual-meta data) in Data URI is saved into the scene, which also promotes rich copy-paste. In both cases will the text get rendered immediately (onto a plane geometry, hence the name ‘_canvas’). The enduser can access visual-meta(data)-fields only after interacting with the object.

NOTE: this is not to say that XR Browsers should not load HTML/PDF/etc-URLs thru src-metadata, it is just that text/plain;charset=utf-8;visual-meta=1 is the minimum requirement.

omnidirectional XR annotations

  +---------------------------------------------------------------+ 
  |                                                               |
  |  index.gltf                                                   |
  |    │                                                          |
  |    ├── ◻ todo                                                 | 
  |    │      └ src:`data:learn about ARC @{visual-meta-start}...`|  
  |    │                                                          |
  |    └── ◻ ARC                                                  |
  |           └── ◻ plane                                         |
  |                   └ src: `data:ARC was revolutionary          |
  |                          @{visual-meta-start}                 |
  |                          @{glossary-start}                    |
  |                          @entry{                              |
  |                           name = {ARC},                       |
  |                           description = {Engelbart Concept:   |
  |                             Augmentation Research Center,     |
  |                             The name of Doug's lab at SRI.    |
  |                           },                                  |
  |                          }`                                   |
  |                                                               |
  +---------------------------------------------------------------+

Here we can see an 3D object of ARC, to which the enduser added a textnote (basically a plane geometry with src). The enduser can view/edit visual-meta(data)-fields only after interacting with the object. This allows the 3D scene to perform omnidirectional features for free, by omni-connecting the word ‘ARC’:

HYPER copy/paste

The previous example, offers something exciting compared to simple textual copy-paste. , XR Fragment offers 4D- and HYPER- copy/paste: time, space and text interlinked. Therefore, the enduser in an XR Fragment-compatible browser can copy/paste/share data in these ways:

Plain Text (with optional visual-meta)

In contrast to markuplanguage, the (dictated/written) text needs no parsing, stays intact, by postponing metadata to the appendix.

This allows for a very economic XR way to:

NOTE: visualmeta never breaks the original intended text (in contrast to forgetting a html closing-tag e.g.)

Embedding 3D content

Here’s an ascii representation of a 3D scene-graph with 3D objects () which embeds remote & local 3D objects () (without) using queries:

  +--------------------------------------------------------+  +-------------------------+ 
  |                                                        |  |                         |
  |  index.gltf                                            |  | ocean.com/aquarium.fbx  |
  |    │                                                   |  |   │                     |
  |    ├── ◻ canvas                                        |  |   └── ◻ fishbowl        |
  |    │      └ src: painting.png                          |  |         ├─ ◻ bass       |
  |    │                                                   |  |         └─ ◻ tuna       |
  |    ├── ◻ aquariumcube                                  |  |                         |       
  |    │      └ src: ://rescue.com/fish.gltf#q=bass%20tuna |  +-------------------------+
  |    │                                                   |    
  |    ├── ◻ bedroom                                       |   
  |    │      └ src: #q=canvas                             |
  |    │                                                   |   
  |    └── ◻ livingroom                                    |      
  |           └ src: #q=canvas                             |
  |                                                        |
  +--------------------------------------------------------+

An XR Fragment-compatible browser viewing this scene, lazy-loads and projects painting.png onto the (plane) object called canvas (which is copy-instanced in the bed and livingroom). Also, after lazy-loading ocean.com/aquarium.gltf, only the queried objects bass and tuna will be instanced inside aquariumcube. Resizing will be happen accordingly to its placeholder object (aquariumcube), see chapter Scaling.

List of XR URI Fragments

Security Considerations

TODO Security

IANA Considerations

This document has no IANA actions.

Acknowledgments

TODO acknowledge.