stream: IETF area: Internet status: informational author: Leon van Kammen date: 2023-04-12T00:00:00Z workgroup: Internet Engineering Task Force value: draft-XRFRAGMENTS-leonvankammen-00
This draft offers a specification for 4D URLs & navigation, to link 3D scenes and text together with- or without a network-connection.
The specification promotes spatial addressibility, sharing, navigation, query-ing and tagging interactive (text)objects across for (XR) Browsers.
XR Fragments allows us to enrich existing dataformats, by recursive use of existing proven technologies like URI Fragments and BibTeX notation.
Almost every idea in this document is demonstrated at https://xrfragment.org
How can we add more features to existing text & 3D scenes, without introducing new dataformats?
Historically, there’s many attempts to create the ultimate markuplanguage or 3D fileformat.
However, thru the lens of authoring, their lowest common denominator is still: plain text.
XR Fragments allows us to enrich/connect existing dataformats, by recursive use of existing technologies:
NOTE: The chapters in this document are ordered from highlevel to lowlevel (technical) as much as possible
XR Fragments strives to serve (nontechnical/fuzzy) humans first, and machine(implementations) later, by ensuring hasslefree text-vs-thought feedback loops.
This also means that the repair-ability of machine-matters should be human friendly too (not too complex).
“When a car breaks down, the ones without turbosupercharger are easier to fix”
Let’s always focus on average humans: the ‘fuzzy symbolical mind’ must be served first, before serving the greater ‘categorized typesafe RDF hive mind’).
Humans first, machines (AI) later.
definition | explanation |
---|---|
human | a sentient being who thinks fuzzy, absorbs, and shares thought (by plain text, not markuplanguage) |
scene | a (local/remote) 3D scene or 3D file (index.gltf e.g.) |
3D object | an object inside a scene characterized by vertex-, face- and customproperty data. |
metadata | custom properties of text, 3D Scene or Object(nodes), relevant to machines and a human minority (academics/developers) |
XR fragment | URI Fragment with spatial hints like #pos=0,0,0&t=1,100 e.g. |
src | (HTML-piggybacked) metadata of a 3D object which instances content |
href | (HTML-piggybacked) metadata of a 3D object which links to content |
query | an URI Fragment-operator which queries object(s) from a scene like #q=cube |
visual-meta | visual-meta data appended to text/books/papers which is indirectly visible/editable in XR. |
requestless metadata | opposite of networked metadata (RDF/HTML requests can easily fan out into framerate-dropping, hence not used a lot in games). |
FPS | frames per second in spatial experiences (games,VR,AR e.g.), should be as high as possible |
introspective | inward sensemaking (“I feel this belongs to that”) |
extrospective | outward sensemaking (“I’m fairly sure John is a person who lives in oklahoma”) |
◻ |
ascii representation of an 3D object/mesh |
(un)obtrusive | obtrusive: wrapping human text/thought in XML/HTML/JSON obfuscates human text into a salad of machine-symbols and words |
fragment | type | example | info |
---|---|---|---|
#pos |
vector3 | #pos=0.5,0,0 |
positions camera to xyz-coord 0.5,0,0 |
#rot |
vector3 | #rot=0,90,0 |
rotates camera to xyz-coord 0.5,0,0 |
#t |
vector2 | #t=500,1000 |
sets animation-loop range between frame 500 and 1000 |
#...... |
string | #.cubes #cube |
object(s) of interest (fragment to object name or class mapping) |
xyz coordinates are similar to ones found in SVG Media Fragments
key | type | example (JSON) | info |
---|---|---|---|
name |
string | "name": "cube" |
available in all 3D fileformats & scenes |
class |
string | "class": "cubes" |
available through custom property in 3D fileformats |
href |
string | "href": "b.gltf" |
available through custom property in 3D fileformats |
src |
string | "src": "#q=cube" |
available through custom property in 3D fileformats |
Popular compatible 3D fileformats: .gltf
, .obj
, .fbx
, .usdz
, .json
(THREEjs), COLLADA
and so on.
NOTE: XR Fragments are file-agnostic, which means that the metadata exist in programmatic 3D scene(nodes) too.
Here’s an ascii representation of a 3D scene-graph which contains 3D objects ◻
and their metadata:
+--------------------------------------------------------+
| |
| index.gltf |
| │ |
| ├── ◻ buttonA |
| │ └ href: #pos=1,0,1&t=100,200 |
| │ |
| └── ◻ buttonB |
| └ href: other.fbx | <-- file-agnostic (can be .gltf .obj etc)
| |
+--------------------------------------------------------+
An XR Fragment-compatible browser viewing this scene, allows the end-user to interact with the buttonA
and buttonB
.
In case of buttonA
the end-user will be teleported to another location and time in the current loaded scene, but buttonB
will
replace the current scene with a new one, like other.fbx
.
Here’s an ascii representation of a 3D scene-graph with 3D objects ◻
which embeds remote & local 3D objects ◻
(without) using queries:
+--------------------------------------------------------+ +-------------------------+
| | | |
| index.gltf | | ocean.com/aquarium.fbx |
| │ | | │ |
| ├── ◻ canvas | | └── ◻ fishbowl |
| │ └ src: painting.png | | ├─ ◻ bass |
| │ | | └─ ◻ tuna |
| ├── ◻ aquariumcube | | |
| │ └ src: ://rescue.com/fish.gltf#q=bass%20tuna | +-------------------------+
| │ |
| ├── ◻ bedroom |
| │ └ src: #q=canvas |
| │ |
| └── ◻ livingroom |
| └ src: #q=canvas |
| |
+--------------------------------------------------------+
An XR Fragment-compatible browser viewing this scene, lazy-loads and projects painting.png
onto the (plane) object called canvas
(which is copy-instanced in the bed and livingroom).
Also, after lazy-loading ocean.com/aquarium.gltf
, only the queried objects bass
and tuna
will be instanced inside aquariumcube
.
Resizing will be happen accordingly to its placeholder object aquariumcube
, see chapter Scaling.
We still think and speak in simple text, not in HTML or RDF.
The most advanced human will probably not shout <h1>FIRE!</h1>
in case of emergency.
Given the new dawn of (non-keyboard) XR interfaces, keeping text as is (not obscuring with markup) is preferred.
Ideally metadata must come later with text, but not obfuscate the text, or in another file.
Humans first, machines (AI) later (core principle
This way:
This allows recursive connections between text itself, as well as 3D objects and vice versa, using BibTeX-tags :
+--------------------------------------------------+
| My Notes |
| |
| The houses seen here are built in baroque style. |
| |
| @house{houses, <----- XR Fragment triple/tag: phrase-matching BibTeX
| url = {#.house} <------------------- XR Fragment URI
| } |
+--------------------------------------------------+
This allows instant realtime tagging of objects at various scopes:
scope | matching algo |
---|---|
textual | text containing ‘houses’ is now automatically tagged with ‘house’ (incl. plaintext src child nodes) |
spatial | spatial object(s) with "class":"house" (because of {#.house} ) are now automatically tagged with ‘house’ (incl. child nodes) |
supra | text- or spatial-object(s) (non-descendant nodes) elsewhere, named ‘house’, are automatically tagged with ‘house’ (current node to root node) |
omni | text- or spatial-object(s) (non-descendant nodes) elsewhere, containing class/name ‘house’, are automatically tagged with ‘house’ (too node to all nodes) |
infinite | text- or spatial-object(s) (non-descendant nodes) elsewhere, containing class/name ‘house’ or ‘houses’, are automatically tagged with ‘house’ (too node to all nodes) |
This empowers the enduser spatial expressiveness (see the core principle): spatial wires can be rendered, words can be highlighted, spatial objects can be highlighted/moved/scaled, links can be manipulated by the user.
The simplicity of appending BibTeX ‘tags’ (humans first, machines later) is also demonstrated by visual-meta in greater detail.
NOTE: infinite matches both ‘house’ and ‘houses’ in text, as well as spatial objects with
"class":"house"
or name “house”. This multiplexing of id/category is deliberate because of the core principle.
The src
-values work as expected (respecting mime-types), however:
The XR Fragment specification bumps the traditional default browser-mimetype
text/plain;charset=US-ASCII
to a green eco-friendly:
text/plain;charset=utf-8;bib=^@
This indicates that bibs and bibtags matching regex ^@
will automatically get filtered out, in order to:
It’s concept is similar to literate programming, which empower local/remote responses to:
This significantly expands expressiveness and portability of human tagged text, by postponing machine-concerns to the end of the human text in contrast to literal interweaving of content and markupsymbols (or extra network requests, webservices e.g.).
For all other purposes, regular mimetypes can be used (but are not required by the spec).
To keep XR Fragments a lightweight spec, BibTeX is used for text/spatial tagging (not a scripting language or RDF e.g.).
Applications are also free to attach any JSON(LD / RDF) to spatial objects using custom properties (but is not interpreted by this spec).
+--------------------------------------------------------------+ +------------------------+
| | | author.com/article.txt |
| index.gltf | +------------------------+
| │ | | |
| ├── ◻ article_canvas | | Hello friends. |
| │ └ src: ://author.com/article.txt | | |
| │ | | @friend{friends |
| └── ◻ note_canvas | | ... |
| └ src:`data:welcome human @...` | | } |
| | +------------------------+
| |
+--------------------------------------------------------------+
The enduser will only see welcome human
and Hello friends
rendered spatially.
The beauty is that text (AND visual-meta) in Data URI promotes rich copy-paste.
In both cases, the text gets rendered immediately (onto a plane geometry, hence the name ‘_canvas’).
The XR Fragment-compatible browser can let the enduser access visual-meta(data)-fields after interacting with the object (contextmenu e.g.).
The mapping between 3D objects and text (src-data) is simple:
Example:
+------------------------------------------------------------------------------------+
| |
| index.gltf |
| │ |
| └── ◻ rentalhouse |
| └ class: house |
| └ ◻ note |
| └ src:`data: todo: call owner |
| @house{owner, |
| url = {#.house} |
| }` |
+------------------------------------------------------------------------------------+
3D object names and/or classes map to name
of visual-meta glossary-entries.
This allows rich interaction and interlinking between text and 3D objects:
“When a car breaks down, the ones without turbosupercharger are easier to fix”
Unlike XML or JSON, the typeless, unnested, everything-is-text nature of BibTeX tags is a great advantage for introspection.
It’s a missing sensemaking precursor to (eventual) extrospective RDF.
BibTeX-appendices are already used in the digital AND physical world (academic books, visual-meta), perhaps due to its terseness & simplicity.
In that sense, it’s one step up from the .ini
fileformat (which has never leaked into the physical world like BibTex):
characteristic | UTF8 Plain Text (with BibTeX) | RDF |
---|---|---|
perspective | introspective | extrospective |
structure | fuzzy (sensemaking) | precise |
space/scope | local | world |
everything is text (string) | yes | no |
paperfriendly | bibs | no |
leaves (dictated) text intact | yes | no |
markup language | just an appendix | ~4 different |
polyglot format | no | yes |
easy to copy/paste content+metadata | yes | up to application |
easy to write/repair for layman | yes | depends |
easy to (de)serialize | yes (fits on A4 paper) | depends |
infrastructure | selfcontained (plain text) | (semi)networked |
freeform tagging/annotation | yes, terse | yes, verbose |
can be appended to text-content | yes | up to application |
copy-paste text preserves metadata | yes | up to application |
emoji | yes | depends on encoding |
predicates | free | semi pre-determined |
implementation/network overhead | no | depends |
used in (physical) books/PDF | yes (visual-meta) | no |
terse non-verb predicates | yes | no |
nested structures | no (but: BibTex rulers) | yes |
${visual-meta-start}
) according to the tagbibs spectext/plain;charset=utf-8;tag=^@
Here’s an XR Text (de)multiplexer in javascript, which ticks all the above boxes:
xrtext = {
decode: (str) => {
// bibtex: ↓@ ↓<tag|tag{phrase,|{ruler}> ↓property ↓end
let pat = [ /@/, /^\S+[,{}]/, /},/, /}/ ]
let tags = [], text='', i=0, prop=''
var bibs = { regex: /(@[a-zA-Z0-9_+]+@[a-zA-Z0-9_@]+)/g, tags: {}}
let lines = str.replace(/\r?\n/g,'\n').split(/\n/)
for( let i = 0; !lines[i].match( /^@/ ); i++ ) text += lines[i]+'\n'
bibtex = lines.join('\n').substr( text.length )
bibtex.replace( bibs.regex , (m,k,v) => {
tok = m.substr(1).split("@")
match = tok.shift()
tok.map( (t) => bibs.tags[match] = `@${t}{${match},\n}\n` )
})
bibtex = Object.values(bibs.tags).join('\n') + bibtex.replace( bibs.regex, '')
bibtex.split( pat[0] ).map( (t) => {
try{
let v = {}
if( !(t = t.trim()) ) return
if( tag = t.match( pat[1] ) ) tag = tag[0]
if( tag.match( /^{.*}$/ ) ) return tags.push({ruler:tag})
t = t.substr( tag.length )
t.split( pat[2] )
.map( kv => {
if( !(kv = kv.trim()) || kv == "}" ) return
v[ kv.match(/\s?(\S+)\s?=/)[1] ] = kv.substr( kv.indexOf("{")+1 )
})
tags.push( { k:tag, v } )
}catch(e){ console.error(e) }
})
return {text, tags}
},
encode: (text,tags) => {
let str = text+"\n"
for( let i in tags ){
let item = tags[i]
if( item.ruler ){
str += `@${item.ruler}\n`
continue;
}
str += `@${item.k}\n`
for( let j in item.v ) str += ` ${j} = {${item.v[j]}}\n`
str += `}\n`
}
return str
}
}
The above (de)multiplexes text/metadata, expands bibs, (de)serializes bibtex (and all fits more or less on one A4 paper)
above can be used as a startingpoint for LLVM’s to translate/steelman to a more formal form/language.
str = `
hello world
@hello@greeting
@{some-section}
@flap{
asdf = {23423}
}`
var {tags,text} = xrtext.decode(str) // demultiplex text & bibtex
tags.find( (t) => t.k == 'flap{' ).v.asdf = 1 // edit tag
tags.push({ k:'bar{', v:{abc:123} }) // add tag
console.log( xrtext.encode(text,tags) ) // multiplex text & bibtex back together
@{references-start}
@misc{emilyHegland/Edgar&Frod,
author = {Emily Hegland},
title = {Edgar & Frode Hegland, November 2021},
year = {2021},
month = {11},
}
The above BibTeX-flavor can be imported, however will be rewritten to Dumb BibTeX, to satisfy rule 2 & 5, as well as the core principle
@visual-meta{
version = {1.1},
generator = {Author 7.6.2 (1064)},
section = {visual-meta-header}
}
@misc{emilyHegland/Edgar&Frod,
author = {Emily Hegland},
title = {Edgar & Frode Hegland, November 2021},
year = {2021},
month = {11},
section = {references}
}
The previous example, offers something exciting compared to simple copy/paste of 3D objects or text. XR Fragment allows HYPER-copy/paste: time, space and text interlinked. Therefore, the enduser in an XR Fragment-compatible browser can copy/paste/share data in these ways:
Include, exclude, hide/shows objects using space-separated strings:
#q=cube
#q=cube -ball_inside_cube
#q=* -sky
#q=-.language .english
#q=cube&rot=0,90,0
#q=price:>2 price:<5
It’s simple but powerful syntax which allows css-like class/id-selectors with a searchengine prompt-style feeling:
src
value (prevents sharing of scene-tampered URL’s).cube
and foo
in #q=cube foo
are matched against 3D object names or custom metadata-key(values)cube
and foo
in #q=cube foo
are matched against tags (BibTeX) inside plaintext src
values like @cube{redcube, ...
e.g.#
equals #q=*
.
like .german
match class-metadata of 3D objects like "class":"german"
.
like .german
match class-metadata of (BibTeX) tags in XR Text objects like @german{KarlHeinz, ...
e.g.For example:
#q=.foo
is a shorthand for#q=class:foo
, which will select objects with custom propertyclass
:foo
. Just a simple#q=cube
will simply select an object namedcube
.
operator | info |
---|---|
* |
select all objects (only useful in src custom property) |
- |
removes/hides object(s) |
: |
indicates an object-embedded custom property key/value |
. |
alias for "class" :".foo" equals class:foo |
> < |
compare float or int number |
/ |
reference to root-scene. Useful in case of (preventing) showing/hiding objects in nested scenes (instanced by src ) (*) |
* =
#q=-/cube
hides objectcube
only in the root-scene (not nestedcube
objects)
#q=-cube
hides both objectcube
in the root-scene AND nestedskybox
objects |
» example implementation » example 3D asset » discussion
Here’s how to write a query parser:
foo:1
and foo
(reference regex: /^.*:[><=!]?/
)-foo
,-foo:1
,-.foo
,-/foo
(reference regex: /^-/
)/foo
(reference regex: /^[-]?\//
).foo
(reference regex: /^[-]?class$/
)foo:1
(reference regex: /^[0-9\.]+$/
).foo
into class:foo
:
rules
id
to true
or false
(false=excluder -
)root
to true
or false
(true=/
root selector is present)store.foo = {id:false,root:true}
e.g.An example query-parser (which compiles to many languages) can be found here
reserved = gen-delims / sub-delims
gen-delims = "#" / "&"
sub-delims = "," / "="
Example:
://foo.com/my3d.gltf#pos=1,0,0&prio=-5&t=0,100
Demo | Explanation |
---|---|
pos=1,2,3 |
vector/coordinate argument e.g. |
pos=1,2,3&rot=0,90,0&q=.foo |
combinators |
Since XR Text contains metadata too, the user should be able to set up tagging-rules, so the copy-paste feature can :
class:secret
e.g.)This document has no IANA actions.
TODO acknowledge.