Interaktionstaxonomie AR: Unterschied zwischen den Versionen
		
		
		
		
		
		Zur Navigation springen
		Zur Suche springen
		
				
		
		
	
 (Die Seite wurde neu angelegt: „This page is a sort of lookup table of different modes of interaction using Augmented reality.  *Taxonomie AR-Interaktionsräume und -möglichkeiten  **Input -…“)  | 
				Keine Bearbeitungszusammenfassung  | 
				||
| (3 dazwischenliegende Versionen desselben Benutzers werden nicht angezeigt) | |||
| Zeile 1: | Zeile 1: | ||
This page is a sort of lookup table of different modes of interaction using Augmented reality.  | This page is a sort of lookup table of different modes of interaction using Augmented reality.  | ||
Eventually, this should be converted into a card game.   | |||
== Input - technical ==  | |||
=== exposed in unity: ===  | |||
Touch  | *Touch  | ||
Multitouch  | *Multitouch  | ||
Parsing Textfield  | *Parsing Textfield  | ||
Timer  | *Timer  | ||
Time  | *Time  | ||
GPS (location)  | *GPS (location)  | ||
Gyroskop  | *Gyroskop  | ||
Accelerometer  | *Accelerometer  | ||
Light Input  | *Light Input  | ||
ARCore (position)  | *ARCore (position)  | ||
Image Marker 2D (pixelcluster - ARCore, Vuforia)  | *Image Marker 2D (pixelcluster - ARCore, Vuforia)  | ||
Object Marker 3D - non-moving (Object Targets - Vuforia)  | *Object Marker 3D - non-moving (Object Targets - Vuforia)  | ||
=== networking: ===  | |||
WiFi Network strength  | *WiFi Network strength  | ||
Cellphone Network Coverage  | *Cellphone Network Coverage  | ||
Network SSIDs  | *Network SSIDs  | ||
Network commands (OSC)  | *Network commands (OSC)  | ||
Bluetooth Beacon  | *Bluetooth Beacon  | ||
non-trivial:  | === non-trivial: ===  | ||
Gesture recognition  | *Gesture recognition  | ||
Voice  | *Voice  | ||
Object Classification (ML - Tensorflow)  | *Object Classification (ML - Tensorflow)  | ||
Face Tracking (OpenCV - ML Tensorflow)  | *Face Tracking (OpenCV - ML Tensorflow)  | ||
Body Segmentation (ARKit only - for now)  | *Body Segmentation (ARKit only - for now)  | ||
== Input - human ==  | |||
*  | === bodily interaction ===  | ||
Shake phone  | *Shake phone  | ||
Turn around  | *Turn around  | ||
*tap on screen (on object)  | |||
*move device to frame Image  | |||
Walk to Location  | *Walk to Location  | ||
Turn on/off Flashlight  | *Turn on/off Flashlight  | ||
Talk to   | *Talk to device  | ||
Talk to other Person  | *Talk to other Person  | ||
*  | === cognitive framing ===  | ||
cast magic  | *cast magic  | ||
start tricorder analysis  | *start tricorder analysis  | ||
throw virtual pokeball  | *throw virtual pokeball  | ||
go to the virtual representation of the Planet Pluto  | *go to the virtual representation of the Planet Pluto  | ||
stand inside the particle stream  | *stand inside the particle stream  | ||
do a sensor sweep  | *do a sensor sweep  | ||
engage warp drive  | *engage warp drive  | ||
run away from spiders  | *run away from spiders  | ||
throw rock at spider  | *throw rock at spider  | ||
magnify spider  | *magnify spider  | ||
send message in a bottle  | *send message in a bottle  | ||
remove your own shadow (skill level five thousand)  | *remove your own shadow (skill level five thousand)  | ||
level up (from skill level fourthousandninehundredandninentynine)  | *level up (from skill level fourthousandninehundredandninentynine)  | ||
answer a question  | *answer a question  | ||
ask a question    | *ask a question    | ||
answer a riddle  | *answer a riddle  | ||
solve a puzzle  | *solve a puzzle  | ||
...  | ...  | ||
== Output - technical ==  | |||
=== direct ===  | |||
Text appears/disappears on screen  | *Text appears/disappears on screen  | ||
Mesh appears/disappears on screen  | *Mesh appears/disappears on screen  | ||
Skinned Mesh Animation starts/stops  | *Skinned Mesh Animation starts/stops  | ||
Play/Stop Animation  | *Play/Stop Animation  | ||
Move Camera  | *Move Camera  | ||
UI appears/disappears  | *UI appears/disappears  | ||
Sound gets played/stopped  | *Sound gets played/stopped  | ||
Video starts/stops  | *Video starts/stops  | ||
Phone vibrates/stops vibrating  | *Phone vibrates/stops vibrating  | ||
Send Network message  | *Send Network message  | ||
Camera turns on/off  | *Camera turns on/off  | ||
App closes/starts  | *App closes/starts  | ||
=== indirect ===  | |||
Simulation gets updated  | *Simulation gets updated  | ||
Method of Interaction changes  | *Method of Interaction changes  | ||
World State changes (Portal opens, etc…)  | *World State changes (Portal opens, etc…)  | ||
== Output - human ==  | |||
*  | === on-device ===  | ||
receive a question  | *receive a question  | ||
receive information  | *receive information  | ||
receive new mode of interaction  | *receive new mode of interaction  | ||
receive new task  | *receive new task  | ||
receive message from other human  | *receive message from other human  | ||
hear/read someone’s thoughts  | *hear/read someone’s thoughts  | ||
listen to music  | *listen to music  | ||
see a specific place  | *feel vibration  | ||
watch a movie  | *see a specific place  | ||
look at a different time  | *watch a movie  | ||
look at a different place  | *look at a different time  | ||
*look at a different place  | |||
*  | === in context ===  | ||
be in the center of the universe  | *be in the center of the universe  | ||
stand on Mars  | *stand on Mars  | ||
Travel through Wormhole  | *Travel through Wormhole  | ||
Move Towards Black hole  | *Move Towards Black hole  | ||
feel Neutrinos  | *feel Neutrinos  | ||
Aktuelle Version vom 1. November 2019, 09:24 Uhr
This page is a sort of lookup table of different modes of interaction using Augmented reality. Eventually, this should be converted into a card game.
Input - technical
exposed in unity:
- Touch
 - Multitouch
 - Parsing Textfield
 - Timer
 - Time
 
- GPS (location)
 - Gyroskop
 - Accelerometer
 - Light Input
 - ARCore (position)
 - Image Marker 2D (pixelcluster - ARCore, Vuforia)
 - Object Marker 3D - non-moving (Object Targets - Vuforia)
 
networking:
- WiFi Network strength
 - Cellphone Network Coverage
 - Network SSIDs
 - Network commands (OSC)
 - Bluetooth Beacon
 
non-trivial:
- Gesture recognition
 - Voice
 - Object Classification (ML - Tensorflow)
 - Face Tracking (OpenCV - ML Tensorflow)
 - Body Segmentation (ARKit only - for now)
 
Input - human
bodily interaction
- Shake phone
 - Turn around
 - tap on screen (on object)
 - move device to frame Image
 - Walk to Location
 - Turn on/off Flashlight
 - Talk to device
 - Talk to other Person
 
cognitive framing
- cast magic
 - start tricorder analysis
 - throw virtual pokeball
 - go to the virtual representation of the Planet Pluto
 - stand inside the particle stream
 - do a sensor sweep
 - engage warp drive
 - run away from spiders
 - throw rock at spider
 - magnify spider
 - send message in a bottle
 - remove your own shadow (skill level five thousand)
 - level up (from skill level fourthousandninehundredandninentynine)
 - answer a question
 - ask a question
 - answer a riddle
 - solve a puzzle
 
...
Output - technical
direct
- Text appears/disappears on screen
 - Mesh appears/disappears on screen
 - Skinned Mesh Animation starts/stops
 - Play/Stop Animation
 - Move Camera
 - UI appears/disappears
 - Sound gets played/stopped
 - Video starts/stops
 - Phone vibrates/stops vibrating
 - Send Network message
 - Camera turns on/off
 - App closes/starts
 
indirect
- Simulation gets updated
 - Method of Interaction changes
 - World State changes (Portal opens, etc…)
 
Output - human
on-device
- receive a question
 - receive information
 - receive new mode of interaction
 - receive new task
 - receive message from other human
 - hear/read someone’s thoughts
 - listen to music
 - feel vibration
 - see a specific place
 - watch a movie
 - look at a different time
 - look at a different place
 
in context
- be in the center of the universe
 - stand on Mars
 - Travel through Wormhole
 - Move Towards Black hole
 - feel Neutrinos