first draft sent

This commit is contained in:
aj 2020-05-07 21:59:13 +01:00
parent e0944dfdea
commit 196958fe3a
2 changed files with 384 additions and 180 deletions

View File

@ -237,7 +237,16 @@ University of Surrey
\end_layout \end_layout
\begin_layout Abstract \begin_layout Abstract
abstract \begin_inset Flex TODO Note (inline)
status open
\begin_layout Plain Layout
PLACEHOLDER
\end_layout
\end_inset
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -268,10 +277,6 @@ LatexCommand lstlistoflistings
\end_inset \end_inset
\end_layout
\begin_layout List of TODOs
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -337,17 +342,15 @@ While some present natural extensions to existing technology as seen in
\noun on \noun on
IKEA Place IKEA Place
\noun default \noun default
.
\begin_inset Flex TODO Note (Margin)
status open
\begin_layout Plain Layout \begin_inset CommandInset citation
reference? LatexCommand cite
\end_layout key "ikea-place"
literal "false"
\end_inset \end_inset
.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -397,20 +400,21 @@ holoportation
\begin_layout Standard \begin_layout Standard
This project aims to extend this suite to support multi-source holoportation, This project aims to extend this suite to support multi-source holoportation,
receiving multiple scenes concurrently analogous to the move from traditional receiving multiple scenes concurrently in a many-to-one configuration ready
phone calls to group conference calls. for composite presentation.
In doing so the implementation of holoportation could be seen to be generalised In doing so the implementation of holoportation could be seen to be generalised
, extending the possible applications of the suite. , extending the possible applications of the suite.
\begin_inset Flex TODO Note (Margin)
status open
\begin_layout Plain Layout
examples?
\end_layout \end_layout
\end_inset \begin_layout Standard
One application would be support for experiences akin to conference-calls
with multiple actors capturing their own environments for composition at
the server.
This could be seen to be applicable both to productivity software similar
to existing conference-call software and to entertainment experiences with
multiple locations combined for consumption by the public.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -719,9 +723,9 @@ In support of the lab's ongoing research into the
\noun on \noun on
LiveScan LiveScan
\noun default \noun default
suite and the area of theory in which it resides, investigations were made suite and it's network behaviour, investigations were made into the effect
into the effect of deliberately limiting the delivered frames per second of deliberately limiting the delivered frames per second on effective display
on effective display latency. latency.
Preliminary data for one method of doing so was gathered and is presented Preliminary data for one method of doing so was gathered and is presented
here in place of proper evaluation of the completed multi-source capabilities. here in place of proper evaluation of the completed multi-source capabilities.
\end_layout \end_layout
@ -761,11 +765,11 @@ Kinect
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
The significance of 3D video like that captured and relayed using the The significance of the renders captured and relayed by the
\noun on \noun on
LiveScan LiveScan
\noun default \noun default
suite is related to the development of new technologies able to immersively suite is related to the development of technologies able to immersively
display such video content. display such video content.
\end_layout \end_layout
@ -815,13 +819,14 @@ literal "false"
\end_inset \end_inset
, the collection and transmission of 3D holograms have applicability to , the generic collection and transmission of 3D holograms has applicability
all forms of XR and as such the state of this space is investigated. to all forms of XR and as such the state of this space is investigated
with an emphasis on handheld AR.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
Finally, existing examples of holoportation are presented including those Finally, existing examples of holoportation are presented including those
display multi-source capabilities. displaying multi-source capabilities.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -859,8 +864,8 @@ ons outside of such environments.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
While capable of high-quality results, visual hull reconstruction requires While capable of high-quality results, these implementations require a tightly
a tightly controlled environment and high-performance hardware. controlled environment and high-performance hardware.
As such, it is briefly presented here to contextualise the use of depth As such, it is briefly presented here to contextualise the use of depth
aware cameras in the aware cameras in the
\noun on \noun on
@ -874,24 +879,38 @@ Visual Hull Reconstruction in a Lab Environment
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
A visual hull defines a 3D representation of an object constructed through With a selection of 2D videos from multiple viewpoints,
the volume intersection of multiple 2D silhouettes \emph on
shape-from-silhouette
\emph default
describes a method to reconstruct depth information through the application
of Epipolar geometry
\begin_inset CommandInset citation \begin_inset CommandInset citation
LatexCommand cite LatexCommand cite
key "visual-hull-laurentini" key "sfs-over-time,sfs-video-cm"
literal "false" literal "false"
\end_inset \end_inset
. .
The result of this is referred to as a visual hull
\begin_inset CommandInset citation
LatexCommand cite
key "visual-hull-laurentini,laurentini-solids-revolution"
literal "false"
\end_inset
.
\end_layout
\begin_layout Standard
A visual hull defines a 3D representation of an object constructed through
the volume intersection of multiple 2D silhouettes.
With knowledge of the relative positions of each view point, corresponding With knowledge of the relative positions of each view point, corresponding
silhouettes of an object can be triangulated to form a 3D object. silhouettes of an object can be triangulated to form a 3D object.
In doing so depth information lost in projection to a 2D image can be reconstru In doing so depth information lost in projection to a 2D image can be reconstru
cted leading to the moniker cted.
\emph on
shape-from-silhouette
\emph default
.
\begin_inset Flex TODO Note (Margin) \begin_inset Flex TODO Note (Margin)
status open status open
@ -954,7 +973,7 @@ literal "false"
\noun default \noun default
employ a combination of all of these techniques with a surface constructed employ a combination of all of these techniques with a surface constructed
from both infra-red depth information, from infra-red,
\emph on \emph on
shape-from-silhouette shape-from-silhouette
\emph default \emph default
@ -1052,7 +1071,11 @@ literal "false"
and the rear-facing LIDAR of the 2020 iPad Pro, the most commonly used and the rear-facing LIDAR of the 2020 iPad Pro, the most commonly used
camera in computer vision research is the camera in computer vision research is the
\noun on \noun on
Microsoft Kinect. Microsoft Kinect
\noun default
due its availability and price
\noun on
.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -1419,34 +1442,7 @@ XR Implementations
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
\begin_inset Flex TODO Note (inline) While XR applications have been demonstrated without dedicated hardware
status open
\begin_layout Plain Layout
Mobile AR examples
\end_layout
\end_inset
\end_layout
\begin_layout Standard
\begin_inset Note Comment
status open
\begin_layout Plain Layout
Although VR and AR headsets have accelerated the development of XR technology,
they are not the only way to construct XR experiences.
\begin_inset CommandInset citation
LatexCommand citeauthor
key "roomalive"
literal "false"
\end_inset
\begin_inset CommandInset citation \begin_inset CommandInset citation
LatexCommand cite LatexCommand cite
key "roomalive" key "roomalive"
@ -1454,55 +1450,56 @@ literal "false"
\end_inset \end_inset
demonstrate , the majority utilise dedicated headsets or handheld devices.
\emph on
RoomAlive
\emph default
, an AR experience using depth cameras and projectors (referred to as
\emph on
procams
\emph default
) to construct experiences in any room.
This is presented through games and visual alterations to the user's surroundin
gs.
A strength of the system is it's self-contained nature, able to automatically
calibrate the camera arrangements using correspondences found between each
view.
Experience level heuristics are also discussed regarding capturing and
maintaining user attention in an environment where the experience can be
occurring anywhere, including behind the user.
\begin_inset Flex TODO Note (Margin)
status open
\begin_layout Plain Layout
Link with work
\end_layout
\end_inset
\end_layout
\begin_layout Plain Layout
A point is also made about how the nature of this room based experience
breaks much of the typical game-user interaction established by virtual
reality and video games.
In contrast to traditional and virtual reality game experiences where the
game is ultimately in control of the user or user avatar, AR experiences
of this type have no physical control over the user and extra considerations
must be made when designing such systems.
\end_layout
\end_inset
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
Traditional media consumption is not the only area of interest for developing Consumer applications pioneer the spaces of video gaming
interactive experiences, an investigation into the value of AR and VR for \begin_inset CommandInset citation
improving construction safety is presented by LatexCommand cite
key "pokemonGO"
literal "false"
\end_inset
, education
\begin_inset CommandInset citation
LatexCommand cite
key "ar-anatomy,ar-education"
literal "false"
\end_inset
and commerce
\begin_inset CommandInset citation
LatexCommand cite
key "ar-commerce,ikea-place"
literal "false"
\end_inset
but the use of XR in health-care
\begin_inset CommandInset citation
LatexCommand cite
key "ar-adrenalectomy"
literal "false"
\end_inset
and dangerous work environments
\begin_inset CommandInset citation
LatexCommand cite
key "ar/vr-construction"
literal "false"
\end_inset
presents the opportunities for life-saving results.
\end_layout
\begin_layout Standard
An investigation into the value of AR and VR for improving construction
safety is presented by
\begin_inset CommandInset citation \begin_inset CommandInset citation
LatexCommand citeauthor LatexCommand citeauthor
key "ar/vr-construction" key "ar/vr-construction"
@ -1527,7 +1524,7 @@ g to reduce the effect of memory on safety.
status open status open
\begin_layout Plain Layout \begin_layout Plain Layout
Link with work Link with work?
\end_layout \end_layout
\end_inset \end_inset
@ -1558,23 +1555,47 @@ Kinect
cameras and a virtual reality headset. cameras and a virtual reality headset.
Users are placed in a virtual space constructed from 3D renders of the Users are placed in a virtual space constructed from 3D renders of the
physical environment around the user. physical environment around the user.
Virtual manipulation of the space can then be achieved with visual, spatial
\begin_inset Note Comment
status open
\begin_layout Plain Layout
Virtual manipulation of the space can then be achieved with visual, spatial
and temporal changes supported. and temporal changes supported.
Objects can be scaled and sculpted in realtime while the environment can Objects can be scaled and sculpted in real-time while the environment can
be paused and rewound. be paused and rewound.
The strength of mixed reality comes with the immersion of being virtually The strength of mixed reality comes with the immersion of being virtually
placed in a version of the physical surroundings, tactile feedback from placed in a version of the physical surroundings, tactile feedback from
the environment compounds this. the environment compounds this.
\begin_inset Flex TODO Note (Margin)
status open
\begin_layout Plain Layout
Link with work
\end_layout \end_layout
\end_inset \end_inset
Acquisition uses multiple
\noun on
Kinect
\noun default
v2 sensors with the
\noun on
RoomAliveToolkit
\noun default
\begin_inset CommandInset citation
LatexCommand cite
key "roomalive"
literal "false"
\end_inset
for calibration.
This calibration process utilises a projected series of Gray codes visible
by each sensor to localise each.
The
\noun on
LiveScan
\noun default
calibration process removes the need for additional projector hardware
by using a set of printed calibration markers to localise each sensor.
\end_layout \end_layout
\begin_layout Subsubsection \begin_layout Subsubsection
@ -1590,7 +1611,7 @@ name "subsec:Handheld-Augmented-Reality"
\begin_layout Standard \begin_layout Standard
This project deals primarily with augmented reality facilitated through This project deals primarily with augmented reality facilitated through
mobile phones, specifically from Selinis' work mobile phones, specifically based on Selinis' work
\begin_inset CommandInset citation \begin_inset CommandInset citation
LatexCommand cite LatexCommand cite
key "livescan3d-android" key "livescan3d-android"
@ -1615,7 +1636,7 @@ literal "false"
Android Android
\noun default \noun default
. .
As such, the state of handheld AR is briefly presented here. As such, the state of handheld AR development is briefly presented here.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -1658,8 +1679,8 @@ literal "false"
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
These frameworks provide native AR environment's in which important prerequisite These frameworks provide native AR environments in which important prerequisites
s including rear-camera pass-through, device motion tracking and plane tracking including rear-camera pass-through, device motion tracking and plane tracking
are implemented with the performance expected of an OS-level library. are implemented with the performance expected of an OS-level library.
\end_layout \end_layout
@ -1733,11 +1754,11 @@ Unity
\noun on \noun on
Hololens Hololens
\noun default \noun default
application to a handheld application to
\noun on \noun on
Android Android
\noun default \noun default
target. .
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -1953,7 +1974,7 @@ status open
\begin_inset Graphics \begin_inset Graphics
filename ../media/telepresence-stereoscopic.png filename ../media/telepresence-stereoscopic.png
lyxscale 30 lyxscale 30
width 40col% width 30col%
\end_inset \end_inset
@ -2011,9 +2032,10 @@ The
Microsoft Research Microsoft Research
\noun default \noun default
paper demonstrates a system using 8 cameras surrounding a space. paper demonstrates a system using 8 cameras surrounding a space.
Each camera captured both near infra-red and colour images to construct Each camera captured both stereo near infra-red and monocular colour images
a colour-depth video stream, a more complex camera configuration than in with additional structured light information to construct a colour-depth
the others cited. video stream, a more complex camera configuration than many of the others
cited.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -2247,6 +2269,16 @@ name "fig:World-in-Miniature-group-by-group"
\end_layout \end_layout
\begin_layout Standard
In comparison to
\noun on
LiveScan
\noun default
these provide domain-specific applications, the implementation developed
within this project aims to provide a general many-to-one application of
the concept to suit the existing philosophy of the suite.
\end_layout
\begin_layout Subsection \begin_layout Subsection
Summary Summary
\end_layout \end_layout
@ -2304,8 +2336,7 @@ Xbox Kinect
v2 camera to record and transmit 3D renders over an IP network. v2 camera to record and transmit 3D renders over an IP network.
A server can manage multiple clients simultaneously in order to facilitate A server can manage multiple clients simultaneously in order to facilitate
multi-view configurations, it is then responsible for displaying the renderings multi-view configurations, it is then responsible for displaying the renderings
in real-time and/or transmitting composite renders to a user experience in real-time and/or transmitting holograms to a user experience or UE.
or UE.
This architecture can be seen in figure This architecture can be seen in figure
\begin_inset CommandInset ref \begin_inset CommandInset ref
LatexCommand ref LatexCommand ref
@ -2463,7 +2494,7 @@ Kinect
status open status open
\begin_layout Plain Layout \begin_layout Plain Layout
Extend Extend?
\end_layout \end_layout
\end_inset \end_inset
@ -2543,7 +2574,7 @@ OpenGL
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
This structure can be seen in figure The structure up to reconstruction at the server can be seen in figure
\begin_inset CommandInset ref \begin_inset CommandInset ref
LatexCommand ref LatexCommand ref
reference "fig:server-structure" reference "fig:server-structure"
@ -2553,7 +2584,7 @@ noprefix "false"
\end_inset \end_inset
. , aspects related to the transmission to user experiences are omitted.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -3095,17 +3126,29 @@ Twitch
\noun on \noun on
Instagram Instagram
\noun default \noun default
's live functionality. 's live streaming functionality.
\begin_inset Flex TODO Note (Margin)
status open
\begin_layout Plain Layout
reference?
\end_layout \end_layout
\end_inset \begin_layout Standard
Another advantage of the suite lies in the required computation power with
both client and server able to run on a single fairly powerful computer.
While systems such as
\noun on
Microsoft's Mixed Media Studios
\noun default
present excellent quality mesh-based renders with extensive post-processing,
there is no expectation of this system running locally on consumer-grade
hardware in real-time.
\noun on
LiveScan3D
\noun default
extends access to such acquisition and reconstruction technology in much
the same way that the
\noun on
Kinect
\noun default
did.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -3656,7 +3699,7 @@ Point3f
\begin_layout Standard \begin_layout Standard
Finally, static methods generate common rotation transformations about each Finally, static methods generate common rotation transformations about each
axis given an arbitrary angle. axis given an arbitrary angle employing Euler angles.
This provided a foundation on which to define how the This provided a foundation on which to define how the
\noun on \noun on
OpenGL OpenGL
@ -4246,6 +4289,21 @@ name "subsec:Mobile-AR"
\end_layout \end_layout
\begin_layout Standard
Here the multi-source updates made to the mobile AR application are presented.
In order to complete this update two objectives must be achieved,
\end_layout
\begin_layout Itemize
The network and rendering behaviour must become source ID-aware, able to
differentiate and render separate scenes
\end_layout
\begin_layout Itemize
The touch input management must be restructured to support the individual
manipulation of separate holograms
\end_layout
\begin_layout Standard \begin_layout Standard
The architecture of the mobile AR application can be divided into two areas The architecture of the mobile AR application can be divided into two areas
of concern. of concern.
@ -4361,7 +4419,7 @@ s is scaled in population to match the size of the hologram and then each
\end_layout \end_layout
\begin_layout Subsubsection \begin_layout Subsubsection
Design Considerations Implementation
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -4382,8 +4440,8 @@ does prefab need defining?
\end_inset \end_inset
was created with the intention of encapsulating the necessary components was created with the intention of encapsulating the necessary components
required to represent a whole source including its presentation touch input required to represent a whole source including its presentation and touch
management. input management.
The The
\noun on \noun on
PointCloudRenderer PointCloudRenderer
@ -4478,7 +4536,7 @@ LiveScan
When encountering adverse network conditions in a single source scenario, When encountering adverse network conditions in a single source scenario,
the desired action could be to wait until transmissions from the client(s) the desired action could be to wait until transmissions from the client(s)
can resume. can resume.
With only a single stream, the alternative would be to quit the experience. With only a single stream, the alternative would be to halt the experience.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -4498,7 +4556,7 @@ simulcast
\noun on \noun on
NFL's RedZone NFL's RedZone
\noun default \noun default
where multiple games can be watched simultaneously dividing the screen. where multiple games can be watched simultaneously, dividing the screen.
Were one of the games to experience transmission issues, it could be considered Were one of the games to experience transmission issues, it could be considered
beneficial to the experience to remove the game from display and wait for beneficial to the experience to remove the game from display and wait for
the conditions to improve, especially in a commercial context. the conditions to improve, especially in a commercial context.
@ -4520,9 +4578,9 @@ stale
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
This could be achieved using a separate thread that periodically iterates This was achieved using a separate thread that periodically iterates through
through the last frame of each source and compares an associated timestamp the last frame of each source and compares an associated timestamp to the
to the current time. current time.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
@ -4716,6 +4774,19 @@ The global settings object was not removed but instead had it's function
Evaluation and Discussion Evaluation and Discussion
\end_layout \end_layout
\begin_layout Standard
\begin_inset Flex TODO Note (inline)
status open
\begin_layout Plain Layout
no individual manipulation of mobile holograms
\end_layout
\end_inset
\end_layout
\begin_layout Standard \begin_layout Standard
The server display's control scheme could be more intuitive as the directions The server display's control scheme could be more intuitive as the directions
of movement are in relation to the fixed axes of the display space instead of movement are in relation to the fixed axes of the display space instead
@ -6349,25 +6420,30 @@ The literature review contextualises the
\noun on \noun on
LiveScan LiveScan
\noun default \noun default
suite within the wider spaces of XR, 3D video and multi-source holoportation suite within the wider spaces of XR, volumetric video and multi-source
itself. holoportation itself.
Previous examples of holoportation are presented and their aims of achieving Previous examples of holoportation are presented and their aims of achieving
telepresence are discussed. telepresence are discussed.
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
The results of the project are The results of the project are presented with the limitations also discussed.
\begin_inset Note Comment
\begin_inset Flex TODO Note (inline)
status open status open
\begin_layout Plain Layout \begin_layout Plain Layout
laid out showing good progress through the required areas of development PLACEHOLDER
\end_layout \end_layout
\end_inset \end_inset
.
Of these areas of concern, the display element has been extended in order \begin_inset Note Comment
status open
\begin_layout Plain Layout
Of these areas of concern, the display element has been extended in order
to allow the rendering of multiple environments simultaneously with a dynamic to allow the rendering of multiple environments simultaneously with a dynamic
sub-system of geometric transformations. sub-system of geometric transformations.
The transformations are responsive to user input allowing arbitrary placement The transformations are responsive to user input allowing arbitrary placement
@ -6377,13 +6453,18 @@ laid out showing good progress through the required areas of development
intuitive. intuitive.
\end_layout \end_layout
\begin_layout Standard \begin_layout Plain Layout
The next steps for the project leading up to its completion are presented, The next steps for the project leading up to its completion are presented,
the initial and current plans for the remaining work is described and additiona the initial and current plans for the remaining work is described and additiona
l stretch goals are defined for any additional time. l stretch goals are defined for any additional time.
How the work will be presented in a final report is also described. How the work will be presented in a final report is also described.
\end_layout \end_layout
\end_inset
\end_layout
\begin_layout Section \begin_layout Section
Conclusions Conclusions
\end_layout \end_layout
@ -6399,6 +6480,23 @@ Kinect
\end_layout \end_layout
\begin_layout Standard \begin_layout Standard
\begin_inset Flex TODO Note (inline)
status open
\begin_layout Plain Layout
PLACEHOLDER
\end_layout
\end_inset
\end_layout
\begin_layout Standard
\begin_inset Note Comment
status open
\begin_layout Plain Layout
At roughly halfway through the time allowed for this project the native At roughly halfway through the time allowed for this project the native
display has successfully been extended to meet the deliverable specification. display has successfully been extended to meet the deliverable specification.
This has resulted in the This has resulted in the
@ -6409,18 +6507,23 @@ OpenGL
arbitrary placement and orientation within the display space. arbitrary placement and orientation within the display space.
\end_layout \end_layout
\begin_layout Standard \begin_layout Plain Layout
From this point the network layer of the suite will be developed to also From this point the network layer of the suite will be developed to also
match the specification, allowing connected clients to be grouped into match the specification, allowing connected clients to be grouped into
sources for polling and processing. sources for polling and processing.
\end_layout \end_layout
\begin_layout Standard \begin_layout Plain Layout
Following the development of the two, testing methodologies will be defined Following the development of the two, testing methodologies will be defined
and carried out to gather quantitative results for the final product. and carried out to gather quantitative results for the final product.
A final report on the results will be available in May 2020. A final report on the results will be available in May 2020.
\end_layout \end_layout
\end_inset
\end_layout
\begin_layout Standard \begin_layout Standard
\begin_inset Newpage newpage \begin_inset Newpage newpage
\end_inset \end_inset
@ -6504,16 +6607,6 @@ name "fig:LiveScan-server-UI"
\end_inset \end_inset
\begin_inset Flex TODO Note (Margin)
status open
\begin_layout Plain Layout
add before picture
\end_layout
\end_inset
\end_layout \end_layout
\begin_layout Itemize \begin_layout Itemize

View File

@ -245,13 +245,15 @@
@inproceedings{roomalive, @inproceedings{roomalive,
abstract = {RoomAlive is a proof-of-concept prototype that transforms any room into an immersive, augmented entertainment experience. Our system enables new interactive projection mapping experiences that dynamically adapts content to any room. Users can touch, shoot, stomp, dodge and steer projected content that seamlessly co-exists with their existing physical environment. The basic building blocks of RoomAlive are projector-depth camera units, which can be combined through a scalable, distributed framework. The projector-depth camera units are individually autocalibrating, self-localizing, and create a unified model of the room with no user intervention. We investigate the design space of gaming experiences that are possible with RoomAlive and explore methods for dynamically mapping content based on room layout and user position. Finally we showcase four experience prototypes that demonstrate the novel interactive experiences that are possible with RoomAlive and discuss the design challenges of adapting any game to any room.}, abstract = {RoomAlive is a proof-of-concept prototype that transforms any room into an immersive, augmented entertainment experience. Our system enables new interactive projection mapping experiences that dynamically adapts content to any room. Users can touch, shoot, stomp, dodge and steer projected content that seamlessly co-exists with their existing physical environment. The basic building blocks of RoomAlive are projector-depth camera units, which can be combined through a scalable, distributed framework. The projector-depth camera units are individually autocalibrating, self-localizing, and create a unified model of the room with no user intervention. We investigate the design space of gaming experiences that are possible with RoomAlive and explore methods for dynamically mapping content based on room layout and user position. Finally we showcase four experience prototypes that demonstrate the novel interactive experiences that are possible with RoomAlive and discuss the design challenges of adapting any game to any room.},
author = {Jones, Brett and Sodhi, Rajinder and Murdock, Michael and Mehra, Ravish and Benko, Hrvoje and Wilson, Andy and Ofek, Eyal and MacIntyre, Blair and Raghuvanshi, Nikunj and Shapira, Lior}, author = {Jones, Brett and Sodhi, Rajinder and Murdock, Michael and Mehra, Ravish and Benko, Hrvoje and Wilson, Andy and Ofek, Eyal and MacIntyre, Blair and Raghuvanshi, Nikunj and Shapira, Lior},
booktitle = {UIST '14 Proceedings of the 27th annual ACM symposium on User interface software and technology}, booktitle = {Proceedings of the 27th Annual ACM Symposium on User Interface Software and Technology},
doi = {10.1145/2642918.2647383},
isbn = {978-1-4503-3069-5}, isbn = {978-1-4503-3069-5},
month = {October}, month = {October},
pages = {637--644}, pages = {637--644},
publisher = {ACM}, publisher = {ACM},
series = {UIST '14},
title = {RoomAlive: Magical Experiences Enabled by Scalable, Adaptive Projector-Camera Units}, title = {RoomAlive: Magical Experiences Enabled by Scalable, Adaptive Projector-Camera Units},
url = {https://www.microsoft.com/en-us/research/publication/roomalive-magical-experiences-enabled-by-scalable-adaptive-projector-camera-units https://doi.org/10.1145/2642918.2647383}, url = {http://doi.acm.org/10.1145/2642918.2647383},
urldate = {2020-03-27}, urldate = {2020-03-27},
year = {2014} year = {2014}
} }
@ -308,8 +310,9 @@
@online{arkit, @online{arkit,
author = {Apple}, author = {Apple},
date = {2017-06-05}, date = {2017-06-05},
organization = {Apple},
title = {ARKit}, title = {ARKit},
url = {https://developer.apple.com/augmented-reality/arkit/}, url = {https://developer.apple.com/augmented-reality/arkit},
urldate = {2020-03-27} urldate = {2020-03-27}
} }
@ -542,3 +545,111 @@
year = {2013} year = {2013}
} }
@article{sfs-over-time,
author = {Cheung, Kong Man and Baker, Simon and Kanade, Takeo},
doi = {10.1007/s11263-005-4881-5},
journal = {International Journal of Computer Vision},
keywords = {3D Reconstruction; Shape-From-Silhouette; Visual Hull; Across Time; Stereo; Temporal Alignment; Alignment Ambiguity; Visibility},
month = {May},
number = {3},
pages = {221--247},
title = {Shape-From-Silhouette Across Time Part I: Theory and Algorithms},
url = {https://www.ri.cmu.edu/publications/shape-from-silhouette-across-time-part-i-theory-and-algorithms},
urldate = {2020-05-07},
volume = {62},
year = {2005}
}
@inproceedings{laurentini-solids-revolution,
author = {{Laurentini}, A.},
booktitle = {[1992] Proceedings. 11th IAPR International Conference on Pattern Recognition},
doi = {10.1109/ICPR.1992.201662},
pages = {720--724},
title = {The visual hull of solids of revolution},
url = {https://ieeexplore.ieee.org/document/201662},
urldate = {2020-05-07},
year = {1992}
}
@phdthesis{sfs-video-cm,
address = {Pittsburgh, PA},
author = {Cheung, Kong Man},
keywords = {Temporal Shape-From-Silhouette; Visual Hull Alignment; Human Kinematic Modeling; Markeless Motion Tracking; Motion Rendering and Transfer},
month = {October},
number = {CMU-RI-TR-03-44},
school = {Carnegie Mellon University},
title = {Visual Hull Construction, Alignment and Refinement for Human Kinematic Modeling, Motion Tracking and Rendering},
url = {https://www.ri.cmu.edu/publications/visual-hull-construction-alignment-and-refinement-for-human-kinematic-modeling-motion-tracking-and-rendering/},
urldate = {2020-05-07},
year = {2003}
}
@article{ar-adrenalectomy,
author = {Lin, Mao-Sheng and Wu, JungleChi-Hsiang and Wu, Hurng-Sheng and Liu, JackKai-Che},
doi = {10.4103/UROS.UROS_3_18},
journal = {Urological Science},
month = {05},
title = {Augmented reality-Assisted single-incision laparoscopic adrenalectomy: Comparison with pure single incision laparoscopic technique},
url = {https://www.researchgate.net/publication/324480263_Augmented_reality-Assisted_single-incision_laparoscopic_adrenalectomy_Comparison_with_pure_single_incision_laparoscopic_technique},
urldate = {2020-05-07},
volume = {29},
year = {2018}
}
@article{ar-anatomy,
abstract = {Although cadavers constitute the gold standard for teaching anatomy to medical and health science students, there are substantial financial, ethical, and supervisory constraints on their use. In addition, although anatomy remains one of the fundamental areas of medical education, universities have decreased the hours allocated to teaching gross anatomy in favor of applied clinical work. The release of virtual (VR) and augmented reality (AR) devices allows learning to occur through hands-on immersive experiences. The aim of this research was to assess whether learning structural anatomy utilizing VR or AR is as effective as tablet-based (TB) applications, and whether these modes allowed enhanced student learning, engagement and performance. Participants (n = 59) were randomly allocated to one of the three learning modes: VR, AR, or TB and completed a lesson on skull anatomy, after which they completed an anatomical knowledge assessment. Student perceptions of each learning mode and any adverse effects experienced were recorded. No significant differences were found between mean assessment scores in VR, AR, or TB. During the lessons however, VR participants were more likely to exhibit adverse effects such as headaches (25\% in VR P < 0.05), dizziness (40\% in VR, P < 0.001), or blurred vision (35\% in VR, P < 0.01). Both VR and AR are as valuable for teaching anatomy as tablet devices, but also promote intrinsic benefits such as increased learner immersion and engagement. These outcomes show great promise for the effective use of virtual and augmented reality as means to supplement lesson content in anatomical education. Anat Sci Educ 10: 549--559. {\copyright} 2017 American Association of Anatomists.},
author = {Moro, Christian and {\v S}tromberga, Zane and Raikos, Athanasios and Stirling, Allan},
doi = {10.1002/ase.1696},
eprint = {https://anatomypubs.onlinelibrary.wiley.com/doi/pdf/10.1002/ase.1696},
journal = {Anatomical Sciences Education},
keywords = {gross anatomy education; health sciences education; undergraduate education; medical education; virtual reality; augmented reality; mixed reality; computer-aided instruction; oculus rift; tablet applications},
number = {6},
pages = {549--559},
title = {The effectiveness of virtual and augmented reality in health sciences and medical anatomy},
url = {https://anatomypubs.onlinelibrary.wiley.com/doi/abs/10.1002/ase.1696},
urldate = {2020-05-07},
volume = {10},
year = {2017}
}
@article{ar-commerce,
abstract = {This study evaluates the effectiveness of augmented reality (AR) as an e-commerce tool using two products --- sunglasses and watches. Study 1 explores the effectiveness of AR by comparing it to a conventional website. The results show that AR provides effective communication benefits by generating greater novelty, immersion, enjoyment, and usefulness, resulting in positive attitudes toward medium and purchase intention, compared to the web-based product presentations. Study 2 compares the paths by which consumers evaluate products through AR versus web with a focus on interactivity and vividness. It is revealed that immersion mediates the relationship between interactivity/vividness and two outcome variables --- usefulness and enjoyment in the AR condition compared to the web condition where no significant paths between interactivity and immersion and between previous media experience and media novelty are found. Participants' subjective opinions about AR are examined through opinion mining to better understand consumer responses to AR.},
author = {Yim, Mark Yi-Cheon and Chu, Shu-Chuan and Sauer, Paul L.},
doi = {10.1016/j.intmar.2017.04.001},
issn = {1094-9968},
journal = {Journal of Interactive Marketing},
keywords = {Augmented reality; Interactivity; Vividness; Immersion; Novelty; Previous media experience},
pages = {89--103},
title = {Is Augmented Reality Technology an Effective Tool for E-commerce? An Interactivity and Vividness Perspective},
url = {http://www.sciencedirect.com/science/article/pii/S1094996817300336},
urldate = {2020-05-07},
volume = {39},
year = {2017}
}
@online{ikea-place,
author = {IKEA},
month = sep,
organization = {Inter IKEA Systems B.V.},
title = {IKEA Place},
url = {https://apps.apple.com/ie/app/ikea-place/id1279244498},
urldate = {2020-05-07},
year = {2017}
}
@article{ar-education,
abstract = {Augmented reality (AR) is an educational medium increasingly accessible to young users such as elementary school and high school students. Although previous research has shown that AR systems have the potential to improve student learning, the educational community remains unclear regarding the educational usefulness of AR and regarding contexts in which this technology is more effective than other educational mediums. This paper addresses these topics by analyzing 26 publications that have previously compared student learning in AR versus non-AR applications. It identifies a list of positive and negative impacts of AR experiences on student learning and highlights factors that are potentially underlying these effects. This set of factors is argued to cause differences in educational effectiveness between AR and other media. Furthermore, based on the analysis, the paper presents a heuristic questionnaire generated for judging the educational potential of AR experiences.},
author = {Radu, Iulian},
doi = {10.1007/s00779-013-0747-y},
issn = {1617-4917},
journal = {Personal and Ubiquitous Computing},
number = {6},
pages = {1533--1543},
risfield_0_da = {2014/08/01},
title = {Augmented reality in education: a meta-review and cross-media analysis},
url = {https://link.springer.com/article/10.1007/s00779-013-0747-y},
urldate = {2020-05-07},
volume = {18},
year = {2014}
}