Module 6 2 AR Selection
Module 6 2 AR Selection
• The MetaSpark Studio platform offers the user flexible tooling to easily build
and share AR experiences at scale. Users can start developing from templates
or build their own custom AR experiences with code.
• It is easy to build interactive AR experiences with the platform, allowing for
target tracking, depth mapping, occlusion and audio and haptic integrations.
Selection of AR Platform
• 8th Wall provides powerful tools to create WebAR experiences which are optimized
for the browser. 8th Wall supports great AR features such as face effects, world
tracking, and image targets, and allows for source code control through a cloud
editor.
Selection of AR Platform
Platform
Tool Feature Set Ease of Use Best For
Compatibility
High-quality visuals, multi-user
ARKit iOS Apple developers iOS apps
experiences
Environmental understanding, multi-
ARCore Android, iOS Android Studio/Unity Cross-platform apps
user
echo3D Multiple Cloud-based, multi-user, animations Drag and drop Web and mobile AR
Unity Multiple 3D applications, multi-platform Requires coding Complex 3D AR apps
Object/image recognition, cloud
Vuforia Multiple Compatible tools Interactive experiences
storage
Wikitude Multiple Image/object/place recognition Simple editor Professional apps
Spark AR iOS, Android Social media effects User-friendly Social media effects
Web, iOS,
8th Wall Web-based AR, multi-user Unity & JavaScript Web-based AR
Android
Kudan iOS, Android SLAM, multi-user Drag and drop Mobile games, demos
ARToolKit Multiple Basic tracking, custom markers Unity compatible Basic AR projects
Selection of AR Platform
https://www.microsoft.com/en-in/hololens/hardware#document-experiences
Selection of AR Platform
Source :https://pilotinstitute.com/drone-imu-calibration-explained/#key-takeaways
Use case- Drone Control in AR
Drone IMU Calibration
Accelerometer
Accelerometers are responsible for measuring linear acceleration, including changes in velocity and
tilt angles. They utilize the principles of inertia to detect variations in motion.
https://www.youtube.com/watch?v=i2U49usFo10&t=107s
Gyroscope
• Every drone is constantly being subjected to a number of forces coming from different directions.
These forces, such as wind, affect the drone’s yaw, pitch, and roll making the drone very hard to
control.
• Integrated gyroscopes can almost instantly detect changes in the position of a drone and
compensate for it in such a way, that it basically seems unaffected, as it re-adjusts its position
hundreds of time every second or can hover calmly in place.
• Modern gyroscopes are manufactured with components between 1 to 100 micrometers in size and
often include sensors for multiple axes in a single package.
https://www.youtube.com/watch?v=cquvA_IpEsA&t=212s
Magnetometers
Magnetometers determine the drone’s heading or orientation to Earth’s magnetic field.
Use case- Drone Control in AR
Drone IMU Calibration
Source: Konstantoudakis, K.;Christaki, K.; Tsiakmakis, D.; Sainidis,D.; Albanis, G.; Dimou, A.; Daras, P.Drone Control in AR: An IntuitiveSystem for Single-Handed Gesture control, Drone Tracking, and Contextualized
Camera Feed Visualization in Augmented Reality.Drones 2022, 6, 43. https://doi.org/10.3390/drones6020043
Tracking
▪ Tracking in Augmented Reality (AR) refers to the technology and techniques
used to align virtual objects with the real world in a way that makes them
appear stable and interactable.
▪ Effective tracking ensures that virtual elements remain consistent with real-
world objects, positions, and movements. Here are the primary types of
tracking used in AR:
Marker-based Tracking
• Relies on physical triggers such as predefined images, patterns, or QR codes
(called markers) placed in the environment.
• The AR system recognizes these markers through the device’s camera, and
virtual objects are overlaid on or around the marker.
• Market Detection
• Scans environment and searches for predefined markers(high contrast patterns) that the
system recognizes
• Pose Estimation
• Computer vision techniques calculate the position and orientation of the marker relative
to the camera’s viewpoint ensuring accurate alignment of contents
• Content Overlay
• With the determined marker’s pose, digital content is overlaid onto the physical marker
in real-time blending the virtual and real worlds
Source: https://www.youtube.com/watch?v=d8XJ83AO-HI
Marker-based Tracking –Use case
Retail : markers on product packaging can activate detailed visualizations of
product features, assembly instructions, or interactive demonstrations, enhancing
consumer engagement and satisfaction
• https://www.linkedin.com/posts/abdulsalam-faheedat-opeyemi_ar-ar-
markerlessar-activity-7191383388786671618-
o5Ck?utm_source=share&utm_medium=member_desktop
Markerless Tracking
GPS/Location-based Tracking
• Uses GPS data from the device to place virtual objects at specific
geographical locations.
• This is often combined with accelerometer and compass data for
better accuracy.
• Use cases:
• Location-based AR apps, tourism guides, and social AR apps that show
content in real-world locations.
GPS/Location-based Tracking
Projection-based Tracking
▪ uses projectors to overlay AR content directly over real-life objects
using projection mapping techniques.
▪ Projection-based AR is appealing and interactive where light is blown
onto a surface and the interaction is done by touching the projected
surface.
Projection-based Tracking-Use case
https://www.youtube.com/shorts/y-LGVeiQ1l8
Projection-based Tracking
Augmented Reality Training for Industrial Assembly Work
https://www.youtube.com/watch?v=mpSG94HTMpQ
Superimposition based Tracking
▪ In this type of AR, object recognition plays a vital role because an app
cannot replace an original object with an augmented one if it cannot
identify the original object.
▪ This type of AR has been popularized by social platforms such as
Instagram, Facebook, and Instagram using filters.
Superimposition based Tracking
▪:
Source : https://zolak.tech/blog/augmented-reality-for-furniture-industry
Types of AR-Use Cases
https://digitalpromise.org/initiative/360-story-lab/360-production-
guide/investigate/augmented-reality/getting-started-with-ar/types-of-ar/
NIKE has taken over a huge LED billboard in the Shinjuku area of Tokyo, Japan to show a forced
perspective illusion celebrating the Air Max shoe.
Source:https://www.youtube.com/watch?v=mcl2nIeJh_A&list=TLGGzBkRwcLgNJQwNTExMjAyNA&t=14s
Challenges and Advances:
▪ Lighting conditions: AR tracking can be sensitive to lighting; poor
lighting can affect camera-based tracking.
▪ Occlusion: Properly occluding virtual objects behind real-world
objects remains a challenge in achieving realism.
▪ Environmental complexity: The more detailed or dynamic an
environment is, the more sophisticated the tracking system needs to
be to avoid jitter or misplacement.
AR Computer Vision
▪ AR Computer Vision is a subset of computer vision technology
applied to Augmented Reality (AR).
▪ It enables devices to interpret and understand the physical
environment in real-time, allowing virtual objects to be overlaid
seamlessly onto the real world.
▪ This integration is essential for creating immersive AR experiences.
▪ Image Recognition
• Identifies and classifies specific images or patterns from the camera feed.
This could include logos, QR codes, landmarks, or any predefined objects.
• Using techniques like feature detection ,the system matches key points in
the image with a database to recognize the object or image.
• Use cases: AR shopping (recognizing products), educational apps
(identifying objects), AR games, and content linked to brand logos.
Cont.,
▪ Object Detection and Recognition
• Identifies and locates specific objects within a scene, such as furniture, human
faces, or even hand gestures.
• By applying machine learning models (like Convolutional Neural Networks -
CNNs) trained on specific objects, the system can classify and locate objects in
the camera’s view. This is used to place virtual objects in relation to real-world
items.
• Use cases: AR filters (detecting faces for masks), virtual try-ons (clothes,
makeup), object-based information overlays in educational apps, and AR in
robotics.
▪ Simultaneous Localization and Mapping (SLAM)
• Allows devices to map an unknown environment while tracking their
position in it.
• SLAM combines data from cameras and inertial sensors (gyroscope,
accelerometer) to continuously update a map of the environment while
simultaneously tracking the camera's position within it. This is essential for
markerless AR.
• Use cases: AR navigation, virtual object placement in physical spaces (e.g.,
furniture apps), interactive AR games, and spatial understanding.
Cont.,
▪ Depth Sensing
• Determines the distance of objects from the camera, enabling 3D spatial
awareness.
• Techniques like stereo vision (using two cameras), structured light
(projecting a known pattern onto a surface), or time-of-flight (using light
pulses) are used to measure depth.
• Newer AR devices like iPhones with LiDAR sensors can measure depth
more accurately, allowing virtual objects to interact more realistically with
the physical world.
• Use cases: Accurate virtual object placement, occlusion (making virtual
objects disappear behind real ones), AR measuring tools, and immersive
games.
Cont.,
▪ Surface Detection and Plane Detection
• Identifies horizontal and vertical surfaces where virtual objects can be
anchored, such as tables, floors, or walls.
• The system looks for patterns in pixel movement (optical flow) and texture in
the environment to detect flat surfaces. Once a surface is detected, virtual
objects can be placed with stability.
• Use cases: Furniture apps, AR art installations, and object placement in AR
games.
▪ Pose Estimation
• Determines the orientation and position of an object or person relative to the
camera.
• Using key point detection and geometric algorithms, pose estimation
identifies how an object is positioned in 3D space relative to the camera.
• This is used for aligning virtual objects with real ones or for creating realistic
interactions with the user.
• Use cases: AR gaming (aligning weapons or tools in first-person perspective),
virtual avatars, motion tracking, and gesture-based interactions.
Cont.,
▪ Feature Detection and Tracking
• Detects and tracks key visual features (edges, corners, etc.) in the real-world
environment to maintain consistent positioning of virtual elements.
• Algorithms like Harris Corner Detection, Shi-Tomasi, or FAST are used to
detect stable visual features, which can then be tracked over time to anchor
AR objects in place.
• Use cases: AR annotations on real-world objects, interactive AR games, and
tracking moving objects in AR experiences.
▪ Occlusion Handling
• Ensures that virtual objects appear behind real-world objects when
appropriate, creating a more realistic and immersive experience.
• Using depth information from depth sensors or advanced algorithms, the AR
system can determine when a real object is closer to the camera than a virtual
one and thus should occlude the virtual object.
• Use cases: Realistic AR scenes where virtual characters walk behind real
objects, or AR measuring apps where virtual measurements disappear behind
physical objects.
Cont.,
▪ Scene Understanding and Semantic Segmentation
• Identifies and categorizes different parts of the environment (e.g.,
distinguishing between the floor, walls, furniture, people, and objects).
• Machine learning models, particularly CNNs or vision transformers, are
used to segment the camera feed into different object classes (e.g., sky,
road, person, car). This enables more advanced interactions with the
environment.
• Use cases: AR city navigation (placing virtual signs and information on
roads and buildings), interactive games, and contextual AR overlays for
educational purposes.
Mapping
▪ Mapping in Augmented Reality (AR) refers to the process of creating a
digital representation of the physical environment, allowing virtual
objects to be overlaid, interact, and stay anchored in the real world.
▪ This is critical for immersive AR experiences, as it enables devices to
understand and respond to the geometry, layout, and objects in a scene.
▪ Plane Detection
• Plane detection is used to find flat surfaces (both horizontal and vertical) in
the environment, such as floors, tables, and walls. Once detected, these planes
serve as anchors for virtual objects.
• Using visual cues, such as parallel lines, textures, and depth information, the
AR system identifies flat surfaces in the environment. Planes are essential for
creating stable virtual content, ensuring objects don’t “float” or move
unrealistically.
• Use cases: Virtual furniture placement (e.g., IKEA Place), object placement in
AR gaming, and virtual art on walls.
Cont.,
▪ Simultaneous Localization and Mapping (SLAM)
• SLAM is the foundational technology for most AR mapping systems. It
allows AR devices to build a map of the environment in real time while
simultaneously determining their own position within it.
• This is done using data from the device’s camera and inertial sensors (like
accelerometers and gyroscopes).
• Process:
• Feature Detection: The system identifies key features in the environment,
such as edges, textures, and distinct objects.
• Tracking: These features are tracked across frames as the device moves.
• Map Building: As the device moves through the environment, it creates a
3D map of the surroundings, updating with new features.
• Use cases: Markerless AR, indoor navigation, AR gaming, and
placing virtual objects in physical spaces (e.g., furniture apps).
Cont.,
▪ Depth Mapping
• Depth mapping provides a sense of distance between the camera and
objects in the environment.
• Devices equipped with depth sensors, such as LiDAR or structured light
sensors, can create detailed 3D maps by calculating the distance to
surfaces.
• Depth is measured by projecting infrared light or other patterns onto the
environment and analyzing how the pattern deforms when it hits
surfaces. This allows for a precise understanding of the 3D structure of the
scene.
• Use cases: Accurate object placement, occlusion (where virtual objects
disappear behind real objects), AR measurement apps, and enhancing
realism in AR games.
Cont.,
▪ Environment Understanding
• Beyond simple plane detection, some AR systems can classify different
parts of the environment (walls, ceilings, floors, objects) and understand
their semantic meaning.
• Using machine learning and computer vision, the AR system analyzes the
scene to distinguish between different objects and surfaces.
• This can involve detecting furniture, appliances, or even recognizing
specific rooms.
• Use cases: Home design apps, where furniture is placed intelligently, AR-
based interior design, or advanced gaming experiences that interact with
the user’s surroundings.
Cont.,
▪ 3D Object Mapping and Reconstruction
• In some cases, AR needs to map not only surfaces but also complex objects
within the environment. This involves recognizing and reconstructing 3D
objects to place virtual elements correctly.
• The AR system scans the object from multiple angles, generating a 3D model.
This can be done in real time, allowing for dynamic interactions between
virtual objects and real-world objects.
• Use cases: AR games where virtual objects interact with physical objects (e.g.,
a virtual character jumping on a real sofa), or educational apps where objects
can be scanned and enhanced with virtual annotations.
Cont.,
▪ Marker-based Mapping
• Marker-based AR uses predefined images, QR codes, or fiducial markers
as anchors to position virtual objects in the physical world.
• These markers are scanned by the camera and serve as a reference point
for positioning the virtual content.
• When the AR system detects a marker, it calculates the marker’s position,
orientation, and scale in 3D space. This allows virtual objects to be placed
accurately relative to the marker.
• Use cases: Interactive product displays, museum guides, or simple AR
games.
Cont.,
▪ Occlusion and Collision Mapping
• Occlusion occurs when virtual objects are hidden behind real-world
objects, enhancing realism.
• Collision mapping ensures that virtual objects don’t pass through physical
ones, providing more realistic interaction.
• Using depth sensors or computer vision, the system determines the depth
of real-world objects and calculates when a virtual object should be
obscured.
• Collision mapping involves detecting when virtual and real-world objects
should not occupy the same space.
• Use cases: Games where virtual characters interact with real-world
objects, AR-based design apps, or simulations that involve both real and
virtual elements.
Cont.,
▪ Spatial Mapping
• Spatial mapping involves creating a detailed, comprehensive 3D model of
the environment.
• This allows AR applications to place virtual objects more accurately,
accounting for walls, furniture, and other obstacles.
• Using data from cameras, depth sensors, or even external sensors, the
system creates a mesh or point cloud representation of the environment.
• This allows for more precise interactions between virtual objects and the
physical world.
• Use cases: Industrial AR (e.g., factory maintenance), architecture and
construction apps, and immersive AR experiences in entertainment.
Cont.,
▪ Cloud-based AR Mapping
• Cloud-based AR mapping allows multiple users to share the same AR
experience by synchronizing mapping data through the cloud. This makes
it possible for virtual objects to be placed in shared physical spaces,
accessible to different users.
• AR data, such as maps and anchors, are uploaded to the cloud and shared
across multiple devices. This enables multi-user AR interactions, where
everyone sees the same virtual content in the same location.
• Use cases: Multi-user AR games, collaborative design apps, or tourism
apps with shared AR experiences.
References
▪ https://medium.com/echo3d/top-10-ar-developer-tools-every-
developer-should-know-32fef1471883
▪ https://aimagazine.com/top10/top-10-augmented-reality-ar-
platforms
▪ https://daily.dev/blog/ar-tools-for-developers
▪ https://program-ace.com/blog/types-of-ar/