Balboa Park Project Hero

Balboa Park Alive!

Designing an interactive augmented reality app to educate young kids and families about local biodiversity.

Duration

January 2024 - Present

Role

UX Design
UX Engineering

Client

University of California, San Diego
Funded by Dr. Seuss Foundation

Tools Used

Figma
Unity
Lightship ARDK
Mapbox SDK
MediaPipe Hand Tracking

Background

Balboa Park Alive! is a series of immersive, interactive phone-based installations that leverage Augmented Reality (AR) to explore biodiversity of San Diego/Tijuana.

Balboa Park Alive App Screens

This app integrates Niantic Lightship ARDK, Mapbox, and mobile hand tracking to facilitate immersive interactions with 3D digital depictions of the creatures and flora, making this region a leading biodiversity spot on the planet. Distinct from similar conservation-focused mobile AR applications, Balboa Park Alive! places emphasis on first-person experiential learning and research-backed methods of exploration. By using directed motion and visual cues, it encourages families to interact with the environment and bond with each other.

The Situation

Balboa Park Alive! provides location-based experiences to explore biodiversity using AR. I joined the team during the alpha stage, where the app was already functional. After initial testing with a small group of families and young kids in the Balboa Park region, the team found that the app lacked usability, proper flow, and design consistency.

My responsibility was to make the app more usable, functional, and immersive. To do this, I needed to understand all the details thoroughly. I began by exploring various aspects of the app, which can be seen below:

Selecting the AR Experience

• The map displays various location markersthroughout Balboa Park

• Tapping on a marker reveals relevantinformation about that location

• After tapping "Start," the camera opens tobegin the AR experience

Experiencing AR in action

• It prompts the user to plant a seed by swipingup on it

• After planting, it prompts the user to selecta pollinator

• The app then shows if the chosen pollinatoris a match or not

Upon examining their current app, I found myself with numerous questions, two of them were...

How might we create an immersive & interactive experience for young kids?

How might we enable young kids to explore, find and learn more about nature?

Approach

Approach

Research

Design Audit

Design Audit

Competitive Analysis

Competitive Analysis

Ethnographic Field Study

Ethnographic Field Study

Design Audit

I started with a detailed design audit, focusing on four key areas: Service Design, Design Research, UX Design, and Visual Design. My goal was to identify and highlight potential enhancements to optimize the AR experience for children. This included increasing interaction clarity, reinforcing visual feedback, improving accessibility, and integrating gamification elements to create a more engaging and educational user journey.

Design Audit

After dividing the application into different design domains, I gained a comprehensive understanding of its current state. This process helped me identify key insights and pinpoint weak areas that needed improvement. By focusing on Service Design, UX Design, Research, and Visual Design, I could clearly see where enhancements were needed to optimize the AR experience for our users.

Competitive Analysis

After reviewing the current application flow across different design domains, I had a clear understanding of the areas needing improvement. To proceed, I began researching competitors in the learning game industry and analyzing their solutions.

Competitive Analysis

To gain a deeper understanding, I selected two competitors with the same target group, platform, and similar features. I analyzed their strengths, weaknesses, and product strategies by creating screen flows and examining them across the same design domains, articulating how we could improve our own product by learning from their approaches.

Ethnographic Field Study

After auditing the current app, I asked my team to help organize an ethnographic field study to observe users in a natural environment. Since this is a location-based experience, we set up in Balboa Park and spent time watching families and young kids as they launched the location-based AR experiences. This allowed us to see the challenges and frustrations encountered in real-time, providing valuable context for our research. To gather more insights, we interviewed a few parents and their kids, asking about their opinions, experiences with the current product, and ideas for improvement.

Study in Balboa Park Region

Study performed in the Balboa Park Region

Observing Users

Observing users in natural environment

Analysis

To gain a deeper understanding, I selected two competitors with the same target group, platform, and similar features. I analyzed their strengths, weaknesses, and product strategies by creating screen flows and examining them across the same design domains, articulating how we could improve our own product by learning from their approaches.

Research Insights

Persona

Based on our insights and observations from the research methods, I created a persona that captured the essence of our users and their characteristics. I also formulated behavior patterns, tech interactions, learning methods to understand more.

User Persona

Cognitive Task Analysis & Journey Map

I created an outline for Cognitive Task Analysis (CTA). CTA was essential for breaking down the steps and identifying critical pain points, which informed how the UI could be more intuitive and engaging. It highlighted the need for clear guidance and feedback throughout the AR experience to keep young users engaged and reduce frustration.

Cognitive Task Analysis

Design

Information Architecture

The app will consist of two main tabs: Map and My Journey. They contain the primary functionalities of the app.

To ensure the application is easy to use, I began by designing the application flow. Based on previous research, I knew our users are impatient and prefer fewer steps. This led me to reduce the number of screens from four to just one by adding a 'My Journey' tab to the 'Map' tab. This tab contains the same information in a list view format, making it a one-click experience.

Information Architecture

After getting clarity with the information architecture, I was ready to move on to designing the user interface for the app. But before diving into sketches, I researched various types of Game UI layouts and explored which type of layout is used for specific information.

As the AR experience contains UI elements—both created using AR and integrated into the real world—these are categorized into four segments: Non-diegetic, Diegetic, Meta, and Spatial UI.

Non-diegetic UI: Elements like menus and HUDs that exist outside the game world, not visible to the characters (In our case, all the buttons, popovers that appear on the screen).

Diegetic UI: Elements that exist within the game world and are visible to the characters, like signs or in-game devices (In our case, the planting of seed, Plant, Pollinator Character).

Meta UI: Elements that represent the player's status or effects, like damage indicators or alert symbols, which can be both diegetic and non-diegetic (In our case, hints appearing on screen while planting seed and in general).

Spatial UI: Elements placed in the 3D space but not part of the game world, like waypoint markers or augmented overlays.

Game Layout

Blueprints

To explore some top-level layout arrangement decisions, I started by creating blueprints for the most important screen of the application—the AR Experience Screen, followed by other screens.

Blueprints over wireframes?
I chose blueprints over wireframes to test the layout arrangement because blueprints focus purely on spatial relationships and the placement of elements without getting into design specifics. This approach helps in making layout decisions by allowing for easy iteration and adjustment of the overall structure, ensuring the layout is functional and intuitive before moving on to more detailed designs.

Blueprint

Low Fidelity Screens

These UI blueprints helped me visualize and evaluate different types of layout arrangements, giving a clear picture of where and how the information would be placed.

With that understanding, I moved on to creating low-fidelity mockups in Figma, refining the layout further.

AR Experience Screen

AR Experience Screen Wireframes

The first iteration has a static sheet that pops up to choose the flower and plant. It felt like the touch targets were small, especially on smaller screens. Based on my team's feedback, I iterated to make the touch targets larger and added a draggable component to make it collapsible. This also aligns with Fitts's Law, which suggests that larger touch targets are easier and faster to interact with.

Home Screen

Home Screen Wireframes

Since the home screen serves as the launchpad for all location-based AR experiences, it needed to be both functional and easy to understand. The current version of the app only allows users to launch experiences by finding them on the map. However, I realized that our primary users, young kids, might feel overwhelmed with this UI. Therefore, I considered a two-tabbed navigation to present the same information in different visual formats. This also helped reduce the number of screens from four to just one, allowing users to launch the experience from My Journey with just one click if they are at that location.

One key UX principle that influenced my decision was Hick's Law, which states that the time it takes to make a decision increases with the number and complexity of choices. By providing two tabs—Map and My Journey—I simplified the decision-making process, reducing cognitive load. This approach ensures that users can easily navigate and find the AR experiences without feeling overwhelmed.

Splash Screen

The Splash screen is designed to create an engaging first impression while the app loads. I incorporated playful elements that hint at the AR experience awaiting users.

Splash Screen Wireframes

Visual Design

This app needs to have a degree of immersion that is heavily driven by usability and, hence, the UI. Since it is based in the Balboa Park region in San Diego, I took some pictures of the park and drew inspiration from nature itself.

Visual Design System

This approach helped me develop a tonal palette with different colors. To ensure accessibility, I used the Adobe Color to test and select accessible color tones. I settled on brown as the primary color for the app, adding an earthy, warm feeling.

Showcase

I was selected along with my team to present our project at the SIGGRAPH '24 conference, the premier event for computer graphics and immersive technologies worldwide. During the conference, I successfully showcased the project to around 100+ attendees. Additionally, we leveraged this opportunity to test our app's usability, gathering valuable feedback and insights from the audience. Read more about the conference here.

SIGGRAPH Conference Logo
Team Testing Session
User Testing at Conference

More Work Coming Soon

In Publication

SIGGRAPH 2024

International Conference

Attended by 9K+ attendees

Our paper titled 'Balboa Park Alive: Exploring Biodiversity through AR' was accepted and presented at SIGGRAPH 2024.

SIGGRAPH 2024

SIGGRAPH 2025 Blogs

Interactive Media

Blog Post

Discover how this AR app brings biodiversity to life, turning nature into an adventure that sparks curiosity and environmental action.

SIGGRAPH 2025 Blogs

Want to get in touch?
I'd love to connect with you

Mail
LinkedIn
Medium
GitHub