Personal Markup: Redesigning for Usability and Flexibility

A mobile-first annotation tool for field workers to quickly and safely perform data collection and communicate across teams.

Company

Esri

Timeline

6 months

Responsibilities

UX/UI Design • Accessibility • Documentation • Developer Handoff

Team

Product Designer I • Sr. Principal Product Designer • 2 QA Engineers • 2 Developers

Overview

Personal Markup is a mobile-first, freeform tool that allows field workers to sketch notes, annotations, and visual markings directly on a map—supporting rapid communication between field teams and back-office stakeholders without requiring GIS admin setup.

Unlike structured data capture, Markup supports flexible, in-the-moment workflows, making it especially valuable for planning work areas, documenting site conditions, and sharing context quickly across teams.

A large number of organizations rely on this feature—roughly a third of the Field Maps user base.

This redesign was delivered as part of the larger mobile platform rebuild, which meant improvements needed to stay tightly scoped while addressing the most impactful usability issues.

Problem


Through centralized research and ongoing customer feedback led by Product Managers, the team identified recurring pain points in the existing Markup experience:

  • Accidental drawing when users attempted to pan the map

  • Difficulty moving or selecting sketched items

  • Poor discoverability of tools

  • Inconsistent interaction patterns between iOS and Android

  • Limited clarity around drawing modes and actions

  • Friction when managing sketches after creation

These issues made a core workflow feel unpredictable and harder to learn—especially for non-GIS users.

Because Markup was being redesigned alongside a broader app rebuild, the team made a deliberate decision:

Rather than introducing new capabilities, focus on improving clarity, predictability, and control within existing functionality.

Goals

To reduce accidental interactions and restore expected map behaviors.

Research

Insights came from multiple sources:

  • Centralized research findings shared across teams

  • Continuous customer feedback surfaced by Product Managers

  • Usage patterns from existing Markup workflows

  • Design audits of current UI and interaction models

These inputs helped categorize opportunities into must / should / could, allowing the team to prioritize the most impactful usability fixes within rebuild constraints.

Design Approach

We approached the redesign in two phases, starting on iOS before establishing parity on Android.

Enhancing Existing Patterns

The first exploration focused on improving the current experience using familiar platform conventions:

  • Introducing explicit tool selection before drawing

  • Adding instructional cues to guide user actions

  • Separating color and type selection so users could configure attributes before or after drawing

  • Restoring standard map panning with one finger

  • Adding clear undo and editing controls

The goal was to align Markup with users’ existing mental models: choose a tool first, then act—rather than drawing immediately on touch.

This phase emphasized clarity, predictability, and reducing cognitive load.

Exploring New Interaction Models

In parallel, we researched patterns from other drawing and annotation tools to explore alternative interaction approaches. This led to experimenting with custom UI that provided clearer tooling and flexibility while still respecting native platform behaviors.

Through multiple rounds of iteration and stakeholder feedback, we converged on a redesigned flow that balanced:

  • Familiar mobile interactions

  • Explicit tooling and labeling

  • Flexible drawing modes

  • Platform consistency across iOS and Android

Key Design Decisions

Rather than expanding Markup with new features, we prioritized usability within existing capabilities:

  • Require explicit tool selection before drawing

  • Add instructional text and labels for key actions

  • Allow users to choose type and color independently

  • Restore standard pan/zoom behavior outside of drawing mode

  • Introduce clearer undo and edit affordances

  • Improve post-markup management (layer visibility, reordering, sharing)

  • Support stylus input on larger form factors, enabling precise drawing while preserving finger-based navigation

These decisions focused on making the experience more approachable for non-GIS users while preserving power for advanced workflows.

Outcome

The redesigned Markup experience delivered:

  • Clearer tool hierarchy and labeling

  • More predictable gesture behavior

  • Reduced accidental drawing

  • Improved navigation between map and markup states

  • Stronger post-creation management of sketches

  • Stylus support for precise field annotations

These changes translated complex spatial annotation into simpler, task-oriented interactions—making Markup more reliable and approachable in real-world field conditions.

Implicit & Explicit Gestures
Pan/Zoom, drawing different forms, selecting multiple items, and moving them on the map.
Post-Markup Management
Viewing drawn item details and actions, toggling visibility across maps, and sharing settings.
Stylus Recognition on Tablet
Users can draw with the stylus and pan and zoom with their fingers.

Platform Parity

Designs were finalized first on iOS, then adapted for Android to achieve parity while respecting platform conventions. Where native components differed, we made intentional choices between visual parity and implementation parity—favoring long-term maintainability and reuse over one-off custom solutions.

This ensured the experience felt native on each platform while remaining structurally consistent.

Metrics

Following the release of the rebuilt Markup experience in October 2025:

  • Thousands of unique users engaged with the new Markup workflows in the first two months

  • Ongoing telemetry and customer feedback continue to inform iteration and refinement

The redesign improved usability while staying aligned with the broader platform rebuild, demonstrating how targeted human-centered improvements can ship successfully within large system modernization efforts.

Reflection

This project reinforced the importance of designing within constraints. Working on Personal Markup highlighted how meaningful UX improvements don’t always require new features—often they come from clarifying intent, restoring familiar patterns, and reducing friction in existing workflows.

By partnering closely with Product and Engineering, we balanced human-centered design with platform realities, delivering a more predictable and accessible experience while contributing to a scalable foundation for future growth.

Personal Explorations: Designing for Assistive Spatial Intelligence

Taking it Further

The ideas I discuss further are not part of shipped work, but my own explorations thinking about more intelligent and system-level consolidation of workflows to support field efforts. And so, what if a sketch could do more than one thing?

The personal markup redesign began as a way to make freeform annotation faster and more intuitive. From research however, it was clear that markup wasn’t just a drawing tool, but a way for users to express intent like identifying a hazard, outlining work zones, organizing tasks, etc.

I used this to ideate on shifting markup from capturing information to interpreting and acting on it.

Feature Level - Assistive Linking of Objects Within a System Layer

I explored a system that detects nearby features and surfaces non-blocking suggestions directly within the existing workflow. The user creates a markup and the system identifies any assets in proximity that might be relevant to link based on a distance threshold and the name/type of markup.

Users can link, add and view any assets or dismiss suggestions that may not be relevant. This behavior keeps the interaction inline, non-disruptive, and reversible.

Rather than auto-linking, the system interprets intent signals like proximity and context and proposes actions:

  • Suggestions appear only when confidence is meaningful

  • The user remains the decision-maker

  • The system supports, but does not override

This preserves markup as a user-owned layer, while reducing manual effort.

Area & Workflow Level - Facilitating Intelligent Task Coordination

Since Field Maps also introduced spatially enabled task creation and completion with the new release, I took the opportunity to explore markup as a context-aware system that understands intent at an area level, not just individual features.

As part of its functionality, users can view spatially marked "tasks" on the map and mark their progress and completion. Each task can be marked with a priority, an assignee, due date, status, etc. and supervisors and field workers use this as a to-do list for field work.

In this exploration, when a user draws and labels a region (e.g., “Work Area 1” or “Team 1”), the system interprets it as a meaningful spatial container.

Within this boundary, the system can:

  • Detect tasks and assets contained in the area

  • Understand assignment status (assigned, unassigned, completed)

  • Evaluate priority and proximity or other metrics based on the type of task sorting enabled by the user

  • Summarize workload across the region

This transforms markup into a queryable interface for operational data.

Assistive Workload Intelligence for Field Workers

This sequence shows how a user-defined markup evolves into a coordination layer.When a region is defined as a work area, the system interprets it as a spatial boundary and begins to aggregate and analyze tasks within it, which removes the cognitive load of interpreting larger list-based filters.

Progressive Disclosure of Insight

  • Detection
    The system parses the area and evaluates contained tasks

  • Summary
    Key signals surface inline:

    • Assigned vs unassigned

    • Priority

    • Progress

  • Contextual Action

    • Suggestions to guide next steps (pick up unassigned tasks)

  • Scalability

    • Fewer tasks = direct preview

    • More tasks =a summarized view with “View All”

This keeps the experience scannable and reuses existing UI patterns in the app.

Planning Work Areas with Assistive Insight for Supervisors

Planning Work Areas with Assistive Insight for Supervisors

This design explores how field supervisors can use markup not just to define space, but to plan and validate work distribution in real time.

They may outline an area for assignment to a team, and the system interprets the region as a planning boundary, surfacing feedback by analyzing the tasks within it and providing an assignment action.This transforms area creation from a static action into a feedback-driven planning loop. Here too, the suggestion is immediate, contextual and reversible.

Rather than adding complexity, the system leverages what users already do, and augments it with meaningful, relevant insight.

Go Back