Back to home

Making density navigable

Designing an exploration tool in a world of excels and dashboards.

What I did

Data VisualizationFrontendInteraction designResearchUsability testing
Role
UX Designer + UX Engineer
Team
1 frontend, 1 graph specialist, 3 data scientists
Timeframe
Nov 2024- Mar 2025
cocacola-logo-3

Impact

One place to look

Glint, PowerBI and Excel data unified into one exploration interface, so HR doesn't bounce between systems to answer a single question.

Time to insight

HR partners answer leadership questions same-day, instead of filing analyst requests and waiting weeks.

Adoption

200+ active users from launch, with growth driven by organic word-of-mouth across HR teams.

HR engagement data at a 42,000-person company lived across three tools: Glint, PowerBI, and Excel. Each one showed a fragment.

To answer questions like "how are managers at the same level performing under this leader?" or "what's the trend on accountability for this manager, and when did it hit its lowest?" — HR partners filed a request and waited. One data analyst told us she'd spent weeks building a single dashboard before handing it off.

The problem wasn't access. Every tool was built to display answers, not explore questions. So we built one that did.

THE PROBLEM

Three tools, zero context

Glint showed survey scores. PowerBI showed org-level trends. Excel held the manual cross-cuts data scientists ran on request. Each tool was useful in isolation. None let an HR partner notice a pattern and follow it.

Cross-analysis meant filing a request. Most questions never got asked.

px-tools

THE DESIGN

Three zones, one clear purpose each

Traditional dashboards are organized around what data exists. This needed to be organized around how exploration actually works: drill down, notice patterns, ask new questions.

Three zones, each with one job. They never compete for attention.

Filter panel
Filter + Legend

Control what you see

Data Panel
Navigate

Move through data spatially

Drilling data panel
Drill down

Examine details without losing context

THE HARDEST DESIGN PROBLEM

Five variables, without turning it back into a dashboard

Exploration needs control. But every control we added pulled the tool back toward dashboard logic — show everything, all the time, in case someone needs it.

Five variables had to coexist: who you're looking at, which question, granularity, comparison point, view mode. Any change had to feel like the same thought, refined — not like switching screens.

One control panel, everything reactive in real time. No "apply filters" button, no reload.

panel left

DETAIL THAT MATTERS

What we almost shipped

The first version showed ring averages for every organizational level — first-line, second-line, all the way up — visible at all times. The thinking was that more context would help users orient.

In testing, it didn't. Multiple averages competed for attention, and users couldn't tell which one mattered for the team they were looking at.

We stripped it back to one contextual average — for the level the user is currently exploring. Same data, less to parse.

mistake

How we knew it worked

78%

Task completion in usability testing

4.6/5

Ease of use

4.8/5

Likelihood to reuse

What I'd do differently

Visualization alone didn't close the loop. Users still needed to communicate findings — reports, slides, screenshots — and we hadn't designed for that. I saw it from the beginning but didn't push hard enough to scope phase 2 from kickoff: report generation, slide export, screenshot sharing.

Not to build it then, but so the team had it in mind when making architecture decisions. The tool enabled exploration. It didn't enable action.

© Git / Made with ❤️ in Next / 2026