Simplifying Research Workflows

Researchers at Princeton University struggle with inefficient data management systems, which hampers the potential of their research and makes it difficult to maintain the sustainability and accessibility of their data.

In this case study, I focus on the major aspects of my contributions to the project, highlighting the key areas I was responsible for.

Role:

Product Designer

Duration:

8 Months

Team:

1 Product Owner
1 Project Manager
1 UX Researcher
3 Product Designers
1 RDM Specialist
2 Developers

Deliverables:

High Fidelity Prototypes
Research Report
Design System
Product Roadmap
Case Study Website
Executive Presentation

Context:

With the rise of open science and the adoption of FAIR principles—guidelines designed to make data easier to find, access, and reuse—the scientific community is shifting towards more transparent and collaborative research, with research data management at its core.

Research Data Management (RDM) is the process of providing proper labeling, storage, and access to data at every stage of a research project.

The Problem

Over 200,000+ research datasets at Princeton remain poorly organized, difficult to access, and disconnected from one another.

Research data is complex, often containing diverse materials, code, and large datasets that must be well-organized and tagged for future use. However, current systems are fragmented, hard to navigate, and lack integration with institutional data storage.

Improper research data management is a critical obstacle that keeps groundbreaking discoveries locked away. Researchers need efficient, organized systems to unlock and share this valuable knowledge for progress and innovation.

80%
scientific data are lost within two decades.
$10.6 bn
wasted every year due to improper data management.
65%
researchers rely on unsustainable storage solutions.

The most pressing challenge in managing research data for researchers at Princeton is the interface.

Researchers currently access institutional data via a command-line interface, which, while powerful, requires memorizing many commands. This complexity adds to their workload, hindering efficient data management.

Research Insights

Few Key Insights That Informed My Design Decisions

01

Speed and accuracy are critical under tight deadlines, with researchers requiring fast, error-free tagging solutions to maintain efficiency.

02

Tagging systems must reflect domain-specific structures, enabling researchers to organize metadata according to their field’s complex categories and ontologies.

03

Researchers often work with massive datasets and need precise search capabilities, expressing frustration with interfaces that lack granular control or require endless scrolling.

04

Not all researchers have the same level of technical literacy, but those with high technical skills don’t want to lose efficiency when transitioning to a more accessible interface.

05

Research frequently involves iterative refinement of datasets, making it crucial for researchers to save and reuse filters to efficiently scale their tagging and filtering processes.

"How might we create a metadata tagging system that empowers researchers to efficiently organize and tag their datasets at scale, while being intuitive enough for non-technical users and flexible enough for advanced users?"

Final SOLUTION

Designing for Speed, Accuracy & Usability

Minimizing context switching for speed and accuracy

A researcher may need to tag files with metadata efficiently under tight deadlines. This can be difficult because switching between different sections or screens to select files and apply tags can slow them down and increase the likelihood of errors.

The two-panel design solves this problem by allowing users to view and select files on the left while applying tags on the right, enabling a seamless workflow.

Leveraging mental models to simplify tagging

Researchers need to tag files using metadata tags that match the specific structure of their field. It can be tricky for them because they’re often dealing with really complex categories and relationships, that can cause redundant tags that mean the same thing.

The tree-like structure organizes metadata naturally, allowing users to easily explore and select tags, staying consistent and efficient in their tagging process.

Efficient file search & filtering for managing large datasets

Researchers work with massive datasets, and they need to quickly find specific files to apply metadata tags. This can be overwhelming because scrolling through endless lists to locate the right files is time-consuming and frustrating.

The search bar, quick filters, and advanced filter button help users efficiently narrow down file lists—searching by keywords, accessing common criteria, and creating customized queries for precise filtering.

Balancing usability for all technical levels

Researchers with varying levels of technical literacy need to create complex filters to find specific files. Creating such filters can feel daunting for less technical users, while highly technical users might find visual interfaces slow or restrictive.

The visual query builder solves this by offering an intuitive drag-and-drop interface for creating filters without requiring coding skills.

For technical users, the code switch provides a direct way to write filters using syntax they’re comfortable with, ensuring they can work efficiently without unnecessary steps.

Streamlining iterative refinement with saved queries

Researchers often refine their datasets iteratively, needing to revisit and adjust their filters multiple times. Without a way to save their work, they’d have to recreate complex filters from scratch every time, which is time-consuming and prone to errors.

The saved queries feature lets users store and reuse filters, saving time, ensuring consistency, and streamlining the tagging and filtering process. 

Together, these design elements enable researchers to efficiently apply custom filters and tag files at scale.

IMPACT

Making Data Management Efficient

  • The design introduces a structured way to manage metadata, replacing ad-hoc, manual, and inconsistent methods. Researchers now have a repeatable and scalable process for tagging and filtering data, likely transforming how they organize their work.
  • For non-technical researchers who previously couldn’t fully engage in metadata management, the system opens new opportunities for participation. They now contribute directly to organizing and tagging, fostering inclusivity in team workflows.
  • The system allows users to tag and organize 5x more files per hour compared to the Open Science Framework (OSF), a widely used research data management tool. This can significantly help researchers working with large datasets.
  • Researchers who spend significant time on manual metadata management could spend about 50% less time managing metadata, allowing them to focus on analysis or experimentation.
How Did we get here?

Planning Research Enabled Fast & Structured Progress

Literature Review
Survey
Heuristic Evaluation
User Interviews
Co-Design Sessions

Before diving deeper into the problem space, we held a kickoff with the Product Team and development leads at TigerData. Our team gained a clear understanding of the product requirements and the vision set for the product, which allowed us to define our problem statement more effectively by setting goals and early constraints.

Kickoff meeting with the Product and Development teams from TigerData at Princeton.

With the problem defined, we were able to set clear research goals and desired outcomes for our research phase using Francine Gemperle’s framework. This framework was crucial in shaping a concise research plan.

Leveraging Francine Gemperle’s outcome-based research framework to outline a research plan.

Following this plan, our team conducted a literature review of existing solutions, a series of interviews with researchers and librarians, a survey distributed to over 3,000 researchers across different universities, a heuristic evaluation of tools currently in use, and co-design sessions. Each team member contributed equally to the execution of this research plan under the lead of a UX Researcher from our team.

Survey distributed to 3000+ researchers across different universities.
User interviews with researchers and librarians to gather insights.
Co-design sessions where participants used basic shapes and text to visualize their ideas.
Ideation and Concept Validation

Using My Technical Background to Validate Ideas

Crazy 200s
Sketching
Storyboarding
Worst Possible Ideas
User Enactment

Our team went all-out in the ideation phase, generating over 200 rapid sketches that explored various combinations of technologies, user groups, and problems. This process resulted in a wide range of ideas—some promising, others less so.

We discussed all our ideas, identified underlying themes, and grouped them accordingly.

To refine our concepts, we used methods like Worst Possible Ideas, Storyboarding, and User Enactment to identify the underlying themes across all our ideas. With a wealth of concepts, we then decided which to keep, combine, or discard.

Using Worst Possible Ideas and Storyboarding to conceptualize our ideas.

Leveraging my technical background, I understood how TigerData’s backend worked and identified potential issues with certain designs. For example, the middleware couldn’t send WRITE commands to the cloud for security reasons, preventing even simple features like renaming. I raised these concerns with my team, and we discussed alternative solutions with the Product and Development teams.

Learn More

Find more about the overall project through below deliverables.

I took part in building a design system for this project along with solely designing and developing a website for early adopters.

Interactive prototype of the full product.