Blog

November 5, 2025

A Practical AI Framework for Evaluating Mobile UI

As someone deep into AI frameworks for my dissertation, I’m always looking for research that’s not just theoretically sound — but actually usable.

A recent paper [1] on evaluating Mobile User Interfaces (MUIs) stood out to me. It tackles a question that most design teams quietly struggle with:
How can we objectively score UI design without endless manual reviews or coding effort?

The Old Bottlenecks

Most UI evaluation falls into two unscalable approaches:

  • Manual evaluation: Accurate but painfully slow. It relies on users or experts to test and score each design.
  • Source-code analysis: Faster but only works if you can read and process code, which locks out many designers.

The researchers propose something smarter: a deep learning–based framework that evaluates UI screenshots directly. No users. No source codes. Just images.

Inside the Framework

The framework unfolds in three clear stages—each addressing a major real-world bottleneck.

1. Data Preparation & Preprocessing

They began with the RICO dataset—9,677 labeled mobile UI screenshots. Each image was standardized, resized, and normalized to feed into the model. This foundational step ensures consistency—a must when your input data comes from different devices, apps, and visual styles.

2. Balancing the Data

The dataset had a major issue: imbalance. About 76% were “good” UIs, and only 24% were “bad.”
That’s a recipe for a biased model that always predicts “good.”

Their fix was smart—Borderline-SMOTE (Synthetic Minority Oversampling Technique). It creates new “bad UI” samples near the decision boundary—exactly where the classifier tends to make mistakes.

That single decision was transformative.
Before SMOTE: 74% accuracy.
After SMOTE: 93% accuracy.

A reminder that robust preprocessing often matters more than fancy models.

3. The Hybrid Model: DenseNet201 + KNN

Here’s where the real innovation lies.

  • DenseNet201 — a 201-layer convolutional neural network—extracts deep visual features. It “sees” everything from edges and colors to layout and hierarchy.
  • Instead of adding a complex neural classifier, they use K-Nearest Neighbors (KNN) for classification. It simply measures how “close” a new design is to known examples.

This hybrid deep + classical ML setup keeps the system both interpretable and powerful.

My Take: Practical Tool vs. Diagnostic Insight

So, what’s actually new here?

The Strength:
It’s a practical model. Designers can upload screenshots and get an objective “good/bad” score—instantly. No code. No test users. Just actionable feedback.

The clever use of SMOTE also shows strong methodological rigor—something many applied AI studies overlook.

The Gaps:
However, the model only gives a general score. It won’t tell you why a UI is bad—whether it’s poor spacing, contrast, or hierarchy.
Also, its training is limited to the RICO dataset, so it still needs validation on larger, more diverse sets of UIs.

And finally, for an imbalanced dataset, I would have liked to see F1-score and AUC reported—not just accuracy, precision, and recall.

Beyond UI: The Framework’s Real Potential

What excites me most isn’t the model itself—it’s the framework logic.
This architecture can automate any visual evaluation task where humans currently verify “does this look right?”

A few real-world parallels:

  • Insurance Claims: An AI system could instantly reject blurry or incomplete damage photos, prompting the field agent to retake them. That alone can cut claim delays by days.
  • E-commerce & Retail: Online marketplaces could automatically score product images against brand standards, flagging low-quality uploads at scale.
  • Manufacturing: Cameras on the production line could detect scratches or misprints in real time—without relying on manual inspection.

This paper’s biggest contribution isn’t about app design at all. It’s about designing a replicable AI workflow that scales human judgment.

Conclusion

This study is a great example of how data discipline and architectural simplicity can outperform brute-force modeling.
By pairing a balanced dataset with a hybrid AI pipeline, the researchers created a system that’s not just accurate but genuinely usable.

More importantly, it sets a precedent. The future of applied AI lies in frameworks that free experts from repetitive visual checks so they can focus on higher-value analysis and creative strategy.

That’s what practical AI should do: make our workflows lighter, faster, and smarter—not just louder.

Reference

[1] M. Soui and Z. Haddad, “Deep Learning-Based Model Using DensNet201 for Mobile User Interface Evaluation,” International Journal of Human–Computer Interaction, vol. 39, no. 9, pp. 1981–1994, 2023, doi: 10.1080/10447318.2023.2175494.

Blogging Blog, Infotainment Review
About Dipo Tepede

I am a Project Management coach. I specialize in making delegates pass any Project Management certification at first try. I successfully achieve this fit through practical application of the knowledge and integration of our Project Management eLearning school at www.pmtutor.org. Welcome to my world.....