Testing AI Design Tools While Solving Real Problems
AI design tools are everywhere. Promises of 10x productivity. Revolutionary workflows. But which ones actually work? Rather than theorize, I decided to test them systematically across an entire design process - from concept to interactive prototype. Not with a fictional brief, but while solving a real problem: helping people avoid overdrafts when managing multiple financial commitments. The real question: Where does AI genuinely accelerate design work, and where does it just create more work?
I identified a pain point affecting many adults aged 25–45: managing payments scattered across multiple platforms.
The scenario: You have Netflix, Spotify, and Disney+. You're paying off a laptop in 3 installments and a coat in 4. Different dates. Different merchants. No consolidated view. Then payday hits, everything charges at once, and you're overdrawn.
Design challenge : How might we reduce overdraft risk for users managing multiple recurring financial commitments?
I benchmarked UI patterns across Revolut, Klarna, Monzo, Rocket Money, and Floa - examining dashboards, subscription flows, alerts, and payment schedules. This informed the design requirements:
This became my testing ground. Every stage, I documented what worked and what didn't.
Tool tested: Make Design System
Approach: Generate initial system, then formalize manually in Figma
Key decisions:
Verdict: AI design systems are excellent starting points but need human refinement for consistency and accessibility. Saved approximately 3 hours on initial setup.
Tools tested: ChatGPT and Claude
Approach: Conversational brainstorming for IA structure, user flows, navigation patterns
Developed:
Verdict: AI excels at generating alternatives and surfacing edge cases I missed. The dialogue format forces more thorough thinking than solo work. Claude slightly better for structured reasoning.
Learning: Use AI as thinking partner, not oracle. Question its suggestions. Push back. The conversation is where value lives.
The test: Uizard vs. Visily & Make
I created identical wireframes of dashboard in Uizard and Make and other wireframes in Visily, then compared output quality and efficiency.
Uizard:
Visily:
Make:
Learning: Not all AI tools are created equal. The quality gap was substantial enough to impact entire project timelines. Tool selection matters as much as skill.
Tools tested: Make + Build.io plugin + Manual Figma
Hybrid approach:
Strategic decision: Created key screens with clean autolayout and components entirely by hand to establish a vision inorder to prepare detailed prompts for prototyping.
Learning: Optimal workflow combines AI speed with human precision. Use AI for exploration; humans for quality control and strategic screens.
Tool tested:Lovable
The strategic opportunity:While building the prototype, I identified a critical gap in the competitive landscape. Existing apps consolidate financial data but provide zero personalized recommendations. They show you the problem but don't help you solve it
Product strategy shift:Leverage AI not just as a design tool, but as a core product differentiator. Transform passive financial tracking into proactive financial guidance
Using ChatGPT for product strategy:going beyond nudges and notifications
Three strategic AI capabilities:
Experience the working application built with Lovable. Click below to interact with the full prototype.
Built with Lovable • Full interactive experience
Learning: Lovable was ideal for this phase—not just for visual prototyping, but for testing complex AI interaction patterns. Its strength isn't making static mockups; it's building functional experiences that demonstrate product value. This revealed AI's dual role in modern product development: as a design accelerator AND as a strategic feature that creates competitive advantage.
In a real case scenario:
Open-banking: Access via PSD2 APIs through an ACPR-authorised AISP; this demo never sees or stores credentials.
Scope: Read-only budgeting and subscription detection. No money movement.
Consent: Access is time-limited by law; users re-confirm to keep syncing.
Insights: Advisory only; users can correct or dismiss.
SEPA: Refund rights—8 weeks (no questions) / up to 13 months if unauthorised.
Cancellation: Fluxia would deep-links to the merchant; France’s “3-click cancel” applies where supported.
Automated decisions: None with significant effects; human review available.
Privacy: Users can export or delete their data at any time.
Not AI replacing designers. AI augmenting designers.
The most effective workflow:
The future of design isn't about choosing sides—AI or human. It's about strategic integration.
The designers who will thrive aren't those who resist AI or adopt everything blindly. They're designers who:
This project proved: Understanding tool strengths and limitations is now as important as mastering design principles.
Beyond the learnings, this project produced:
Continue testing tools like ProtoPie, Make. Refine the AI-augmented workflow. Document patterns for when to use which tools. Share findings with design teams navigating similar questions.
The goal isn't to be an early adopter or a skeptic. It's to be systematic, critical, and strategic about integrating new tools that genuinely improve design work.