What should I look for first when evaluating a UX consultancy?
Prioritize evidence of outcomes: case studies with measurable impact, senior team involvement, and a repeatable methodology that spans research, design, and validation. Tip: Ask for pre/post metrics from two similar projects to verify outcome consistency.
How do I compare vendors with similar case studies?
Compare who actually did the work, decision cadence, research rigor, and how findings drove design changes. Tip: Request a sample research-to-design trace showing how one key insight led to a shipped change.
What signals show a good culture fit with our team?
Look for open communication, facilitation skill, and transparent risk management. Tip: Run a 60–90 minute workshop pilot to test collaboration before you commit.
How much domain experience should a UX partner have?
Sector knowledge helps, but the ability to transfer UX patterns matters more. Tip: Ask for 3 examples of cross-domain pattern reuse that saved time or reduced risk.
What differentiates a senior-led engagement from a junior-heavy one?
Senior leads make sharper scoping choices, faster synthesis, and clearer stakeholder alignment. Tip: Insist on named senior leads with defined hours and participation at key milestones.
Should I prioritize local, regional, or global UX partners?
Choose based on access to users, time zone overlap, and governance. Tip: Map your research windows and stand-ups; ensure the partner’s core team overlaps at those moments.
How many references should I check?
Three is ideal: one similar scope, one similar culture, one stretch complexity. Tip: Ask each reference what they would change if restarting the project.
What proof points matter more than awards?
Before/after metrics, reduced rework, adoption, and decision velocity. Tip: Ask for a single slide summarizing problem, intervention, and measured impact.
How do I avoid choosing on portfolio polish alone?
Probe the research backbone and validation rigor behind polished visuals. Tip: Request two unpolished artifacts (e.g., synthesis wall, test script) from a completed project.
When is a specialized boutique better than a large firm?
Boutiques can be faster and senior-heavy; large firms scale well and add breadth. Tip: Match complexity and speed needs to the staffing model actually proposed.
Who will actually work on my project day-to-day?
Confirm named individuals, roles, weekly hours, and responsibilities. Tip: Ask for a RACI showing who drives research, design decisions, and stakeholder alignment.
How do multidisciplinary teams collaborate effectively?
Clear rituals: research readouts, design crits, decision logs, and developer handoffs. Tip: Request a sample cadence calendar and example decision log.
What seniority mix is appropriate for complex work?
Complexity rises with unknowns; increase senior research and strategy time. Tip: Ask for a risk-adjusted staffing plan tied to uncertainty reduction milestones.
How do you ensure knowledge continuity when team members rotate?
Use living docs, decision logs, recorded readouts, and paired handovers. Tip: Require a continuity plan with owner, artifacts list, and backup roles.
Do your team members teach, mentor, or publish?
Practitioner-educators tend to articulate rationale clearly and coach stakeholders. Tip: Ask for one recent deck or article illustrating their thinking.
How do you integrate with our product and engineering teams?
Through joint discovery, design/dev pairing, and acceptance criteria aligned to test insights. Tip: Pilot a design-dev spike to validate the working model.
What design systems experience do you bring?
We create and evolve systems with tokens, accessibility, and governance. Tip: Request a sample contribution model and audit report of a past system.
Can you support service design as well as product UX?
Yes—service blueprints align backstage operations to frontstage experiences. Tip: Ask for a sample blueprint that drove measurable operational change.
How do you staff research for hard-to-reach audiences?
Use layered recruitment, mixed methods, and ethical incentives. Tip: Review a previous recruitment plan for a similarly constrained cohort.
What languages and regions can you cover?
We support multilingual research and global time zones with local moderators. Tip: Validate moderator bios and localization plans for key markets.
What’s your start-to-finish process in plain language?
Align on outcomes → plan research → synthesize insights → co-create concepts → iterate design → validate → hand off → measure. Tip: Ask for a one-page process map with sample artifacts per phase.
How do you tailor scope without undercutting quality?
We reduce breadth, not rigor—fewer journeys, same depth of research and validation. Tip: Demand an explicit trade-off table for any scope change.
What does a strong discovery phase produce?
Aligned goals, risks, user segments, and a research plan with success metrics. Tip: Gate the project: continue only after a discovery readout and sign-off.
How often will we see work in progress?
Expect weekly or bi-weekly touchpoints with demoable artifacts. Tip: Schedule fixed readouts in the contract to protect visibility and momentum.
How do you prevent design-by-committee?
Evidence-led decisions with a named decision owner and criteria set in advance. Tip: Require a decision charter and stick to it.
How do you handle technical constraints early?
We co-define constraints with engineering and prototype within real limits. Tip: Include engineering in concept reviews from week one.
What do handoff deliverables include?
Annotated flows, acceptance criteria, tokens, component specs, and accessibility notes. Tip: Ask for a sample handoff package from a shipped project.
How do you ensure traceability from insight to design?
We link insights to requirements, patterns, and final UI decisions. Tip: Review a traceability matrix for one previous feature.
What tools will we use for collaboration?
Commonly Figma, Miro, secure repositories, and your PM stack. Tip: Align tool access and permissions during kickoff to avoid delays.
How do you close projects to ensure internal adoption?
Through enablement sessions, playbacks, and a 30–60–90 day measurement plan. Tip: Make enablement a contracted deliverable with attendance targets.
What research methods do you use and when?
Interviews, field studies, usability tests, surveys, analytics, and diary studies—selected based on decision needs. Tip: Ask for a method-to-decision map for your brief.
How do you ensure research quality and ethics?
IRB-informed practices, informed consent, data minimization, and secure storage. Tip: Request the consent template and data retention policy.
How are accessibility requirements built into design?
WCAG-driven acceptance criteria, semantic structure, keyboard flows, and contrast checks. Tip: Require an accessibility checklist per screen or component.
Do you test with assistive technologies?
Yes—screen readers, magnification, voice control, and switch devices as appropriate. Tip: Include at least one session with AT users per critical flow.
How large do usability studies need to be?
Start lean (5–8 per segment) and iterate; expand for riskier flows. Tip: Budget for two rounds rather than one large test.
Can you use our existing analytics and VOC data?
Absolutely—we triangulate quant and qual to target high-impact issues. Tip: Provide baseline funnels and VOC themes before discovery.
How do you measure success post-launch?
Define leading indicators (task success, errors) and lagging ones (conversion, retention). Tip: Agree on 3–5 KPIs with thresholds before design begins.
What quality gates do you use before release?
Test coverage review, accessibility audit, heuristic pass, and stakeholder sign-off. Tip: Make a go/no-go checklist part of the SOW.
How do you handle research with regulated audiences?
Compliance-ready protocols, anonymization, and sector-specific consent. Tip: Ask for a redacted protocol from a regulated study.
Do you support longitudinal learning after release?
Yes—continuous discovery and periodic health checks feed the roadmap. Tip: Schedule quarterly UX health reviews tied to product OKRs.
Fixed price or time-and-materials — which is better?
Discovery uncertainty favors T&M; well-defined build phases suit fixed price. Tip: Use hybrid: T&M for discovery, fixed for well-bounded sprints.
How do you estimate accurately without overpadding?
Bottom-up tasking, historical velocity, and risk-adjusted reserves. Tip: Ask for a three-scenario estimate (base, stretch, lean) with assumptions.
What’s your change management process?
Documented change requests, impact analysis, and client approval gates. Tip: Add a no-surprises clause: changes surfaced within 48 hours of discovery.
Who owns IP and research data?
Clients typically own IP and anonymized data; we retain process know-how. Tip: Clarify IP, raw files, and derivative rights in the MSA and SOW.
How do you handle confidentiality and security?
NDAs, least-privilege access, encrypted storage, and vetted vendors. Tip: Request a brief security checklist and tool inventory.
What risks do you flag early in UX programs?
Recruitment feasibility, dependency timing, and decision bottlenecks. Tip: Include a living risk log with owners and mitigations.
How are invoices tied to milestones?
Billing aligns to phase gates and accepted deliverables. Tip: Link payments to mutually defined acceptance criteria.
How do you ensure transparency on progress and burn?
Weekly dashboards show percent complete, risks, and budget burn. Tip: Require a sample dashboard in the proposal.
What’s your policy on subcontractors?
We disclose, vet, and integrate subs with the same standards and NDAs. Tip: Ask for named subs and their bios before kickoff.
How do we exit cleanly if priorities change?
Short termination windows, knowledge transfer, and asset packaging. Tip: Ensure the contract specifies a structured wind-down plan.