Boagworld: UX, Design Leadership, Marketing & Conversion Optimization

Paul Boag, Marcus Lillington

Boagworld: The podcast where digital best practices meets a terrible sense of humor! Join us for a relaxed chat about all things digital design. We dish out practical advice and industry insights, all wrapped up in friendly conversation. Whether you're looking to improve your user experience, boost your conversion or be a better design lead, we've got something for you. With over 400 episodes, we're like the cool grandads of web design podcasts – experienced, slightly inappropriate, but always entertaining. So grab a drink, get comfy, and join us for an entertaining journey through the life of a digital professional.

  1. Why UX Teams Need a Maturity Audit Right Now

    APR 23

    Why UX Teams Need a Maturity Audit Right Now

    UX is under pressure. A proactive maturity audit gives you a voice before leadership makes decisions about your team without you. Something uncomfortable is happening in organizations right now. UX teams are being quietly reassessed. AI has disrupted the field, leadership expectations have gone unmet, and there's a growing sense that UX hasn't delivered what it promised. The conversations are happening, but often not with the people who actually do UX work. If you're in a UX role, decisions about your team's future might be forming in rooms you're not in. That's the situation I've been thinking about lately, and it's why I want to talk about UX maturity audits. Not as a defensive measure or a tick-box exercise, but as a genuinely useful tool for getting ahead of a conversation that's already underway. The expectation gap is real A lot of the cynicism toward UX right now traces back to one thing: overselling. Leadership was told UX would deliver a hundredfold return on every dollar spent. That figure gets thrown around a lot, and someone took it seriously enough to hire one UX person and wait for the magic to happen. It didn't. That disappointment is partly our industry's fault, though it's not something we often admit openly. We've marketed UX with promises that assume a level of organizational change nobody warned leadership they'd have to make. Hiring one person doesn't transform an organization into a user-centric one. It never did. There's a certain naivety in the idea that a single hire will magically produce amazing experiences, without understanding the breadth of change required for an organization to truly become user-focused. But plenty of people implied it would. The result is a leadership team that feels, not unreasonably, like they were sold something that didn't arrive. Why waiting is a bad idea The natural response to this situation is to keep your head down and hope things settle. Understandable, but a mistake. If leadership is already souring on UX, the absence of any structured conversation about what UX is actually delivering gives that skepticism room to grow unchallenged. Decisions start getting made. Quietly, and without much input from the people who understand what's actually happening. A proactive UX maturity audit changes that dynamic. Instead of waiting to be judged, you're shaping the conversation. You're the one bringing evidence, framing the questions, and defining what success looks like. That's a considerably better position to be in. And it's not just damage control. Even mature, well-functioning UX teams benefit from this kind of review. There's always a next stage. Whether it's wider adoption, better integration with product teams, or moving toward something more democratized, an audit helps you see where you are and decide where to go. What a solid audit covers A UX maturity audit should cover five areas. Not exhaustively, but enough to give you a real picture. Strategy and leadership. Does UX have a seat at the table? Is there genuine sponsorship from someone with budget and influence, or is UX being practiced in a corner while real decisions happen elsewhere? Culture and capability. How widely does the organization understand what UX actually involves? Are there training pathways and career development? Or is it just a job title a few people happen to have? Research and design processes. Is UX practice consistent, or does it depend entirely on who's available? Are designers and researchers involved early, or called in after the big decisions are already made? Outcomes and measurement. Can the team point to specific improvements in user outcomes? Are there agreed definitions of what success looks like, and is anyone actually tracking it? Cross-functional integration. Is UX embedded across teams, or sitting in its own silo waiting for people to come to it? None of these are particularly complicated questions. The hard part is being honest about the answers. The difference between a real audit and a survey An audit that just collects opinions tells you what people think, which is interesting but not necessarily accurate. A good audit looks for evidence. That means checking whether research plans actually exist. Whether findings get used or disappear into a folder. Whether design systems are maintained or quietly falling apart. Whether the team can point to specific recent changes that improved user outcomes rather than just shipped features. But the more revealing question is often why these things aren't happening, because the answer usually points straight to the organizational problems that stop UX from gaining traction in the first place. A missing research plan isn't just an admin gap. It's often a signal that no one with authority has made space for it, or that the team has learned it wouldn't be taken seriously anyway. The questions worth asking aren't simply "how good is our UX?" They're "how well is UX supported here? How consistently is it practiced? What would move us forward?" This shifts the audit from a performance review to a diagnostic tool. Diagnostics are much easier to have productive conversations about. Where to start It's worth being honest about one thing before you dive in: this isn't something you can do half-heartedly. A UX maturity audit that gets treated as a side project, or squeezed into the gaps between real work, tends to produce polite summaries that nobody acts on. It needs management buy-in from the outset, not as an afterthought once the findings are ready. There's also a strong argument for bringing in someone external to run it. Not because your internal team lacks the ability, but because independence matters here. People will say different things to an outsider. And an external reviewer is less likely to be seen as someone with a stake in the outcome, which means their conclusions carry more weight when they land on a senior leader's desk. The right person for this isn't someone who will sit in judgment of the UX team's output. The question isn't whether the work is good. The question is whether the organization has created the conditions for good work to be possible. That's a different kind of assessment, and it requires someone who understands enough about how UX actually functions to read the environment accurately rather than just counting deliverables. Given where things are right now, that feels like a fairly important prerequisite. Find The Latest Show Notes

    6 min
  2. AI Is Showing UI Designers the Door

    APR 21

    AI Is Showing UI Designers the Door

    So this month Marcus and I get into a slightly uncomfortable question. If AI can knock out decent interfaces from a text prompt, where does that leave the people whose day job is opening Figma and making screens look nice? We start with Google Stitch, which has been getting a lot of attention lately. Then we zoom out into something I have become mildly obsessed with, which is building AI skills. Not prompt snippets, but reusable, documented processes that let you get consistent work out of AI without drowning it in context. App of the Month This month’s tool is Google Stitch (v2), Google’s AI UI generator. You describe what you want, it produces an interface, and you can do some light manual tweaking. It is not a full replacement for Figma. The editing controls are basic. The bigger story is what it represents. We are now at the point where a decent, usable UI can be generated fast enough that the real value shifts from "can you draw the screens" to "can you judge what good looks like." That is where experience, and yes, taste, starts to matter. If you want to compare approaches, I mentioned Figr again, which I still prefer for the quality of what it produces. Are UI Designers Becoming Vinyl? The question Stitch raises is not "can AI design interfaces". It clearly can. The question is what happens to the job market when "good enough" becomes cheap, fast, and widely available. I found myself telling 2 different clients recently that they could probably skip hiring a UI designer. They had tight budgets, tight timelines, and already had solid brand guidelines or a design system. In those situations, I could push the work through AI, iterate it a bit, and get something perfectly serviceable. That line of advice made me feel a bit grubby. Not because it was wrong for those clients, but because it hints at a bigger shift. My worry is that UI design becomes like vinyl records. Most people will not need it. A small number will care deeply and pay for it. The middle ground shrinks. Marcus made the important caveat here. Some designers will still be in demand because they bring something AI cannot easily fake. A distinctive visual style. Creative judgment. Brand thinking. The ability to make something feel like it came from a real point of view, not a model averaging the internet. We also talked about where UI designers can expand their value, because "I make pretty screens" is not a great long-term career plan. Broaden into UX and problem solving. Look past the interface and into the business problem, user needs, and research. Own the stuff between screens. AI still tends to think screen by screen. Humans are better at flows, journeys, and the messy reality of how people actually get from A to B. Lean into information architecture. For websites especially, the structure and content model matter as much as the visual design. We used a music analogy that will probably annoy some people, which makes it perfect. AI tools can generate "background" output that is fine for low-stakes use. They will not replace great musicians. But they will reduce the number of gigs available. AI Skills As a Career Asset After we finished terrifying UI designers, we moved on to something more useful. I think a lot of roles are going to need an AI toolkit. Not a handful of clever prompts, but a proper library of reusable skills. When I say "AI skills," I mean documented processes that an AI can follow reliably. Think SOPs you can run repeatedly, not prompt snippets you copy and paste. I now have around 60 skills in my library, and it is growing constantly. Outside of the Boagworld website, it might be the most valuable business asset I have. The reason is consistency and context management. AI can produce terrible output when you dump too much information on it at once. Skills let you break work into focused chunks and chain them. We talked about 3 levels of skills: Company-level skills Standard processes that keep things consistent. Proposals. Expense claims. Holiday booking. The sort of stuff that should not depend on one person remembering every step. Team or discipline skills For example, UX teams can create skills for personas, journey mapping, surveys, and top task analysis. That helps remove bottlenecks and lets colleagues do decent work without reinventing the wheel. Individual skills This is where it gets interesting for your career. These are the skills that capture how you do something, including all the weird little bits you have learned over the years. A key point here is that the value is not only in having the skill. It is in creating it. Writing down a process forces you to surface assumptions and explain what "good" looks like. We also got into AI agents. If you describe your skills well, an agent can chain them to complete bigger jobs. I gave a sales example where a meeting transcript can be turned into a CRM entry, follow-up tasks, company research, and a draft proposal with very little manual effort. That is exciting. It is also mildly terrifying if you are attached to the idea of being indispensable. For more on AI Skills read: Your AI Toolkit Is Your Competitive Edge Read of the Month I mentioned an article that helped me connect a few threads in my own work. UX, conversion rate optimization, and design leadership can look like 3 different things until you realize they all operate on the same system. The piece is called "How CRO and UX Work Together to Increase Website Conversion". It frames CRO and UX as two sides of the same coin. CRO asks, "Did they convert?" UX asks, "Was it easy and enjoyable?" I would add that UX also cares about what happens after conversion, because retention is often where the real money is. The shared foundation is data. Analytics, event tracking, heat maps, session recordings. The same signals can tell you where people struggle and where the biggest conversion wins are likely to be. It also reinforced something I believe strongly. CRO and UX should not sit in separate silos. Both work best when they cover the entire journey, not just one page at a time. Marcus’ Joke "I just purchased an original Van Gogh coffee table. I know it’s original because there’s a bit of veneer missing." Find The Latest Show Notes

    53 min
  3. Website Rebuilds, AI Tools, and UX in 2026

    MAR 17

    Website Rebuilds, AI Tools, and UX in 2026

    This month, Paul and Marcus get into a tool that has made Paul cancel his Figma subscription, walk through how Paul has completely changed the way he approaches website rebuilds thanks to AI, and round things off with the latest thinking from Nielsen Norman Group on where UX is heading in 2026. App of the Week: figr.design Paul has been road-testing AI design tools as part of a workshop he ran on AI and UI, and after going through dozens of them, one stood out: figr.design. What makes it work where others fall short? A few things. It lets you feed in a significant amount of context upfront, things like style guides, design systems, and personas, which means the output is far more tailored than the generic average you often get from AI design tools. Iteration is also genuinely fast. You can queue up a whole list of changes and it processes them all in one go, rather than making you wait between each tweak. The prototypes it produces are more realistic than what you would typically get out of Figma. Text fields you can actually type in, accordion states that open and close, button states, fully responsive layouts. Not exactly revolutionary in theory, but refreshingly functional in practice. Export to Figma is available when you need it. The main limitation is that you cannot manually adjust elements yourself. Everything goes through the conversational interface. Paul has also been looking at a tool called Inspector, which runs locally and connects to the Claude API so you pay as you go rather than a flat monthly token allocation. It has been a bit fiddly to set up but worth keeping an eye on. For anyone regularly using Figma for wireframing and prototyping, it is worth giving figr.design a proper look. The shift Paul describes, from hunching over Figma to leaning back and having a conversation with the tool, is a fairly good summary of where this kind of work is heading. Rebuilding a Website in 2026 Paul has fundamentally changed how he approaches website rebuilds, and the shift is largely down to AI making a genuinely hard problem, getting good content onto a website, a lot easier. The old problem Website rebuilds have traditionally meant migrating existing content into a new design. Which sounds fine until you remember that most of that content was written by subject matter experts who know their field but have never thought about writing for the web. The result is pages that lecture rather than help, that bury the things users actually want to know, and that rarely arrive on time, because the content phase is almost always where projects stall. Why things are different now AI has changed three things meaningfully. First, generating content is no longer the enormous manual effort it used to be. Second, doing the research that informs good content, finding out what users actually ask, worry about, and need, is much simpler with tools like Perplexity. Third, AI-powered search engines are pushing toward a more question-oriented approach to content anyway, which makes getting this right more important than it used to be. How Paul works now Here is the process Paul walks through for a rebuild project. 1. Online research Using Perplexity, Paul researches the audience. For a well-known client, he'll ask specifically about them. For a smaller or niche client, he looks at the sector. He is looking for the questions people are asking, the tasks they are trying to complete, their objections, goals, and pain points. This takes about 10 minutes. 2. Personas The research output goes into AI, which identifies patterns and segments it into a set of personas. A couple of hours of back and forth to get these right. 3. Company overview Paul records his kickoff meeting with the client and points AI at the transcript. Out comes a clean summary of what the company does, its products and services, and how it talks about itself. An hour for the meeting, plus 10 minutes for the summary creation. 4. Top task analysis and information architecture If time and budget allow, Paul runs a formal top task analysis, collecting and prioritizing the questions users most want answered. For card sorting, he uses UX Metrics. If there is no time for that, AI brainstorms the top tasks from the personas and company overview. Either way, those tasks get fed into an AI-generated information architecture. 5. Building out the IA Paul builds the IA in the CMS or in Notion, assigning the relevant tasks and questions to each page. Stakeholders can see the structure and understand what each page is there to do before a word of copy is written. 6. Getting stakeholders to contribute Rather than asking stakeholders to write content (a recipe for delays), Paul asks them to do two simpler things for each page: bullet-point answers to the questions assigned to that page, and any other talking points they want included. Bullets only. No pressure to write. 7. Writing the content with AI This is where it all comes together. Paul sets up an AI project with four inputs: A web copywriting best practice guide covering readability, structure, and scanning A company-specific style guide built from existing brand materials The audience personas The company overview For each page, he drops in the questions and stakeholder bullet points, and the AI drafts the content using all of that context. Paul recommends Claude for writing tasks. The result is copy that actually reflects the company's voice and addresses what users need, rather than generic filler. 8. Review and refinement Stakeholders review the draft and leave comments, ideally directly in Notion where AI can read the page, take in the comments, and rewrite accordingly. One more pass by stakeholders and it is ready to go. Paul has been using this approach on half a dozen projects and reckons you can work through a full site's worth of content in about a week (depending on size) once the setup is done. For clients, it is a service worth paying for because it takes the content burden off them while producing noticeably better results than migrating whatever was already there. One thing Paul is careful to flag: this does not mean starting from absolute scratch every time. Old articles, compliance pages, event databases, templated content that just has to be there, all of that can still come across. The point is to treat migration as the exception rather than the default. Read of the Week: State of UX 2026 The Nielsen Norman Group article Design Deeper to Differentiate confirmed, in Marcus's words, most of what Paul has been saying for the past year. Paul took this as further evidence he is always right! A few of the key points from the article: UX has stabilized after the 2023-24 downturn, but teams are leaner. UX practitioners are now expected to cover more ground and demonstrate business impact rather than just shipping deliverables. AI fatigue has set in, both among designers tired of the "you're being replaced" narrative, and among users who have grown skeptical of AI features that add sparkle without actually improving anything. The article argues that trust is now the central design problem for AI-powered products, covering transparency, control, consistency, and what happens when things go wrong. UI quality is becoming commoditized. If your value is primarily in making interfaces look good and work correctly, the ceiling on that work is dropping. Real differentiation lives in service design, content strategy, complete user flows, and the connective tissue that links everything together over time. The hard-to-automate skills, taste, contextual understanding, critical thinking, and judgment, are where humans still add the most value. To thrive, the article suggests UX practitioners need to position themselves as strategic problem-solvers with a broad toolkit rather than deliverable-focused specialists doing what it calls "design theater." Paul agreed with all of it. Marcus mostly agreed too, while noting that it must be genuinely difficult to be a UX specialist inside a large organization right now, particularly in teams that have cut so far back that one person is expected to cover the entire discipline. The answer, in Marcus's entirely unbiased view, is to hire Headscape! Marcus' Joke I stole a neck brace from the hospital. I feel kind of bad, but at least I can hold my head up high. Find The Latest Show Notes

    1 hr
  4. From Agency Work to Product Success

    FEB 17

    From Agency Work to Product Success

    This episode we're joined by Stu Green, a product designer, agency founder, and serial app builder who's sold not one but two successful SaaS products. We dig into the realities of building your own product versus running an agency, the role AI plays in modern product development, and whether the flood of AI-built apps is a threat or an opportunity for professionals. Plus, we check out Bleet, an app that turns your meeting transcripts into social media content, and Paul shares how AI-powered personas are changing the way he approaches user research. App of the Week: BleetYou know you should be posting on LinkedIn. You've told yourself that every week for the past 6 months. But then you sit down, stare at the blank post box, and realize you have absolutely no idea what to write about. So you close the tab and promise yourself you'll do it tomorrow. You won't. Bleet is an app built by Stu Green (and collaborator Nick) that solves this by mining the conversations you're already having. It takes your meeting recordings and transcripts, extracts the key topics using AI, and helps you turn them into social media posts. And the thing that sets it apart from just asking ChatGPT to write something for you is that it pulls your actual words and phrases from the conversation, piecing them together into posts that genuinely sound like you rather than generic AI slop. How It WorksYou connect your meeting recordings or transcripts (or even just speak a thought into the app), and Bleet will surface a list of topics you covered. From there, you pick the ones you want to post about and hit "create." You can dial in how much creative liberty the AI takes, from near-verbatim to lightly polished. So you sit down for 10 minutes once a week, pick a handful of topics, schedule them up, and you're done. A single meeting can generate enough content for almost a week of daily posts. What About Client Confidentiality?The number one concern people raise is about sharing sensitive client information. Bleet strips out client names, specific people, and identifiable details. It focuses on the general topic and the ideas discussed, not the specifics of who said what in which meeting. And of course, you review everything before it goes anywhere, so if something feels too close to the bone, you just skip it or edit it. Topic of the Week: Building Products vs. Running AgenciesStu Green has lived both lives. He's run agencies, built products from scratch, and sold 2 SaaS businesses. So what's the difference between building for clients and building for yourself? Quite a lot, as it turns out. Start by Solving Your Own ProblemBoth of Stu's successful apps, a project management tool and HourStack (a time management app), started the same way: he needed something that didn't exist. The project management tool grew out of running his own consultancy. HourStack came from juggling small children and fragmented work hours, and wanting a way to visualize and stack little blocks of productive time. If you're genuinely your own best customer, there's a good chance others like you exist. And if even 2 or 5 or 10 of them show up, you've got the start of something real. The Myth of "I One-Shotted This"AI has made it dramatically easier to build apps, but Stu is refreshingly honest about the gap between a demo and a product. Sure, he cloned entire apps in a single prompt and it looked great. But behind that impressive facade? Hours of iteration, hosting setup, video infrastructure, S3 servers, and a stack of decisions that require real product-building experience. The people posting "I built this in one shot" on X are technically telling the truth, but they're showing you the Hollywood set, not the house behind the door. Getting from prototype to something you can actually charge money for still takes professional knowledge. You need to know what questions to ask, which answers are good, and when you're being led down a rabbit hole. Two Tiers of AI ToolsPaul and Stu landed on a useful mental model: there are essentially 2 categories of AI building tools. Tools for everyone: Platforms like Lovable or Figma Make that let anyone create a basic app or prototype. Great for personal use, proof of concepts, and quick experiments.Tools for professionals: Things like Cursor and Claude Code that enhance a developer's ability to build production-quality software faster and better, but still require real expertise to use well.Think of it like desktop publishing in the '90s. When it arrived, everyone panicked that graphic designers were finished. Instead, regular people made terrible flyers with Comic Sans, and the professionals used the same tools to produce better work, faster. AI-built apps are following the same pattern. The 3-Stage Development ModelPaul offered a framework for thinking about where AI fits in the build process: Prototype and proof of concept: Anyone can do this with AI tools. Great for validating ideas quickly and cheaply.The production build: This still needs a professional. Security, scalability, accessibility, solid architecture: these are non-negotiable if people are paying to use your product.Post-launch iteration: Once a professional has laid a strong foundation, less technical people can step back in and make tweaks and improvements with AI assistance, because they're working within a well-built structure.A Revenue-Sharing Model Worth ConsideringStu floated an interesting agency model: instead of charging a client the full upfront cost to build their app, what if you took partial ownership? The client pays a smaller retainer and upfront fee, you build and host the product, and you share in the revenue. If the app takes off, everyone wins. If it doesn't, your exposure is limited. The key is picking partners carefully. They need to bring the marketing and audience side of the equation, because your job is the infrastructure and development. It's a model that silverorange, a Canadian agency, used successfully with e-commerce clients years ago, and it still holds up. When to SellStu sold both his apps when they hit what he calls "the plateau," that point where growth flattens and your churn rate starts catching up with new customer acquisition. At that stage, you either invest heavily to push through (hiring, scaling infrastructure, customer success teams) or you sell to someone who wants a product with proven recurring revenue. For Stu, as a creative who'd rather build new things than manage database consultants and customer support, selling was the obvious choice. He used brokers both times, people who handle the paperwork, the letter of intent, and protect both sides of the deal. They take a cut, but they also sent chocolates, so it all evens out. Finding the Right IdeasWith everyone building apps now, how do you pick the ones worth pursuing? Stu's answer is to not go it alone. Find partners who are excited enough about the idea to invest their time and audience. If you pitch an idea and nobody wants in, that's useful information. If someone does, you've got both validation and a distribution channel on day one. He tested this with an AI running coach concept, reaching out to local running coaches in Jacksonville. When they responded with polite indifference, he moved on rather than sinking months into a product nobody was asking for. Read of the Week: AI-Powered PersonasPaul shared his latest obsession: using AI to breathe new life into user personas. He's written 2 articles for Smashing Magazine that walk through the process: Functional Personas With AI: A Lean, Practical Workflow: How to build genuinely useful personas that focus on what people are trying to do, not just demographic data.AI In UX: Achieve More With Less: Broader lessons from using AI across user research, design, development, and content creation.The approach: take all your research (surveys, interviews, call logs, analytics) plus deep online research from tools like Perplexity, feed it into AI, and generate highly detailed personas, far more detailed than the traditional single-page variety. Then load those personas into a project in ChatGPT, Claude, or Gemini, with instructions to answer questions from the persona's perspective. The result is something you can consult in every meeting, on every decision. A product team can upload photos of next season's lineup and ask "what would our audience think?" A web team can test wireframes against the personas. Real user research still matters, of course, but this approach makes research-informed thinking available at a frequency and scale that traditional methods never could. Marcus's Joke"I tried to steal spaghetti from the shop, but the female guard saw me and I couldn't get pasta." Courtesy of comedian Masai Graham. And yes, it's exactly as bad as you think. Find The Latest Show Notes

    1 hr
  5. The UX Reckoning: What 2026 Holds for Our Industry

    JAN 13

    The UX Reckoning: What 2026 Holds for Our Industry

    In this episode, we kick off 2026 with a candid look at where the UX industry stands and where it's heading. We dig into a thought-provoking article from Nielsen Norman Group, share our hopes (and fears) for the year ahead, and explore a fantastic design pattern catalog focused on building user trust. Plus, we discuss why generalists might just be the unicorns the industry needs right now. Topic of the Week: Preparing for 2026 and the UX ReckoningWe spent a good chunk of this episode discussing an article from the Nielsen Norman Group that, while technically published in early 2025, remains just as relevant today. Written by Kate Morin, Sarah Gibbons, and others at NNGroup, it tackles the challenges facing our industry head-on. UX Is Back on the Chopping BlockLet's not sugarcoat it. It's been a tough time for UX professionals. Layoffs have hit hard, particularly in the US, and there's a palpable sense of doom and gloom floating around LinkedIn and other professional spaces. We've seen this before, though. We set up Headscape right in the middle of the dot-com bust, after being laid off ourselves. It wasn't fun, but times like these have a way of separating the wheat from the chaff. Economic downturns tend to clear out people who jumped into UX because they saw easy opportunities, leaving behind those with genuine understanding and passion for the work. And despite all the negativity online, the World Economic Forum actually ranked UX design as one of the 8th fastest-growing industries. So the discipline itself isn't dying. There's just been a mismatch between the number of people entering the field and the reality of what the market can absorb. The Rebranding Debate Is a Red HerringSome people are suggesting we rebrand UX to "product design" or "experience design" to solve our problems. We don't think that's the answer. The word "design" does carry some baggage. In many business minds, it's seen as a luxury rather than a business-critical function. So when budgets get tight, "design" gets cut while "conversion optimization" and "customer retention" survive. That's a perception problem, not a naming problem. The real issue is that there are too many low-quality UX practitioners who've been churned out through bootcamps. They've been taught a process to follow, and they follow it come what may. That's not their fault; they were taught that way. But six months of bootcamp doesn't prepare you for the messy, contextual reality of actual UX work. The AI ReckoningThe negativity around AI on LinkedIn has been phenomenal lately. There's anger about "AI slop" and a general feeling that it's no good for anything. Paul posted about using AI to help create personas and do online research, and got absolutely slated for it. AI is just a tool. Like any tool, if you use it badly, you get bad results. If you use it well, it can be genuinely helpful. The good news is that we're finally moving past the "AI for AI's sake" phase. We're starting to see thoughtful integration of AI into products and services, AI that actually solves real user needs. Every technology goes through the same cycle. Remember video recorders? First, we were just amazed the technology worked at all. Big analog buttons, you started recording and stopped recording, and that was it. Then manufacturers added more and more features until the things became unusable with their tiny buttons and complicated preset systems. Then someone invented a code you could enter from the Radio Times to set recording times automatically. And finally, Sky came along with "press a button and it records." AI is going through that exact same evolution right now. Shallow UX Is Suffering (and That's Okay)Templates, processes, production-line UX: that stuff is really struggling, and it will continue to struggle. AI can do that now. You're not going to make money or build a career by blindly following the double diamond and churning out deliverables. What you need going forward are distinctly human skills: critical thinking, taste, knowing whether something is heading in the right direction, and navigating messy organizational dynamics. Those are the skills that matter. Soft skills like relationship building, facilitation, and empathy are going to be far more valuable than whether you can use Figma. Stop Worshipping Templates and ProcessesUX is messy. You can't box it up the same way on every project. Templates and checklists are great starting points, but they're not a substitute for thinking. Context is everything. There's no such thing as best practice. When someone from Google or Facebook says you need a 6-week discovery phase with facilitated usability testing of at least 6 people, and sure, that probably worked great for their situation, with their team, their product, and their stakeholders. But it doesn't mean it's right for your startup or your client with a third of the budget and massive internal politics. If you've been taught a linear process, shift your mindset. Don't have a process. Have a toolkit of techniques you use as and when appropriate. You don't always need a discovery phase; sometimes a quick phone call is enough. You don't always need journey mapping; sometimes that's just not appropriate. Don't Lose the Human ConnectionBe careful that all these AI-powered conveniences don't cost you your connection with actual users. It's tempting to just run surveys, do unfacilitated remote testing, or let AI do online research. But you don't build real empathy that way. When you sit down to write copy or design an interface, you want to be able to picture the person in your head. You want to feel who they are, what they'd say, what they'd struggle with. The more levels of abstraction between you and your users, the harder that becomes. Even if it's just talking to one or two customers, make sure you're seeing them as real people. Become a Unicorn (a.k.a. Generalist)For years, we've been told to specialize. But now? We need to become more comfortable wearing multiple hats. You might be doing wireframing, user research, strategy work, and training, all in the same week. You might need to understand adjacent fields like marketing, business strategy, data modeling, or product management. AI can help extend your capabilities. Maybe you know a bit about accessibility or SEO, but not enough to do a full audit. With AI's help, you can now be better in those areas. Still not as good as a specialist, but better than you would have been alone. Stop focusing so much on outputs (wireframes, reports) and start focusing on outcomes. Elevate your thinking from tactical to strategic. If you want to dig deeper into this, check out Paul's free email course. It's 30+ emails on thinking more strategically and holistically about UX. Read of the Week: Design Patterns Catalog by Projects by IfWe stumbled across a brilliant resource from an agency called IF. They've created a design patterns catalog with a particular emphasis on building trust through transparency, user control, and thoughtful approaches to consent and data sharing. This is increasingly important, especially as AI becomes more prevalent. It's not about slapping a testimonial on a page and calling it done. It's about baking trust into the experience itself. The catalog is beautifully illustrated and well-explained, making it a great scannable reference. Paul found this while working on Bleet, a tool that automatically extracts advice from your recorded meetings and turns it into social media content. The trust challenge there is obvious: you're uploading client meetings with confidential information, so finding patterns for building that trust was essential. Marcus's Joke"I dropped a tub of margarine on my foot two weeks ago. I can't believe it's not better." Find The Latest Show Notes

    52 min
  6. Surviving Crisis: Lessons from Higher Ed's Financial Storm

    12/23/2025

    Surviving Crisis: Lessons from Higher Ed's Financial Storm

    In this episode, we welcome back Andrew Millar from the University of Dundee to discuss the current state of higher education, vibe coding platforms for non-developers, and the importance of community-driven conferences like Scottish Web Folk. App of the Week: Bolt.newThis week we're looking at Bolt.new, a vibe coding platform designed specifically for non-developers. Unlike tools like Cursor that are built for developers to pair program with AI, Bolt is aimed at people like marketers, designers, and small business owners who want to create functional applications without ever touching code. Paul has been using Bolt to build practical tools for his own business, including a custom top task analysis app, WordPress plugins, JavaScript extensions, and CSS animations. The platform handles everything from the database to publishing and hosting, making it genuinely accessible for non-technical users. However, we'd caution against treating these tools as production-ready for enterprise use. They're excellent for prototyping, internal tools, and small-scale applications, but they likely won't pass rigorous quality control in larger organizations. Think of them like desktop publishing was in the early days. They democratize creation but don't eliminate the need for professional expertise. For production-ready code, the real value comes when developers use AI pair programming tools where they can review, understand, and quality-check the output. The future likely involves professionals using these tools to increase productivity rather than replacing expertise entirely. Topic of the Week: The State of Higher Education and Digital TransformationAndrew Millar, who runs the digital team at Dundee University, joins us to paint an honest picture of the current higher education landscape. It's not pretty, but his candid insights offer valuable lessons for anyone navigating organizational crisis, whether in universities or elsewhere. The Perfect Storm Facing UniversitiesHigher education has always claimed poverty, but the structural problems have become impossible to ignore. Universities face two fundamental financial challenges: funding per student hasn't kept pace with inflation over the past decade, and research grants typically only cover around 80% of actual costs, leaving institutions to make up the difference. International students became the solution to plug this gap. They could be charged higher fees and effectively cross-subsidized teaching for domestic students and research activities. This worked until a perfect storm hit: COVID disruptions, international conflicts, hostile government rhetoric toward international students, and for Dundee specifically, the Nigerian economy's collapse, which dramatically reduced one of their key international markets. Dundee found themselves with a 30 million pound deficit. Within a year, the principal resigned, the entire executive changed, the Scottish government stepped in with emergency funding, and 500 staff members have left from a workforce of around 3,000. The Three Phases of Crisis ManagementAndrew outlined three distinct phases organizations go through during financial crisis, and his framework offers practical guidance for anyone facing similar situations. Phase 1: Cut, Cut, Cut When crisis hits, budgets get slashed, often multiple times. Andrew recommends categorizing everything into three buckets: what's absolutely critical to keep the lights on, what will hurt but won't cause lasting harm, and what's easy to eliminate. This is actually an opportunity to clear out legacy systems and processes that nobody uses but somehow persist. The challenge is that during this phase, people aren't open to change or new ways of working. They just want to see the existing stuff cut. Don't waste energy trying to introduce innovations here. Focus on strategic pruning. Phase 2: The Great Spaghetti Flying Contest This is where everyone becomes an expert on how to solve the crisis. Phrases like "we should at least try it" and "isn't it good to test ideas?" fly around constantly. The problem is that these are the exact phrases digital teams have been using for years to encourage experimentation, now thrown back at them by people with competing priorities. Governance structures become critical here. You can clarify requests (ensuring they're truly worth pursuing), compromise on scope, or clog them up in committees until priorities become clearer. When your escalation paths have collapsed, as they did at Dundee when leadership departed, you're left justifying decisions without backup. The key insight: never say "computer says no" via email. Have conversations. Explain your reasoning. When people understand the constraints, they typically accept them. Email refusals just get escalated to whoever shouts loudest. Phase 3: The Big Squeeze With less money, fewer people, less institutional knowledge, and no clear strategy, this phase is when things get really difficult. But paradoxically, it's also when people become more open to change. They've accepted that old ways aren't working and are more receptive to credible, evidence-based proposals for doing things differently. Digital Team Transformation and the Hub-and-Spoke ModelAndrew's team has evolved significantly since their original digital transformation work. They reduced the number of people managing the corporate website from 350 to about 20 while maintaining quality. Now they're moving toward a hub-and-spoke model, with centralized governance but distributed execution. The ideal version of this model, which IBM pioneered, has people embedded in individual departments but reporting into the central digital function. This creates healthy tension, since they need to keep their central manager happy while also serving their local colleagues. It maintains standards while building subject matter expertise across the organization. One emerging priority is what Andrew calls "generative engine optimization," ensuring content is structured so AI tools can accurately surface and represent it. As more users get information through AI intermediaries without ever visiting your website, getting this right becomes critical. The Value of Community: Scottish Web FolkThe conference that inspired this episode, Scottish Web Folk, emerged partly out of necessity. When travel budgets got cut, Dundee created their own event. It's now grown to over 150 attendees with strong sponsor support, all while maintaining its community-first ethos. The conference bans sales pitches from sponsors, limiting them to 30 seconds of self-promotion. Instead, it emphasizes knowledge sharing between suppliers and institutions. This approach keeps sponsors coming back because they recognize that embedding themselves in the community pays long-term dividends. For any digital team, hosting events like this builds internal credibility and external relationships simultaneously. It positions you as thought leaders within your organization while creating the networks that sustain careers and enable collaboration across institutional boundaries. Marcus's Joke"I started dating a zookeeper, but it turned out she was a cheetah." That's a wrap for this episode. See you in the new year! Find The Latest Show Notes

    1h 4m
  7. E-commerce UX Secrets: What 200,000 Hours of Research Reveals About Conversion

    11/18/2025

    E-commerce UX Secrets: What 200,000 Hours of Research Reveals About Conversion

    If you run an e-commerce site or work on digital products, this conversation is packed with research-backed insights that could transform your conversion rates. Apps of the WeekBefore we get into our main discussion, we want to highlight a couple of tools that caught our attention recently. UX-Ray 2.0We talked about this last week, but it deserves another mention. UX-Ray from Baymard Institute is an extraordinary tool built on 150,000 hours (soon to be 200,000 hours) of e-commerce research. You can scan your site or a competitor's URL, and it analyzes it against Baymard's research database, providing specific recommendations for improvement. What makes UX-Ray remarkable is its accuracy. Baymard spent almost $100,000 just setting up a test structure with manually conducted UX audits of 50 different e-commerce sites across nearly 500 UX parameters. They then compared these line by line to how UX-Ray performed, achieving a 95% accuracy rate when compared to human experts. That accuracy is crucial because if a third of your recommendations are actually harmful to conversions, you end up wasting more time weeding those out than you saved. Currently, UX-Ray assesses 40 different UX characteristics. They could assess 80 parameters if they dropped the accuracy to 70%, but they chose quality over quantity. Each recommendation links back to detailed guides explaining the research behind the suggestion. For anyone working in e-commerce, particularly if you're trying to compete with larger players, this tool is worth exploring. There's also a free Baymard Figma plugin that lets you annotate your designs with research-backed insights, which is brilliant for justifying design decisions to stakeholders. SnapWe also came across Snap this week, which offers AI-driven nonfacilitated testing. The tool claims to use AI personas that go around your site completing tasks and speaking out loud, mimicking user behavior. These kinds of tools do our heads in a bit. On one hand, we're incredibly nervous about them because they could just be making things up. There's also the concern that they remove us from interacting with real users, and you don't build empathy with an AI persona the way you do with real people. But on the other hand, the pragmatic part of us recognizes that many organizations never get to do testing because management always says there's no time or money. Tools like this might enable people who would otherwise never test at all. At the end of the day, it comes down to accuracy and methodology. Before using any such tool, you should ask them to document their accuracy rate and show you that documentation. That will tell you how much salt to take their output with. E-commerce UX Best Practices with Christian HolstOur main conversation this month is with Christian Holst, Research Director and Co-Founder of Baymard Institute. We've been following Baymard's work for years, and having Christian on the show gave us a chance to dig into what nearly 200,000 hours of e-commerce research has taught them about conversion optimization. The Birth of Baymard InstituteChristian shared the story of how Baymard started about 15 years ago. His co-founder Jamie was working as a lead front-end developer at a medium-sized agency, and he noticed something frustrating about design decision meetings. When the agency prepared three different design variations, the decision often came down to who could argue most passionately (usually the designer who created that version), the boss getting impatient and just picking one, or the client simply choosing their favorite. Rarely did anyone say they had large-scale user experience data to prove which design would actually work better. They realized they could solve this problem by testing general user behavior across sites and looking for patterns that transcend individual websites. If they threw out the site-specific data and only looked for patterns across sites, they could uncover what are general user behaviors for specific UI components and patterns. It started with just checkout flows. It wasn't even clear they would ever move beyond that. But now, 15 years later, Baymard has a team of around 60 people, with 35 working full-time on conducting new research or maintaining existing research. The Role of Research-Backed GuidelinesOne important point Christian emphasized is that Baymard's research isn't meant to replace your own internal testing. You should always do your own data collection and usability testing. The point of having a large database of user behavior and test-based best practices is that when you're redesigning something, you have maybe 100 micro decisions to make. You can't run internal tests for every single one of those decisions. Even Fortune 500 companies that have the budget don't have the time to wait for results on every micro decision. So what happens is you collect research on the two or three big things that are site-specific or unique to your brand or customer demographic. But all the generic stuff (how to design an expand and collapse feature, how the quantity field should work, how the phone field should be designed in a checkout flow) these are extremely standardized UI components where users have standardized expectations. You shouldn't squander your internal test resources on testing things that are completely generic. That's where pre-made research comes in. It removes 97 of the micro decisions so you can focus your resources on what's unique and important to your brand. Common E-commerce Conversion KillersWe asked Christian what kills conversion the most on e-commerce sites. While it depends on each site's specific issues, there are some concrete things Baymard has consistently seen sites fail at that are surprisingly easy to fix. The Order Review TrapIn countries where you have an order review step (where users review the whole order before pressing "place order"), there's a really dangerous trap. The order review step and the order confirmation step look very similar in users' minds. Both are textual pages that appear after entering credit card data. Both show a summary of information. In testing, Baymard consistently sees some users misinterpret the order review step for a confirmation step. This is a critical error because these users will exit the page thinking they've completed their order. They don't even realize the abandonment occurred. It's the worst type of checkout abandonment that can happen. A very simple trick is to take the "place order" button that you usually have at the bottom of the page and duplicate it so there's also one at the top of the page. One audit client did this and got a $10 million return on investment from just duplicating that button. It won't affect 10% of users, but if it prevents one out of 200 users from abandoning, that's half a percent of all your site revenue you've recovered. Error Recovery ExperienceChristian called this "the least sexy but most important topic" in checkout flows. The general error recovery experience in checkout flows has improved over the 15 years Baymard has been researching, but it's still way too poor. When a validation error occurs, users struggle with three things: Understanding that an error actually happenedUnderstanding where the error isUnderstanding how to resolve itBest practices for error recovery: Provide visual styling for each field that's wrongHave a description at the top of the page that outlines all errorsUse conditional logic: if there's only one error, scroll them to that field. If there are multiple errors, scroll to the top where they can see the overviewBaymard sees users who fix one error, resubmit, and then get frustrated when the page reloads with another error they didn't see. They sometimes conclude the page is broken. When Baymard surveys users, 6% say they've abandoned a checkout flow in the past quarter due to perceived technical errors. Most of these aren't actual technical errors, the page is just extremely complicated to use. Adaptive Error MessagesInstead of saying "phone number is invalid," tell users exactly why. Your technical system knows exactly which validation rule was triggered. If the phone number is wrong because it includes a special character, tell them: "Special characters cannot be used. You don't need to include the country code." If it's too long or too short, say that specifically. This helps users recover faster. Ideally, much of this should be fixed in the backend. Postcodes are a great example, some people put a space in UK postcodes, some don't. Some write it all uppercase, some use mixed case. Why isn't this fixed in the backend? There should be something tidying it up and dropping it into the database in the correct format. Product Data and ImageryOne area where Baymard has seen genuine improvements is around product data and product imagery. Most sites took a long time to figure out that the content on the product details page is crucial to user experience. When users land on a product details page, 90-95% of what they do as a first action is look at the image. But they also use images for tasks where it's a terrible idea. For instance, if trying to figure out whether a speaker has the right connection, instead of going to the specification sheet, they look for images showing the speaker from the back to see the connections. If they can't see it, they conclude it doesn't have the connection and abandon that product. Users are extremely visually driven, even trying to use images to solve problems where it's a poor strategy. Sites need really good imagery from multiple angles, detailed videos showing what goes on visually, and proper product descriptions. Building Trust in E-commerceWe asked Christian about building trust beyond the lazy approach of just shoving social proof and awards on the site. His insights were revealing: Social Proof On and Off-Site Social proof is important both on your page and off it. If people are in doubt whether

    58 min
  8. Freelancing for Small Businesses: Real World Budget Constraints and High Stakes

    10/21/2025

    Freelancing for Small Businesses: Real World Budget Constraints and High Stakes

    Welcome to Episode 27 of the Boagworld Show, where we dive into a side of web work that doesn't get nearly enough attention. This month, we're exploring life as a freelancer working with small businesses. We're joined by Paul Edwards, a fellow member of the Agency Academy who has spent two decades serving clients that don't have massive budgets or sprawling marketing teams. If you've ever wondered how best practice advice translates to the real world of limited resources and high stakes, this conversation is for you. App of the Week: Baymard UX-RayBefore we get into our main conversation, we need to talk about an extraordinary tool that just launched. Baymard UX-Ray is built on the Baymard Institute's 150,000 hours of ecommerce research. If you're not familiar with Baymard, they've been conducting rigorous usability research for years, building an enormous repository of what actually works in ecommerce design. What makes UX-Ray remarkable is how it applies all that research. You can input your own site or a competitor's URL, and the tool scans it against Baymard's research database. It then provides specific recommendations for improvement, each one linked back to detailed guides explaining the research behind the suggestion. Now, we'll be honest. Tools like this can feel a bit depressing when you first encounter them. Another thing that AI can do that used to be our job, right? But the reality is more nuanced. You still need expertise to ask the right questions, to know when to ignore advice that doesn't fit your situation, and to implement recommendations effectively. What UX-Ray really does is democratize access to quality research, allowing smaller teams and solo practitioners to benefit from insights that would otherwise require a massive research budget. For anyone working in ecommerce, particularly if you're trying to compete with larger players, this tool is worth exploring. Life as a Freelancer Serving Small BusinessesOur main conversation this month centers on something we don't discuss enough in the UX and web design community. Most of the advice you read online, most of the case studies and best practice articles, come from people working with large organizations. We're guilty of this too. Between the two of us, we've worked with clients like Doctors Without Borders, GlaxoSmithKline, and major universities. That shapes our perspective in ways we don't always recognize. Paul Edwards brings a different lens. He's spent 20 years as a freelancer, and while he's worked with organizations of varying sizes, the common thread through his client list isn't scale. It's circumstance. His clients typically have small or nonexistent marketing teams. They're often time-poor and lack technical expertise. Most importantly, they have skin in the game in a way that corporate clients rarely do. The Origin StoryPaul's freelance journey started dramatically. On November 5, 2005, he had a tantrum at his job as a commercial manager for a civil engineering company and quit on the spot. No savings, no business plan, no real idea what he was doing. He just knew he'd been teaching himself web design with Dreamweaver and Fireworks, and he thought maybe he could make a go of it. What followed was the classic freelancer trajectory. He worked his friends and family network, which led him into academia and international development work. He found himself building sites for projects funded by the Bill and Melinda Gates Foundation, DFID, and the World Bank. These weren't necessarily well-funded projects despite the prestigious funders, but they gave him experience working with agencies across Europe and projects in Africa focused on critical issues like hygiene and sanitation. What Makes Small Business Work DifferentWhen you're working with a small business owner, the stakes are fundamentally different. As Paul put it, the number of clicks their campaign generates directly affects how much money they take home at the end of the month and the security of their family. That changes everything about the relationship. This isn't to say working with large organizations is easy or that the work doesn't matter. But in a corporation, success and failure are distributed across many people and many factors. When you're working with someone who owns their business, your work has an immediate, visible impact on their livelihood. The opportunity cost of failure is enormous. The credit for success is also more direct, which can be incredibly motivating. Paul's business has evolved toward more retainer and time bank arrangements over project work. This shift happened gradually but has been transformative. For clients, it guarantees access to his expertise when they need it. For Paul, it provides income stability. But there's another benefit that often gets overlooked. When you have long-term retainer clients, especially small ones with staff turnover, you become a point of continuity in their organization. One of Paul's retainer clients had a marketing department of two people. Both left within a year. Paul was literally the only person who understood the history of their digital presence, their past campaigns, and their strategic direction. That kind of institutional knowledge is incredibly valuable, and it's something freelancers can uniquely provide. The Budget RealityWe had to ask about budget because it's the elephant in every room. When you're working with smaller clients, you simply have fewer resources to work with. So how do you adapt all the best practice advice that assumes you have time for extensive user research, iterative testing, and comprehensive documentation? Paul's answer was illuminating. He doesn't find himself frustrated by advice that doesn't apply to his situation. He just doesn't apply it. As a generalist, he's always picked and chosen what's relevant, learning what he needs for each specific job and disregarding the rest. He can't let his head explode trying to take in everything, so he focuses ruthlessly on what matters for the work at hand. The reality is that best practice often needs to be adapted regardless of client size. A lot of what gets labeled as essential process work serves organizational needs as much as user needs. In a large organization, you might conduct extensive research partly to align compliance, get legal on board, and protect your client contact from political fallout. In a small business where you're talking directly to the decision maker, you can move faster and iterate post-launch. That doesn't mean cutting corners on things that matter. Paul still does discovery and research work, but he structures it differently. Rather than one large project with research baked in, he often does pre-project discovery as separate billable work. This allows him to flex the scope based on what the client has in-house, what they lack, and what will actually move the needle for them. Filtering Clients and Managing RiskOne of the most valuable parts of our conversation was Paul's approach to client selection. He's learned through hard experience that taking on a client who isn't a good fit costs far more in stress and lost time than the revenue is worth. Every single time he's taken someone on when his gut said no, it's been worse than if he hadn't brought that money in. So Paul has developed a risk scoring process. He researches Companies House filings and financial accounts. He Googles potential clients thoroughly. He makes sure to be himself from the first conversation, explaining that he's blunt and tends to say what he thinks. Some people say they want that but really don't, and it's better to discover the mismatch early. When things do go wrong, which is rare after 20 years, Paul offboards as quickly and graciously as possible. He sees it as partly his fault for misjudging the fit, so he tries not to burn bridges. He'll help them find someone else to work with and exits professionally. We wondered whether this kind of risk management is more necessary when working with smaller organizations. After all, you know Oxford University will eventually pay their bills, even if slowly. Paul's experience is that payment risk exists at all scales, but small businesses can have more volatile finances. However, most of his clients pay within 48 hours, which is remarkable. The key is that by moving toward retainer and time bank models where time is paid upfront, a lot of payment anxiety simply disappears. The Generalist Advantage and AI's RoleOur conversation kept circling back to the value of being a generalist, and how AI is amplifying that advantage. Paul described AI as helping him get out of his own way. If he knows 90 percent of what's needed to help a client but lacks that final 10 percent, he used to decline the work. The opportunity cost of getting it wrong felt too high. Now, AI helps him bridge that last 10 percent with confidence. He shared a perfect example. A trade business client, selling into the architectural sector, wanted help with their Google Ads campaign. Paul had dabbled in PPC but wasn't an expert. The client was willing to pay him to learn, which was fortunate, and AI supported that learning process. It helped him analyze the massive amounts of data that PPC campaigns generate, identify trends, and fill knowledge gaps. The result was a completely new campaign with much lower spend, a huge increase in relevant clicks, and better funnel positioning. The client was so pleased they sent him a Christmas hamper, a first in 20 years. This is what the return of the generalist looks like. AI isn't replacing expertise. It's allowing people with broad knowledge and good judgment to tackle problems that previously required specialists. You still need to know enough to ask good questions, to recognize when something feels off, and to verify AI's suggestions. But you can now say yes to opportunities that would have been too risky before. What Large Organizations Can LearnNear the end of our conversation, P

    1 hr
4.9
out of 5
96 Ratings

About

Boagworld: The podcast where digital best practices meets a terrible sense of humor! Join us for a relaxed chat about all things digital design. We dish out practical advice and industry insights, all wrapped up in friendly conversation. Whether you're looking to improve your user experience, boost your conversion or be a better design lead, we've got something for you. With over 400 episodes, we're like the cool grandads of web design podcasts – experienced, slightly inappropriate, but always entertaining. So grab a drink, get comfy, and join us for an entertaining journey through the life of a digital professional.

You Might Also Like