Diese Präsentation wurde erfolgreich gemeldet.
Die SlideShare-Präsentation wird heruntergeladen. ×

Myths, Lies and Illusions of AB and Split Testing

Anzeige
Anzeige
Anzeige
Anzeige
Anzeige
Anzeige
Anzeige
Anzeige
Anzeige
Anzeige
Anzeige
Anzeige

Hier ansehen

1 von 71 Anzeige

Myths, Lies and Illusions of AB and Split Testing

Herunterladen, um offline zu lesen

What are the common assumptions about AB (split) testing that are wrong? What are the lies told by vendors, consultants and the stuff you have convinced yourself about. What is illusory - what can you trust - what's it really all about. 20 top myths debunked after asking fellow CRO professionals what is on THEIR top list.

What are the common assumptions about AB (split) testing that are wrong? What are the lies told by vendors, consultants and the stuff you have convinced yourself about. What is illusory - what can you trust - what's it really all about. 20 top myths debunked after asking fellow CRO professionals what is on THEIR top list.

Anzeige
Anzeige

Weitere Verwandte Inhalte

Diashows für Sie (20)

Andere mochten auch (15)

Anzeige

Ähnlich wie Myths, Lies and Illusions of AB and Split Testing (20)

Anzeige

Aktuellste (20)

Myths, Lies and Illusions of AB and Split Testing

  1. 1. Myths, Lies and Illusions of Split Testing @OptimiseOrDie
  2. 2. @OptimiseOrDie • UX, Analytics, Split Testing and Growth Rate Optimisation • Started doing testing & CRO 2004 • Split tested over 45M visitors in 19 languages • 67+ mistakes I MADE with AB testing • Like riding a bike… • Optimise your optimisation? Get in touch!
  3. 3. Myths, Lies and Illusions of Optimisation 1. Optimisation and Testing are the same 2. Optimisation and Testing are easy 3. Statistical significance tells you when to stop 4. High traffic always makes tests shorter 5. It’s all about Conversion Rates 6. You don’t need to test the test 7. Your tests will give you the quoted lift 8. High traffic pages give the best tests 9. It’s about getting all your tests to win 10. Optimisation is only about websites @OptimiseOrDie 11. Segmentation will tell you what happened 12. You can’t run multiple simultaneous tests 13. Testing is great for settling arguments 14. You can spot trends early on in a test 15. More test volumes = better results 16. Your tests tell you truths that last forever 17. You can test even on low traffic sites 18. Other people’s tests are ‘Best Practice’ 19. Doesn’t involve changing the way you work 20. Testing makes you a Data Scientist
  4. 4. #fail @OptimiseOrDie
  5. 5. @OptimiseOrDie 26.6M
  6. 6. @OptimiseOrDie28.4M
  7. 7. Oppan Gangnam Style! @OptimiseOrDie6.9M
  8. 8. You naughty boy…
  9. 9. @OptimiseOrDie1. Testing & Optimisation are the same thing Testin g Optimise
  10. 10. What is CRO/Optimisation? • “Using Analytics data and Customer feedback to improve the performance of your website.” • “Finding out why visitors aren’t converting and then resolving these issues.” • “Running loads of crappy split tests randomly until the heat death of the universe” @OptimiseOrDie
  11. 11. What is my definition? “A structured, systematic and continuous application of techniques that are used to discover, quantify and prioritise issues. These can be turned into hypotheses to drive experiments and opportunity in the following business outcomes:” • Increased revenue or profitability • Increasing LTV, loyalty, NPS/Sat scores, Kristoffer Potential • Removing cost from the business or contact center • Higher productivity or labour flexibility • Delighting customers • Reduced development effort @OptimiseOrDie
  12. 12. What Optimisation is NOT! • A way to change things you (or others) hate • A methodology for running split tests • A guarantee of increased conversion • A methodology for looking at analytics data • A rescue for silo-bound or non-agile design or development processes • A way to trick people into buying • A bolt-on – it IS the process @OptimiseOrDie
  13. 13. Optimisation is a: • Way of joining the worlds of Customer insight, UX, Analytics, Split testing and Business Strategy • Overarching Design and development process which prioritises work around opportunity • Strategic, not tactical, response to wasted development effort or product change • Way to create a deep and meaningful connection between the team, customers, business and the outcomes of making product changes • The killer app to remove ego, opinion, assumptions, cherished notions or ‘we just do it that way’ from decision making. • More powerful method than UX research or analytics alone, in guiding the directionality of product change @OptimiseOrDie
  14. 14. Optimisation includes these: • Qualitative research • Analytics, Quant analysis and insight • UX inspection and discovery • Competitive Intelligence • Priority based opportunity • VOC, Surveys and Customer Satisfaction (NPS) • Call tracking & Call centre optimisation • AB testing • Multivariate testing • Photography optimisation • EEG / ECG / Galvanic response @OptimiseOrDie • Web performance tuning • Forms analytics • Eye tracking • Market research • Big & Unstructured data analysis • PPC optimisation • Session replay analysis • Customer journey mapping • Ethnographic (diary) study research • Cross device, platform and channel insight • Email optimisation
  15. 15. 2. Optimisation & Testing are EASY
  16. 16. Surely you just add Javascript? • This work is harder than anything I’ve ever done! • You all have expensive & limited resources for testing - like Airport take-off slots • You MUST make use of these efficiently • You HAVE to balance resource cost with opportunity by prioritising carefully @OptimiseOrDie • You’re doing the equivalent of drug trials! • For large OR small companies, the instrumentation, analytics, tools setup, plan design, test methodology and analysis cannot be ‘done later on’. It’s not complex but it is vital. • The best companies have a structural shell (process, methodology, management) around their activities • If you’re not optimising the process continuously (Kaizen), you won’t increase your velocity of iterations. • Optimise the Optimisation
  17. 17. 3. Statistical significance tells you when to stop @OptimiseOrDie
  18. 18. The 95% Stopping Problem @OptimiseOrDie • Many people use 95, 99% ‘confidence’ to stop • This value is unreliable and moves around • Nearly all my tests reach significance before they are actually ready • You can hit 95% early in a test (18 minutes!) • If you stop, it could be a false result • Read this Nature article : bit.ly/1dwk0if • Optimizely have changed their stats engine • This 95% thingy – must be LAST on your stop list • Let me explain
  19. 19. The 95% Stopping Problem @OptimiseOrDie Scenario 1 Scenario 2 Scenario 3 Scenario 4 After 200 observations Insignificant Insignificant Significant! Significant! After 500 observations Insignificant Significant! Insignificant Significant! End of experiment Insignificant Significant! Insignificant Significant! “You should know that stopping a test once it’s significant is deadly sin number 1 in A/B testing land. 77% of A/A tests (testing the same thing as A and B) will reach significance at a certain point.” Ton Wesseling, Online Dialogue
  20. 20. The 95% Stopping Problem @OptimiseOrDie “Statistical Significance does not equal Validity” http://bit.ly/1wMfmY2 “Why every Internet Marketer should be a Statistician” http://bit.ly/1wMfs1G “Understanding the Cycles in your site” http://mklnd.com/1pGSOUP
  21. 21. 4. High Traffic always means short tests @OptimiseOrDie
  22. 22. Business & Purchase Cycles @OptimiseOrDie • Customers change • Your traffic mix changes • Markets, competitors • Be aware of all the waves • Always test whole cycles • Don’t exclude slower buyers • When you stop, let test subjects still complete! Start Test Finish Avg Cycle
  23. 23. • TWO BUSINESS CYCLES minimum (week/mo) • 1 PURCHASE CYCLE minimum • 250 CONVERSIONS minimum per creative (e.g. checkouts) • 350 & MORE! if response is very similar • FULL WEEKS/CYCLES never part of one • KNOW what marketing, competitors and cycles are doing • RUN a test length calculator - bit.ly/XqCxuu • SET your test run time , RUN IT, STOP IT, ANALYSE IT • ONLY RUN LONGER if you need more data • DON’T RUN LONGER just because the test isn’t giving the result you want! @OptimiseOrDie How Long? Simple Rules to follow
  24. 24. 5. It’s ALL about Conversion Rate @OptimiseOrDie
  25. 25. It’s all about the business @OptimiseOrDie • You’re optimising a business here, not a page or site • Tricking, pushing or persuading people at a superficial level to take an action is not a viable strategy • Your optimisation strategy is a series of steps, not a tool • Testing is about learning, not converting. • Tests that fail to tell you anything (regardless of outcome) are a failure themselves • If you don’t shift the business goals, your optimisation and testing budget will be threatened
  26. 26. 6. You don’t need to test the test – just go @OptimiseOrDie Browser testing www.crossbrowsertesting.com www.browserstack.com www.spoon.net www.saucelabs.com www.multibrowserviewer.com Mobile devices www.appthwack.com www.deviceanywhere.com www.opendevicelab.com Read this article bit.ly/1wBccsJ
  27. 27. 7. The test result gives the promised lift @OptimiseOrDie
  28. 28. The result is a range @OptimiseOrDie • Version A is 3% conversion • Version B is 4% conversion • Yay! That’s a 25% lift • Let’s tell everyone • When it goes live, you get 5.7% • That’s because it was A RANGE • 3% +/- 0.5 • 4% +/- 0.4 • Actual result was 3.5% for A • Actual result was 3.7% for B
  29. 29. AB Testing Visualisation Tool @OptimiseOrDie abtestguide.com/calc/
  30. 30. 8. Testing is best on high traffic pages @OptimiseOrDie Think like the CEO of a department store! If you can’t refurbish the entire store, which floors or departments will you invest in optimising? Wherever there is: • Footfall • Low return • Opportunity
  31. 31. Ecommerce Bounce Search or Category Product Page Add to basket View basket Checkout Complete @OptimiseOrDie 8. Testing is best on high traffic pages
  32. 32. @OptimiseOrDie 8. It’s all about prioritisation!
  33. 33. 9. It’s all about WINNING test results @OptimiseOrDie
  34. 34. Failing is good @OptimiseOrDie • Tests that are ‘about the same’ are a failure • They’re also very hard to call • That means you have to be BOLD not conservative • A test that comes out negative is NOT a failure • If a ‘negative’ test teaches you something, it’s a success! • If you hit 40/50/60% failed tests, that’s fine • If you aren’t failing regularly, you’re not BOLD enough • Success is about the number of tests you finish each month, and what you learn from them
  35. 35. We believe that doing [A] for People [B] will make outcome [C] happen. We’ll know this when we observe data [D] and obtain feedback [E]. (reverse) @OptimiseOrDie
  36. 36. 10. Optimisation is only for Websites @OptimiseOrDie
  37. 37. Optimisation is just for websites @OptimiseOrDie • Service Design (Airbnb) • Onboarding flows with emails • Email templates • Apps – testing, debugging, tracking • Phone tracking and call centre optimisation • Social, Display, TV, Video and other advertising • Print adverts, Direct Mail • In-store promotions • Product manuals, guides, interfaces • EVERYTHING has elasticity – just find it • Even Multi-variate call centre scripts
  38. 38. 11. Segmentation will explain things @OptimiseOrDie
  39. 39. Segmentation explains stuff @OptimiseOrDie • Beware of small sample sizes • A = 350 conversions • B = 300 conversions • A Conversions for Safari = 20 • B Conversions for Safari = 25 • Only needs 2 people to change that from 25% lift to 14%
  40. 40. 12. You can’t run concurrent split tests @OptimiseOrDie
  41. 41. Oh yes you can, with GA! @OptimiseOrDie • If you push events or variables into GA, you can report on behaviour for A or B (or any variations). • If you do it that way, you can easily run multiple tests on different page targets simultaneously. • You grab AAA, AAB, ABA, ABB and analyse. • Test subjects get a recipe of tests, so one caveat • If you pick things that clash or jar the experience • If you have $5 delivery but then $2 in another test • Apart from that, it’s fine to run these • They tell you about the right experience recipe across several pages being optimised in concert
  42. 42. 13. Testing is great for settling arguments @OptimiseOrDie
  43. 43. 14. You can spot trends early in a test @OptimiseOrDie
  44. 44. You can spot trends early @OptimiseOrDie • Tests are volatile in the early stages • Watch but shut your mouth and wait • Keep an eye on any odd behaviour • Be patient!
  45. 45. 15. More tests = Better results @OptimiseOrDie
  46. 46. More tests = Better results @OptimiseOrDie • Increasing volume without optimising velocity • Not optimising the process means it scales badly • If you put garbage in, you’ll get garbage out • Ramping up before you have it tuned is crazy • Get the team, process, project management, methodology, toolkits, selling and PR nailed first • THEN optimise and scale • I know that one UK retailer is doing 140 per month! • That’s LOW compared to some companies I work with • Big companies doing 2 a month? Meh.
  47. 47. 16. Each test is a truth FOREVER @OptimiseOrDie
  48. 48. Tests are a Truth Forever @OptimiseOrDie • Traffic changes • Prices change • Product mix changes • Advertising evolves and changes • Markets are different • Customers have changed • Competitors or regulatory landscape moves • Things happen outside of your control • You need to revisit tests or ITERATE • Always be trying to beat the new winner, not basking in the glory of a test you ran 9 months ago • The lift may have vanished • Schrodinger’s AB test
  49. 49. 17. Testing is fine on Low Traffic sites @OptimiseOrDie
  50. 50. Testing is fine on low traffic sites @OptimiseOrDie • Yes, if you estimate your minimum testing unit • This is the time to AB test a sample of say 250 conversions, with the rules I set earlier. • A payday loan company? 2 months minimum! • Run the calculations • Check the test length • If it takes like, 8 million years, what can you do? • Read this or download my AB testing decks: bit.ly/1umy5Y6
  51. 51. 18. Other people’s tests are Best Practice @OptimiseOrDie “STOP copying your competitors They may not know what the f*** they are doing either” Peep Laja, ConversionXL
  52. 52. Tests you see online? @OptimiseOrDie • Your customers are not the same • Your site is not the same • Your advertising and traffic is not the same • Your UX is not the same • How the f*** do you think it guarantees a result? • Use them to inform or suggest ideas • They’re like the picture on meal packets • Serving Suggestion Only
  53. 53. 19. Testing makes you a Data Scientist @OptimiseOrDie
  54. 54. 20. Testing and Innovation is a plugin @OptimiseOrDie
  55. 55. WE’RE ALL WINGING IT
  56. 56. Guessaholics Anonymous
  57. 57. 2004 Headspace What I thought I knew in 2004 Reality
  58. 58. 2014 Headspace What I KNOW I know Me, on a good day
  59. 59. Rumsfeldian Space
  60. 60. @OptimiseOrDie #1 Culture & Team #2 Toolkit & Analytics investment #3 UX, CX, Service Design, Insight #4 Persuasive Copywriting #5 Experimentation (testing) tools The 5 Legged Optimisation Barstool
  61. 61. @OptimiseOrDie Marketing ? Variation Heritability of good ideas Selection based on death
  62. 62. Business Future Testing? Congratulations! Today you’re the lucky winner of our random awards programme. You get all these extra features for free, on us. Enjoy. Mr D. Vader
  63. 63. Thank You! Mail : sullivac@gmail.com Deck : Linkedin : linkd.in/pvrg14
  64. 64. READ STUFF
  65. 65. READ STUFF
  66. 66. READ STUFF
  67. 67. #5 : FIND STUFF @OptimiseOrDie @danbarker Analytics @fastbloke Analytics @timlb Analytics @jamesgurd Analytics @therustybear Analytics @carmenmardiros Analytics @davechaffey Analytics @priteshpatel9 Analytics @cutroni Analytics @avinash Analytics @Aschottmuller Analytics, CRO @cartmetrix Analytics, CRO @Kissmetrics CRO / UX @Unbounce CRO / UX @Morys CRO / Neuro @UXFeeds UX / Neuro @Psyblog Neuro @Gfiorelli1 SEO / Analytics @PeepLaja CRO @TheGrok CRO @UIE UX @LukeW UX / Forms @cjforms UX / Forms @axbom UX @iatv UX @Chudders Photo UX @JeffreyGroks Innovation @StephanieRieger Innovation @BrianSolis Innovation @DrEscotet Neuro @TheBrainLady Neuro @RogerDooley Neuro @Cugelman Neuro @Smashingmag Dev / UX @uxmag UX @Webtrends UX / CRO
  68. 68. #5 : LEARN STUFF @OptimiseOrDie Baymard.com Lukew.com Smashingmagazine.com ConversionXL.com Medium.com Whichtestwon.com Unbounce.com Measuringusability.com RogerDooley.com Kissmetrics.com Uxmatters.com Smartinsights.com Econsultancy.com Cutroni.com www.GetMentalNotes.com
  69. 69. #12 : The Best Companies… • Invest continually in analytics instrumentation, tools, people • Use an Agile, iterative, cross-silo, one team project culture • Prefer collaborative tools to having lots of meetings • Prioritise development based on numbers and insight • Practice real continuous product improvement, not SLEDD* • Are fixing bugs, cruft, bad stuff as well as optimising • Source photos and content that support persuasion and utility • Have cross channel, cross device design, testing and QA • Segment their data for valuable insights, every test or change • Continually reduce cycle (iteration) time in their process • Blend ‘long’ design, continuous improvement AND split tests • Make optimisation the engine of change, not the slave of ego * Single Large Expensive Doomed Developments
  70. 70. THE FUTURE OF TESTING : www.conductrics.com

Hinweis der Redaktion

  • This is the title of my talk today.
    Wonderful picture, isn’t it?

    It’s from an IBM advert from 1951 and is a great piece of work, especially the copywriting. The whole message here is “Buying an IBM computer gets you the same power as 150 extra engineers”. And not a feature in sight – the trick is they’re not selling the computer, they’re selling what the computer will do for your business and your life.

    And what am I talking about today? Well the fact that most split tests being run these days are just bullshit – the slide rules don’t add up for a lot of companies.
    Many C level execs I’ve spoken to complain about the variability of return or success on this kind of testing.
    There’s a reason for this -
  • And here’s a boring slide about me – and where I’ve been driving over 400M of additional revenue in the last few years. For the sharp eyed amongst you, you’ll see that Lean UX hasn’t been around since 2008. Many startups and teams were doing this stuff before it got a new name, even if the approach was slightly different. For the last 4 years, I’ve been optimising sites using a blend of these techniques.
  • And here are some of the clients I’ve been working for.
    Dull bit is now officially over.
  • And don’t worry – if it’s not working for you – and looks like this, it’s OK – you’re just doing it the wrong way.

    Although I admire AB testing companies - all of them - for championing the right to test and making it easy for anyone to implement - there's a problem. Democratisation of testing brings with it a large chunk of stupidity too.

    When YouTube first appeared, did anyone think "Oh boy, there's only ever going to be high quality content to see on here. Seriously. No”
  • And this crappy AB testing is basically the equivalent of funny cat videos
  • People taking videos of themselves playing video games
  • And like, wow, there are 6.9 million Gangnam Style videos. Just incredible.
    But hidden in those big numbers, YouTube will always have a tiny percentage of really great stuff, very little good stuff and a long tail of absolute bollocks.
    And the same is true of split testing - there's some really well run stuff, getting very good results and there's a lot of air guitar going on.
  • So – the first myth. Some people think that testing and optimisation are the same thing.

    AB testing is just one of the techniques that I’ll use to optimise a business and there are many more. Let’s talk about definitions.
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • It has taken me a long time to find out where all the bear traps are hidden. Mainly from screwing up tests and figuring out what was wrong, through lots of testing time.

    And most companies and teams are stepping on these bear traps without even realising. And they wonder why the test results aren’t replicated in the bank account results. Hah.

    I have a list now of about 60 ways to easily break, skew, bias or screw up your tests completely. But here are some real biggies to watch for:
  • I once explained to my daughter – you know, when adults like look really in control and making decisions and appearing not to suffer from indecision? Don’t believe it for a minute – we’re just better at winging it cause we’re older.

    And this is the huge hole that’s gnawing at the hear of many digital operations. The inability to understand what you can and can’t be confident about – but nobody wants to admit they’re guessing a lot of the time.
  • There is one answer to this trap I call taking a visit to Guessaholics Anonymous - to surrender to the higher power of testing and innovation by using consumer psychological insight and data to guide your hand. To recognise you’re powerless at deciding what’s best or second guessing what will win.

    It's actually liberating to not be sitting in a meeting room, arguing about the wording of a bloody button for 4 fucking hours, ever again.
  • And this was the state of my head in 2004. The inability to understand what you can and can’t be confident about – but nobody wants to admit they’re fucking guessing a lot of the time.

    And it took me a long time to figure out I didn’t know anything really – it was all assumptions and cherished notions. It was pretty crushing to test my way to this realisation but MUCH I’m happier now.
  • Now I think I know this much - but I might know a wee bit more than I think I do – but I’m erring on the side of caution.

    That’s because I'm always questioning everything I do through the lens of that consumer insight and testing.

    Without customers and data driven insights, you can’t shape revenue and delight. They’ll give you the very psychological insights you need to apply levers to influence them, if you only ask questions. Everything else is just a fucking guess.

    Even with tests, if the only inputs you’ve got are ego and opinion, they’re going to be lousy guesses and you’re wasting your experiments.
  • And now a bit about something I call Rumsfeldian Space – exploring the unknowns. This is vital if you want to make your testing bold enough to get great results.
  • You need to inhabit the contextual and emotional landscape of the consumer to really shape product or service experience. The only way to do this is have teams and cultures that create a direct and meaningful connection between teams and the customer, in the impact that every change has on the outcome.

    Every atom of every piece of copy, design, error message, email, website, support, help content, absolutely bloody everything you do - has to be framed within knowledge and empathy with the consumer fears, worries, barriers, pain but also the real problems we solve by designing products not as features but as life enhancing. And this is the best marketing of all, like the IBM ad.

    Business Model Optimisation requires a watchmakers eye – a complete understanding of the watch from macro to micro - the flow of delight and money that can be shaped inside every customer experience, website, and interaction - at a component and a service design level.

    Most people have 1 or 2 legs at most. The best companies I've worked with are doing all of these.
  • Darwin did NOT say 'survival of the fittest' – that was actually another guy called Herbert Spencer. What Darwin actually pushed was that the key ingredients were heritability of traits, variation and selection based on survival. If only your marketing programme was quite as ruthless eh?

    And if you want variation and innovation, the survival of good ideas in favour of bad and knowledge that you pass on – you need a culture of adaptability, improvement and change. Agile is about a shared mind-set across managers, leaders and everyone in the team.

    There’s a Harvard survey about how the *most* productive teams communicate. Not in meetings but all the time - deskside, IM, phone, skype, GitHub, agile tools, apps - these are the telegraph wires of the collaborative, participative and mission oriented teams.

    My key insight of the last 10 years in building and leading teams is that agile, open, flat, cross-silo, participative, flexible and collaborative environments produce customer connected products of high quality. Autoglass NPS higher than Apple.
  • I hope you enjoyed it as much as I did writing it. All my details are here and slides will be uploaded shortly.

    Thank you for your time today.
  • So – what’s driving this change then? Well there have been great books on selling and persuading people – all the way back to ‘Scientific Advertising’ in 1923.
    And my favourite here is the Cialdini work – simply because it’s a great help for people to find practical uses for these techniques.
    I’ve also included some analytics and testing books here – primarily because they help so MUCH in augmenting our customer insight, testing and measurement efforts.
    There are lots of books with really cool examples, great stories and absolutely no fucking useful information you can use on your website – if you’ve read some of these, you’ll know exactly what I mean. These are the tomes I got most practical use from and I’d recommend you buy the whole lot – worth every penny.
  • So – what’s driving this change then? Well there have been great books on selling and persuading people – all the way back to ‘Scientific Advertising’ in 1923.
    And my favourite here is the Cialdini work – simply because it’s a great help for people to find practical uses for these techniques.
    I’ve also included some analytics and testing books here – primarily because they help so MUCH in augmenting our customer insight, testing and measurement efforts.
    There are lots of books with really cool examples, great stories and absolutely no fucking useful information you can use on your website – if you’ve read some of these, you’ll know exactly what I mean. These are the tomes I got most practical use from and I’d recommend you buy the whole lot – worth every penny.
  • So – what’s driving this change then? Well there have been great books on selling and persuading people – all the way back to ‘Scientific Advertising’ in 1923.
    And my favourite here is the Cialdini work – simply because it’s a great help for people to find practical uses for these techniques.
    I’ve also included some analytics and testing books here – primarily because they help so MUCH in augmenting our customer insight, testing and measurement efforts.
    There are lots of books with really cool examples, great stories and absolutely no fucking useful information you can use on your website – if you’ve read some of these, you’ll know exactly what I mean. These are the tomes I got most practical use from and I’d recommend you buy the whole lot – worth every penny.
  • These are all people on twitter who cover hybrid stuff – where usability, psychology, analytics and persuasive writing collide. If you follow this lot, you’ll be much smarter within a month, guaranteed.
  • And here are the most useful resources I regularly use or share with people. They have the best and most practical advice – cool insights but with practical applications.
  • In my opinion, these are the attributes of companies doing great things with optimisation and continuous improvement.
  • This is the future of testing. A machine learning system that will test out variants and tell you what’s driving response to all your experiments.
    Know if that offer works because of someone’s age or past spending patterns – let the tool explain to you where the value is and let it exploit these patterns as an intelligent agent under your control.

×