Zach’s ugly mug (his face) Zach Leatherman

The Art of Deception, Lighthouse Score Edition

A white lighthouse sits next to a darkened ocean. Dark clouds in the background. A few smaller waves crash.
October 14, 2021

A few very interesting discussions on Twitter have led me to understand that some folks are talking about Lighthouse scores in a way that is—in my opinion—not as forthright as it could be (intentionally or not). Let’s level set a bit and talk a bit about the different flavors of wiggle room:

Super Fast Hardware

Here’s a screenshot of a Lighthouse result from this morning, October 14th, 2021, run on an old MacBook Air (2012) using Chrome 86. Lighthouse Score for Mobile on Old Hardware: 64 on Performance, 97 on Accessibility, 100 on Best Practices, 100 on SEO

Here’s the same result on my MacBook Air (M1, 2020) using Chrome 94: Lighthouse Score for Mobile: 94 on Performance, 89 on Accessibility, 93 on Best Practices, 100 on SEO

It’s incredible to me the variability effect that your hardware can have: from a 64 on Performance to a 94—that’s a thirty point swing!

Don’t Do This™ Evil Tip: When running Lighthouse, only use the best, beefiest, latest and greatest, most expensive hardware and network connections.

As additional context, I built a project called Speedlify, which is a self-hosted dashboard for performance monitoring and comparison. We use it on the Eleventy Leaderboards.

Speedlify has two common modes of operation: on a hosted CI/CD server or in DIY mode on your local machine. These two methods often provide different Performance scores! Running on a hosted server is typically more resource constrained and is more challenging to score well in Lighthouse’s Performance category.

For disclosure purposes, the Eleventy Leaderboards run in DIY mode primarily because the scale of the number of sites tested goes well beyond the build-time limit of the build server—but that does also mean that the scores are likely higher than if we were to able to run the project in hosted mode.

Normal Statistical Variability

Network conditions can vary. Maybe your computer was doing a resource intensive task while you were completing your test. Here are two additional runs of the same site on my MacBook Air (M1, 2020) using Chrome 94. It’s notable that these results offer a slightly higher performance score compared to the first run above. Lighthouse Score for Mobile: 99 on Performance, 89 on Accessibility, 93 on Best Practices, 100 on SEO Lighthouse Score for Mobile: 100 on Performance, 89 on Accessibility, 93 on Best Practices, 100 on SEO
Don’t Do This™ Evil Tip: When running Lighthouse, run it a bunch of times and pick the highest score.

In Speedlify, we attempt to smooth out this issue by running each test multiple times (by default 3) and selecting the median run using an algorithm from Lighthouse based on First Contentful Paint, Time to Interactive, and Largest Contentful Paint.

Speedlify improvements sparked by a discussion on Twitter with Patrick Hulce, who works on Lighthouse.

Mobile versus Desktop

Mobile scores are more difficult to score a perfect 100, particularly in the performance category. As Andy Davies states, this is “by design as mobile uses a simulated slower network, and CPU.” Importantly, and as previously discussed, the performance conditions are relative to the hardware/software of the current machine.

When a user shares a screenshot of a perfect Four Hundo score, Lighthouse (as it stands) makes it impossible to visually distinguish whether or not that score was taken under the Mobile or Desktop mode.

For example, here’s two Lighthouse scores of the same site taken back-to-back on the same hardware. One is a mobile result and one is desktop (you can click through to see a broader view). Note that structurally the screenshots are the same. Lighthouse Score for Mobile, Performance score of 90 Lighthouse Score for Desktop, Performance score of 100

MacBook Air (M1, 2020) using Chrome 94.

Don’t Do This™ Evil Tip: Always share your Desktop Score. Never reveal that it is a Desktop score—keep any discussion of the testing mode ambiguous.

There has been some discussion about adding a visual indicator to make the mode more obvious, which would help greatly!

Some more related Patrick Hulce discussion on Twitter.

I feel as though I should also mention—in a perfect world—if a web benchmark were to start from scratch with a new Lighthouse, the slow hardware simulation, network throttling, viewport size testing should be built into a single mode. I’d love it if the next version of Lighthouse ran Mobile mode, then Desktop mode, and displayed both scores together or combined them somehow. Get rid of the separation and it would clear up a bunch of the confusion in a very clean way.

Lab Data versus Field Data

This is perhaps the most nefarious distinction, because it is the most complex and as such offers the most effective kind of wiggle room: confusion.

Lab data is taken in a controlled environment. Field data is gathered from the recorded measurements of the performance of real visitors. Related: Why lab and field data can be different (and what to do about it).

Most of the methods we’ve talked about so far are only reporting lab data. But having field data is great, too! The caution I’d offer here is when someone focuses too closely on Field Data and never mentions Lab Data. But why? Isn’t it better to measure the real world? Why does it matter what happens in the lab?

Let us consult this classic blog post from Chris Zacharias: Page Weight Matters, in which Chris discusses a case study on the YouTube web site in which they decreased the page weight and the measured field data results got worse!

I had decreased the total page weight and number of requests to a tenth of what they were previously and somehow the numbers were showing that it was taking LONGER

Correspondingly, entire populations of people simply could not use YouTube because it took too long to see anything.

Large numbers of people who were previously unable to use YouTube before were suddenly able to.

If you have great field data: you may exist in the same realm as pre-optimized YouTube! The point being is that takes a holistic view of both field data and lab data to make good performance decisions!

Don’t Do This™ Evil Tip: If you have a wealthy, first-world, limited San Francisco-heavy audience with good hardware, make sure you shout about your field data! Field data is the most important thing! Pay no attention to the mobile/throttled/average hardware Lab data hiding behind the curtain.

Another way to say it:

  • If your field data is good and your lab data is bad, you may have built yourself a site for the wealthy western web.
  • If your field data is bad and your lab data is good (and assuming you aren’t doing any of the things we discussed to fudge your lab data scores), don’t fret! Your world wide web site may be reaching a global audience!
  • If you have both good field data and good lab data then you are a unicorn—I applaud you and celebrate your success. I love that for you. Please share how you successfully banished your third-party JavaScript to the shadow realm.


You might walk away from this article thinking: wow, Lighthouse scoring could be improved! I agree, but I also think that it’s been a net-win for performance discussions with other stakeholders in a professional setting. I genuinely hope they solve the Performance variability problem and add visual indicators to show you the mode in which a test ran (Desktop or Mobile).

But mostly, this is a plea to y’all: please don’t game your Lighthouse score. I hope an increased awareness of these tricks will decrease the frequency at which we see them appear in the wild. Stay safe out there, y’all.

< Newer
A New Eleventy Mascot from Geri Coady
Older >
Who Pays for Web Frameworks?

Zach Leatherman IndieWeb Avatar for a builder for the web at IndieWeb Avatar for He is the creator and maintainer of IndieWeb Avatar for https://www.11ty.devEleventy (11ty), an award-winning open source site generator. At one point he became entirely too fixated on web fonts. He has given 79 talks in nine different countries at events like Beyond Tellerrand, Smashing Conference, Jamstack Conf, CSSConf, and The White House. Formerly part of Netlify, Filament Group, NEJS CONF, and NebraskaJS. Learn more about Zach »


Fynn BeckerJoel G GoodmanJim NielsenAmit GharatYoav WeissChris HeilmannAlexanderBenjamin Griescaptain.dynamiteXoan /ʃαn/ 🐟Pablo Lara HDiana LeMarc Filleul 🇫🇷Fresh Frontend LinksAlex PopoutsisBrad FrostJacky EfendiTim VereeckeWeb DirectionsOne Kaitou of the Dead 🐍Friday Front-EndВеб-стандартыMat Casner : Designer : Freelancer : CoachChiyana Simões Silvestar Bistrović 🤘Adam ShiverIrina the Nerd 🐻Nebojsa (Cookie Duster)James Evers-SwindellHN Front PageWinson TangHacker NewsAngsuman ChakrabortyFrontend Daily 🚀Front-End FrontGregory HammondiwbjpSienna Web DesignsIndieWeb Avatar for https://softwaredevblog.techIndieWeb Avatar for https://softwaredevblog.techIndieWeb Avatar for https://www.reddit.comIndieWeb Avatar for


yashMayankFasterizeWilliam WijayaValentino Gagliardi 🇮🇹night of the living dezPiotr NalepaJens GrochtdreisMatthew PayneMike Bifulco 😷💉 (get vaxx'd)Matt SecoskeJavier Diaz ☀️Matthew RoachHector PiñeroAndy DaviesArnaud TanielianOne Kaitou of the Dead 🐍Niklas StåhleMichael GoodingSidNicolas ForgeotTim VereeckePatrick HulceAdam AhmedOptimizing MattersGeorge LiusozonomeJacky EfendiRaphaël AméaumeHugo Nogueiratigersway.netManda Putra ⠕Brad FrostAlex PopoutsisJoseph CurtisOleksandr ShutRyan BrooksDiana LeBogdan CerovacLuca DegasperiSamuel HauserSkullinsworthcaptain.dynamiteJohn Kemp-CruzRyan MulliganbertrandkellerBenjamin Gries𝕕𝔾𝕣𝕒𝕞𝕞𝕒𝕥𝕚𝕜𝕠MWDelaneyHeatherphil, who actually grew a pumpkin this year 🎃Søren Birkemeyer 🦊Chris HannabyScott JehlHolger BartelFynn BeckerSanti CrosSam TancharoensuksavaiMichael ScharnaglTim GilesOsmelrossKristofer KoishigawaBrett JankordAlex (He/Him)Prince WilsonDana BOOerly 💀Max BöckTim ChaseJim NielsenBrantley Harris 🪐Eric WallaceThord D. Hedengren⚡️

1 Bookmark

IndieWeb Avatar for
  1. Santi Cros


    Great article Zach. One thing I do is to check the Lighthouse score on so my computer doesn't affect the results :)

  2. Santi Cros


    And still there's some variation even if you run it repeatedly:s

  3. Zach Leatherman


    Yeah! Great point. I should add a note about that and… as a way to avoid hardware variability (though these don’t solve the other variability issues, in my experience).

  4. Zach Leatherman


    jinx! 🏆

  5. Brett Jankord


    I’ve been a fan of using to collect metrics and averaging out scores over multiple runs. It allows us to reduce some variability around network speeds and hardware as we offload it to Google. They throttle it down to fast 3g speeds and 4x slowdown on CPU.

  6. Jonathan Holden


    Interesting about speed... But how do you account for the different a11y scores in the first 2 images?

  7. bkardell


    Yeah, it would be great to ack that in the document somehow - I was suprised it wasn't mentioned because they do seem to vary. I would guess the answer is "one has more DOM loaded at the time of measure" or "the screen size has diff responsive elements hidden/show… Truncated

  8. Zach Leatherman


    Good question! I didn’t update anything on the old MacBook before running the test—so the versions of Chrome are different. Worth re-running on a newer version of Chrome with the old hardware to test though

  9. Eleventy 🎈 2.0.0-canary.14


    I just mean that statistical variability might be the only thing swapping you between 99 and 100 on performance in lighthouse. A bit more detail here:…

  10. Zach Leatherman (Away)


    per some of the other discussions in this thread you may be interested in this blog post… ok byeeeeeee

  11. Mayank


    you have a blogpost for everything! am i misremembering or did you also have one comparing the per-component bundle cost of various frameworks?

  12. Eleventy 🎃🎈 v2.0.0-canary.16


    I think this thread would likely be helpful to you!

  13. Mayank


    yes, this is it! 🏆 thank you

Shamelessly plug your related post

These are webmentions via the IndieWeb and

Sharing on social media?

This is what will show up when you share this post on Social Media:

How did you do this? I automated my Open Graph images. (Peer behind the curtain at the test page)