• 1 Post
  • 9 Comments
Joined 2 months ago
cake
Cake day: March 6th, 2026

help-circle
  • We solved the tiered testing problem like this, for reference (k6 + GitHub Actions):

    • Every PR: 30-second smoke test, 5 VUs, just checks nothing is catastrophically broken
    • Nightly: 10-minute average load test against main, results posted to Slack
    • Pre-release: Full soak test, stress test, spike test — run manually with sign-off required

    The nightly run is where we actually catch regressions. The PR check is just a safety net, not a real performance signal.

    The staging environment problem feels like the hardest one

    It is. We partially solved it by tagging every load test run with the git SHA and keeping a historical baseline database. Alerting triggers when p95 increases by more than 20% vs the rolling 30-day average. Not perfect but it surfaces real problems.




  • Very similar journey here , moved from JMeter about two years ago.

    The threshold-as-code approach is the feature I’d never give up. Having performance acceptance criteria living in the same repo as the tests completely changed our relationship with stakeholders. Now we have actual conversations about what p95 latency should be before a feature ships, rather than after it’s already slow in production.




  • I’m a QA consultant based in Canada specialising in a11y. Honest answer to your question:

    Most of it needs to stay manual. Automated tools (Axe, Lighthouse etc.) catch maybe 30–40% of real accessibility issues — the rest requires human judgment and actual assistive technology.

    What you can automate:

    • Axe-core integrated into your Playwright suite catches low-hanging fruit on every PR
    • Custom linting rules for missing alt text, empty labels, bad heading structure

    What you cannot automate:

    • Whether a screen reader actually conveys the right meaning
    • Logical focus order
    • Whether error messages are actually helpful when announced

    Your exhaustion is valid. This work is skilled and time-intensive. Push back on anyone who tells you an overlay or an automated scanner “handles” accessibility.


  • We did a formal test audit at my last company (fintech startup, based in the UK). Here’s roughly what we did:

    1. Pulled test run history from the last 90 days out of our CI system
    2. Tagged anything with >15% flake rate as “quarantine candidates”
    3. Ran a coverage diff to see what was actually protected by the remaining tests
    4. Deleted ruthlessly — killed about 600 tests

    The hardest part was getting stakeholder buy-in to delete tests. People see test count as a metric of quality. You have to reframe it: fewer, trustworthy tests beat many unreliable ones.





page 1