Benchmarking Quantum Hardware: How to Measure Qubit Performance and Interpret Results
benchmarkinghardwareperformance

Benchmarking Quantum Hardware: How to Measure Qubit Performance and Interpret Results

DDaniel Mercer
2026-05-15
25 min read

A practical guide to benchmarking quantum hardware, interpreting qubit metrics, and running reproducible cross-provider tests.

Quantum hardware benchmarking is where abstract promises meet operational reality. If you are a developer, platform engineer, or administrator, the question is not whether a quantum processor is “advanced” in a press release sense; it is whether the device can run your circuit, at your depth, with reproducible results, on a schedule you can trust. That means measuring more than a single headline number and learning how to interpret coherence times, gate fidelities, readout errors, queue delays, and the practical throughput of a noisy hardware strategy. It also means understanding how benchmarking fits into broader measurement discipline and how to compare providers without being misled by marketing-friendly averages.

This guide is written for teams evaluating a quantum cloud platform, testing quantum developer tools, or building internal standards for quantum performance tuning. We will cover the metrics that matter, how to run repeatable experiments, what the common numbers actually mean, and how to build a benchmarking workflow that stands up across providers, backends, and time. Along the way, we will connect methodology to real-world development practice, just as engineers do when validating systems in other domains such as end-to-end validation pipelines or operating under constrained resources like memory scarcity.

1. What Quantum Hardware Benchmarking Is Actually Trying to Prove

Benchmarking is not a single number; it is a claim about behavior

A useful benchmark answers a practical question: if I submit this workload, what quality of outcome should I expect, how often will it fail, and how much time will it cost me? In quantum computing, that means benchmarking must capture both the device physics and the system behavior around it. A chip with excellent single-qubit fidelity but poor readout stability may look good in a lab note, yet be unreliable for production experiments. Conversely, a device with modest fidelity but excellent calibration discipline and predictable access may outperform a “better” chip in day-to-day developer use.

Developers often underestimate how many layers sit between a circuit and its result. Hardware quality, compiler/transpiler decisions, calibration freshness, qubit placement, queue latency, and run scheduling all shape the output you see. A robust benchmark therefore tests the whole stack: device, control plane, and workflow. This is why benchmarking should be approached the way teams approach safety-critical software validation: define acceptance criteria, pin the environment, run repeated tests, and compare like with like.

Why marketers’ numbers and engineers’ numbers are not the same thing

Hardware vendors often highlight best-case fidelities, fast reset times, or impressive coherence values. Those figures are useful, but only when interpreted in context. A high T1 value does not automatically mean your algorithm will run well, because T2 dephasing, crosstalk, connectivity limits, and compiler overhead may dominate. In other words, a hardware spec sheet is only the start of the analysis. Real benchmarking asks how those specs translate into circuit success probability, effective depth limits, and repeatable throughput on your target workload.

This distinction is similar to what product teams learn when evaluating operational tools in other domains: the headline metric matters, but so do reliability, ergonomics, and workflow fit. For an analogy, see how a domain-specific product can succeed only when it aligns with the user’s actual workflow, much like Garmin’s nutrition tracking highlighted user-market fit over raw feature count. Quantum hardware should be judged the same way: by fit for purpose, not just by isolated lab metrics.

Benchmark categories you should always separate

There are at least four benchmark categories worth separating in your program. First are physics-based characterizations like T1 and T2, which describe how long the qubit stays useful. Second are gate-level metrics like one-qubit and two-qubit fidelity, which show how accurately the device executes operations. Third are readout metrics, which show how reliably the device reports states back to you. Fourth are application-level or system-level benchmarks, which tell you whether your actual workload can complete with a useful answer. A mature benchmarking program tracks all four, because each answers a different operational question.

If you are building internal standards for quantum hardware evaluation, it can help to think like platform teams do when they design reliable production systems. Compare that mindset with the practical guidance in embedding trust in AI adoption or the operational rigor seen in hybrid fire systems: different failure modes require different controls. Quantum benchmarking is no different.

2. The Core Metrics: What to Measure and What They Mean

T1 and T2: coherence time is necessary, not sufficient

T1 is the energy relaxation time: how long it takes an excited qubit to decay toward its ground state. T2 is the dephasing time: how long the qubit preserves phase information. In practice, T1 and T2 set a ceiling on how long a quantum computation can maintain useful information before noise overwhelms the signal. But these are not “performance scores” in the same way a benchmark score is. They are limits, and limits alone do not tell you the whole story.

Interpret T1 and T2 as budget constraints. A circuit with more layers, more entangling gates, or longer idle windows is spending coherence time faster. If T2 is low relative to your average gate duration and compilation overhead, your algorithm may degrade even if the gate fidelity looks decent. The most productive way to use coherence data is to estimate the practical depth your workload can support before error mitigation or different circuit design becomes necessary.

Gate fidelity: the most cited number and the easiest to misread

Single-qubit and two-qubit gate fidelities are usually the first numbers people compare across quantum hardware benchmarks. They matter because they directly affect the probability that your circuit executes as intended. However, they are often reported as averages, and averages can hide the qubits or couplers that perform worst. A device with a strong median but several weak edges may be far less useful than a slightly lower-average device with uniform performance and stable calibration.

Two-qubit gates deserve special attention because they tend to be the largest source of error in many architectures. If your workload needs repeated entanglement, then the quality and topology of the two-qubit gates can matter more than the nominal qubit count. When you evaluate hardware, do not stop at the platform-level average; inspect the per-qubit and per-edge distributions. That is the difference between a marketing story and a deployment-ready assessment.

Readout fidelity, assignment error, and measurement symmetry

Readout fidelity tells you how accurately the hardware reports a qubit’s state at measurement time. Assignment error is the rate at which the device maps a true state to the wrong classical bit. In many practical workloads, readout errors can dominate the tail of your result distribution, especially when circuits are shallow and the answer depends on a few critical bits. If you ignore readout fidelity, you may blame the algorithm for failures caused by measurement noise.

You should also inspect asymmetry in readout errors. Some devices are better at reading |0⟩ than |1⟩, or vice versa, and that imbalance can bias experiments in ways that are not obvious from a single average figure. For developers, the key lesson is to calibrate your interpretation of results, not just the device itself. Measurement correction can help, but only if your benchmark captures the raw error structure in the first place.

Pro Tip: Never compare a vendor’s “best qubit” to another vendor’s “platform average.” Compare like-for-like: same circuit family, same depth, same transpilation constraints, same number of shots, and the same reporting window.

3. Additional Metrics That Matter in Real Usage

Connectedness, crosstalk, and qubit map quality

Even a strong qubit can become a poor system resource if its neighbors are noisy or if routing forces the compiler into extra swaps. Connectivity determines whether your circuit can be mapped efficiently, and crosstalk tells you whether operations on one qubit disturb another. These effects do not always show up in simple single-qubit benchmarks, but they can dramatically affect application performance. In distributed systems terms, this is the difference between a fast node and a fast cluster.

If your team is used to thinking about systems engineering under constrained resources, the logic will feel familiar. The article on architecting for memory scarcity offers a useful analogy: local improvements do not help if system-level pressure still limits throughput. Likewise, a promising qubit is not enough if the coupling graph or calibration instability makes the device impractical for your circuit structure.

Queue time, job success rate, and real throughput

Benchmarking should not stop at execution quality. For teams working in the cloud, the real cost of a quantum run includes queue time, job retry rate, cancellation rate, and the time it takes to receive usable results. This is where “real-world throughput” matters. A device that can execute 100 circuits quickly in isolation may still be poor operationally if only a fraction of jobs complete or if calibration windows frequently interrupt access.

Think in terms of developer experience. How long does it take from submission to useful output? How many shots can you reasonably get in a maintenance window? How often do calibration updates force revalidation? Teams that build internal cloud usage standards already think this way for other services, and the same approach should be applied to trusted operational platforms and quantum backends alike.

Error rates, mitigation overhead, and usable fidelity

Raw error rates are only part of the story, because many quantum workflows use error mitigation. That introduces overhead in extra shots, additional calibration circuits, and more complex post-processing. Usable fidelity is the performance you get after accounting for that overhead. A backend that looks slightly worse on raw metrics may in fact deliver better usable performance if it is more stable, easier to calibrate, or less variable across runs.

This is why reproducibility matters so much. If you are comparing vendors or tracking a device over time, record the whole experiment recipe: transpiler settings, seeds, backend version, calibration timestamp, shot count, and mitigation options. Benchmarking without provenance is just anecdote at scale. If you want a practical model for disciplined experimentation, study the same workflow rigor seen in validation pipelines and adapt it to quantum runs.

4. How to Design Reproducible Quantum Benchmarks

Freeze the environment before you measure anything

Reproducibility begins with environment control. Pin your SDK version, transpiler version, compiler options, and circuit seed. Record the backend name, device calibration window, and any provider-side changes announced during the test period. If possible, run the same benchmark suite multiple times across a short window and then across a longer window so you can distinguish normal variance from structural drift.

In a practical sense, your benchmark harness should behave like any other test harness in software engineering. Treat circuits as test cases, backends as environments, and measurement outcomes as assertions with tolerances. This is similar to the engineering discipline used in CI/CD validation systems, where repeatability is a feature, not an afterthought. Quantum hardware is noisier than conventional compute, but your testing process should be stricter, not looser.

Use benchmark families, not single circuits

No single benchmark tells you everything. A balanced suite should include basic calibration-sensitive circuits, entanglement-heavy circuits, and application-inspired workloads. Common patterns include randomized benchmarking style tests, quantum volume-style evaluations, mirror circuits, and algorithmic microbenchmarks. Each reveals different failure modes, and together they provide a more honest picture of performance.

For developers, the most useful benchmark family is often the one that mirrors the shape of your workload. If you are building variational algorithms, use layered parameterized circuits. If you are exploring chemistry, stress the entangling patterns and ansatz depth that matter to your model. If your job is mostly experimentation and platform comparison, keep a stable reference suite and rerun it periodically so you can track drift.

Control for shots, seeds, and sampling noise

Quantum results are stochastic, so you need enough shots to separate true signal from sampling noise. Too few shots can make a backend look either better or worse than it is. Use confidence intervals, not just point estimates, and report variance over repeated runs. Also beware of cherry-picking: if you only publish the run that happened to look best, you are measuring luck rather than hardware.

One practical tactic is to run each benchmark with fixed seeds across multiple shot budgets. This shows you how sensitive your metric is to sample size and whether the device’s rank changes as you collect more data. That kind of sensitivity analysis is important in any performance program, much like pricing and timing analysis for high-end GPU procurement. In both cases, buying decisions should be based on distributions, not just anecdotes.

5. A Practical Benchmarking Workflow Across Providers

Step 1: Define the workload class you care about

Start by classifying the workload: gate-model prototype, hybrid variational workload, algorithmic benchmark, or application-inspired test. Different hardware strengths surface under different workloads, so the same platform can rank differently depending on what you are running. A provider that excels at shallow circuits may be less competitive on deeper entangling workloads, while another might shine in readout stability but lag in coupling efficiency. Your benchmark should reflect the workload class that actually matters to your team.

This is a common mistake in technology evaluation: people ask “which platform is best?” when the real question is “best for what?” The same principle appears in practical product and service comparisons across industries, whether you are evaluating feature fit or planning operational upgrades in another environment. Benchmarks should be purpose-built, not generic theater.

Step 2: Build a provider-neutral test harness

A neutral harness keeps your circuit definitions, transpilation strategy, and reporting format consistent while allowing backend-specific execution. Use a common intermediate representation when possible, and normalize outputs into a single schema with fields for backend, version, timestamp, circuit family, depth, qubit count, shots, result distribution, and mitigation method. This makes it much easier to compare runs across a quantum cloud platform without rewriting analysis logic for every vendor.

It also reduces the risk of accidental bias. If one provider gets a more favorable mapping or a later calibration window, your comparison is no longer apples-to-apples. Treat provider neutrality as a requirement, not a preference. This is the same reason reliable engineering teams design platform-agnostic checks before they optimize for one vendor-specific environment.

Step 3: Separate compile-time performance from hardware performance

Transpilation quality can dominate your benchmark outcome. If one backend requires far more SWAPs or deeper routing than another, the hardware score may really be measuring compiler burden. That is why you should report both logical circuit metrics and physical circuit metrics. Logical metrics describe the intended algorithm; physical metrics describe the compiled circuit actually run on hardware.

Where possible, report two versions of the benchmark: one with a fixed compiler budget and one with each provider’s recommended optimization settings. The first shows raw platform comparability, while the second shows best-case user experience. That dual view is valuable for developers, because it separates intrinsic device quality from toolchain excellence. In practice, both matter when choosing where to invest engineering time.

6. Interpreting Results Without Getting Misled

Look for rank stability, not just the top score

A benchmark suite should tell you whether the same backend stays near the top across repeated runs. A device that wins once and falls apart later is not a dependable choice for ongoing development. Rank stability matters because many teams need repeatability more than peak performance. If your project requires weekly or daily runs, consistency often beats a slightly higher but brittle best-case result.

Stability analysis also reveals when a platform has become operationally safer. If variance shrinks over time, the backend may be maturing even if its headline numbers move only modestly. Conversely, if the mean improves while variance widens, your user experience may actually get worse. That nuance is often missed in quick comparisons, which is why robust benchmarking practice is closer to observability than to simple scoring.

Distinguish statistical significance from practical significance

In quantum benchmarking, a small numerical difference may not matter operationally. A 1% fidelity gain is important if it prevents a circuit from crossing a failure threshold, but irrelevant if the result is still too noisy to use. Always ask whether the measured difference changes your decision. If it does not affect circuit depth, usable accuracy, or throughput, it may not be worth optimizing for immediately.

This distinction is useful when presenting results to stakeholders. Engineering teams should be able to explain not only that Backend A beat Backend B, but also what the difference means for the target workload and what tradeoffs it introduces. That level of interpretation turns benchmark data into decision support rather than vanity metrics. It also builds trust, which is crucial when adopting emerging platforms and tools.

Translate hardware metrics into developer outcomes

Developers care about outcomes: can I run a longer circuit, get a more stable estimate, or finish my experiment in a shorter time window? That means every low-level metric should be mapped to a practical consequence. For example, a shorter T2 can imply shallower circuit limits, a worse readout fidelity can increase measurement correction overhead, and lower two-qubit fidelity can reduce convergence quality in variational algorithms. If you cannot translate the number into a workflow impact, the metric is probably under-specified for your use case.

When you need a practical example of translating system metrics into user value, look at how product narratives are built in other technical fields, such as trust-centered adoption patterns or orchestrating specialized AI agents. The best technical communication always connects instrument readings to operational choices.

7. A Comparison Table for Common Quantum Hardware Metrics

The table below summarizes the metrics teams most often use in hardware benchmarking, what they mean, and how to interpret them in practice. Use it as a checklist when reviewing provider documentation or designing your own test suite.

MetricWhat It MeasuresWhy It MattersCommon PitfallPractical Interpretation
T1Energy relaxation timeSets a limit on how long a qubit can hold excitationAssuming a longer T1 guarantees better algorithm performanceUseful for estimating maximum circuit duration
T2Phase coherence timeMeasures how long phase information survivesIgnoring it when assessing circuit depthCritical for phase-sensitive and deeper circuits
Single-qubit fidelityAccuracy of one-qubit gatesShows base operation qualityOverweighting it relative to two-qubit performanceImportant, but not enough for entangling workloads
Two-qubit fidelityAccuracy of entangling gatesUsually the main limiter for multi-qubit workloadsComparing averages without topology contextKey metric for most practical algorithms
Readout fidelityAccuracy of measurement resultsAffects final answer quality and correction overheadIgnoring asymmetric errors between 0 and 1Essential for interpreting output distributions
Queue timeDelay before execution startsAffects developer throughput and experiment cadenceMeasuring only hardware execution timeVital for cloud usage planning
Calibration stabilityHow often performance stays consistent over timePredicts whether benchmarks are reproducibleTaking one snapshot as representativeNeeded for operational reliability

8. How to Benchmark Across Quantum Cloud Providers Fairly

Control for backend topology and compilation strategy

Cross-provider benchmarking is difficult because devices differ in qubit topology, gate sets, access policies, and calibration schedules. Fair comparison means you must either normalize these differences or explicitly report them. A ring topology is not directly comparable to a heavy-hex layout without considering routing cost, and a backend with excellent fidelity but limited connectivity may still lose in a workload that needs broad entanglement. That is why every report should include mapping details, transpilation depth, and swap count.

If you are comparing vendors for enterprise adoption, also consider operational aspects such as API stability, billing transparency, reservation options, and data retention rules. Those platform concerns mirror lessons from procurement and operational tooling in other sectors, including practical guidance on trusted adoption and platform comparison strategies seen in other technology markets. Benchmarking is not just a physics exercise; it is a purchasing process.

Use the same circuit library and analysis notebook everywhere

One of the simplest ways to increase fairness is to keep a single benchmark repository and run it unchanged across providers. The same circuits, the same parameters, the same analysis notebooks, and the same acceptance thresholds should be reused whenever possible. This eliminates avoidable drift and makes result deltas meaningful. A version-controlled benchmark suite is the quantum equivalent of a repeatable test harness in software engineering.

You should also document any provider-specific differences that cannot be removed, such as native gate names or access to pulse-level controls. When those differences are material, note them as part of the evaluation rather than hiding them. That transparency makes your results more useful to colleagues and future you. It also prevents false equivalence between platforms that solve the same problem in different ways.

Build a scoring model that reflects your priorities

Not every team values the same tradeoffs. If your priority is algorithmic accuracy, weight fidelity and stability more heavily. If your priority is high experiment velocity, queue time and successful job completion may matter more. A scorecard should reflect your actual deployment goals, not an abstract notion of “best hardware.”

Many organizations are tempted to optimize for the most impressive number in isolation, but the right way to compare a quantum developer stack is to assign weights that mirror reality. A research team might accept longer queue times for slightly better fidelity, while a platform team might do the opposite to maximize throughput. Make the tradeoff explicit, and your decisions will be easier to defend.

9. Common Benchmarking Mistakes and How to Avoid Them

Cherry-picking runs or reporting only the best outcome

Quantum hardware is noisy, which makes cherry-picking especially dangerous. If you only present the best run, you understate the operational risk. The right approach is to report mean, median, variance, and confidence intervals across multiple repeats. If your provider comparison cannot survive that scrutiny, it is not ready for decision-making.

Similarly, avoid benchmarking on a day when the device was unusually stable and then assuming that result generalizes. A single favorable window can hide calibration drift, queue spikes, or silent firmware changes. The best practice is to run at multiple times and preserve the full experimental log. That record becomes your evidence base when stakeholders ask why the numbers changed.

Ignoring compiler effects and circuit depth inflation

Many benchmark comparisons accidentally measure compiler burden instead of hardware quality. If one backend causes substantially more routing overhead, the logical workload may no longer be the same as the physical workload. That is why you must report compiled depth, two-qubit gate count, and swap count alongside the final score. Otherwise, the benchmark is incomplete at best and misleading at worst.

This problem is easy to avoid when you treat compilation as part of the benchmark, not a pre-processing step you can ignore. The same lesson shows up in broader systems work where hidden transformation layers reshape outcomes, much like platform tooling can reshape operational results in other domains. In quantum computing, invisible overhead is still overhead.

Confusing hardware quality with task suitability

Not every high-quality device is the best fit for every algorithm. Some workloads are more sensitive to readout than to T1, while others are dominated by entangling gate errors or by connectivity. If you benchmark only one style of circuit, you may select the wrong platform for your intended use case. To avoid this, map each benchmark to a concrete user story: chemistry simulation, optimization prototype, educational demo, or calibration research.

That approach keeps the evaluation honest and useful. It also prevents teams from overspending time trying to make one backend do everything when a different backend would better match the workload. The best benchmark result is not always the highest number; it is the clearest decision signal.

Start with a minimal, repeatable test suite

Begin with a small suite that includes one simple calibration circuit, one entangling benchmark, one readout-heavy circuit, and one application-inspired workload. Run them daily or weekly, depending on your access model, and store the results in a queryable format. This creates a baseline that can reveal drift, regressions, or improvements over time. Once the core suite is stable, expand it gradually rather than building a sprawling benchmark library too early.

If you need inspiration for how to grow a reusable technical program without losing clarity, look at the way operational content is structured in other engineering contexts, such as validation pipelines or specialized orchestration. The pattern is consistent: keep the first version narrow, then scale intentionally.

Automate data collection and store provenance

Manual benchmarking does not scale well. Use scripts to submit jobs, capture metadata, and archive outputs alongside the compiled circuit and analysis settings. Store the provider, backend ID, calibration snapshot, seed, number of shots, and any mitigation parameters. This will save time later when you need to reproduce a surprising result or explain a comparison to stakeholders.

Provenance is not bureaucracy; it is the basis of trust. Without it, you cannot determine whether a result reflects hardware changes, compiler changes, or test drift. That is why serious teams treat benchmarking like a controlled experiment rather than a one-off report. Good records turn an experiment into an asset.

Review benchmark results as an operational KPI, not just a science metric

Once your benchmark suite is established, use it to make actual decisions. Which backend should host this experiment class? Which provider’s uptime and calibration patterns fit your team’s cadence? Which SDK changes improved outcome stability? These are operational questions, and they should be reviewed with the same seriousness as other platform metrics.

A strong benchmark program becomes a living signal for platform choice, cost control, and development velocity. It helps your organization move from curiosity-driven exploration to repeatable execution. That is exactly what teams need when they are serious about quantum computing for developers and looking for practical pathways to adoption.

11. Final Takeaways for Developers and Admins

Benchmark the whole stack, not just the qubit

Quantum hardware performance is a layered problem. The qubit matters, but so do compilation, connectivity, readout, queueing, and calibration stability. If your benchmark program measures only one of those layers, you will make decisions on an incomplete picture. A reliable benchmarking strategy should therefore examine both microscopic physics and macroscopic workflow outcomes.

Use repeated, comparable, provenance-rich tests

Reproducibility is the difference between noise and signal. Fix your harness, log your metadata, and compare runs over time. Where possible, benchmark the same circuit families across providers with the same reporting format. That is how you make results portable and meaningful.

Optimize for the workload you actually care about

There is no universal “best” quantum backend. The right choice depends on your depth, topology, shot budget, tolerance for queue time, and error profile. Benchmarks become useful when they guide real choices: where to prototype, where to deploy, and when to revisit a provider decision. That practical mindset is what separates casual experimentation from serious quantum engineering.

For related reading on how to design circuits for imperfect devices, see Designing Quantum Algorithms for Noisy Hardware. For broader context on how trust and platform reliability shape adoption, revisit Why Embedding Trust Accelerates AI Adoption and Internal Linking Experiments That Move Page Authority Metrics—and Rankings.

FAQ: Quantum Hardware Benchmarking

1. What is the most important metric when comparing quantum hardware?

There is no single most important metric for every workload. For many practical circuits, two-qubit gate fidelity and readout fidelity matter more than headline qubit count. If your workload is depth-sensitive, then T1 and T2 become more important because they constrain how long the circuit can remain coherent. The best metric depends on the algorithm and the deployment pattern.

2. How many shots should I use in a benchmark?

Use enough shots to make the output distribution stable for the metric you are measuring. For small circuits, that may be thousands of shots; for more complex studies, you may need more. The important thing is to keep shot count consistent across comparisons, and to report confidence intervals so readers can judge the noise level. More shots reduce sampling error, but they do not fix hardware noise.

3. Can I fairly compare two providers with different qubit topologies?

Yes, but only if you account for topology in your analysis. You should report transpilation depth, swap count, and any routing overhead introduced by the mapping. If one provider forces much higher physical depth, then a raw score comparison is not fair unless you explicitly include that cost. Ideally, benchmark the logical workload and the compiled workload separately.

4. Why do my benchmark results change from day to day?

Quantum devices are calibrated frequently, and performance can drift as hardware conditions change. Queue patterns, backend updates, compiler version changes, and shot noise can also affect results. That is why reproducible benchmarking must record provenance and run repeated tests over time. Day-to-day variation is normal; what matters is whether the variation is understood and bounded.

5. Do error mitigation techniques make benchmarking less trustworthy?

No, but they change what you are measuring. With mitigation, you are no longer measuring raw hardware output only; you are measuring hardware plus correction overhead. That can be valuable, because it reflects how real users will try to get useful results. Just make sure to report both raw and mitigated outcomes so readers can interpret the tradeoff.

6. What should admins track beyond qubit metrics?

Admins should track queue time, job success rate, calibration stability, access policies, billing behavior, and backend versioning. Those operational metrics strongly influence whether a platform is usable for a team, even if the hardware itself is strong. A backend that looks excellent on paper can still be a poor choice if access is unreliable or too slow for your workflow.

Related Topics

#benchmarking#hardware#performance
D

Daniel Mercer

Senior Quantum Content Strategist

Senior editor and content strategist. Writing about technology, design, and the future of digital media. Follow along for deep dives into the industry's moving parts.

2026-05-15T21:09:22.964Z