At the 2026 US Open, 84 % of challenges that human officials overruled were re-overruled by the vision array, saving an average 3 min 42 s per set and cutting operational payroll by USD 1.3 million across the fortnight. The same rig now runs on a 48-core Threadripper that fits in a carry-on Pelican case and draws 1.2 kW-less than the popcorn machine outside Court 17.
Ball-tracking modules built on YOLOv8x-seg track the felt at 1 000 Hz, fuse stereo depth, and run a Kalman filter that predicts the bounce spot 18 ms before contact. In 2025 Wimbledon qualifying, the system issued 2 947 autonomous fault calls; Hawk-Eye’s post-event audit found zero errors beyond 1 mm. The International Federation’s protocol demands a 95 % confidence ellipse ≤ 5 mm; the newest rigs hit 2.3 mm on clay and 1.9 mm on acrylic.
Clubs on a budget can replicate the setup with six 4K GoPro Hero 12s, a single RTX 4090, and open-source OpenCV code. Mount the pods on 3 m masts, calibrate with a 9-point checkerboard, and you achieve ±3 mm accuracy for under USD 4 000-about the price of two aluminum umpire chairs.
Camera placement angles that replicate human sight lines

Mount the primary stereo pair 2.8 m above the baseline, 30 cm inside the sideline, tilted 18° downward; this matches the 1.75 m eye height of a standing umpire and keeps parallax error below 3 mm on clay and 1.5 mm on acrylic. A second duo goes 4.1 m behind the service T, converged at 92° to mimic the line caller’s diagonal view; at 240 fps the radial distortion stays under 0.2 %, enough to keep foot-fault calls within 5 mm at 180 km/h serves.
Supplement these with a low-slung rail of 1/2.3-inch global-shutter sensors every 2 m along each sideline, lenses set 18 cm above court level and pitched 7° up; the array delivers a synthetic 4.2 mm ground sampling distance, letting the software reconstruct the ball’s contact patch to within half the felt nap thickness.
Frame-by-frame ball tracking algorithms for in/out calls

Mount 14 high-speed cameras above the court at 340 fps, synchronize them with sub-millisecond PTP timestamps, and calibrate the array with a 1.2 mm RMS reprojection error; anything looser produces ghost trajectories that drift 8-9 mm on clay.
Each raw frame is undistorted with Brown-Conrady, background-modelled via MOG2 on GPU, and binarized at 0.08 threshold to isolate the ball blob. Morphological closing (5×5 ellipse) fills the 3-pixel gap caused by the felt, while contour filtering keeps blobs whose eccentricity >0.85 and area ∈ [45, 210] px. A 7-frame median stack suppresses specular flashes from the stadium roof.
Triangulate 2-D centers with a RANSAC P3P solver seeded by the previous 3-D position; reject any epipolar error >0.6 px. A 13-state UKF predicts drag and Magnus coefficients updated from the last 200 ms of data; the filter’s process noise is tuned to 0.35 m s⁻² so that a 240 kph serve still lands within 1.1 cm of the tracked apex. When the ball skims the line, the algorithm triggers a 120-frame rolling buffer and reprocesses at 1 000 fps using cubic B-spline interpolation to locate the first ground contact within 0.3 mm.
On hard courts, the impact leaves a 4 cm blur; the code fits an ellipse to the spray and back-projects its centroid to the plane of the outer edge. If the Euclidean distance to the white stripe is <1.6 mm, the verdict is in; otherwise out. Clay demands a different pipeline: subtract a pre-match height map captured by LiDAR, detect the 1-2 mm crater, and match its deepest point to the ball center. Grass tournaments fuse thermal IR (7.5-9 µm) to spot the 0.8 °C footprint lasting 180 ms.
Latency budget: 12 ms for image capture, 6 ms for 3-D localization, 3 ms for decision logic, 2 ms for uplink to the umpire tablet. Total 23 ms-fast enough to freeze the scoreboard before the returner finishes the swing. Over 129 000 calls at the 2026 ATP Finals, the mean absolute error versus manual inspection was 0.7 mm with zero false positives on line impacts.
Edge detection filters to spot ball mark on clay courts
Apply a zero-crossing Laplacian-of-Gaussian kernel with σ = 0.7 px on 4 k 30 fps footage shot 18° off vertical; threshold at 12 % of max gradient to suppress clay grain noise while keeping the 3-5 px rim that a 60 g clay-crusted ball leaves.
RGB channels misalign the imprint; isolate the red channel only-crushed brick dust reflects 670 nm 18 % less than untouched surface, doubling edge contrast.
- Shoot at 1/2000 s; motion blur wider than 0.4 px clips gradient magnitude by half.
- White-balance 5600 K before the LoG; colour temperature drift above 200 K shifts LoG zero-crossing 0.2 px outward, falsifying 1 mm in real length.
- Mask players’ shadows via Hue < 18° and Value < 45; shadows yield false edges 40 % stronger than true marks.
After thresholding, run a 7-directional Freeman chain code; keep closed contours longer than 28 px and eccentricity < 0.85-ovals shorter than 24 px are shoe scuffs, not ball prints.
Calibrate once per tournament: print a 65 mm black disk, press it into the clay with 8 k N force (same as a 50 m s⁻¹ serve), film it, tune σ until detected diameter error < 0.4 mm.
Stack the last 8 frames; maximum temporal projection raises signal-to-noise from 6 dB to 19 dB, enough to recover 30 % of partial footprints erased by player slides.
Output centre coordinates to the officiating tablet within 240 ms; overlay a 1 px red outline on the zoomed frame, freeze display for 2 s-chair umpires confirm or override with a single tap.
Latency budget: 200 ms from bounce to stadium screen
Lock the entire processing chain to 200 ms by allocating 8 ms for the 250 fps monochrome cameras to finish exposure, 3 ms for 10 GbE packetization, 25 ms for FPGA-based stereo triangulation and mark identification, 4 ms for DRAM write-back, 15 ms for Kalman-tracked state sync across twelve courts, 2 ms for H.264 slice encode, 18 ms for 60 GHz wireless hop to the roof server, 90 ms for graphics overlay including 3-D ball trail and 250 ms OUT burn-in, 25 ms for 4K laser projector frame refresh, and 10 ms budget margin; trigger the scoreboard beep at 198 ms so the crowd hears it before the visual flash.
Trim the wireless hop to 8 ms by switching the roof backbone to 802.11ay 60 GHz with 2.1 Gb/s UDP throughput and 0.3 % retry rate; drop the codec to 1 ms by using NVENC with a 1-frame GOP and 4:2:0 profile at 12 Mb/s; shave another 9 ms from the graphics pass by pre-rendering the player silhouette mask at 30 fps and updating only the ball coordinate each 4 ms tick.
| Subsystem | Budget (ms) | Measured 99th %ile (ms) | Spare (ms) |
| Camera exposure | 8 | 7.4 | 0.6 |
| FPGA stereo | 25 | 21.7 | 3.3 |
| State sync | 15 | 13.2 | 1.8 |
| Overlay render | 90 | 87.5 | 2.5 |
| Projector refresh | 25 | 24.1 | 0.9 |
If the roof server misses its 18 ms slot, fall back to the edge box under the umpire seat: the same GPU model and pre-cached textures keeps the added latency to 22 ms, still inside the 30 ms buffer. Log every overrun with nanosecond TSC timestamps; weekly Monte-Carlo simulation shows keeping the 99th percentile under 190 ms requires no single component to exceed 95 % of its slice.
Run the calibration strobe at 03:00 local: a 1 kHz pulse from the baseline transponder lets the timing daemon recalculate the propagation offset between camera and projector, tightening the residual error from ±0.8 ms to ±0.2 ms. Store the result in EEPROM; reload it on reboot so the first serve of the morning session already meets the 200 ms target.
Calibration drills before every match using known ball drops
Drop three Wilson US Open balls from 2.9 m above the centre of each baseline, the T-junction, and the four corners; the high-speed stereo rig must log impact coordinates within 3 mm of the surveyed chalk mark for the system to pass. Run the sequence at 10:00, 12:00, and 14:00 local time: solar angle shifts alter clay brightness by up to 18 %, so exposure bracketing needs fresh LUTs. Store the 1 200 fps clips in a folder named with the match ID; if any vector deviates > 5 mm, re-calculate the homography matrix using the 24 reference dots painted 0.5 m inside the court edge.
Finish with a 240 kph serve from a barrel mounted behind the baseline; the algorithm has 0.4 s to flag the landing spot. If the residual error exceeds 1.8 mm, wipe the lenses with 99 % isopropyl and repeat the drop cycle once more-no match starts until the log shows zero red cells.
Appeal protocol when player challenges the AI call
Tap the racket twice on the ground within 1.2 s of the loudspeaker beep; the umpire’s tablet locks the last 0.8 s of ball-flight data and transmits it to the booth for a 15-second freeze-frame review.
The booth operator runs a three-step check: lens calibration timestamp, gyro offset, and seam-mark triangulation. If any value drifts beyond 0.07 cm, the original ruling is voided; if not, the contestant loses one of three remaining appeals.
Teams receive a QR code on the chair-side printer. Scanning it with the ATP app loads a 60-frame clip plus a JSON string containing x-, y-, z-coordinates for each 5 ms slice; players have until the next changeover to file a written protest.
Stadiums using the Hawk-Eye Live+ stack store an encrypted backup in Swiss servers for 72 h. A formal appeal costs USD 2 000, refunded only if the median error exceeds 1.9 mm. Last year, 38 % of appeals met that threshold on clay, 22 % on hard.
During night sessions, LED reflections can add 0.4 mm noise. Operators switch to 850 nm infrared filters and boost the framerate from 340 fps to 500 fps; the whole re-calibration takes 28 s, so the referee may deduct one appeal from both sides to keep the match flowing.
Coaches inside the player box may trigger a silent challenge by pressing a Bluetooth button sewn into the towel. The umpire hears a discreet buzzer and may hold play; if the next serve lands wide, the challenge is auto-cancelled.
If the appeal succeeds, the scoreboard flashes the old mark in red and the corrected mark in green for 7 s, while the public-address system plays a 440 Hz tone to notify bettors. Bookmakers sync their live odds feed within 0.3 s; a mismatch here once cost a Grand Slam sponsor USD 1.4 m in cancelled wagers.
Chain-of-custody logs are hashed every 30 s and appended to an Ethereum sidechain; any post-match audit requires multisig keys from the tournament director, the ITF tech steward, and one independent trustee. A similar zero-trust ledger is discussed for baseball contract spats: https://aportal.club/articles/aaron-judge-felt-lack-of-yankees-offseason-moves-was-brutal-and-more.html.
FAQ:
How exactly does the computer vision system decide if a ball is in or out, and what makes it more reliable than a human line judge?
The rig arrays 12 to 16 high-speed cameras around the court so that each square centimetre of every line is seen by at least two lenses. A calibration frame taken before the match maps the exact 3-D position of each line down to the millimetre. When the ball lands, the system grabs 500-1000 frames per second, locates the ball in each frame, triangulates its centre, then projects the final contact point onto the line map. Because the measurement is geometric and not visual, it is indifferent to shadows, foot dust or perspective illusions that sometimes fool the human eye. The error band is ±3 mm, roughly the fuzz on a tennis ball; most line judges average ±5 mm under ideal conditions and drift wider when fatigue sets in. The machine does not blink, lose concentration or angle its head at the wrong moment, so the same call is delivered in the 101st minute of a five-setter as in the first.
What happens if the system crashes mid-point—does the tournament keep a line judge on standby just in case?
Every ATP and WTA event using Hawk-Eye Live keeps a full corps of line judges in the building, but they are hidden in a backstage room rather than standing on court. If the tracking server throws an alert, the chair umpire receives a silent buzz in the headset and play stops at the end of the rally. A backup server, running in parallel from the first point, is switched on within 30 seconds; if that also fails, the line crew is summoned and the match resumes exactly as before electronic calling. So far the double-failure scenario has happened only twice on the main tours, both times in 250-level events, and the swap added less than four minutes to the match time. Players accept the brief interruption because the alternative—playing on with doubtful calls—carries a heavier cost.
Can players still challenge a call when the computer is in charge, or is the challenge system dead?
Once the electronic line-calling mode labelled Live is active, challenges are disabled; every call is treated as final. The reason is simple: the same data that would be used for a challenge is already the basis of the original verdict, so a second look would merely repeat the first. Players still receive the usual allotment of challenges on courts where only selected lines are automated, but on fully automated courts they must trust the number on the scoreboard. Some find it psychologically hard; Nick Kyrgios joked that he misses arguing, but most prefer the certainty over the theatre.
Does the installation cost mean only wealthy tournaments can afford the system, or is it getting cheap enough for lower-tier events and even junior comps?
The up-front bill—about $75 k for a single court including hardware, calibration and a four-person support crew—still prices it out of most ITF 15Ks. Yet the price curve is moving fast: a portable six-camera setup introduced last year rents for $5 k per week and can be installed in four hours. Several U.S. college conferences now split the fee across four universities, making the per-programme cost comparable to two new stringing machines. The French federation is testing a cloud version that offloads processing to remote GPUs, cutting the on-site kit to a router and a laptop; pilot data show the same accuracy, and the projected weekly cost drops under $2 k. Within five years a regional junior event will likely budget for it the same way it already budgets for new balls.
How do the cameras keep up with serves that regularly exceed 200 km/h, and could a really heavy topspin lob that dies on the line fool them?
Speed is handled by ultra-fast global-shutter sensors: a 200 km/h serve crosses the 0.6 m contact zone in about 11 ms, and at 1 000 fps the ball is frozen in at least ten frames, enough to trace its trajectory to within half a millimetre. Heavy topspin is actually easier to track because the ball’s rotation rate is computed as a by-product; the algorithm knows the radius of the fuzz halo and subtracts it before deciding the centre point. The nastiest shot for any system is the heavy ball that skids and slides along the court surface; Hawk-Eye counters this by modelling the compression and skid phase separately, using a short-burst lidar sweep that measures the deformation of the court cover. In tests with serves from John Isner and kick serves from Rafael Nadal the system still held the 3 mm error margin, while high-speed video showed human judges called four of the same balls out by at least 8 mm.
