Discovering sooner algorithms for matrix multiplication stays a key pursuit in pc science and numerical linear algebra. For the reason that pioneering contributions of Strassen and Winograd within the late Sixties, which confirmed that normal matrix merchandise might be computed with fewer multiplications than beforehand believed, varied methods have emerged. These embrace gradient-based strategies, heuristic methods, group-theoretic frameworks, graph-based random walks, and deep reinforcement studying. Nonetheless, considerably much less focus has been positioned on matrix merchandise with inherent construction, corresponding to when the second matrix is the transpose or similar to the primary, or when matrices possess sparsity or symmetry. This oversight is notable, provided that expressions like AA^T seem ceaselessly in domains corresponding to statistics, deep studying, and communications, representing vital constructs like Gram and covariance matrices. Furthermore, XX^T is repetitive in massive language mannequin coaching algorithms like Muon and Shampoo.
Earlier research have explored structured matrix multiplication utilizing varied theoretical and machine learning-based strategies. Illustration idea and the Cohn–Umans framework have been employed to design environment friendly multiplication schemes for structured matrices. Reinforcement studying has additionally proven promise—fashions have discovered to find or rediscover recognized algorithms like Strassen’s. Current work has centered on optimizing the computation of XX^T over finite fields and sophisticated domains. Amongst these, probably the most environment friendly recognized methodology for real-valued XX^T is Strassen’s algorithm, who apply Strassen’s algorithm recursively on 2×2 block matrices, successfully translating the structured downside again into the area of normal matrix multiplication.
Researchers from the Chinese language College and the Shenzhen Analysis Institute of Huge Information have developed RXTX, an algorithm for effectively computing XX^T the place X belongs to R^n*m. RXTX reduces the variety of required operations—multiplications and additions—by roughly 5% in comparison with the present main strategies. Not like many algorithms that solely present advantages for giant matrices, RXTX delivers enhancements even for small sizes (e.g., n = 4). The algorithm was found by means of machine learning-based search and combinatorial optimization, leveraging the precise construction of XX^T for constant-factor acceleration.
The RXTX algorithm improves matrix multiplication by decreasing the variety of operations in comparison with earlier strategies like recursive Strassen and Strassen-Winograd. It makes use of 26 normal matrix multiplications and optimized addition schemes, leading to fewer whole operations. Theoretical evaluation exhibits RXTX performs fewer multiplications and mixed operations, particularly for bigger matrices. Sensible assessments on 6144 × 6144 matrices utilizing a single-thread CPU present RXTX is about 9% sooner than normal BLAS routines, with speedups noticed in 99% of runs. These outcomes spotlight RXTX’s effectivity for large-scale symmetric matrix merchandise and its benefit over conventional and state-of-the-art recursive algorithms.
The proposed methodology integrates RL with a two-tier Blended Integer Linear Programming (MILP) pipeline to find environment friendly matrix multiplication algorithms, notably for computing XX^T. The RL-guided Massive Neighborhood Search generates a big set of potential rank-1 bilinear merchandise, that are candidate expressions. MILP-A explores all linear mixtures of those merchandise to precise the goal outputs, whereas MILP-B identifies the smallest subset that may characterize all targets. This setup mirrors the AlphaTensor strategy however simplifies it by decreasing the motion area considerably, specializing in lower-dimensional tensor merchandise and leveraging MILP solvers like Gurobi for fast computation.
For instance, to compute XX^T for a 2×2 matrix X, the aim is to derive expressions like x_1^2 + x_2^2 or x_1x_3 + x_2x_4. The RL coverage randoMLy samples hundreds of bilinear merchandise utilizing coefficients from {−1, 0, +1}. MILP-A finds mixtures of those merchandise that match the specified expressions, and MILP-B selects the fewest wanted to cowl all targets. This framework enabled the invention of RXTX, an algorithm that performs 5% fewer multiplications and general operations than prior strategies. RXTX is environment friendly for giant and small matrices and demonstrates a profitable fusion of ML-based search and combinatorial optimization.
Take a look at the Paper. All credit score for this analysis goes to the researchers of this mission. Additionally, be at liberty to observe us on Twitter and don’t neglect to affix our 95k+ ML SubReddit and Subscribe to our Newsletter.

Sana Hassan, a consulting intern at Marktechpost and dual-degree scholar at IIT Madras, is obsessed with making use of know-how and AI to deal with real-world challenges. With a eager curiosity in fixing sensible issues, he brings a contemporary perspective to the intersection of AI and real-life options.