Part of Advances in Neural Information Processing Systems 34 (NeurIPS 2021)
Xiangming Meng, Tomoyuki Obuchi, Yoshiyuki Kabashima
We theoretically analyze the typical learning performance of ℓ1-regularized linear regression (ℓ1-LinR) for Ising model selection using the replica method from statistical mechanics. For typical random regular graphs in the paramagnetic phase, an accurate estimate of the typical sample complexity of ℓ1-LinR is obtained. Remarkably, despite the model misspecification, ℓ1-LinR is model selection consistent with the same order of sample complexity as ℓ1-regularized logistic regression (ℓ1-LogR), i.e., M=O(logN), where N is the number of variables of the Ising model. Moreover, we provide an efficient method to accurately predict the non-asymptotic behavior of ℓ1-LinR for moderate M,N, such as precision and recall. Simulations show a fairly good agreement between theoretical predictions and experimental results, even for graphs with many loops, which supports our findings. Although this paper mainly focuses on ℓ1-LinR, our method is readily applicable for precisely characterizing the typical learning performances of a wide class of ℓ1-regularized M-estimators including ℓ1-LogR and interaction screening.