We study an online linear regression setting in which the observed feature vectors are corrupted by noise and the learner can pay to reduce the noise level. In practice, this may happen for several reasons: for example, because features can be measured more accurately using more expensive equipment, or because data providers can be incentivized to release less private features. Assuming feature vectors are drawn i.i.d. from a fixed but unknown distribution, we measure the learner's regret against the linear predictor minimizing a notion of loss that combines the prediction error and payment. We first study the case in which the mapping between payments and noise covariance is known and prove order-optimal regret bounds in the interaction length (up to log-factors). We then derive order-optimal bounds also when the noise covariance is unknown and prove that the regret rate is worse than the case of known covariances. Our analysis leverages matrix martingale concentration, showing that the empirical loss uniformly converges to the expected one for all payments and linear predictors.
Online Linear Regression with Paid Stochastic Features / N. Merlis, K. Jang, N. Cesa Bianchi (PROCEEDINGS OF THE ... AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE). - In: AAAI-26 Technical Tracks 29 / [a cura di] S. Koenig, C. Jenkins, M. E. Taylor. - [s.l] : AAAI Press, 2026 Mar 14. - ISBN 978-1-57735-906-7. - pp. 24370-24378 (( 40. AAAI Conference on Artificial Intelligence : Thirty-Eighth Conference on Innovative Applications of Artificial Intelligence : Sixteenth Symposium on Educational Advances in Artificial Intelligence : January 20 – 27 Singapore 2026 [10.1609/aaai.v40i29.39618].
Online Linear Regression with Paid Stochastic Features
K. Jang;N. Cesa Bianchi
2026
Abstract
We study an online linear regression setting in which the observed feature vectors are corrupted by noise and the learner can pay to reduce the noise level. In practice, this may happen for several reasons: for example, because features can be measured more accurately using more expensive equipment, or because data providers can be incentivized to release less private features. Assuming feature vectors are drawn i.i.d. from a fixed but unknown distribution, we measure the learner's regret against the linear predictor minimizing a notion of loss that combines the prediction error and payment. We first study the case in which the mapping between payments and noise covariance is known and prove order-optimal regret bounds in the interaction length (up to log-factors). We then derive order-optimal bounds also when the noise covariance is unknown and prove that the regret rate is worse than the case of known covariances. Our analysis leverages matrix martingale concentration, showing that the empirical loss uniformly converges to the expected one for all payments and linear predictors.| File | Dimensione | Formato | |
|---|---|---|---|
|
39618-Article Text-43709-1-2-20260314.pdf
accesso aperto
Tipologia:
Publisher's version/PDF
Licenza:
Altro
Dimensione
183.38 kB
Formato
Adobe PDF
|
183.38 kB | Adobe PDF | Visualizza/Apri |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.




