iconLogo
Published:2026/1/7 2:45:37

オンライン学習、情報制限下でもイケる!✨

  1. タイトル & 超要約(15字以内) 情報少ない時でも賢く学習!IT業界に革命☆

  2. ギャル的キラキラポイント✨ ×3

    • ● 情報少ないときでも、良い感じに学習できるってすごくない?😎
    • ● リアルタイムデータ分析(リアルタイムデータぶんせき)とか、未来感ヤバくない?🚀
    • ● メモリ(めもり)容量少なくてもOK!コスパ最強!💰
  3. 詳細解説

    • 背景 ビッグデータ時代(じだい)到来!でも情報(じょうほう)全部使えるわけじゃないじゃん?😥 メモリも限られてるし…そんな状況(じょうきょう)でも、賢く学習できる方法(ほうほう)を研究してるんだって!

    • 方法 スライディングウィンドウモデルってのがあって、新しい情報が大事な時に役立つらしい!🧐 専門家(せんもんか)のアドバイス(あどばいす)を参考にしつつ、限られた情報とメモリで、効率(こうりつ)よく学習するアルゴリズムを開発!

続きは「らくらく論文」アプリで

Online Learning with Limited Information in the Sliding Window Model

Vladimir Braverman / Sumegha Garg / Chen Wang / David P. Woodruff / Samson Zhou

Motivated by recent work on the experts problem in the streaming model, we consider the experts problem in the sliding window model. The sliding window model is a well-studied model that captures applications such as traffic monitoring, epidemic tracking, and automated trading, where recent information is more valuable than older data. Formally, we have $n$ experts, $T$ days, the ability to query the predictions of $q$ experts on each day, a limited amount of memory, and should achieve the (near-)optimal regret $\sqrt{nW}\text{polylog}(nT)$ regret over any window of the last $W$ days. While it is impossible to achieve such regret with $1$ query, we show that with $2$ queries we can achieve such regret and with only $\text{polylog}(nT)$ bits of memory. Not only are our algorithms optimal for sliding windows, but we also show for every interval $\mathcal{I}$ of days that we achieve $\sqrt{n|\mathcal{I}|}\text{polylog}(nT)$ regret with $2$ queries and only $\text{polylog}(nT)$ bits of memory, providing an exponential improvement on the memory of previous interval regret algorithms. Building upon these techniques, we address the bandit problem in data streams, where $q=1$, achieving $n T^{2/3}\text{polylog}(T)$ regret with $\text{polylog}(nT)$ memory, which is the first sublinear regret in the streaming model in the bandit setting with polylogarithmic memory; this can be further improved to the optimal $\mathcal{O}(\sqrt{nT})$ regret if the best expert's losses are in a random order.

cs / stat.ML / cs.DS / cs.LG