● 顔の動きとか、色んな情報(信号源)を組み合わせて、呼吸を測るよ!😲 ● 信号の良し悪し(品質)をチェックして、一番良い呼吸データを出すようにしてる!😎 ● 医療、エンタメ、セキュリティ…色んな分野で、新しいサービスとか作れちゃうかも!😳
背景 動画から呼吸数を測る技術は、めっちゃ色んなとこで使えそうじゃん?🤔でも、動画ってノイズ(雑音)がいっぱいあるから、正確に測るのが難しかったんだよね💦 この研究は、その問題を解決するために、信号の品質に注目したんだって!
方法 顔の色(rPPG)の変化、体の動き、AI(深層学習)を使った方法とか、色んな情報(信号源)を組み合わせるよ!✨それぞれの情報の信頼度をチェックして、一番良いデータを選んだり、ノイズを消したりするんだって!😳
続きは「らくらく論文」アプリで
Video-based respiratory rate (RR) estimation is often unreliable due to inconsistent signal quality across extraction methods. We present a predictive, quality-aware framework that integrates heterogeneous signal sources with dynamic assessment of reliability. Ten signals are extracted from facial remote photoplethysmography (rPPG), upper-body motion, and deep learning pipelines, and analyzed using four spectral estimators: Welch's method, Multiple Signal Classification (MUSIC), Fast Fourier Transform (FFT), and peak detection. Segment-level quality indices are then used to train machine learning models that predict accuracy or select the most reliable signal. This enables adaptive signal fusion and quality-based segment filtering. Experiments on three public datasets (OMuSense-23, COHFACE, MAHNOB-HCI) show that the proposed framework achieves lower RR estimation errors than individual methods in most cases, with performance gains depending on dataset characteristics. These findings highlight the potential of quality-driven predictive modeling to deliver scalable and generalizable video-based respiratory monitoring solutions.