A Multi-Modal Dataset for Ground Reaction Force Estimation Using Consumer Wearable Sensors
arXiv:2603.28784v1 Announce Type: new Abstract: This Data Descriptor presents a fully open, multi-modal dataset for estimating vertical ground reaction force (vGRF) from consumer-grade Apple Watch sensors with laboratory force plate ground truth. Ten healthy adults aged 26--41 years performed five activities: walking, jogging, running, heel drops, and step drops, while wearing two Apple Watches positioned at the left wrist and waist. The dataset contains 492 validated trials with time-aligned inertial measurement unit (IMU) recordings (approximately 100 Hz) and force plate vGRF (Force\_Z, 1000 Hz). The release includes raw and processed time series, trial-level metadata, quality-control flags, and machine-readable data dictionaries. Trial-level matching manifests link recordings across mod
View PDF HTML (experimental)
Abstract:This Data Descriptor presents a fully open, multi-modal dataset for estimating vertical ground reaction force (vGRF) from consumer-grade Apple Watch sensors with laboratory force plate ground truth. Ten healthy adults aged 26--41 years performed five activities: walking, jogging, running, heel drops, and step drops, while wearing two Apple Watches positioned at the left wrist and waist. The dataset contains 492 validated trials with time-aligned inertial measurement unit (IMU) recordings (approximately 100 Hz) and force plate vGRF (Force_Z, 1000 Hz). The release includes raw and processed time series, trial-level metadata, quality-control flags, and machine-readable data dictionaries. Trial-level matching manifests link recordings across modalities using stable identifiers. Of the 492 validated trials, 395 are triad-complete, containing wrist, waist, and force plate data, enabling cross-sensor analyses and reproducible model evaluation. Dataset quality is characterised through a three-phase cross-sensor plausibility and consistency framework, repeatability analysis of peak vGRF (intraclass correlation coefficient 0.871--0.990), and systematic checks of force ranges and trial completeness. Monte Carlo sensitivity analysis showed that correlation-based validation metrics were robust to single-sample timing perturbations at the IMU sampling resolution. All data are released under CC BY 4.0, with analysis scripts archived alongside the dataset and mirrored on GitHub. This resource supports reproducible research in wearable biomechanics, benchmarking of machine learning models for vGRF estimation, and investigation of sensor placement effects using widely available consumer wearables.
Subjects:
Signal Processing (eess.SP); Artificial Intelligence (cs.AI)
Cite as: arXiv:2603.28784 [eess.SP]
(or arXiv:2603.28784v1 [eess.SP] for this version)
https://doi.org/10.48550/arXiv.2603.28784
arXiv-issued DOI via DataCite
Submission history
From: Parvin Ghaffarzadeh [view email] [v1] Thu, 19 Mar 2026 13:53:49 UTC (3,996 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
modelbenchmarkrelease
Average-Case Reductions for $k$-XOR and Tensor PCA
arXiv:2601.19016v2 Announce Type: replace-cross Abstract: We study the computational properties of two canonical planted average-case problems -- noisy planted $k$-XOR and Tensor PCA -- by formally unifying them into a family of planted problems parametrized by tensor order $k$, number of entries $m$, and noise level $\delta$. We build a wide range of poly-time average-case reductions within this family, across all regimes $m \in [1, n^k]$. In the denser $m \geq n^{k/2}$ regime, our reductions preserve proximity to the computational threshold, and, as a central application, reduce conjectured-hard $k$-XOR instances with $m \approx n^{k/2}$ to conjectured-hard instances of Tensor PCA. Additionally, we give new order-reducing maps at fixed densities (e.g., $5\to 4$ for $k$-XOR with $m \appro

Optimus: A Robust Defense Framework for Mitigating Toxicity while Fine-Tuning Conversational AI
arXiv:2507.05660v2 Announce Type: replace Abstract: Customizing Large Language Models (LLMs) on untrusted datasets poses severe risks of injecting toxic behaviors. In this work, we introduce Optimus, a novel defense framework designed to mitigate fine-tuning harms while preserving conversational utility. Unlike existing defenses that rely heavily on precise toxicity detection or restrictive filtering, Optimus addresses the critical challenge of ensuring robust mitigation even when toxicity classifiers are imperfect or biased. Optimus integrates a training-free toxicity classification scheme that repurposes the safety alignment of commodity LLMs, and employs a dual-strategy alignment process combining synthetic "healing data" with Direct Preference Optimization (DPO) to efficiently steer mo

PRISM: Robust VLM Alignment with Principled Reasoning for Integrated Safety in Multimodality
arXiv:2508.18649v2 Announce Type: replace Abstract: Safeguarding vision-language models (VLMs) is a critical challenge, as existing methods often suffer from over-defense, which harms utility, or rely on shallow alignment, failing to detect complex threats that require deep reasoning. To this end, we introduc PRISM (Principled Reasoning for Integrated Safety in Multimodality), a System 2-like framework that aligns VLMs through a structured four-stage reasoning process explicitly designed to handle three distinct categories of multimodal safety violations. Our framework consists of two key components: a structured reasoning pipeline that analyzes each violation category in dedicated stages, and PRISM-DPO, generated via Monte Carlo Tree Search (MCTS) to refine reasoning quality through Direc
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Releases

UniMark: Artificial Intelligence Generated Content Identification Toolkit
arXiv:2512.12324v3 Announce Type: replace Abstract: The rapid proliferation of Artificial Intelligence Generated Content has precipitated a crisis of trust and urgent regulatory demands. However, existing identification tools suffer from fragmentation and a lack of support for visible compliance marking. To address these gaps, we introduce the \textbf{UniMark}, an open-source, unified framework for multimodal content governance. Our system features a modular unified engine that abstracts complexities across text, image, audio, and video modalities. Crucially, we propose a novel dual-operation strategy, natively supporting both \emph{Hidden Watermarking} for copyright protection and \emph{Visible Marking} for regulatory compliance. Furthermore, we establish a standardized evaluation framewo

I just shipped my first major update to a Chrome extension. Here's what I changed and why.
Building in public means being honest about mistakes. Here's one I made with Prompt Helix and how I fixed it in v1.0.2. Prompt Helix is a Chrome extension that extracts webpage content and sends it directly to your chosen AI. No copy-pasting. No tab switching. Click, ask, get an answer in context. I launched it in February and have been iterating since. The mistake I made with the free tier. When I launched I gave away too much for free. OpenAI and Claude completely free with no daily caps. It felt generous and user-friendly. In reality it meant there was no reason to ever create an account or pay. Someone could install it and use it every day forever without seeing a single upgrade prompt. Classic freemium mistake. I only realised this when I looked at my Clerk dashboard and saw 60 instal


Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!