What 82,000 Apple Watch Wearers Are Teaching Scientists About Fitness and Health

Written by appleinc | Published 2026/03/11
Tech Story Tags: apple-watch-health-research | apple-wearable-data | health-data-study | healthkit-activity-tracking | heart-rate-exercise | wearable-health-dataset | physiological-monitoring | wearable-sensor-data

TLDRThe Apple Heart & Movement Study tracked health and activity data from 82,000+ U.S. participants using the Apple Watch and Apple iPhone. Researchers collected millions of sensor readings—including 1.1 million ECG recordings—alongside surveys and health records. The study shows how large-scale wearable data could unlock personalized exercise insights and improve understanding of links between activity, heart health, and disease risk.via the TL;DR App

Authors:

  1. James Truslow
  2. Angela Spillane
  3. Huiming Lin
  4. Katherine Cyr
  5. Adeeti Ullal
  6. Edith Arnold
  7. Ron Huang
  8. Laura Rhodes
  9. Jennifer Block
  10. Jamie Stark
  11. James Kretlow
  12. Alexis L. Beatty
  13. Andreas Werdich
  14. Deepali Bankar
  15. Matt Bianchi
  16. Ian Shapiro
  17. Jaime Villalpando
  18. Sharon Ravindran
  19. Irida Mance
  20. Adam Phillips
  21. John Earl
  22. Rahul C. Deo
  23. Sumbul A. Desai
  24. Calum A. MacRae

Abstract

Physical activity or structured exercise is beneficial in a wide range of circumstances. Nevertheless, individual-level data on differential responses to various types of activity are not yet sufficient in scale, duration or level of annotation to understand the mechanisms of discrete outcomes nor to support personalized recommendations. The Apple Heart & Movement Study was designed to passively collect the dense physiologic data accessible on Apple Watch and iPhone from a large real-world cohort distributed across the US in order to address these knowledge gaps.

Longstanding associations of exercise with lower incident disease rates for many disorders have been replicated in large studies with research-grade, as well as consumer-grade, wearables1,2,3. Differential individual responses to exercise are emerging as potential predictors of clinical outcomes in many disorders including diabetes, hip fractures, cancer and rates of cognitive decline4,5,6,7.

Tailored medical advice on activity remains variable, largely as a result of the limited scope of interventional studies to date, and as a consequence of the wide range of exercise capacity and the heterogeneity of responses to comparable activities1. Evidence suggests that most individuals do not meet population recommendations for activity, and it has been proposed that more individualized recommendations might be more effective8. Better understanding of the relationships between specific activities and individual physiologic adaptation requires granular documentation of the attributes of different activities and their effects across a range of individuals with discrete response outcomes8,9,10. The convergence of wearable technologies, electronic health records, and modern data science makes such studies feasible.

Apple Watch (Watch) is a multi-sensor wearable which combines passively tracked physiologic metrics (e.g. activity, gait, and heart rate metrics) and incorporates ‘at the wrist’ annotation of events through user-input such as logging workout types. The Apple Heart & Movement Study (AH&MS) was designed to enable longitudinal collection of sensor, activity and health data from individuals to explore the relationships between activity, wellness and health. The study makes possible the principled incorporation of complex physiological models established through deeper phenotyping (such as event follow-up) via Institutional Review Board (IRB)-approved, direct participant outreach.

The current manuscript describes the study design and baseline data from individuals who provided informed consent to participate. The study is ongoing, and data collection continues to evolve with the addition of new sensors, new questionnaires, and other data. Research app, the mobile application participants use to enroll and interact with the study, enables frequent modifications to the study (with IRB approval) facilitating adaptation to new circumstances, such as a pandemic or new data types. Data are time stamped and versioned (both hardware and software) so specific analyses can be framed within the relevant context and App changes can be controlled for. We highlight the core features of the study noting the utility of this approach to incorporate and complement more traditional study frameworks in health and wellness research.

Full methods and summary data from participants followed for at least the initial year are available online. The cohort in the current manuscript was observed until 2021-11-13, two years after the study launch, so that each participant has been observed for at least one full year, and no more than two years. After applying all selection criteria, the initial cohort consisted of 82,809 participants. The detailed characteristics of the participants are reported in Table 1 and Supplementary Table 6. The study cohort is 72% White, 74% male at birth, 74% self-identified as male with a mean age at enrollment is 39.3 years (± 13.1 years). 80% of participants are part-time or full-time employed, 62% college-graduate, 52% married. Current smokers make up 5.3% of the cohort. Mean Body Mass Index (BMI) is 28.4 kg/m2 (±6.5 kg/m2). The most common prevalent diseases reported by participants were allergies (26.0%), depression (26.0%), and anxiety disorders (24.1%) but despite the age of the study participants, other medical conditions are reported at notable rates (Table 2) with 61% of participants currently taking at least one medication (Supplementary Table 7).

To demonstrate the range of the HealthKit data shared by the cohort during a single week, we aggregated results over a representative 7-day period to average out weekly cycles in participant activity (Table 3). The most common activity type was walking, which was shared at least once by 20.0% of the cohort. A total of 25,304 (30.6%) people in the cohort shared at least one workout during the week of observation, averaging, among them, 6.54 workouts per person.

Workouts are a special case of exercise tracking, within the general class of HealthKit samples, many other types of which exist (Supplementary Table 8). Among these, the sample types that are most commonly shared are those generated by everyday Apple Watch-wear and passively collected by sensors and software native to Watch. Accordingly, step count, heart rate and stand hours are shared with the study during the specific week represented in about half of the cohort. Less commonly shared sample types include: ‘Mindfulness’ sessions (shared by 5.5% during the specific week), which record a mindful session that is guided by Watch but requires active participant engagement; and high heart-rate event (shared by 2.4% during the specific week), which is passively collected by Watch but which is not a frequent event for healthy participants. Other data supplied by connected third party sensors are less frequently shared. For comparison, participant confirmed workouts are included, when initiated by a participant or confirmed from an auto-detected workout.

The cohort included 66,752 people (80.6%) who, for at least one day in their initial year post-enrollment, had an Apple Watch capable of recording an ECG paired with Research app. A single-lead ECG can be recorded at any time through the ECG app by holding the watch crown for 30 seconds. Within this subset, there were 55,740 people (83.5% of those wearing an ECG-capable Watch) who recorded and shared an ECG in their first-year post-enrollment, for a total of 1,132,473 ECGs (see Table 4). 25,402 ECGs (2.2%) were classified as showing atrial fibrillation, representing 1641 participants (2.0% of the cohort (see Supplementary Table 9). Of these participants, 477 (29.1%) had reported that they were known to have atrial fibrillation, suggesting that symptoms were often a driver for the specific recordings.

The Apple Health app allows users to download clinical health records from participating institutions by signing into their healthcare provider’s portal and choosing to share Fast Healthcare Interoperability Resources (FHIR) data with HealthKit. To date, the proportion of participants who have been able to share these data types is modest (~10%) as a consequence of local FHIR compliance and the process required. In the current cohort, 8408 people shared at least one such record with our study.

To measure participant engagement with the study over time, we present two very basic indices which complement the more detailed reports of survey data and HealthKit sharing online. These are: 1) how often a participant’s Apple Watch shares the HealthKit sample type “Stand Hour” with Research app; 2) how often a participant’s Research app uploads any kind of data to Apple’s secure study servers. Figure 1 shows these two indices of participation on discrete time scales. Panel a shows the fraction of the cohort who did not participate on any given day post-enrollment (00:00:00 to 23:59:59 UTC). Panel b shows the fraction that never participates at any time after a given day post-enrollment and can be regarded as a measure of the incidence of cohort dropout over time, specifically a measure of the fraction that becomes indefinitely inactive, according to that index of participation. Thus, for both of these indices, the decline in daily participation over time is largely attributable to the accumulation of permanently inactive participants over time, and less a consequence of degradation of study engagement among active participants. For example, 44% of the cohort does not share a Stand hour sample on day 365 post-enrollment, but this is not much larger than the 34% of the cohort that has already stopped sharing Stand hour samples at all times after day 365.


Blue lines display the proportion of participants who have not contributed a Stand Hour sample, measured by their Watch. Red lines display the proportion of participants who have not uploaded any data at all from their Research app. Panel (a) shows the fraction of the cohort that do not participate in these two ways on a given day, post-enrollment. Panel (b) shows the fraction that no longer participates at any time after a given day, post-enrollment as a measure of the cumulative incidence of dropout from the study.

AH&MS also includes a series of 16 surveys sent to participants and outlined in Table 5. Except for the 5 surveys that are triggered by rare events detected by Watch, almost all the surveys have a participation rate greater than 70%. Figure 2 shows the response rates vs. time since enrollment for two surveys delivered with the highest frequency - the monthly Stress Scale survey and the quarterly Changes in Health survey. A decreasing trend of response rate over time is clearly visible in Fig. 2, starting at 69.55% and gradually dropping to 32.48% after one year. As expected, due to the burden of survey completion on participants, this one-year decline is larger in both absolute and relative terms than the decline in active users as measured by Research app uploads in Fig. 1 Panel b, which shows only 28% of the cohort becoming indefinitely inactive at one year. A similar decreasing trend of response rate over time is observed in the Changes in Health survey.

For the Stress Scale survey, percentage is calculated among 42,181-person sub-cohort since only people enrolled after 2020-05-01 are considered to avoid any frequency change in survey delivery. For Changes in Health survey, the whole 82,809-person cohort is considered.

AH&MS is similar in scale to several other studies of wearable data and has already demonstrated the potential to overcome many of the constraints in prior studies of activity and physiologic adaptation1. The combination of multiple independent sensors, granular ‘at the wrist’ annotation of physiologic data, and the access to health conditions and health outcomes within Apple’s ecosystem is distinctive and brings high dimensionality to a large cohort without discrete medical indications. AH&MS enables the longitudinal investigation of a broad range of validated physical performance attributes combined with dense contextual metadata and extensive outcome metrics. The sustained survey engagement of participants is considerably improved from prior studies of serial health questionnaires. Among the features which are likely to prove of greatest utility are dynamic recruitment strategies, the passive collection over time and trending of validated reference biometrics (such as VO2max, heart rate recovery, etc), a rapidly modifiable study App11,12, the availability of participant generated high-resolution annotation, and consented access to clinical health record and claims data.

We anticipate that the AH&MS study design will directly address the need to understand relationships between longitudinal real-world wearable trajectories and the cross-sectional clinical or biological research data collected in typical biomedical research studies. Many existing epidemiologic cohorts lack longitudinal objective data on core lifestyle characteristics, while collecting rigorous outcome data for specific causes of mortality and morbidity. AH&MS collects and trends extensive behavioral and activity data elements, in addition to expanded demographic, anthropometric and general lifestyle data elements from HealthKit. Combining these datatypes with research sensor and usage data allows for unique exposure and outcomes assessments.

AH&MS has been designed to allow direct comparisons with traditional epidemiologic studies and randomized clinical trials through shared minimal common datasets. The ability to contact individual participants also enables deep phenotyping based on structured sampling. Longitudinal trends in complex physiologic metrics can be compared to interpolated external measurements, which in the past have been measured in cross-sectional fashion in highly selected populations. Models trained on sub-cohorts that have been deeply phenotyped can then be deployed across the entire AH&MS or in any traditional clinical study format.

The availability with Apple Watch of real-time ‘at the wrist’ annotation of activities and associated passive recording offers a granular ground truth to AH&MS. Research app is highly configurable to enable rapid modification and to accommodate emerging or secondary research questions such as the study response to the SARS Co-V-2 pandemic.

AH&MS also has ongoing access to consented data from participant clinical health records and data from “Blue Button” surfacing of records through FHIR application programming interfaces (APIs), and both these data types are accumulating after initial delays due to low healthcare utilization during COVID. As the process of incorporating EHR data into AH&MS is simplified, we are also exploring approaches to contemporaneous validation of specific incident diagnoses for the study. These datasets will allow the integration of long-term wearable trajectories with both prevalent and incident healthcare data, laying the foundation for continuous or semi-continuous data trajectories from wellness to disease.

The study has several limitations which must be considered. The study is limited to Apple iPhone and Apple Watch users, and there are limits regarding generalizability to other populations. Naturalistic study design, in both sampling and in analysis, may introduce discrete confounding as may the ongoing addition of new algorithms or new sensors. The loss to follow-up estimated from monthly survey response rates is higher than other retention metrics, though the survey characteristics and delivery cadences have not yet been optimized for response or retention rates. Other forms of missing data are prevalent, and though Watch wear information can assist in interpretation, this missingness must be accounted for in any analyses. These challenges include timing of data loss with respect to phone upgrade, the characteristics of participation prior to study drop-out, frequency of contact, survey length, prior survey completion and many others. The patterns of dropout in the initial year of the study skew participation further in the direction of initial recruitment biases, emphasizing the need for systematic approaches to drive representativeness in participant recruitment and retention. AH&MS extends age, gender, racial and ethnic diversity, but remains incompletely representative and we have added quantitative strategies to both recruit and retain relevant populations. These include local and social media campaigns and friend or family member recruitment. Female study recruitment has steadily grown over time and retention of this demographic is high. We anticipate much more granular understanding of study dropout mechanisms and their prevention as the study progresses.

Understanding physiologic responses to external challenges offers systems-level information on the physiologic set points of the individual and has been shown to lead to much more rigorous discrimination of intrinsic differences between individuals than passive cross-sectional measurement. These details highlight the potential to enable much more specific recommendations on optimal activity patterns for the individual user. The current study has been designed to lay the foundation for a broader and deeper interrogation of health and fitness metrics and to relate these parameters to outcomes in health and fitness.

Methods

Study design

This is a mobile application-based longitudinal cohort study involving the collection of sensor, survey, and health data. Participants were informed about the study through IRB approved materials, including study websites managed by Apple, the American Heart Association (AHA) and Brigham and Women’s Hospital. Participants are asked to complete a series of surveys and to consent to sharing data collected from their Health app (which uses the HealthKit framework) and sensor data obtained directly from iPhone and Apple Watch (Supplementary Table 1). The Research app framework enables participants to opt into and out of sharing specific types of health data with the study. The study began enrolling on November 14, 2019 with a goal of enrolling up to 500,000 participants. The planned duration of this study is 5 years, that is until November 2024, with a potential for extension or additional long-term follow-up. The study was approved by the Advarra Central Institutional Review Board (PRO00036784) and registered to ClinicalTrials.gov (ClinicalTrials.gov Identifier: NCT04198194). There is no compensation for participation.

Survey questions were designed to enable comparison with data from US health studies of similar scale, such as National Health and Nutrition Examination Survey (NHANES) developed by the Centers for Disease Control and Prevention13 and the All of Us Study, a large research program sponsored by the National Institutes of Health14. Questions were modified to support delivery within the mobile app user interface and were standardized, where relevant, across all three of the simultaneously launched Apple studies in 2019, including the Apple Women’s Health Study11 and the Apple Hearing Study12.

The Research app user interface was designed to be simple and intuitive while enabling data collection, through tasks such as survey completion, to be distributed over time, so as to reduce participant burden. The estimated time demand for survey responses was approximately 30 min in the first month, with 10 min per month during ongoing participation as additional surveys are triggered based on responses or the need for additional data collection (such as data on COVID exposures).

This study was designed with participant privacy in mind. Participant data are coded and encrypted while in transit and at rest. Coded data are stored in a system designed to meet the technical safeguard requirements of the Health Insurance Portability and Accountability Act (HIPAA). To maximize participant privacy and the confidentiality of health data and to minimize the risk of unauthorized access to participant data, a discrete workflow was created which enables Apple access to coded study data while restricting access to identifiable information such as name and contact information to a limited number of authorized staff at BWH. This access also allows contact between BWH and study participants through a discrete workflow which is inaccessible to Apple.

Similar to the Apple Women’s Health Study and Apple Hearing Study11,12, eligibility criteria include access to an iPhone with Research app installed, comfort communicating in written and spoken English, residence in the United States, aged at least 18 years old (at least 19 years old in Alabama and Nebraska, at least 21 years old in Puerto Rico), unique use of iCloud account or iPhone, and willingness to provide informed consent to participate in the study. An additional requirement for AH&MS includes use of an Apple Watch (Series 1 or later) paired with an iPhone at the time of enrollment.

Prior to enrollment, participants complete a profile which includes information such as name, date of birth, email, phone number, current region and state of residence. These data are used to confirm eligibility. For AH&MS, Research app is also able to confirm if an Apple Watch is paired to the iPhone. If the requirements for age, location, and Watch pairing status are met, individuals are able to continue to study onboarding, including reading and signing the informed consent form (ICF), HIPAA Authorization, and California Bill of Rights (if applicable). Since study launch, there have been revisions to survey frequency and to questions within the surveys, introduction of new surveys, and updates to the ICF.

Once onboarding is completed, newly enrolled participants are immediately allocated two “tasks”, specifically, the demographic survey along with a brief guide “Using your Apple Watch to Contribute” explaining that both logging workouts on Watch and taking ECGs (for those with Watch Series 4 or higher) are valuable to the study.

Ongoing recruitment

With institutional review board (IRB) approval, a study website hosted by Brigham and Women’s Hospital/Harvard Medical School was also made public. The American Heart Association (AHA) simultaneously launched an informational website to spread awareness of the study (November 2019), including IRB-approved social media and email campaigns (October 2020). The AHA also launched a website on heart.org to raise awareness of the study and broaden ongoing recruitment efforts. In addition, a new Research app feature was introduced and launched in October 2020 to allow updates to be sent directly to participants in the app to encourage continued participation, maximize participant engagement, provide study insights, and increase recruitment efforts.

Survey data—at enrollment and annually thereafter

On enrollment, participants in AH&MS responded to the Research Profile Survey within the Research app and to the Demographics Survey. The data included the year of birth, state of residence, race and ethnicity, marital status, employment status, education level, gender identity, sex assigned at birth, and subjective social status. Within the first month after enrollment participants received the following surveys: a Risk of Falling survey (based on STEADI survey, first 12 questions)15, a Medical History survey, a Medications survey, and a Health Behaviors survey (based on the Alcohol Use Disorders Identification Test (AUDIT-C)16 and All of Us study, NIH)14 delivered through Research app. The Activity Status survey, a questionnaire related to physical activity, was administered at the beginning of month two. Annual surveys are delivered on a staggered timeline to reduce participant burden and anticipated to take approximately 5 min each. Each survey expires approximately 28 days after delivery, except the Demographics survey, which never expires.

Scheduled interim surveys and timing

The participants also received quarterly surveys, specifically, the Mental Health survey (PHQ-2, GAD-2)17, the Activity Status survey (Modified Rosow-Breslau)18, the Perceived Stress Scale survey (PSS-4)19, the Disability Assessment survey (WHODAS 2.0)20, and reported outcomes related to changes in health in the Changes in Health survey. These surveys are also staggered across months two, three, and four, with an estimation of 10 min for survey completion each month.

Triggered surveys and timing

The study design also includes surveys sent to participants only if they meet certain criteria occurring during the period they are enrolled in the study. There were 3 triggered surveys at study launch related to sensor data observed on Apple Watch, with an estimated 5 min required to complete each survey.

The Irregular Rhythm Follow-up survey is administered 3 months after a participant receives an irregular heart rhythm notification while wearing Watch versions that support this functionality, which Watch can detect passively. The Irregular Rhythm Follow-up survey is designed to understand what participants did in response to the notification and is limited to a single administration every 90 days if the trigger criterion is met.

The second triggered survey, called ECG Follow-up, is administered 3 months after a participant receives an atrial fibrillation result during use of the Apple Watch ECG feature. The questions are related to any actions taken and care received after the result is given to the participant. This is designed to only be administered once to each participant whose ECG exhibits evidence of atrial fibrillation.

The Potential Fall survey, which is triggered when Apple Watch detects motion signatures that suggest the participant has experienced an impact, such as a hard fall. This questionnaire is delivered the day after the fall event is detected by the Research app and is designed to verify that the event was a fall, capture the participant’s activities during the fall, and assess if there were resulting injuries. To reduce the burden for participants who fall frequently, the study limits the number of Potential Fall surveys to 4, for events classified as falls with high probability, and limits delivery to 1 survey per month for events classified as falls with low probability.

Participant-approved follow-up to triggered surveys by study staff

A workflow was implemented to collect more detailed information from participants who agree to be contacted regarding events, such as a potential fall, that meet protocol-defined criteria. To contact participants, BWH staff use a secure mobile application and workflow to access contact information including name, phone, and email that is not accessible by Apple. Using an IRB- approved script, answers to detailed questions regarding the event are logged in a formatted structure, that is reviewed for accuracy and removal of any personally identifiable data and then aggregated with sensor and survey data.

Research sensor and usage data

Participants consented to the collection of a set of data (approved by the participant) of derived metrics retrieved from iPhone sensors and a paired Apple Watch. These data included ECG details, heart rate via the optical sensor (PPG), elevation (barometric pressure), motion (accelerometer, gyroscope), speed and distance (derived from GPS) and other sensor data such as the “on-wrist state,” pedometer data, and fall statistics summarized in Supplementary Table 2. Once enrolled, participants can opt into or out of sharing specific types of data with the Study using controls accessible within Research app.

HealthKit data

Using the HealthKit framework to collect both passively and manually added data types which participants have consented to share, provides a central repository for health and fitness data on iPhone and Apple Watch. Under such permission, specific apps write and read data using HealthKit which in turn can be accessed and shared with Research app while maintaining participant privacy and control. HealthKit stores data merged from multiple sources and contains data types such as heart rate, work out data, sleep analysis, and clinical health records (lab tests, diagnoses) from clinical interfaces. For this initial descriptive analysis, we have provided sampling from a randomly chosen (and typical) week to demonstrate the extent of the data collected.

Participation

This study was designed to allow participation in the following complementary ways: (1) response to survey questions in Research app; (2) contribution of HealthKit data; (3) contribution of sensor and usage data, which include sensor-based data streams from SensorKit (SK); and (4) response to direct outreach from study staff if specific IRB-approved criteria are met. In general, a participant is considered to be actively participating when contributing data through any of these methods.

Defining demographic variables

We used the same questionnaire that was used to determine the race and ethnicity in the NIH-sponsored All of Us Study and we classified responses into traditional reporting of race and ethnicity. All three of the Research app studies used the MacArthur Scale of Subjective Social Status21. This scale has been observed to correlate with health status across the lifespan. Notably, the MacArthur Scale is correlated with objective socioeconomic status (SES) but has the benefit of broader applicability as a marker of social status than simple objective measures of SES in non-White populations. We arbitrarily categorized the responses into the following categories: 1 to 4 corresponding to low, 5 to 6 corresponding to middle, and 7 to 10 corresponding to high.

Planned statistical analyses

Statistical analyses are planned for the following 3 categories of study: (1) longitudinal analyses of survey data, (2) longitudinal analyses of passively collected iPhone and Apple Watch data, and (3) analyses of associations between the survey data and passively collected data including clinical health record data. In each category, we will perform exploratory descriptive analyses and formulate more specific hypothesis-driven models.

For all 3 types of analyses, we will use longitudinal extensions of regression methods, such as linear and generalized linear mixed models, statistical learning techniques for high-dimensional data, and functional data analysis methods. For the longitudinal analysis of the survey data, we will quantify the associations among both participant characteristics and risk factors and specific functional outcomes, and how these associations vary across the age range of the study population. For the longitudinal analysis of passive data, we will perform individual-level analyses to identify the possible change points in behaviors over time and how they relate to subsequent health outcomes. For the longitudinal analysis of passive and survey data, the predictors will initially be the daily summary statistics derived from passively collected data and the outcomes of interest will consist of all the items on which survey data are available. We will train machine learning models on objective outcomes defined within the study itself and use these models to classify specific time to event trajectories.

Participants who enrolled but failed to submit the Demographics survey are excluded from the cohort studied in this article. A Welch two-sample, two-sided t-test was performed to compare mean age at enrollment between included and excluded participants. A Pearson’s chi-squared test with simulated p-values was performed to compare the distribution of geographic regions between included and excluded participants. This analysis was performed using R version 3.6.0 (base R).

Data

Baseline Characteristics of Participants

We present characteristics, measured as close to enrollment as possible, for the cohort of study participants who enrolled in the study in its first year, 2019-11-14 through 2020-11-13. The cohort in the current manuscript was observed until 2021-11-13, two years after the study launch, so that each participant has been observed for at least one full year, and no more than two years.

We did not include the following: (1) data used to test Research app for quality assurance purposes (n = 29); 2) cases where eligibility became ambiguous after enrollment (e.g. participant modified dates of birth or address after enrollment to imply age less than local age of consent (n = 100); or, (3) participant did not complete the Demographic survey after enrollment (n = 1751). After applying all selection criteria, the initial cohort as of 2020-11-13 consists of 82,809 participants.

For the characteristics reported in Table 1, the value represented for each participant is the earliest value recorded by Research app for that participant and characteristic. Multiple values may occur when a participant is enrolled for long enough that Research app presents them with a survey for a second or third time, for example a 1st annual Demographics survey and a 2nd annual Demographics survey. Participants are also able to edit, at any time, their date of birth and place of residence in the profile maintained for them by Research app. In cases where a participant edits this information after enrollment, we present only the earliest value that they share. In the event that a participant edits their data, an additional eligibility check is run and the individual may be removed if ineligible (e.g. moves to a new state that has a higher minimum age limit for participation).

Comparison of this 82,809-person cohort to the 1751 participants who did not respond to a Demographic survey is performed in Supplementary Table 34 and 5 on the basis of participants’ age and state of residence at enrollment (these data are obtained by Research app prior to enrollment, regardless of whether a participant submits the Demographics survey). The 1751 excluded participants are 3.2 years younger than the cohort, on average (95% CI = [3.79 y to 2.58 y]).

The study cohort is 72% White, 74% male at birth, 74% self-identified as male (Table 1). Mean age at enrollment is 39.3 years (± 13.1 years). 11% of the cohort is Hispanic; details of the racial makeup of this Hispanic population are in Supplementary Table 6. 80% of participants are part-time or full-time employed, 62% college-graduate, 52% married. Current smokers make up 5.3% of the cohort. Mean BMI is 28.4 kg/m2 (±6.5 kg/m2).

2,684 participants (3.24% of the cohort) withdrew from the study within one year of enrollment. Among those who withdrew, 25% withdrew in their first 13 days, and 50% withdrew within 111 days. Among those who withdrew < 1% were automatic withdrawals triggered by an update to a participant’s state of residence or date of birth that rendered them no longer eligible to continue in the study. More details about withdrawal rates can be found in Supplementary Fig. 1.

The most common prevalent diseases reported by participants were allergies (26.0%), depression (26.0%), and anxiety disorders (24.1%) (Table 2), but despite the relatively young age of the study participants, they reported many other medical conditions at notable rates.

Among all participants in the cohort, 61% report currently taking at least one medication (Supplementary Table 7). The most commonly reported medications were NSAIDs (27%), antidepressants (20%), and either ACEIs or ARBs (11%).

Health-data sharing in a single week

To demonstrate the variety of the HealthKit data shared by the cohort during a single week, we aggregated results over a 7-day period to average out weekly cycles in participant activity (data not shown). We chose the final week in our observation period (2021-11-07 through 2021-11-13) in particular since all participants would have been enrolled for at least one year at that point, and since—at two years after study launch—it was the point closest to the middle of the study’s 5-year period. Comparison of this week to 145 other weeks between 2019-11-14 and 2022-09-01 established that the period chosen is representative of a typical week (See Supplementary Fig. 2A and 2B).

Table 3 shows the most common of the 82 types of workout sample logged into HealthKit and shared with the Research app by at least 100 participants. For each activity type, the table gives the number of participants who shared at least one of that sample type in that week and gives the average number of activities of that type per participant (among those who performed that activity). The most common activity type was walking, which was shared at least once by 20.0% of the cohort. A total of 25,304 (30.6%) people in the cohort shared at least one workout during the week of observation, averaging, among them, 6.54 workouts per person.

Workouts are a special case of exercise tracking, within the general class of HealthKit samples, many other types of which are listed in Supplementary Table 8, which presents HealthKit data from the week starting 2021-11-07 with 100 or more participants contributing to each data type. This sample is restricted to year 1 enrollees who were active during this specific week in year 2. Among these, the sample types that are most commonly shared tend to be those that are generated by everyday Watch-wear and that are passively collected by software and sensors native to Watch. Accordingly, step count, heart rate and stand hours are shared with the study during the specific week represented by Supplementary Table 8 by about half of the cohort. Less commonly shared sample types include: ‘Mindfulness’ sessions (shared by 5.5% during the specific week), which record a mindful session that is typically guided by Watch but which requires active participant engagement; and high heart-rate event (shared by 2.4% during the specific week), which is passively collected by Watch but which is not a frequent event for healthy participants. Other data supplied by connected third party sensors, for example blood glucose (shared by 1.1% during the specific week), are less frequently shared.

For comparison, participant confirmed workouts are included, when initiated by a participant or confirmed from an auto-detected workout. This attribute, and the fact that structured physical exercise is not a frequent part of everyday activities, means that our dataset contains many fewer workout samples than, for example, heart rate samples or step count samples.

ECGs

The cohort included 66,752 people (80.6%) who, for at least one day in their initial year post-enrollment, had an Apple Watch capable of recording an ECG paired with Research app. A single-lead ECG can be recorded at any time through the ECG app by holding the watch crown for 30 s. Within this subset, there were 55,740 people (83.5% of those wearing an ECG-capable Watch) who recorded and shared an ECG in their first-year post-enrollment, for a total of 1,132,473 ECGs (see Table 4). During the defined data collection period for this cohort, there were two tasks that encouraged participants to take an ECG. For all participants with a capable Watch, one task is presented at enrollment that encourages taking an ECG and an additional bi-weekly task was added in April 2021 that continued through January 2022 to take an ECG and record the result via multiple choice response.

25,402 ECGs (2.2%) were classified as showing atrial fibrillation, representing 1641 participants (2.0% of the cohort). Other classifications are shown in Supplementary Tables 9 and 10. This collection period includes both ECG version 1 and ECG version 2, which became available on WatchOS 7.2 and iOS 14.3, originally released in December 2020, with expanded ECG classification capability.

Clinical health records

The Apple Health app allows users to download clinical health records (FHIR format) from participating institutions by signing into their healthcare provider’s portal and choosing to share FHIR data with HealthKit. Study participants may elect to share this data with our study. To date, the proportion of participants who have been able to share these data types is modest (~10%) as a consequence of local FHIR compliance and the process required. In the cohort, 7757 people shared at least one such record with our study in their initial year post-enrollment.

Measures of participation over time

To measure participant engagement with the study over time, we present two very basic indices which complement the more detailed reports of survey data and HealthKit sharing above. These are: (1) how often a participant’s Apple Watch shares the HealthKit sample type “Stand Hour” with the Research app; (2) how often a participant’s Research app uploads any kind of data to Apple’s secure study servers.

Each Stand Hour sample is an estimate by Apple Watch of whether or not the participant has stood and moved for at least 1 min during a given hour of the day. If Watch is not on the wrist and powered on, then no sample is created.

We use the presence of Stand Hours as a proxy for Apple Watch wearing, since this parameter is passively collected and because Watch logs an indication of “Stood” or “Idle” each hour the participant is wearing the device. We interpret the absence of Stand Hour samples on a given day to mean that the participant was not wearing Watch that day. We note, however, that at least two other conditions might result in the absence of a Stand Hour sample: (1) the data upload path from Watch to the Research app and then to the AH&MS servers was not active (e.g. lack of connectivity); or (2) the participant has opted out of sharing Stand Hours with the Research app after enrollment, a user setting which it is not possible to directly ascertain but which our data suggest is unusual for current study participants.

Our second definition of participation is based on a more modest requirement: that a participant’s Research app has uploaded any data to the study servers on any given day. Such an upload might represent a Stand Hour sample or other health and sensor data, but it also might only represent low-level operations of Apple Watch, iPhone, and study servers, such as in a regularly scheduled check-in between Research app and the servers. The presence of one of these uploads shows that the participant has Research app installed on their iPhone and that their iPhone is connected to the internet.

Figure 1 shows these two indices of participation on discrete time scales. Panel a shows the fraction of the cohort who did not participate on any given day post-enrollment (00:00:00 to 23:59:59 UTC). Panel b shows the fraction that never participates at any time after a given day post-enrollment and can be regarded as a measure of the incidence of cohort dropout over time, a measure of the fraction that becomes indefinitely inactive, according to that index of participation (as of the time of writing, 2022-12-01). Note that the observation window for Fig. 1 is long enough that the entire cohort has been enrolled for at least one year, but that participation after one year is not shown. The denominator in every fraction is constant: 82,809 participants.

In Fig. 1a, the fraction of the cohort whose Research app does not upload any data on day 0 is very low, around 1%, and is still relatively low (around 38%) at one-year post-enrollment. The fraction of the cohort which do not share Stand hours follows the same trend over time, but is slightly higher at all times post-enrollment, starting at 5% and increasing to 44%. The lower rate of sharing Stand hours reflects the additional requirements that the participant wears their Watch and enables sharing of Stand hours with Research app (as well as the requirement that they stand and move for 1 min that day).

Figure 1b shows a closely related trend. The fraction of the cohort whose Research app has stopped uploading any data on day 0 is <1%, increasing to 28% at one year post-enrollment. The fraction of the cohort who has indefinitely stopped sharing Stand hours on day 0 is 2%, increasing to 34% at one-year post-enrollment.

Thus, for both of these indices of participation, the decline in daily participation over time is largely attributable to the accumulation of permanently inactive participants over time, and less attributable to a degradation of study engagement among active participants. For example, 44% of the cohort does not share a Stand hour sample on day 365 post-enrollment, but this is not much larger than the 34% of the cohort that has already stopped sharing Stand hour samples at all times after day 365.

Survey responses

Table 5 shows the percentage of participants in the cohort who complete at least one survey in their first year following enrollment, for each of the 16 survey types. Except for the 5 surveys that are triggered by rare events detected by Watch, almost all the surveys have a participation rate greater than 70%.

Figure 2 shows the response rates, vs. time since enrollment for two surveys delivered with the highest frequency—the monthly Stress Scale survey and the quarterly Changes in Health survey.

For the Stress Scale survey, only participants enrolled after 2020-05-01 are considered in order to avoid changes in delivery frequency when this initially quarterly survey became monthly after May 2020. This leads to 42,181 (the denominator for the response rate) participants. The time window we consider here extends from the individual’s enrollment date to the 12th Stress Scale survey expiration date. The scheduled delivery of this monthly survey is as follows: the first Stress Scale survey is delivered on the first Sunday of the month after enrollment and all following surveys are delivered on the first Sunday of subsequent months. Each survey expires 28 days after delivery. A decreasing trend of response rate over time is clearly visible in Fig. 2, starting at 69.55% and gradually dropping to 32.48% after one year. As expected, due to the burden of survey completion on participants, this one-year decline is larger in absolute and relative terms than the decline in active users as measured by Research app uploads in Fig. 1b, which shows only 28% of the cohort becoming indefinitely inactive at one year.

For the Changes in Health survey, the entire cohort is considered, and the time window represented by Fig. 2 extends from enrollment to 400 days later to ensure that only the first 4 quarterly surveys for each participant are counted. The quarterly surveys were distributed on the 3rd, 6th, 9th, 12th months post-enrollment. As with the Stress Scale survey, we observe a decreasing trend of response rate over in Changes in Health survey, starting at 60.69% and dropping to 34.06%.

Changes in Health survey results

The changes in Health survey is designed to monitor various key events relevant to health, including new medical conditions, changes in medications, new injuries, and major lifestyle changes. There were 56,553 (63.8%) participants who completed at least one changes in Health survey.

Table 6 reports the number of participants indicating a change in health by three broad categories: new medical conditions, new surgical procedures, or other changes. Since the survey prompts a participant to give the date of the change, Table 6 displays different totals computed according to whether or not the participant gave a date of the change, and whether or not the date of the change coincided with their time in study. If an event is dated after enrollment, but prior to the quarterly period queried by the survey in which it is reported, then we do not exclude it from any of the counts in Table 6. Dates after the completion of the survey are coded as missing.

After excluding events reported with no date and excluding events dated before enrollment, events in the category “Other Changes” were reported by the greatest number of people (n = 44,333). Events in categories “Medical Conditions” and “Surgical Procedures” were reported by fewer participants (n = 1443 and n = 1294, respectively).

Table 7 displays these events in more detail. All new medical conditions were reported with low frequency (<1% of the cohort). The most common reported new medical condition was arthritis (0.5%). New surgical procedures were also reported at low frequency. The most common new surgical procedure was “other bone surgery” (1.1%). The most commonly reported change overall was “change of insurance”, reported by 16.6% of the cohort in their first year. Additional events included in the Changes in Health survey but not shown in Table 7 were the following: a new or continued pregnancy, other medical emergencies, newly diagnosed pre-diabetes or impaired glucose tolerance, regularly smoking cigarettes, side effect of any medication or drug, or change in lower limb arthritis. Participants reporting respiratory problems were provided with follow-up survey questions to assess the duration and severity of their respiratory problem. Similarly, participants reporting a new or continued pregnancy were provided with follow-up survey questions to assess how far along they were in their pregnancy or the outcome of their pregnancy (vaginal delivery, Cesarean section, miscarriage, or other). Participants reporting a broken bone, accident, or trauma were provided with follow-up survey questions assessing which part of the body was affected or if they needed an assistive device (cane, crutches, leg braces, prosthetics, scooter, walker without wheels, wheelchair, or other assistive device).

Potential fall surveys and follow-up participation

Participants in this cohort submitted 2055 survey responses that met protocol defined criteria for follow-up before 2021-11-14, representing 1735 participants. The study received consent to follow up with 1392 surveys by phone, representing 1179 participants. Callers reached a participant in 967 of those cases, representing 829 participants (47% of the total eligible surveys, 48% of participants with eligible surveys).

Reference demography

Recruitment patterns during this initial period resulted in some skewing of the baseline demographics which should be considered in the context of the current report. The cohort was more likely to be male (74% vs. 49% of US), white (72% vs. 60% of US), and college educated (89% of the cohort vs. 62% of US older than age 25 with > 12 years education). Ongoing recruitment continues to move the study demographics toward the national mean (data not shown) and will be described in detail in subsequent manuscripts.

Compared to traditional epidemiology or disease cohorts at the time of enrollment, the current cohort is similarly skewed, but we anticipate that with ongoing recruitment and strategies designed to correct for representation, the cohort will continue to become more representative over time. Notably, AH&MS does not enforce an upper limit on participant age and the set of participant-shared data in AH&MS is extensive. For example, the cohort has contributed ~19,300 cycling workouts, ~14,700 running workouts, ~137,000,000 heart-rate samples, and ~57,800 VO2max estimates in just a single week.

Comparison of period 2021-11-07 – 2021-11-13 to other 7-day periods

In Supplementary Fig. 2A and Supplementary Fig. 2B we report a cross-sectional description of participant data sharing for the period 2021-11-07 through 2021-11-13, and note that this period was unremarkable compared to other 7-day periods before or after it.

We performed this comparison by collecting weekly counts of 6 data types shared by the cohort between 2019-11-14 and 2022-09-01. We compared the following three passively collected HealthKit sample types: stand hours, VO2,Max, and mindful breathing sessions. We also looked at 3 types of workouts actively annotated: walking, yoga, and traditional strength training.

From this 6-dimensional dataset, we computed the Mahalanobis distance, d, of our sample week from the mean of the other 145 weeks. Assuming that each of the six variables is normally distributed, the square of this Mahalanobis distance, d2, should be χ2-distributed with ν = 6. This allows us to test for a significant difference between our sample week and the other 145 weeks. We found d2 = 1.57. Since  we find no significant difference between our week and the average study week.

We also performed a 6-component PCA on the dataset and plotted the first 2 components to graphically demonstrate the distance of our sample week from the other 145 weeks. See Supplementary Fig. 2C.

Reporting summary

Further information on research design is available in the Nature Research Reporting Summary linked to this article.

Data availability

Data are not publicly available. Any request for data will be evaluated and responded to in a manner consistent with the specific language in the study protocol and informed consent form. Requests for data should be addressed to one of the corresponding authors (CAM).

Code availability

Computer code for all statistical analyses was written in Python and R and may be available for review upon request from one of the corresponding authors (C.A.M.). Any request for code will be evaluated and responded to in a manner consistent with policies intended to protect participant confidentiality and language in the study protocol and in the informed consent form.

References

  1. Arnett, D. K. et al. 2019 ACC/AHA guideline on the primary prevention of cardiovascular disease: executive summary: a report of the American College of Cardiology/American Heart Association Task Force on Clinical Practice Guidelines. J. Am. Coll. Cardiol. 74, 1376–1414 (2019).

  2. Khurshid, S. et al. Wearable accelerometer-derived physical activity and incident disease. NPJ Digit Med. 5, 131 (2022).

  3. Master, H. et al. Association of step counts over time with the risk of chronic disease in the All of Us Research Program. Nat. Med. 28, 2301–2308 (2022).

  4. Alemany, J. A., Delgado-Diaz, D. C., Mathews, H., Davis, J. M. & Kostek, M. C. Comparison of acute responses to isotonic or isokinetic eccentric muscle action: differential outcomes in skeletal muscle damage and implications for rehabilitation. Int J. Sports Med. 35, 1–7 (2014).

  5. Ross, L. M., Slentz, C. A. & Kraus, W. E. Evaluating individual level responses to exercise for health outcomes in overweight or obese adults. Front. Physiol. 10, 1401 (2019).

  6. Shigeta, T. T. et al. Cardiorespiratory and muscular fitness associations with older adolescent cognitive control. J. Sport Health Sci. 10, 82–90 (2021).

  7. Vidoni, E. D. et al. Dementia risk and dynamic response to exercise: a non-randomized clinical trial. PLoS ONE 17, e0265860 (2022).

  8. Ross, R. et al. Precision exercise medicine: understanding exercise response variability. Br. J. Sports Med. 53, 1141–1153 (2019).

  9. Neufer, P. D. et al. Understanding the cellular and molecular mechanisms of physical activity-induced health benefits. Cell Metab. 22, 4–11 (2015).

  10. Roberts, M. D. et al. Physiological differences between low versus high skeletal muscle hypertrophic responders to resistance exercise training: current perspectives and future research directions. Front. Physiol. 9, 834 (2018).

  11. Mahalingaiah, S. et al. Design and methods of the Apple Women’s Health Study: a digital longitudinal cohort study. Am. J. Obstet. Gynecol. 226, 545 e541–545.e529 (2022).

  12. Neitzel, R. L. et al. Toward a better understanding of nonoccupational sound exposures and associated health impacts: Methods of the Apple Hearing Study. J. Acoust. Soc. Am. 151, 1476 (2022).

  13. Chen, T. C., Clark, J., Riddles, M. K., Mohadjer, L. K. & Fakhouri, T. H. I. National Health and Nutrition Examination Survey, 2015-2018: sample design and estimation procedures. Vital-. Health Stat. 2, 1–35 (2020).

  14. All of Us Research Program, I. et al. The “All of Us” research program. N. Engl. J. Med. 381, 668–676 (2019).

  15. Lohman, M. C. et al. Operationalisation and validation of the Stopping Elderly Accidents, Deaths, and Injuries (STEADI) fall risk algorithm in a nationally representative sample. J. Epidemiol. Community Health 71, 1191–1197 (2017).

  16. Saunders, J. B., Aasland, O. G., Babor, T. F., de la Fuente, J. R. & Grant, M. Development of the alcohol use disorders identification test (AUDIT): WHO collaborative project on early detection of persons with harmful alcohol consumption-II. Addiction 88, 791–804 (1993).

  17. Ware, J., Jr. Kosinski, M. & Keller, S. D. A 12-Item Short-Form Health Survey: construction of scales and preliminary tests of reliability and validity. Med. Care 34, 220–233 (1996).

  18. Rosow, I. & Breslau, N. A Guttman health scale for the aged. J. Gerontol. 21, 556–559 (1966).

  19. Cohen, B. G., Colligan, M. J., Wester, W. 2nd & Smith, M. J. An investigation of job satisfaction factors in an incident of mass psychogenic illness at the workplace. Occup. Health Nurs. 26, 10–16 (1978).

  20. Andrews, G., Kemp, A., Sunderland, M., Von Korff, M. & Ustun, T. B. Normative data for the 12 item WHO Disability Assessment Schedule 2.0. PLoS ONE 4, e8343 (2009).

  21. Adler, N. E., Epel, E. S., Castellazzo, G. & Ickovics, J. R. Relationship of subjective and objective social status with psychological and physiological functioning: preliminary data in healthy white women. Health Psychol. 19, 586–592 (2000).

Acknowledgements

The researchers, the Sponsor, Apple Inc. and the American Heart Association wish to thank all of our study participants. Their generosity and willingness to contribute their time and health information have made this research possible.

This paper is available on nature under CC by 4.0 Deed (Attribution 4.0 International) license.


Written by appleinc | Apple designs consumer electronics, software, and services, known for iPhone, Mac, iOS, and an integrated ecosystem.
Published by HackerNoon on 2026/03/11