Digital Medicine

: 2022  |  Volume : 8  |  Issue : 1  |  Page : 10-

Physician perceptions of surveillance: Wearables, Apps, and Chatbots for COVID-19

Alexandra R Linares1, Katrina A Bramstedt2, Mohan M Chilukuri3, P Murali Doraiswamy1,  
1 Department of Psychiatry and Behavioral Sciences, Duke University School of Medicine, Durham, USA
2 Department of Medicine, Bond University Medical Program, Queensland, Australia
3 Department of Family Medicine, University of North Carolina School of Medicine, Chapel Hill, USA

Correspondence Address:
Alexandra R Linares
Department of Psychiatry and Behavioral Sciences, Duke University School of Medicine, DUMC Box: 3018, Durham, NC 27710


Background and Purpose: To characterize the global physician community's opinions on the use of digital tools for COVID-19 public health surveillance and self-surveillance. Materials and Methods: Cross-sectional, random, stratified survey done on Sermo, a physician networking platform, between September 9 and 15, 2020. We aimed to sample 1000 physicians divided among the USA, EU, and rest of the world. The survey questioned physicians on the risk-benefit ratio of digital tools, as well as matters of data privacy and trust. Statistical Analysis Used: Descriptive statistics examined physicians' characteristics and opinions by age group, gender, frontline status, and geographic region. ANOVA, t-test, and Chi-square tests with P < 0.05 were viewed as qualitatively different. As this was an exploratory study, we did not adjust for small cell sizes or multiplicity. We used JMP Pro 15 (SAS), as well as Protobi. Results: The survey was completed by 1004 physicians with a mean (standard deviation) age of 49.14 (12) years. Enthusiasm was highest for self-monitoring smartwatches (66%) and contact tracing apps (66%) and slightly lower (48–56%) for other tools. Trust was highest for health providers (68%) and lowest for technology companies (30%). Most respondents (69.8%) felt that loosening privacy standards to fight the pandemic would lead to misuse of privacy in the future. Conclusion: The survey provides foundational insights into how physicians think of surveillance.

How to cite this article:
Linares AR, Bramstedt KA, Chilukuri MM, Doraiswamy P M. Physician perceptions of surveillance: Wearables, Apps, and Chatbots for COVID-19.Digit Med 2022;8:10-10

How to cite this URL:
Linares AR, Bramstedt KA, Chilukuri MM, Doraiswamy P M. Physician perceptions of surveillance: Wearables, Apps, and Chatbots for COVID-19. Digit Med [serial online] 2022 [cited 2022 May 20 ];8:10-10
Available from:

Full Text


Public health surveillance is the systematic collection and analysis of health-related data to prevent or control disease, followed by its application for public health action.[1] The global scale of the COVID-19 pandemic has accelerated the use of non-traditional, technology-based, public health, and self-surveillance mechanisms to control the spread of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2).[2],[3],[4],[5],[6],[7],[8],[9],[10],[11],[12],[13],[14],[15],[16],[17],[18],[19],[20],[21],[22],[23] Examples of such tools include contact-tracing apps, analyses of global positioning systems and social media data for population movement tracking, fever-sensing infrared thermal detection systems, symptom self-screeners (e.g. chatbots), and smartwatch applications to detect physiological signs of infection.[4],[5],[6],[7],[8],[10]

Digital technologies can rapidly collect, store, analyze, and share numerically encoded information, making them potentially highly useful in a pandemic such as COVID-19. Blue Dot, a Canadian digital health company, reportedly identified the emergence of COVID-19 through the aggregation of big data from sources such as social media and air travel, before even the WHO issued an alert.[14] However, these digital surveillance tools are experimental, and their accuracy across different settings is not fully established.[15],[16],[17],[18],[19],[20] For example, studies have shown that the accuracy of facial recognition technologies differs by race, gender, and age.[22] These tools also come with a number of potential legal and ethical risks,[18],[24],[25],[26],[27],[28] such as privacy concerns, discrimination, and over-reach of the data mission that “highlight the long-standing tensions between individual and collective rights.”[18]

Notably, the morbidity and mortality of the COVID-19 pandemic have heightened worldwide anxiety to an extent that digital public health surveillance has become ubiquitous (e.g. national requirements for downloading contact tracing apps; thermal scanning by employers and private businesses; personal location data collection via QR codes; texting of COVID-19 assay results to patients). COVID-19 is not the world's first pandemic, nor will it be the last. Thus, it is vital to understand the views of physicians, as they are involved in many facets of health data and its application to COVID-19 care. The aim of this report is to characterize the views of physicians regarding the benefits and risks of surveillance technologies.

 Materials and Methods


This study was deemed as exempt research by Duke University Medical Center's Institutional Review Board.

Study sample

To characterize the opinions of physicians on this topic, we analyzed data from a cross-sectional, random, stratified survey of physicians registered with Sermo, a secure digital platform for medical crowdsourcing and anonymous surveys. The Sermo platform is exclusive to verified and licensed physicians and has over 800,000 registered physicians, of all specialties, worldwide.

Following informed consent, the English-language survey sampled physicians between September 9 and September 15, 2020 (before initiation of SARS-CoV-2 vaccination), with a target sample size of 1000 doctors equally divided between the US, EU, and rest of the world (RoW). The survey results were de-identified to create anonymized data for analysis.

Survey instrument

Five questions in the survey [Figure 1] and [Supplemental Table 1] asked physicians their opinions on the benefits and risks/harms of using smartwatch sensor alerts, contact tracing apps, thermal cameras for mass fever screening, chatbots, and social media tracking for public health or self-surveillance. These questions focused on the risk-benefit ratio, and the answer options were “Yes” “No” and “Uncertain” For the questions on sensors (wearables and thermal cameras), we provided accuracy estimates derived from published studies. We specified that informational chatbots do not require regulatory approval in most countries for the question on chatbots, as physicians may not be aware of this. The survey then asked about the level of trust in different organizations (technology companies, government, employer, medical providers, educational universities/nonprofit bodies, or no one) to protect private surveillance data. The answer choices for this question were “Very much” “Somewhat” “Neutral” “Not really” or “Not at all” Respondents were then asked about the impact of current surveillance on future privacy standards. A final question asked physicians to provide brief qualitative comments to elaborate on their views. The results of two survey questions are reported elsewhere.[29],[30]{Figure 1}[INLINE:1]

Data and statistical methods

Descriptive statistics examined physicians' characteristics and opinions by age group, gender, frontline status, and geographic region. To test the effect of age, subjects were grouped as “younger” or “older” by age 49 years. To test the effect of frontline status, physicians more directly involved in COVID-19 care were grouped as frontline (e.g. internal medicine, ICU, ED), whereas the rest were categorized as non-frontline (although we recognize that all physicians may interact with or consult on COVID-19 patients). For geographic analyses, we pooled doctors into three groups based on the location of practice (US, European Union, RoW), while recognizing these subgroups are not homogeneous. The five categories relating to trust were combined into three categories as trusted (“Somewhat” or “Very much” responses), not trusted (“Not really” or “Not at all”), and “Neutral”. Gender analyses were restricted to those who categorized themselves as male or female. ANOVA, t-test, and Chi-square tests with P < 0.05 were viewed as qualitatively different. As this was an exploratory study, we did not adjust for small cell sizes or multiplicity. We used JMP Pro 15 (SAS), as well as Protobi.


Sample characteristics

The final respondent sample consisted of 1004 physicians representing 40 countries in North and South America, Europe, and Asia-Pacific [Supplemental Table 2]. The average age of the sample was 49.1 ± 12.3 years and 49% of respondent physicians were characterized as frontline. Of the sample, 40% were male, 20.6% were female, and 39% opted out of indicating their gender.[INLINE:2]

Utility of surveillance tools

Response rates for the support of various digital surveillance tools are shown in [Figure 2] and [Table 1]. Smart watches were supported by 65.5% of respondents (χ2 = 468.58, P < 0.0001), contact tracing apps were supported by 66.4% (χ2 = 496.85, P < 0.0001), fever cameras were supported by 58.9% (χ2 = 306.34, P < 0.0001), symptom screener chatbots by 47.6% (χ2 = 92.18, P < 0.0001), and social media by 50.9% (χ2 = 145.68, P < 0.0001).{Figure 2}{Table 1}

Age differences

Younger physicians (69.3%) were more likely to support the use of smart watch sensors compared with older physicians (61.3%) (χ2 = 7.06, P = 0.03). Younger physicians (69.7%) were marginally more likely to support the use of contact tracing apps compared to older physicians (62.8%) (P = 0.02). Younger physicians (63.6%) were more likely to support the use of mass fever screenings compared with older physicians (53.6%) (χ2 = 11.62, P = 0.003), whereas older physicians were more likely to be uncertain. Younger physicians (56.1%) were more likely to support the use of social media for population movement tracking versus older physicians (45.2%) (χ2 = 14.34, P = 0.0007), and older physicians tended to be more uncertain. Responses did not significantly differ by physician age for the utility of symptom screener chatbots.

Gender differences

Male physicians (60%) were slightly more likely to support the use of fever cameras than female physicians (51%) (P = 0.012). Responses did not differ by gender for the other surveillance tools.

Frontline status differences

Responses did not differ by frontline status.

Which entity do you trust the most with your personal surveillance data?

Physicians picked “medical providers” as the most trusted entity to protect the privacy of COVID-19 surveillance data, with about 68% of respondents reporting that they trusted their medical provider [Figure 3]. The second most trusted group was “educational/non-profit bodies” with a combined 52% of respondents reporting “somewhat” and “very much” levels of trust. Conversely, the most distrusted group was “technology companies” with only 30% of respondents reporting “somewhat” or “very much” and 46% reporting “not really” or “not at all”. Following technology companies, respondents reported low levels of trust for the “government” with only 36% responding “somewhat” or “very much” Older physicians were more likely to be distrustful of technology companies (48.9%), the government (44.5%), and educational universities/non-profit bodies (26.9%) compared with younger physicians (42.4%, 33.9%, 16.9%, respectively) (P = 0.038, 0.001, <0.001). US physicians (54.1%) were more likely to be distrustful of technology companies, compared with both EU (41.0%) and RoW (41.3%) physicians (P < 0.001). US physicians (51.8%) were also more distrustful of the government, compared with RoW (38.6%) and EU (27.5%) physicians (P < 0.001).{Figure 3}

Effect of current surveillance on future misuse of privacy

The majority of respondents (69.8%) believed that potentially loosening privacy standards to fight the pandemic would lead to misuse of privacy in future (χ2 = 601.50, P < 0.0001) [Figure 4]. Frontline physicians (73.2%) were more likely to voice concern compared with nonfrontline physicians (66.5%) (χ2 = 7.65 P = 0.022). More male physicians (71.9%) believed that a loosening of privacy standards would lead to misuse, compared with female physicians (61.8%) (χ2 = 9.58 P = 0.048).{Figure 4}

Selected qualitative comments by physicians about digital surveillance

Respondents also had an option to provide qualitative comments on digital surveillance. Supportive comments [Supplemental Table 3] included statements such as “the Future is here,” “must be made mandatory,” “anything that prevents deaths is fine, I don't worry about privacy,” and “During the 1940–1941 bombing of London called The Blitz, I believe there were zero residents of London who said 'I have a constitutional right to leave my lights on a night if I feel like it.'” Concerns about efficacy [Supplemental Table 4] included statements like “way too early,” “bad for patient and physician,” and ”data should be analyzed in clinical trials.” Concerns about harms [Supplemental Table 5] included statements such as “Pandora's box,” “creepy, extreme slippery slope,” “any great idea can have unforeseen consequences” and “I fear the behavior of people not technologies.”[INLINE:3][INLINE:4][INLINE:5]


Data is currency. Technology companies know this, governments know this, and so does the public. Like with any currency, data can be accumulated, bought and sold, or even be stolen. Hence, its storage needs to be secure. Health data are a form of personal information that generally people want to keep private, and many regulations have been implemented to safeguard personal data privacy rights.[31],[32] During a pandemic, public health interests allow for broader powers, governments, and health systems to collect, use, store, and share personal information. However, as our survey shows, this creates concern among even the physicians who are part of this process (and concurrently attempting to prevent and treat the implicated illness).

Key findings

Overall, support varied from 48% to 66% for the various surveillance tools. Two-thirds of physicians voiced support for the use of smartwatches in self-monitoring. This appears consistent with recent studies documenting the promise of consumer smart watch-based physiological signals (e.g. heart rate, sleep, activity, skin temperature) for discriminating COVID-19 test positive cases from negative cases, as well as for detecting pre-symptomatic COVID-19 infection.[4],[5] Further, a smartwatch-linked platform, Aura, recently received an EU CE mark for this purpose based on its sensitivity of 94% and ability to detect an infection signal on average 2.64 days after inoculation.[6] The minority of respondents who oppose smart watch-based infection detection technology were likely concerned about the potential for noisy data leading to misdiagnosis and unnecessary testing.[19]

Two-thirds of physicians also voiced support for contact tracing apps, even those that collected personal data. Many countries have implemented contact tracing apps, and physicians are well versed in traditional contact tracing principles for infection control, both of which likely increased physician confidence in their utility. However, since our survey, some studies have questioned the effectiveness (e.g. sensitivity of only 7% in one study) and ethics of digital contact tracing.[20],[21],[22] This suggests that the optimism of respondents in our survey may have been premature.

Physician support was slightly lower (59%) for “fever cameras” but still optimistic, consistent with the utility for mass screening offered by their high negative predictive value,[8] However, the positive predictive value (<20% in one study) of these systems remains low,[8],[9] suggesting the need for further optimization to reduce false positives.

Support for the use of social media tracking (51%) and chatbots (48%) was also slightly lower. Social media tracking is a promising tool that offers real-time data for public health officials to monitor citizen movement or social interactions during lockdowns.[10],[11],[12],[13] However, questions remain about lack of consent, accuracy, and misuse potential. Chatbots, especially those designed using WHO or CDC guidelines, very likely helped large numbers of users (over 200 million messages by some estimates) quickly get reliable information. However, to our knowledge, there are no published accuracy or outcomes data on the utility of chatbots for pandemic self-screening. Hence, there is a need for further research into the effectiveness and potential for the spread of misinformation.[20],[21]

Respondents in our survey also voiced concerns over privacy risks and over-reach of the data mission. Respondents had low trust in technology companies (30%) and governments (36%) to safeguard surveillance data. Trust was highest with medical providers (68%), followed by non-profit organizations. The higher level of trust among EU physicians may be due to the stricter data privacy laws in the EU versus the US.[31] These concerns are legitimate since some technology platforms rely on selling user data to advertisers,[27] and studies have found apps and chatbots share information with a variety of third parties.[20],[27],[28],[29] The risks also go beyond privacy breaches.[27] Historically, surveillance has worsened stigmatization and discrimination against racial or religious minorities who were often falsely blamed for disease outbreaks.[27] Further, some governments have reportedly used the pandemic to rank citizens by health status or analyze personal telecommunications traffic.[27] Hence, surveillance done wrong may “invite mission creep into adjacent fields, such as automated policing and content control.”[24]

“No turning back” is a famous quote used in many settings, and our research makes it pertinent to digital health as well. The fast portability of health data, along with the complexity of legal regulations and voluminous “Terms of Use” documents that are rarely read by users,[30] create a reality of data that has the potential to quickly bounce to all corners of the world. In addition, there is the very real presence of hackers.[31] Therefore, some of the data receivers have motives that have nothing to do with “public interest” Accordingly, the fears and lack of trust we observed are likely well-founded and highlight the need for risk mitigation to harness the full promise of public health surveillance during a pandemic.

Strengths and limitations

This is the first global survey, to our knowledge, to investigate the opinions of physicians about the utility, trust, and risks of commonly used public health digital surveillance tools. Our survey data is from a relatively large and diverse sample of verified practicing physicians. Potential limitations include cross-sectional design, the limited number of respondents from developing countries, inability to control for all possible confounding variables (e.g. personal medical history, socio-political beliefs, local data privacy regulations, knowledge about digital tools), and inability to deduce causality. Further, physician perceptions may change over time if infection risk and prevalence decrease, due to vaccination and herd immunity. Our findings should be interpreted within this context. Nevertheless, they provide a useful baseline for future surveys.

Interpretation and implications

Physicians were optimistic but not equally supportive of all surveillance tools suggesting the need for further research on effectiveness. There was also variation in physician opinions by age group. This may in part reflect differences in physician knowledge about emerging technologies and/or risk-benefit analyses, which would benefit from further education. The low level of trust in technology companies to protect personal data suggests that independent entities (governed by stricter privacy laws) should be the gatekeepers of such data. Current regulations fall short of addressing the risks posed by these new technological developments. It has been said that “data moves at the speed of trust.” During public health emergencies, any data collection through such newer tools should be both time-limited and scope-limited, with decisions made in a transparent way before the launch of surveillance activity.[22],[27] In parallel, we may need to strengthen other data privacy rules to ensure any temporary loosening during public health emergencies does not result in future misuse in normal times. We hope that insights from surveys such as this may spur public health agencies and technology innovators to work together to develop the evidence base and balance individual versus societal versus commercial needs.[24] As aptly noted by one of the survey respondents, “we can learn from films like Spiderman and The Dark Knight – with great power comes great responsibility.”

Financial support and sponsorship

Sermo provided non-financial technical platform support for the study.

Conflicts of interest

PMD has received research grants and/or advisory/ board fees from health and technology companies. PMD owns shares in companies and is a co-inventor on patents. MMC reported personal fees outside the submitted work. KB has received consulting fees outside of the submitted work. ARL has no conflicts of interest to report.


1Lee LM, Thacker SB. Public health surveillance and knowing about health in the context of growing sources of health data. Am J Prev Med 2011;41:636-40.
2Gunasekeran DV, Tham YC, Ting DS, Tan GS, Wong TY. Digital health during COVID-19: Lessons from operationalising new models of care in ophthalmology. Lancet Digit Health 2021;3:e124-34.
3Murray CJ, Alamro NM, Hwang H, Lee U. Digital public health and COVID-19. Lancet Public Health 2020;5:e469-70.
4Mishra T, Wang M, Metwally AA, Bogu GK, Brooks AW, Bahmani A, et al. Pre-symptomatic detection of COVID-19 from smartwatch data. Nat Biomed Eng 2020;4:1208-20.
5Quer G, Radin JM, Gadaleta M, Baca-Motes K, Ariniello L, Ramos E, et al. Wearable sensor data and self-reported symptoms for COVID-19 detection. Nat Med 2021;27:73-7.
6Empatica. Aura Receives First of Its Kind European CE Mark for Early Detection of COVID-19 Empatica; 2021. Available from: [Last accessed on 2021 Jun 04].
7Huang Z, Guo H, Lee YM, Ho EC, Ang H, Chow A. Performance of digital contact tracing tools for COVID-19 response in Singapore: Cross-sectional study. JMIR Mhealth Uhealth 2020;8:e23148.
8Nguyen AV, Cohen NJ, Lipman H, Brown CM, Molinari NA, Jackson WL, et al. Comparison of 3 infrared thermal detection systems and self-report for mass fever screening. Emerg Infect Dis 2010;16:1710-7.
9Martinez-Jimenez MA, Loza-Gonzalez VM, Kolosovas-Machuca ES, Yanes-Lane ME, Ramirez-GarciaLuna AS, Ramirez-GarciaLuna JL. Diagnostic accuracy of infrared thermal imaging for detecting COVID-19 infection in minimally symptomatic patients. Eur J Clin Invest 2021;51:e13474.
10Our Work on COVID-19; 2021. Available from: [Last accessed on 2021 Jun 07].
11Beria P, Lunkar V. Presence and mobility of the population during the first wave of Covid-19 outbreak and lockdown in Italy. Sustain Cities Soc 2021;65:102616.
12Pérez-Arnal R, Conesa D, Alvarez-Napagao S, Suzumura T, Català M, Alvarez-Lacalle E, et al. Comparative analysis of geolocation information through mobile-devices under different COVID-19 mobility restriction patterns in Spain. ISPRS Int J Geo Inf 2021;10:73.
13Morinishi L, Barkume K, Kim E, Reinhart A. Home for the Holidays? The Impact of US Holidays on Social Behaviors and Preventative Measures. Delphi Group; 2021. Available from: [Last accessed on 2021 Jun 7].
14Stieg C. How this Canadian Start-Up Spotted Coronavirus before Everyone Else Knew About It; 2020. Available from: [Last accessed on 2021 Jun 04].
15Sweeney Y. Tracking the debate on COVID-19 surveillance tools. Nat Mach Intell 2020;2:301-4.
16Mbunge E. Integrating emerging technologies into COVID-19 contact tracing: Opportunities, challenges and pitfalls. Diabetes Metab Syndr 2020;14:1631-6.
17Colizza V, Grill E, Mikolajczyk R, Cattuto C, Kucharski A, Riley S, et al. Time to evaluate COVID-19 contact-tracing apps. Nat Med 2021;27:361-2.
18Sekalala S, Dagron S, Forman L, Meier BM. Analyzing the Human Rights Impact of Increased Digital Public Health Surveillance during the COVID-19 Crisis. Health Hum Rights 2020;22:7-20.
19Zhu T, Watkinson P, Clifton DA. Smartwatch data help detect COVID-19. Nat Biomed Eng 2020;4:1125-7.
20Miner AS, Laranjo L, Kocaballi AB. Chatbots in the fight against the COVID-19 pandemic. NPJ Digit Med 2020;3:65.
21Fan X, Chao D, Zhang Z, Wang D, Li X, Tian F. Utilization of self-diagnosis health chatbots in real-world settings: Case study. J Med Internet Res 2021;23:e19928.
22Zimmermann BM, Fiske A, Prainsack B, Hangel N, McLennan S, Buyx A. Early perceptions of COVID-19 contact tracing apps in German-speaking countries: Comparative mixed methods study. J Med Internet Res 2021;23:e25525.
23Grother P, Ngan M, Hanaoka K. Face Recognition Vendor Test (FRVT): Part 3: Demographic Effects. Gaithersburg, MD: USDo Commerce, National Institute of Standards and Technology; 2019.
24Cohen J, Hartzog W, Moy L. The Dangers of Tech-Driven Solutions to COVID-19; 2020. Available from: [Last accessed on 2021 Jun 07].
25Valade P. Jumbo Privacy Review: North Dakota's Contact Tracing App; 2020. Available from: [Last accessed on 2021 Jun 07].
26Mozur P, Zhong R, Krolik A. In Coronavirus Fight, China Gives Citizens a Color Code, With Red Flags. The New York Times; 2020. Available from: [Last accessed on 2021 Jun 07].
27Global Mobile Consumer Survey: US Edition; 2017. Available from: technology-media-telecommunications/us-tmt-2017-global-mobile- consumer-survey-executive-summary.pdf. [Last accessed on 2021 Jun 04].
28Koppel R, Kuziemsky C. Healthcare data are remarkably vulnerable to hacking: Connected healthcare delivery increases the risks. Stud Health Technol Inform 2019;257:218-22.
29Doraiswamy PM, Chilukuri MM, Linares AR, Bramstedt KA. Are we ready for COVID-19's golden passport? Insights from a global physician survey. J Health Soc Sci 2021;6:079-86.
30Doraiswamy PM, Chilukuri MM, Ariely D, Linares AR. Physician perceptions of catching COVID-19: Insights from a global survey. J Gen Intern Med 2021;36:1832-4.
31EU General Data Protection Regulation (GDPR): Regulation (EU) 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing Directive 95/46/EC (General Data Protection Regulation), OJ 2016 L 119/1.
32The Health Insurance Portability and Accountability Act (HIPAA). Washington, D.C.: U.S. Dept. of Labor, Employee Benefits Security Administration. US 1996.