The wearable technology market is booming, withhalf a billion wearablessold globally in 2020. Apps on these devices, or the devices themselves, often claim to monitor our health to spot illnesses, track our workouts to help us reach our fitness goals, or keep an eye on our children’s whereabouts to enhance their safety.
But they’re also divisive. Supporters of wearable technology claim that health trackers should beprescribed by the NHSand could even deliver anearly warningof a possible COVID-19 infection. GPS tracking devices designed to be worn by children, meanwhile, are seen as asafety assetfor parents.
Yet studies have found fitness trackers to be tooinaccurateandmisleadingto be used bymedical professionals, and that, because they’ve been rushed to market, wearables of all kinds are an insecure “Wild West” region of technology that requires urgent regulation.
Ina recent report, we looked at the security risks associated with wearable devices, as well as “smart toys” that can record children in their homes. We found a concerning lack of security – especially for devices aimed at children – which lack even the most basic cybersecurity precautions, leaving them open to abuse.
Fitness trackers and personal data
The 💜 of EU tech
The latest rumblings from the EU tech scene, a story from our wise ol' founder Boris, and some questionable AI art. It’s free, every week, in your inbox. Sign up now!
One key issue with wearables is the data they generate and share. For instance, many fitness trackers rely on data on a person’s location to map their workouts. That’s great if you’re keen to track the distance of your jogs, but it’s not especially sensible if you’re embarking on those jogsfrom a military basein hostile territory.
Strava released their global heatmap. 13 trillion GPS points from their users (turning off data sharing is an option).https://t.co/hA6jcxfBQI… It looks very pretty, but not amazing for Op-Sec. US Bases are clearly identifiable and mappablepic.twitter.com/rBgGnOzasq
— Nathan Ruser (@Nrg8000)January 27, 2018
Beyond that specific example, which caused some embarrassment for the US military in 2018, it’s clear that sharing your location publicly, even in a safe civilian setting, comes with significant risks.
And it’s not just the real-time tracking of your running route that could expose your whereabouts. Because these trackers upload your workouts to an app and share them publicly, it’s possible for predators to use historic running, biking or hiking routes to predict where you might be at a given time. This safety issue isn’t only restricted to workouts. Even something as innocuous assharing a photo through your Apple watchcan give away your geolocation.
Are trackers safe for children?
Even more concerning are devices designed to be worn by children, sales of which are expected to reach$875 million (£620 million)by 2025. These watches are marketed as wearable tech to keep kids safe, tracking their location and alerting parents when the watch’s onboard “SOS” button is pressed – or if the child travels beyond a geofenced area.
Smart watches as safety devices on children’s wrists may sound like aboon for anxious parents, but a2017 surveyof children’s smart watches found that the all-important “SOS” button either got stuck or didn’t work at all in most cases.
Additionally, flaws in some smart watches’ accompanying apps have raisedserious safety concerns.Security researchershave found they could not only easily access children’s historical route data – like their path to and from school – and monitor their geolocation in real time, but they could also speak directly to the child, through the watch, without the call being reported in the parent’s app.
Connected toys
Fears that internet of things devices can give people unauthorised access to children also extend tothe “smart toy” market. Some of these toys contain hidden cameras and microphones which, if hacked, could be used to record the interior of your home, including children’s rooms.
In 2017, German regulators recognised this danger bybanning the saleof the Cayla “smart doll”, labelling it as the kind of “de facto espionage device” that Germany’sTelecommunications Actlegislates against. In an unusual and unsettling move, the regulator went further by asking parents who’d bought one todestroy the dollto prevent illicit surveillance.
Goodbye Spy#Toy:#GermanyBans My Friend#CaylaDoll |https://t.co/RcIiVGTPWy#IoT#Security#Privacy#Surveillancepic.twitter.com/P6gkrDqzkU
— HackRead.com (@HackRead)February 18, 2017
Even if the manufacturers of smart toys and children’s smart watches can guarantee far better security than that which led to the Cayla ban, there remain other surveillance concerns. In 2019, aUNICEF-led reporthighlighted how children’s rights – to creativity, freedom of choice and self-determination – are challenged by smart devices. Present in schools, at home, and on the wrist, this kind of round-the-clock surveillance, the report argues, restricts carefree childhood and hurts kids’ development.
Making trackers safer
Trackers and toys can be made safer. Before we allow these devices to flood the market, it’s essentialwe standardisethe minimum security requirements that manufacturers must comply with – no matter where in the world these devices are made.
Key among these standards should be the removal offactory-default passwordson devices – which, like “admin” or “1234”, are easily guessed or discovered by even the most novice hacker. Manufacturers should also publish avulnerability disclosureto help users understand risks, and make regular software updates in response to vulnerabilities unearthed by security researchers.
Clearly, monitoring people’s health via wearable trackers has the potential to radically improve access to medical care. Likewise, every parent wants their child to be safe, and smart devices, like mobile phones before them, could be a reliable tool for checking in with them. But without safety standards, these devices have the potential to cause more harm than they offset. Regulators must act fast to stop this growing market from leading to significant harms.
Article bySaheli Datta Burton, Research Fellow, Department of Science Technology Engineering and Public Policy,UCLandMadeline Carr, Professor of Global Politics and Cybersecurity,UCL
This article is republished fromThe Conversationunder a Creative Commons license. Read theoriginal article.
Story byThe Conversation
An independent news and commentary website produced by academics and journalists.An independent news and commentary website produced by academics and journalists.
Get the TNW newsletter
Get the most important tech news in your inbox each week.