In a few days, we’ll find out if Tiktok is really going to be pounced out of the U.S. market. [This was written in early January. Answer appears to still be yes when I pushed the publish button on this post as the app was removed from the app stores.] In the meantime, the Tiktok saga has already revealed many things about the data collection industry. (In this post, I shall assume that the allegations by the U.S. government hold up even though there aren’t any actual evidence of wrongdoing.)
What technologies can do > what technologies are doing
With the legislative action against Tiktok, the lawmakers are applying a doctrine of preemptive strike. I’ve always believed in early action to mitigate potential harms from new technologies. There are endless examples of such predictable harms. Just to name a few:
- Self-checkout tech will induce more stealing
- Tolling via reading license plates will lead to more nonpayment and evasion
- Car telematics will raise people’s insurance rates, and the number of parking and speeding tickets issued
- Deep fakes using deep learning technologies will supercharge all kinds of scams
Big Tech has always opposed early action. I don’t know how many times I’ve been told off on these worries. All technology can be misappropriated, Big Tech supporters say, and so the possibility of harm cannot justify restricting any technology. A knife can be used to kill, would you ban knives? Meanwhile, let's focus on the great utility knives bring.
Not banning something, however, does not mean turning a blind eye. (And that is what the industry has been doing regarding deep fakes. Just look at how even the term “deep fake” is fading away according to Google Trends.)
In this chart, blue is for “deepfake” while red is for “deep fake”. Both lines are moving downwards.
By now, we know:
- many retailers have had to dial back on self checkout (link)
- within one week of pushing out congestion pricing in Manhattan, NYPD had to take action on toll evasion (link)
- Texas just sued AllState for compiling telematics data and then raising premiums, dropping coverage, etc. (link)
- Let’s see who’s right on deep-fake scams!
The basis of this Tiktok action is the possibility that the collected data would be exploited for nefarious use. If we apply the Big Tech viewpoint, this possibility of misuse is of no concern. Nevertheless, Big Tech is silent about this legislative action against Tiktok.
What I learned from the Tiktok saga is now we’re supposed to worry about how technologies could be put to nefarious use before someone uses them nefariously!
Social media apps collect highly sensitive data that can be damaging to users.
The government alleges that the data collected by Tiktok could be damaging to its users. By extension, the data collected by any social media app, indeed any app, could harm people.
Since all apps available on major phone brands have to follow guidelines set by Apple’s iPhone or Google’s Android, it’s hard to imagine that Tiktok engages in data collection that goes beyond the capabilities of other social media apps, indeed any mobile app. If you’ve ever opened one of those terms and conditions, you’ve probably been advised that there are hundreds or thousands of data collectors on every website. These are businesses whose raison d’etre is monetizing user data.
Just for giggles, I saved this popup from a UK-based dictionary website (typically, U.S. sites don't even show these popups):
Let's review this quickly, shall we? "We care about your privacy" [therefore] "We and our 777 partners store and access personal data..." Further down, "[We and our 777 partners] use precise geolocation data." Pray tell why a dictionary website needs to know the user's precise location. I commend them for being honest, and speaking in plain English; and if you're alarmed, don't be mad at Collins, they are not the world champion of data collection by any measure. I picked on them just because I was on their site recently.
To break this down further, there are data elements that all apps collect, and then, there are data elements that are proprietary to individual app developers.
If you’ve ever seen a pop-up allowing you to “customize” the amount of tracking on a website, you’ll have heard of such terms as “legitimate use.” Most of the data that have been alleged to be massive invasion of Americans’ privacy – such as location, device ID, device type, IP address, battery usage, etc. – are routinely collected by any IP device. They are a form of “data exhaust”: logging these data elements have been standard even before the Web got commercialized. Everyone classifies such data as “legitimate use” (e.g. necessary for the normal functioning of the app).
Any data element can be “legitimized” under the above framework. If we were to exclude location data from normal logging, then some developer will add a feature on their app that requires location data, thus legitimizing the data element! For example, weather apps are big business because most users are looking for local weather, which requires them to disclose their current location. For weather apps to function, users must authorize the collection of location data, and the app developers to sell them.
Proprietary data are also collected by all apps. If only the app owner (but not others) can see such data, then it must concern something users do inside the app. Some examples might be how much time a user spends on Tiktok, draft versions of videos that never went public, and what kinds of editing occurred on a video. It’s actually hard to come up with many data elements in this category. Let’s take the social network of a particular user, i.e. which other Tiktok user is this person connected to. In most social-media apps, this is public information (by design) so it’s not really proprietary to Tiktok. Further, the businesses that make money from data collection frequently pays app developers to install their scripts inside their apps, in which case even “proprietary” data have been made available to third parties, putting them in the other category.
Location data is sensitive, as it allows bad people to find us. But the idea that banning Tiktok will stop a foreign government from getting location data on Americans is ludicrous, when there are thousands of data collection scripts on every phone. If we don’t want our location data out there, we can stop using apps that require location data. (Even that might not be effective – see point #1 of this post.)
The lesson of this Tiktok saga is that the government has reason to believe that Tiktok collects sensitive, potentially harmful data from its users, an allegation which generalizes to all apps.
Our data can be dangerous even if we’ve done nothing wrong.
One of Big Tech’s favorite slogans about data collection is "If you have something that you don't want anyone to know, maybe you shouldn't be doing it in the first place." Eric Schmidt famously made that comment in 2009 when he was CEO of Google, at the very beginning of this data collection tidal wave. The implication is that if a driver was speeding on the highway, and Google detected it (e.g. via tracing location data on an Android phone), and sold the data to a third party on contract with the police to issue speeding tickets, then it’s the driver’s fault for producing the data in the first place.
The Tiktok saga wholly usurps such logic because the legislators are claiming that most people, who have done nothing wrong, could become victims if their data fall into the wrong hands!
Social-media apps provide the platforms for indoctrination.
The U.S. government alleges that foreign states can exploit Tiktok to spread propaganda. But the real lesson here is the social-media platforms plant ideas into their audiences via exposure and filtering, which makes them not different from traditional mainstream media. One difference is that traditional media are broadcast networks so the same messages are sent to everyone while on social media, different messages (even opposite messages) can be sent to different people, without individuals realizing they got different messages from others.
It beggars belief that these foreign states can’t figure out how to use Facebook, Instagram, Twitter/X, Youtube, etc. to influence those audiences. It’s possible – although unproven – that these U.S.-originated platforms do the government’s bidding in filtering speech, and therefore may be ineffective at spreading opinions unfriendly to said government.
This feature of social media has long been known and exploited - by the marketing and advertising industries. The entire social-media influencer industry is built around the idea that these platforms efficiently deliver marketing messages, including many spams and scams, to users. Savvy politicians have also mastered using social media to “mobilize” constituents; some governments have also censored social media. This Tiktok action confirms the efficacy of social media.
Recent Comments