"Inftoxicity" and other new words to describe malicious information and communication thereof

By Jáchym Fibír @ 2023-12-23T18:40 (+1)

This is a linkpost to https://www.lesswrong.com/posts/Lb4oC4pCimAAcebtc/inftoxicity-and-other-new-words-to-describe-malicious

The only way an act guided by good intention can reliably translate into a good outcome is when the act is done in a truthfully informed manner - which in turn requires true and trustworthy sources of information. 

Therefore, I wanted to share here my proposition to start using new words such as "inftoxic" to describe information or data intentionally created or spread to cause harm, so that people can quickly identify and label them as untrustworthy. 

The post is not meant to present the best way to do this but to see if other people would consider such terminology useful.

Background

We are becoming more and more surrounded by synthetic data (such as text or images generated by LLMs) and as these mix together with real-world, human-generated data, there is great need for ways how to distinguish between the two, as well as generally make sense of all the informational mess.

While writing about the risks arising from generating synthetic data with potential malicious uses, I realized that I was lacking the proper vocabulary to effectively describe "a dataset that was intentionally constructed to mislead". I have tried looking for a suitable adjective to describe such dataset, and having failed, I tried to ask GPT4 to help me find and coin a new word for it

The conversation I had with the LLM was quite engaging at the time, and with some reflection there still seemed to be some valuable takeaways, so I decided to do this little writeup about it. (Read the original conversation only at your own risk, haha!)

Inftoxicity and related terms

After some brainstorming from the GPT, I asked it to elaborate on the word "inftoxic" as an adjective that can be used to describe data, information sources, systems, actors, or actions that had been created with, or are acting in, a malicious intent to damage someone or something, or manipulate or mislead in a negative way. 

This would be in contrast to "biased", which describes something or someone arriving at incorrect or misleading conclusions, yet not necessarily producing a negative outcome or being a result of malicious intent. 

Below is a bunch of derivative words GPT4 was able to flesh out (slightly edited by me). I am intentionally leaving in all ten it generated, both to amuse the reader and inform of this interesting capability for creating novel words and definitions.

Inftoxic - adjective 

Inftox - noun, invariable

Inftoxicity - noun, singular (plural: inftoxicities)

Inftoxinator - noun, singular (plural: inftoxinators)

Inftoxify - verb

Inftoxifiable - adjective

Inftoxication - noun, singular (plural: inftoxications)

Inftoximeter - noun, singular (plural: inftoximeters)

Inftoxology - noun, singular (plural: inftoxologies)

Inftoxical - adjective

Inftoxosphere - noun, singular (plural: inftoxospheres)

Implications and further directions

I was part amused, part impressed and part intrigued by the new terminology above, which I was able to generate within minutes, yet which seemed potentially quite useful to the world at large (especially as we get to more widespread adoption of LLMs and as ByteDance is using GPT4 to create synthetic datasets for its own LLM training). 

I continued the conversation asking about potential strategies for spreading this terminology so that there would be more awareness around information that is purposefully made or disseminated for malicious ends. I am not fully convinced I want to spread this terminology, but there is one approach suggested by GPT4 that I particularly liked:

"Spot the Inftox" Challenge

So, after getting all the way through my first post here, do you think this is worthy of spreading around? Want to help kick off the #SpotTheInftox challenge? Let me know in the comments! :) 🙏