Hey, have you used Tumblr? I ask, because I don’t think that this is always people trying to infiltrate a political discussion to paralyze effective leftist organizing. I do think it totally is sometimes- but sometimes it’s because of how people structure their values and philosophy of engagement with the world, politics and moral actions.
I have become very familiar with how, on Tumblr, the dominant cultural paradigm has a strong tendency to several of those traits purely because of a combination of ways that the internet, and that website, is structured; and, the ambient cultural values of the US informing how they structured their beliefs about morality and politics.
People who are part of this paradigm tend to have a strongly dentological bent, and are obsessed with if an action is good or bad in and of itself; and, especially critically- if there is any part of it that represents any moral compromise, no matter how small. They do not want to ever have to compromise their principles, and frame those principles as actions and behaviors and not ends. They are very focused on maintaining a sense of moral purity and superiority, which naturally leads to inaction due to the inherent compromises present in political action and general life.
Paired with this is a deep desire to prove one’s virtue, which is done by performing it- frequently by finding an acceptable target for harassment or abuse, then heaping unpleasant behavior on them in order to show that bad people are bad and they, a good person, is good. It’s very simplistic and results in people who are constantly vigilant of if anything they do can be construed as wrong, because then it becomes a vector for harassment and attack, and who are constantly trying to discern if someone else is currently vulnerable to the same.
This mixes with a general lack of critical thinking skill, reading comprehension and fact-checking that so defines our modern septic pit of an internet; and, you have a cycle of inaction and abuse that accomplishes very little. It’s very frustrating, and a major contributing factor to me not using Tumblr anymore. I got really burnt out on people who would use, for example, you not reblogging a post supporting a specific political point as proof that you were maliciously against the political point, even if you openly advocated for it, or it was about a marginalized group you were a part of.
I feel like you are identifying a pattern that is very real and important, but I think your conclusions about why it happens may be too narrow. I think there’s a multiplicity of groups of different political and philosophical tendencies that are contributing to this atmosphere. I also feel like sometimes people need a place to vent about how incredibly infuriating US politicians and politics are- I try to keep that to my friends and personal writing, nowadays, but there was a point when I was incredibly bitter about how the Democrats continued to neglect and ignore people in need due to political exigencies. Sure, I get it, and sure, I support them whenever I get a chance to, but damn if it’s not frustrating.
I increasingly feel like there needs to be more sectioning of discussions on platforms to allow constructive discussion and vent-posting to be clearly separated and have that be aggressively enforced.
Hi, once more, I’m happy to have a discussion about this. I have very firm views on it, and enjoy getting a chance to discuss them and work towards an ever greater understanding of the world.
I completely understand the desire to push back against certain kinds of “understandings” people have about LLM due to their potentially harmful inaccuracy and the misunderstandings that they could create. I have had to deal with very weird, like, existentialist takes on AI art lacking a quintessential humanity that all human art is magically endowed with- which, come on, there are very detailed technical art reasons why they’re different, visually! It’s a very complicated phenomenon, but, it’s not an inexplicable cosmic mystery! Take an art critique class!
Anyway, I get it- I have appreciated your obvious desire to have a discussion.
On the subject of understanding, I guess what I mean is this: Based on everything I know about an LLM, their “information processing” happens primarily in their training. This is why you can run an LLM instance on, like, a laptop but it takes data centers to train them. They do not actually process new information, because if they did, you wouldn’t need to train them, would you- you’d just have them learn and grow over time. An LLM breaks its training data down into patterns and shapes and forms, and uses very advanced techniques to generate the most likely continuation of a collection of words. You’re right in that they must answer, but that’s because their training data is filled with that pattern of answering the question. The natural continuation of a question is, always, an answer-shaped thing. Because of the miracles of science, we can get a very accurate and high fidelity simulation of what that answer would look like!
Understanding, to me, implies a real processing of new information and a synthesis of prior and new knowledge to create a concept. I don’t think it’s impossible for us to achieve this, technologically, humans manage it and I’m positive that we could eventually figure out a synthetic method of replicating it. I do not think an LLM does this. The behavior they exhibit and the methods they use seem radically inconsistent with that end. Because, the ultimate goal of them was not to create a thinking thing, but to create something that’s able to make human-like speech that’s coherent, reliable and conversational. They totally did that! It’s incredibly good at that. If it were not for the context of them politically, environmentally and economically, I would be so psyched about using them! I would have been trying to create templates to get an LLM to be an amazing TTRPG oracle if it weren’t for the horrors of the world.
It’s incredible that we were able to have a synthetic method of doing that! I just wish it was being used responsibly.
An LLM, based on how it works, cannot understand what it is saying, or what you are saying, or what anything means. It can continue text in a conversational and coherent way, with a lot of reliability on how it does that. The size, depth and careful curation of its training data mean that those responses are probably as accurate to being an appropriate response as they can be. This is why, for questions of common knowledge, or anything you’d do a light google for, they’re fine. They will provide you with an appropriate response because the question probably exists hundreds of thousands of times in the training data; and, the information you are looking for also exists in huge redundancies across the internet that got poured into that data. If I ask an LLM which of the characters of My Little Pony has a southern accent, they will probably answer correctly because that information has been repeated so much online that it probably dwarfs the human written record of all things from 1400 and earlier.
The problem becomes evident when you ask something that is absolutely part of a structured system in the english language, but which has a highly variable element to it. This is why I use the “citation problem” when discussing them, because they’re perfect for this: A citation is part of a formal or informal essay, which are deeply structured and information dense, making them great subjects for training data. Their structure includes a series of regular, repeating elements in particular orders: Name, date, book name, year, etc- these are present and repeated with such regularity that the pattern must be quite established for the LLM as a correct form of speech. The names of academic books are often also highly patterned, and an LLM is great at creating human names, so there’s no problem there.
The issue is this: How can an LLM tell if a citation it makes is real? It gets a pattern that says, “The citation for this information is:” and it continues that pattern by putting a name, date, book title, etc in that slot. However, this isn’t like asking what a rabbit is- the pattern of citations leads into an endless warren of hundreds of thousands names, book titles, dates, and publishing companies. It generates them, but it cannot understand what a citation really means, just that there is a pattern it must continue- so it does.
A flat-earther has some understanding of what truth is, even if their definition is divergent from the norm. The things they say are deeply inaccurate, but you can tell that they are the result of a chain of logic that you can ask about and follow. It’s possible to trace flat-earth ideas down to sources. They’re incorrect, but they’re arrived at because of an understanding of prior (incorrect) information. A flat-earther does not always invent their entire argument and the basis for their beliefs on the spot, they are presenting things they know about from prior events- they can show the links. An LLM cannot tell you how it arrived at a conclusion, because if you ask it, you are just receiving a new continuation of your prior text. Whatever it says is accurate only when probability and data set size is on its side.