There have been multiple accounts created with the sole purpose of posting advertisement posts or replies containing unsolicited advertising.

Accounts which solely post advertisements, or persistently post them may be terminated.

Beta testing Stad.social

@vidarh

This profile is from a federated server and may be incomplete. Browse more on the original instance.

This new data poisoning tool lets artists fight back against generative AI (www.technologyreview.com)

A new tool lets artists add invisible changes to the pixels in their art before they upload it online so that if it’s scraped into an AI training set, it can cause the resulting model to break in chaotic and unpredictable ways....

vidarh ,
@vidarh@lemmy.stad.social avatar

Trying to detect poisoned images is the wrong approach. Include them in the training set and the training process itself will eventually correct for it.

I think if you build more robust features

Diffusion approaches etc. do not involve any conscious “building” of features in the first place. The features are trained by training the net to match images with text features correctly, and then “just” repeatedly predict how to denoise an image to get closer to a match with the text features. If the input includes poisoned images, so what? It’s no different than e.g. compression artifacts, or noise.

These tools all try to counter models trained without images using them in the training set with at most fine-tuning, but all they show is that models trained without having seen many images using that particular tool will struggle.

But in reality, the massive problem with this is that we’d expect any such tool that becomes widespread to be self-defeating, in that they become a source for images that will work their way into the models at a sufficient volume that the model will learn them. In doing so they will make the models more robust against noise and artifacts, and so make the job harder for the next generation of these tools.

In other words, these tools basically act like a manual adversarial training source, and in the long run the main benefit coming out of them will be that they’ll prod and probe at failure modes of the models and help remove them.

vidarh ,
@vidarh@lemmy.stad.social avatar

They don’t even need to detect them - once they are common enough in training datasets the training process will “just” learn that the noise they introduce are not features relevant to the desired output. If there are enough images like that it might eventually generate images with the same features.

vidarh ,
@vidarh@lemmy.stad.social avatar

Human brains clearly work differently than AI, how is this even a question?

It’s not all that clear that those differences are qualitatively meaningful, but that is irrelevant to the question they asked, so this is entirely a strawman.

Why does the way AI vs. the brain learn make training AI with art make it different to a person studying art styles? Both learn to generalise features that allows them to reproduce them. Both can do so without copying specific source material.

The term “learning” in machine learning is mainly a metaphor.

How do the way they learn differ from how humans learn? They generalise. They form “world models” of how information relates. They extrapolate.

Also, laws are written with a practical purpose in mind - they are not some universal, purely philosophical construct and never have been.

This is the only uncontroversial part of your answer. The main reason why courts will treat human and AI actions different is simply that they are not human. It will for the foreseeable future have little to do whether the processes are similar enough to how humans do it.

vidarh ,
@vidarh@lemmy.stad.social avatar

Society is built to distribute wealth, so that everyone can live a decent life.

As a goal, I admire it, but if you intend this as a description of how things are it’d be boundlessly naive.

vidarh ,
@vidarh@lemmy.stad.social avatar

This idea that copyright and IP shouldn’t exist at all is kinda absurd.

For the majority of human existence, that was the default.

Copyright exists as an explicit tradeoff between the rights of the public to be able to do as they please with stuff introduced into the public sphere, and a legal limitation infringing on the publics liberty for a limited time for the purpose of encouraging the creation of more works for the public benefit. It was not introduced as some sort of inherent right, but as a trade between the public and creators to incentivise them.

Stripping it away from existing artists who has come to depend on it without some alternative would be grossly unfair, but there’s nothing absurd about wanting to change the bargain over time. After all, that has been done many times, and the copyright we have now is vastly different and far more expansive and lengthy than early copyright protection.

Personally, I’d be in favour of finding alternative means of supporting creators and stripping back copyright as a tradeoff. The vast majority of creators earn next to nothing from their works; only a very tiny minority makes a livable wage of art of any form at all, and of the rest the vast majority of profits take place in a very short period of initial exploitation of a work, so we could allow the vast majority to earn more from their art relatively cheaply, and affect the rest to a relatively limited degree, while benefiting from the reduced restrictions.

vidarh ,
@vidarh@lemmy.stad.social avatar

You can see the difference in the process in the results, for example in how some generated pictures will contain something like a signature in the corner

If you were to train human children on an endless series of pictures with signatures in the corner, do you seriously think they’d not emulate signatures in the corner?

If you think that, you haven’t seen many children’s drawings, because children also often pick up that it’s normal to put something in the corner, despite the fact that to children pictures with signatures is a tiny proportion of visual input.

Or how it is at least possible to get the model to output something extremely close to the training data

People also mimic. We often explicitly learn to mimic - e.g. I have my sons art folder right here, full of examples of him being explicitly taught to make direct copies as a means to learn technique.

We just don’t have very good memory. This is an argument for a difference in ability to retain and reproduce inputs, not an argument for a difference in methods.

And again, this is a strawman. It doesn’t even begin to try to answer the questions I asked, or the one raised by the person you first responded to.

That at least proves that the process is quite different to the process of human learning.

Neither of those really suggests that all (that diffusion is different to humans learn to generalize images is likely true, what you’ve described does not provide even the start of any evidence of that), but again that is a strawman.

There was no claim they work the same. The question raised was how the way they’re trained is different from how a human learns styles.

vidarh , (edited )
@vidarh@lemmy.stad.social avatar

To me, that’s not an argument for regulating AI, though, because most regulation we can come up with will benefit those with deep enough pockets to buy themselves out of the problem, while solving nothing.

E.g. as I’ve pointed out in other debates like this, Getty Images has a market cap of <$2bn. OpenAI may have had a valuation in the $90bn range. Google, MS, Adobe all also have shares prices that would trivially allow them to purchase someone like Getty to get ownership of a large training set of photos. Adobe already has rights to a huge selection via their own stock service.

Bertelsmann owns Penguin Random-House and a range ofter publishing subsidiaries. It’s market cap is around 15 billion Euro. Also well within price for a large AI contender to buy to be able to insert clauses about AI rights. (You think authors will refuse to accept that? All but the top sellers will generally be unable to afford to turn down a publishing deal, especially if it’s sugar-coated enough, but they also sit on a shit-ton of works where the source text is out-of-copyright but they own the right to the translations outright as works-for-hire)

That’s before considering simply hiring a bunch of writers and artists to produce data for hire.

So any regulation you put in place to limit the use of copyrighted works only creates a “tax” effectively.

E.g. OpenAI might not be able to copy artist X’s images, but they’ll be able to hire artist Y on the cheap to churn out art in artist X’s style for hire, and then train on that. They might not be able to use author Z’s work, but they can hire a bunch of hungry writers (published books sells ca 200 copies on average; the average full time author in the UK earns below minimum wage from their writing) as a content farm.

The net result for most creators will be the same.

Even wonder why Sam Altmann of OpenAI has been lobbying about the dangers of AI? This is why. And its just the start. As soon as these companies have enough capital to buy themselves access for data, regulations preventing training on copyrighted data will be them pulling up the drawbridge and making it cost-prohibitive for people to build open, publicly accessible models in ways that can be legally used.

And in doing so they’ll effectively get to charge an “AI tax” on everyone else.

If we’re going to protect artists, we’d be far better off finding other ways of compensating them for the effects, not least because it will actually provide them some protection.

vidarh ,
@vidarh@lemmy.stad.social avatar

That’s hilarious, given that if these tools become remotely popular the users of the tools will provide enough adversarial data for the training to overcome them all by itself, so there’s little reason to anyone with access to A100’s to bother trying - they’ll either be a minor nuisance used a by a tiny number of people, or be self-defeating.

vidarh , (edited )
@vidarh@lemmy.stad.social avatar

Yes: Train on more images processed by this.

In other words: If the tool becomes popular it will be self-defeating by producing a large corpus of images teaching future models to ignore the noise it introduces.

There are likely easier “quick fixes” while waiting for new models, but this is the general fix that will work against almost any adversarial attack like this.

There might be theoretical attacks that’d be somewhat more difficult to overcome to the extent of requiring tweaks to the models, but given that there demonstrably exists a way of translating text to images that overcomes any such adversarial method that isn’t noticeable to humans, given that humans can, there will inherently always be a way to beat them.

vidarh ,
@vidarh@lemmy.stad.social avatar

An AI model will “notice them” but ignore them if trained on enough copies with them to learn that they’re not significant.

vidarh ,
@vidarh@lemmy.stad.social avatar

You wouldn’t want to. If you just feed it to the models, then if there are enough of these images to matter the model will learn to ignore the differences. You very specifically don’t want to prevent the model from learning to overcome these things, exactly because if you do you’re stuck with workarounds like that forever, but if you don’t the model will just become more robust to noisy data like this.

vidarh ,
@vidarh@lemmy.stad.social avatar

I doesn’t need to be full on UBI. In a lot of countries grants mechanisms and public purchasing mechanisms for art already make up a significant proportion of income for artists. Especially in smaller countries, this is very common (more so for literary works, movies and music where language provides a significant barrier to accessing a bigger audience, but for other art too). Imagine perhaps a tax/compulsory licensing mechanism that doesn’t stop AI training but instead massively expands those funding sources for people whose data are included in training sets.

This is not stoppable, not least because it’s “too cheap” to buy content outright.

I pointed out elsewhere that e.g. OpenAI could buy all of Getty Images for ~2% of their currently estimated market cap based on a rumoured recent cash infusion. Financing vast amounts of works for hire just creates a moat for smaller players while the big players will still be able to keep improving their models.

As such it will do nothing to protect established artists, so we need expansion of ways to fund artists whether or not inclusion of copyrighted works in training sets becomes restricted.

vidarh ,
@vidarh@lemmy.stad.social avatar

If you work on commission only online, or never went to art school those won’t cover you.

There’s no reason it has to stay like that. And most people in that position are not making a living from art as it is; expanding public funding to cover a large proportion of working artists at a better level than today would cost a pittance.

These large tech companies become so highly valued at the start because of venture capital and then in 5-10 years collapse under their own weight. How many of these have come up and are now close to drowning after pushing out all competitors? Sorry if I’m not excited about an infusion of cash into a large for profit company that is just gobbling up anything anyone posts online without consent to make a quick buck.

MS, Apple, Meta, Google etc. are massively profitable. OpenAI is not, but sitting on a huge hoard of Microsoft cash. It doesn’t matter that many are close to drowning. The point is the amount of cash floating around that enable the big tech companies to outright buy more than enough content if they have to means that regulation to prevent them from gobbling up anything anyone posts online without consent will not stop them. So that isn’t a solution. It will stop new entrants with little cash, but not the big ones. And even OpenAI can afford to buy up some of the largest content owners in the world.

The point was not to make you excited about that, but to illustrate that fighting a battle to restrict what they can train on is fighting a battle that the big AI companies won’t care if they lose - they might even be better off if they lose, because if they lose, while they’ll need to pay more money to buy content, they won’t have competition from open models or new startups for a while.

So we need to find other solutions, because whether or not we regulate copyright to training data, these models will continue to improve. The cat is out of the bag, and the computational cost to improving these models keeps dropping. We’re also just a few years away from people being able to train models competitive to present-day models on computers within reach of hobbyists, so even if we were to ban these models outright artists will soon compete with output from them anyway, no matter the legality.

Focusing on the copyright issue is a distraction from focusing on ensuring there is funding for art. One presumes the survival of only one specific model that doesn’t really work very well even today and which is set to fail irrespective of regulation, while the latter opens up the conversation to a much broader set of options and has at least a chance of providing working possibilities.

vidarh ,
@vidarh@lemmy.stad.social avatar

As long as people aren’t ready for it, then it doesn’t solve the immediate problem that needs to be solved today.

vidarh ,
@vidarh@lemmy.stad.social avatar

It doesn’t need to “develop its own style”. That’s the point. The more examples of these adversarial images are in the training set, the better it will learn to disregard the adversarial modifications, and still learn the same style. As much as you might want to stop it from learning a given style, as long as the style can be seen, it can be copied - both by humans and AI’s.

vidarh ,
@vidarh@lemmy.stad.social avatar

I don’t see these grants or public funding ever covering a private company for one.

Companies are by far the largest recipients of public funding for art in many countries and sectors. Especially for e.g. movie production in smaller languages, but also in other sectors.

And for two, I don’t see AI art ever actually getting to the point where it fully replaces artists.

I do agree it won’t fully replace artists, but not because it won’t get to the point where it can be better than everyone, but because a huge part of art is provenance. A “better Mona Lisa” isn’t worth anything, while the original is priceless, not because a “better” one isn’t possible, but because it’s not painted by Da Vinci.

But that will only help an even narrower sliver than the artists who are making good money today.

It will take time, but AI will eat far more fields than art, and we haven’t even started to see the fallout yet.

Because it is trained to make a homogeneous rendering of what you are looking for

Diffusion models are not trained “for” anything other than matching vectors to denoising to within your own tolerance levels of matching to what you are looking for. Accordingly, you’ll see a whole swathe of models tuned on more specific types of imagery, and tooling to more precisely control what they generate. The “basic” web interfaces are just scratching the surface of what you can do with e.g. Controlnet and the like. It will take time before they get good enough, sure. They are also only 2 years old, and people have only been working on tooling around then for much less than that.

Open AI might be sitting on Microsoft money, but how many other companies has Microsoft gobbled up over the years? Open AI if it starts to struggle will just fall under the Microsoft umbrella and become part of its massive conglomerate, integrated into it. Where are our AR goggles that we are supposed to all be wearing, Microsoft and Google both had those? So many projects grow and die with multiple millions thrown at them. All end up with crazy valuations based on future consumer usage. As we all can’t even afford rent.

OpenAI is just one of many in this space already. They are in the lead for LLMs, that is text-based models. But even that lead is rapidly eroding. They don’t have any obvious lead for diffusion models for images. Having used several, it was first with the recent release of DallE 3 that it got “good enough” to be competitive.

At the same time there are now open models getting close enough to be useful, so even if every AI startup in the world collapsed this won’t go away.

There is also this idea that people wouldn’t willing contribute if just asked.

That’s fine, but that doesn’t fix the financial challenge.

vidarh ,
@vidarh@lemmy.stad.social avatar

So what you are saying is open ai should get the public grants for artists to give to artists?

No. What in the world gave you that idea? I’m saying artists or companies employing artists should get grants, just like is the case for a large number of grants now. I’m saying I’d like to see more of that to compensate for the effects being liberal about copyright would have.

I understand it isn’t trained for anything, I have done training with them. The training leads to homogeneous outcomes. It had been studied as well. You can look it up.

There is no “the training”. There are a huge range of models trained with different intent producing a wide variety in output to the point that some produces output that others will just plain refuse.

Dall-e 3 still isn’t good enough to be competitive.

Dall-E 3 isn’t anywhere near leading edge of diffusion models. It’s OpenAI playing catch up. Now, neither Midjourney or Firefly, nor any of the plethora of Stable Diffusion derived models are good enough to be competitive with everyone without significant effort either, today, but that is also entirely irrelevant. Diffusion models are two years old, and the pace of the progress have been staggering, to the point where we e.g. already have had plenty of book-covers and the like using them. Part of the reason for that is that you can continue training of a decent diffusion model even on a a somewhat beefy home machine and get a model that fits your needs better to an extent you can’t yet do with LLMs.

Asking and crediting would go a long way to help fix the financial challenge. Because it is a start to adding a financial component. If you have to credit someone there becomes an obligation to that person.

If there is a chance crediting someone will lead to a financial obligation, people will very quickly do the math on how cheaply they can buy works for hire instead. And the vast bulk of this is a one-off cost. You don’t need to continue adding images to teach the models already known thing, so the potential payout on the basis of creating some sort of obligation. Any plan for fixing the financial challenge that hinges on copyright is a lost cause from the start because unless it’s a pittance it creates an inherent incentive for AI companies to buy themselves out of that obligation instead. It won’t be expensive.

vidarh ,
@vidarh@lemmy.stad.social avatar

Quick iteration is definitely the big thing. (The eye is fun because it’s so “badly designed” - we’re stuck in a local maxima that just happens to be “good enough” for us to not overcome the big glaring problems)

And yes, if all the inputs are corrupted, the output will likely be too. But 1) they won’t all be, and as long as there’s a good mix that will “teach” the network over time that the difference between a “corrupted cat” and an “uncorrupted cat” are irrelevant, because both will have most of the same labels associated with them. 2) these tools work by introducing corruption that humans aren’t meant to notice, so if the output has the same kind of corruption it doesn’t matter. It only matters to the extent the network “miscorrupts” the output in ways we do notice enough so that it becomes a cost drag on training to train it out.

But you can improve on that pretty much with feedback: Train a small network to recognize corruption, and then feed corrupted images back in as negative examples to teach it that those specific things are particularly bad.

Picking up and labelling small sample sets of types of corruption humans will notice is pretty much the worst case realistic effect these tools will end up having. But each such countermeasure will contribute to training sets that make further corruption progressively harder. Ultimately these tools are strictly limited because they can’t introduce anything that makes the images uglier to humans, and so you “just” need to teach the models more about the limits of human vision, and in the long run that will benefit the models in any case.

vidarh ,
@vidarh@lemmy.stad.social avatar

I feel like you are one of the people who feel that AI is just going to be the future with no real problems to anyone who matters. We can’t stop it, we can’t regulate it in any way whatever; and people should just move out of the way, give up and if they can’t find a place in the new world, die already. Artists don’t matter, writers don’t matter and anyone impacted by this new system doesn’t matter. The algorithm is all that matters.

If I thought that, I wouldn’t have emphasised the need to sort out the funding issue, and argued that just regulation will be insufficient to solve it.

I think it will cause a massive degree of upheaval. I don’t think regulation has any hope in hell of preventing upheaval significant enough that unless a solution is found to ensure better distribution of wealth it will cause violence and uprisings and governments to fall. Not necessarily in and of itself, but in accelerating a process of reducing the monetary value of labour.

I can’t know anything about LLMs, machine learning or anything about this.

I’ve not suggested anything of the sort.

How you can interpret anything I’ve written as suggesting I don’t think there will be problems is beyond me.

You therefore throw out the idea that bias exists due to tagging systems.

I’ve done no such thing.

vidarh ,
@vidarh@lemmy.stad.social avatar

I live in the UK and I don’t drink beer because I generally, ironically, think beer overall tastes like piss, and yet I still know Tsingtao. It has fairly substantial market recognition in a lot of countries.

vidarh OP ,
@vidarh@lemmy.stad.social avatar

Indeed, I don’t see why people clickbait on the Fediverse where there isn’t even any advertising to be had. There’s no benefit.

I posted an article I found interesting, and I avoid editorialising titles. Clearly people here hated that, though I find the reaction pretty bizarre. Oh well, people like what they like.

vidarh OP ,
@vidarh@lemmy.stad.social avatar

The reality of course, is that the reason they want to go after the IRS is because they don’t wan’t them to be able to afford to go after the big fish.

vidarh OP , (edited )
@vidarh@lemmy.stad.social avatar

We actually had to step over still wet seaweed at the bottom of the path (the darker bit at the end of the path) to even get onto the beach at this spot, so the normal tidal variations even during a relatively calm summer weekend still got all the way to foot of the walls. I would not particularly like to see the insurance costs or list of things that are not covered for the closest properties. Most of the properties along that stretch are set above fairly sizable seawalls.

Ryde is facing the Solent, the strait between Isle of Wight and “mainland” England, so I guess it’s more sheltered than some other parts of the island, but they must still get plenty of nasty weather during the winter.

vidarh , (edited )
@vidarh@lemmy.stad.social avatar

Three things: Scale, recency and contrition or perceived lack thereof.

The British Empire is the largest empire there has ever been. At its greatest extent, in 1920, it covered about 1/4 of the entire world, long after having lost many holdings like the US. The second largest, the Mongol Empire, reached almost the same size, but hundreds of years earlier.

In the same time period as the British, the Russian empire covered <20% in 1895, but its proportion of colonial lands to their own was much smaller than for the British Empire and the proportion of the current world population living in those areas is also much smaller. The French colonial empire covered less than 1/10th of the world at its peak in 1920, and was by far the largest recent holding of colonies outside the holding country.

Spain is rarely brought up, I think, in large part because the Spanish empire reached its peak in the early 1800’s and so is “history”. Belgium doesn’t get discussed at much because 98% of their colonial holdings was Leopold II’s personal ownership of the Congo Free State. And then we get to the last bit: Contritition.

Nobody goes around saying the massive scale of gross abuse that happened under Leopold II’s rule of the Congo Free State was a good thing. Few people I’ve met ever defend France’s atrocities in Vietnam. Even the defence of their ownership of Algeria, which was special enough to trigger an attempted coup against Charles de Gaulle when he wanted to let it have independence because many saw it as part of France itself, is relatively muted.

But there’s still mainstream support for the British Empire in the UK. There are still people who insist the British Empire was awesome for the colonies that were exploited because they got English and rails and British legal systems and that somehow outweighs the mass murder and brutal exploitation and erasure of local cultures.

E.g. this survey from 2019, where 32% were proud of the British Empire, 37% were neutral, and only 19% considered it “more something to be ashamed of”. 32% were proud of their country’s history of colonialism and oppression. Critically this was significantly higher than for other colonial powers other than the Dutch. At the same time 33% thought it left the colonies better off vs. only 17% who thought they were worse off.

I’m not British, but I’ve lived in the UK for 23 years, and I’ve experienced this attitude firsthand from even relatively young British people (ok, so all of them have been Tories) - a refusal to accept that the fact that a substantial number of these former colonies had to take up arms to get rid of British rule might perhaps be a little bit of a hint that the colonial rule was resented and wrong.

No other modern empire has left behind such a substantial proportion of the world population living in countries that have either a historical identity tied up to rebelling against British rule, and/or have relatively recently rebelled against British rule, and/or still have substantial reminders, such as Commonwealth membership or the British monarch as their monarch. When a proportion of the British population then keeps insisting this was great, actually, there you have a big part of it.

vidarh ,
@vidarh@lemmy.stad.social avatar

1/4 yes, but also worth mentioning that today far more than 1/4 of the present-day population live in that quarter of the world that has a history of being under British rule in recent history.

Couple that with the UK population being far more likely to be proud of the empire, wish Britain still had an empire, and insist the colonies wee left better off for having been oppressed, the British Empire has a certain stench about it many of the others haven’t, or haven’t anymore because of either age, a greater willingness to admit it was a bad thing, or lack of scale.

vidarh ,
@vidarh@lemmy.stad.social avatar

You see similar issues with French ex-colonies, but since they weren’t as many they don’t appear as much in the news.

Or people aren’t as aware of them. E.g. notably their mandates in Syria and Lebanon after World War 1 where they intentionally stirred divisions on the basis of a theory of wanting to keep it so France as a mediator was needed in order to keep them stable. And then they fucked off and left chaos behind.

vidarh ,
@vidarh@lemmy.stad.social avatar

I think Hong Kong is the rare exception that’s at least possible to reasonably argue, since the alternative was never independence but being ruled by someone granting even fewer freedoms.

vidarh ,
@vidarh@lemmy.stad.social avatar

I interviewed with them once, and they swore up and down that they were cleaning up and divesting of all the harmful stuff, and wanted me to trust they were all about health and a smoke-free future.

Thankfully they were so staggeringly full of bullshit during the interviews that I quickly realized it’d be an absolutely horrifically toxic (groan, yes, sorry) place to work irrespective of my other doubts, and I ended up telling them I didn’t want to continue the process and that I was so unhappy with the assorted bullshit during the process that I didn’t want to ever be approached by them again.

That’s the very long way of saying I’m not the slightest bit surprised it turns out they are in fact still massive asshats, and I’m very happy I caught on early enough.

CIA publicly acknowledges 1953 coup it backed in Iran was undemocratic as it revisits 'Argo' rescue (apnews.com)

While revealing new details about one of the most famed CIA operations of all times — the spiriting out of six American diplomats who escaped the 1979 U.S. Embassy seizure in Iran — the intelligence agency for the first time has acknowledged something else as well....

vidarh ,
@vidarh@lemmy.stad.social avatar

When you are the only one here who is publicly stating your support for a coup that brought a mass-murdering lunatic into power you do not have the moral high ground on anything at all.

vidarh ,
@vidarh@lemmy.stad.social avatar

Nobody will take the authority of someone who is on record in this very thread supporting a mass murderer.

vidarh ,
@vidarh@lemmy.stad.social avatar

“They had the right colour uniform on when they carried out the mass murder, and that makes it better” <- how you come across.

vidarh ,
@vidarh@lemmy.stad.social avatar

So you’re supporting an actual, proven far-right extremist mass murderer over a democratically elected leader who didn’t kill anyone because in your head you imagines crimes he never carried out. That is the typical mindset of a fascist.

vidarh ,
@vidarh@lemmy.stad.social avatar

I’ve literally never done anything of the sort, you liar.

vidarh ,
@vidarh@lemmy.stad.social avatar

I’m neither violent or a conservative, nor will your ongoing attempt to justify mass murder make it so. For someone who keeps calling people fascists, you’re the only one in this thread supporting a far-right extremist authoritarian, mass-murdering dictator.

vidarh ,
@vidarh@lemmy.stad.social avatar

“Better mass murder than the fiction I invented in my head” is the comeback you think it is. All it underlines is that you’re doubling back on the support of violence and destruction of democracy when it fits your extremist ideology.

vidarh ,
@vidarh@lemmy.stad.social avatar

I have never in my life supported any violent coups or putting mass murderers in charge by any means. You have on the other hand repeatedly done so in this thread alone.

vidarh ,
@vidarh@lemmy.stad.social avatar

In reality the dictator you’ve repeatedly expressed support for here killed thousands, while unlike you I’ve never supported any oppressive, mass-murdering government of any kind.

Actual murders trumps your fictional head canon any time.

vidarh ,
@vidarh@lemmy.stad.social avatar

Allende didn’t carry out any mass murder. Pinochet did. Hence your purported “knowledge” about what Allende might have done had your preferred mass murderer not taken power is only in your head.

Trying to set up strawmen by pointing to entirely different regimes that nobody in this thread have expressed support for does not change the fact that you’re still the only one here repeatedly arguing in favor of someone who actually carried out mass murder.

vidarh ,
@vidarh@lemmy.stad.social avatar

And again, actual murders trump your fictional headcanon any time. This “it’s for the good of everyone” argument is the traditional argument of fascists to justify dismantling democracy and hunting down political opponents. Every brutal dictator and their supporters think they are the heroes and project their own brutality on the people they murder while they leave a trail of blood behind them.

vidarh ,
@vidarh@lemmy.stad.social avatar

You’ve only “addressed this” by repeating your imaginary scenario of thought crime as justification for arguing that bringing mass murderer into power was right, just like the supporters of every brutal dictator in history though their favourite mass murderer did what they needed to as well.

It’s not actually “addressing” anything - all you’ve done is doubling down on trying to justify actual, real murders that happen with your fear of something hypothetical. This irrational fear of hypothetical harm and willingness to preemptively harm the other first, driven by a relentless bloodlust lies at the core of the fascist psyche.

vidarh ,
@vidarh@lemmy.stad.social avatar

I’ve addressed your points repeatedly, while you resort to fiction and defamatory lies.

vidarh ,
@vidarh@lemmy.stad.social avatar

I remember growing up in the 1980’s and 90’s when there were still a horrifying amount of people who refused to believe CIA did things like that at all, even in a relatively left-wing country like Norway.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • lifeLocal
  • goranko
  • All magazines