[email protected]

Seems to be purely to post misinformation with repeated claims that Russia is innocent and the US caused the Ukraine situation, that they’re stopping Ukraine from agreeing to Russia’s super amazing peacedeals, etc.

This is the sort of garbage one would expect to find on ML or Hex, is CA intended to be the same low quality instance?

  • antiykns
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    3
    ·
    edit-2
    6 hours ago

    It’s well known that peace activists are just enemy spies. Everyone who protested against the Iraq War was a Saddam fanboy, all hippies were really bolchevik agent, and so on /s

    Seriously, why aren’t you in ukraine if you care about it so much? Ah, that’s why, warmongerers only want other people to serve as cannon fodders

  • KingOfTheCouch@lemmy.ca
    link
    fedilink
    arrow-up
    10
    ·
    18 hours ago

    First let me say: I’m glad you’re making this observation. The shitastic propaganda machines need to be called out.

    Now second, I have a point to make but I should give some context first. I already blocked that community and many others about politics and war and crap because I’m just tired of all of it. That said, before I did, it didn’t seem like endless war was all (or even mostly) anti-ukraine in nature. It did seem anti-US but I wasn’t about to go looking into that statistically - just an observation regarding the titles of posts that did percolate up my feed from time to time.

    Now, I’m not an admin here, but if I were (based on experiences in past adminning forums, chatrooms and reddit subs) I would ask for citations and proof of the problem. It’s all well and good to just say “go look for yourself” but people have jobs and other interests. I regularly would ignore these things, but if someone went to the effort to present proof? Then you had my interest. Like, don’t reply to me with it, edit it back in your original post - like I say, I’ve blocked it and I’ve moved on.

    I would be looking to know if there are multiple accounts posting shit or is it one or two “power users” posting the majority of it. Were the mods of the sub alerted to it? (Obvious moot point if it’s the mods posting…) Showing the response from the mods will really sell what tone they’re aiming for in the sub. Hell - if they have it written in their community header info that they are their to pander to putin, that’s evidence too. “Seems to be”, as you wrote, is highly conjectural.

    Anyway, that was my two cents (and, yes, with the state of the dollar and the fact we have no pennies anymore that means it’s worth fuck all.) I wish you good luck in your mission to expose the assholes of the world - we do need to stay vigilant.

    And also because I might not have been clear: fuuuuuuuuuck russia.

  • Nils@lemmy.ca
    link
    fedilink
    arrow-up
    9
    arrow-down
    1
    ·
    17 hours ago

    If you check the history of lemmy.ca, you will see that this intolerant propaganda/misinformation is cyclical. One user disappears and another comes in. Sadly, it usually takes a few months before something happens to them.

    The mod of the community you shared, created an account and on the same day started to post propaganda and tiptoeing around the instance rules. Before, it was another user posting the same content from the same sources, with the same tactics, until they went a bit too far and were banned.

    They always do the same thing, create an account, create a community with a “normal” name, like geopolitics, and start spreading stuff. I would not be surprised if it is an agency doing this kind of stuff: I cannot imagine someone being so evil to do it willingly, might be either coerced or depending on the money.

    I do not mind when the propaganda is benign, like bots posting random video game or Canadian news, but I draw the line on intolerance.

    Another of those intolerant people tactics (tiptoeing) is that they do not demand for people to get killed in their comments. They construct it as a consequence of the victims’ actions, they deserve to die. For me, that is just as bad, if not worse, because they know the things they are doing are wrong, and they are trying hard to no get caught.

    I hope the instance comes with a better and swift way to deal with these kinds of problems.

  • Rottcodd@lemmy.world
    link
    fedilink
    arrow-up
    22
    arrow-down
    6
    ·
    edit-2
    20 hours ago

    While I agree with your assessment, I’d note that pretty much everyone at this point declares that whatever views they disagree with are “misinformation,” so proactively banning things solely because someone has declared that they’re “misinformation” isn’t a sound strategy.

    And again, I agree with that assessment in this case. But that’s really beside the point.

    • WrenFeathers@lemmy.world
      link
      fedilink
      arrow-up
      11
      arrow-down
      2
      ·
      edit-2
      19 hours ago

      There a huge difference between what one thinks is misinformation, and what is proven to be misinformation though. It shouldn’t be hard for admins to suss the difference.

      • Rottcodd@lemmy.world
        link
        fedilink
        arrow-up
        8
        arrow-down
        5
        ·
        18 hours ago

        There a huge difference between what one thinks is misinformation, and what is proven to be misinformation though.

        Epistemologically, yes. But for all practical purposes, at this point in time, there really isn’t, since anyone can find sources that purportedly “prove” that whatever they want to believe is true and/or that whatever they don’t want to believe is “misinformation.” It makes absolutely no difference what the claim in question is - somebody somewhere online has “proven” that it’s true, and somebody else somewhere online has “proven” that it’s not.

        So what that means is that to avoid the trap of endlessly dueling contradictory claims, somebody is going to have to simply decree what is or is not to be considered to be true - which sources and purported proofs are legitimate and which are not - and that’s where it inevitably goes wrong.

        And in fact, to go all the way back to the start of this thread, that’s exactly how hexbear and ml work. They maintain their bubbles by essentially arbitrarily decreeing that [this] is true and [that] is misinformation. And if you press them on it, they’re more than willing to post links to the “proof.”

      • howrar@lemmy.ca
        link
        fedilink
        arrow-up
        6
        arrow-down
        4
        ·
        18 hours ago

        How can any of this actually be proven to be misinformation? We’re here on our couches reading second/third/nth hand information. None of us were in the rooms where these decisions were made. None of us are on the front lines. The best we can do is make an educated guess on who is a credible source, and that’s especially difficult when everyone involved has an interest in lying about the situation when things don’t go their way.

    • sunzu2@thebrainbin.org
      link
      fedilink
      arrow-up
      6
      arrow-down
      3
      ·
      20 hours ago

      I am with you…

      OP, if you see misinformation correct the record so that everyone else can see there is a dispute and they can make up their own mind.

      Hard Suppression of opinions is not good IMHO

      Let idiots speak and let their idea be challenged.

      Tankies and political subs expose their bias via the modlog and that’s the point.

      Chanllange them and let them use that hammer so everyone can see their idiotic positions what they are

      • running_ragged@lemmy.world
        link
        fedilink
        arrow-up
        2
        arrow-down
        1
        ·
        19 hours ago

        Hard Suppression of opinions is not good IMHO

        Tankies and political subs expose their bias via the modlog and that’s the point.

        If the modlog is that questionable, they should be given a fair chance to provide receipts for their censorship of users trying to correct the record, and if they can’t do that fairly, it should be on the admins to remove a community that is both spreading misinformation and censoring verifiable corrections.

        I feel it shouldn’t be only up to the users filtering / monitoring modlogs of all the communities.

  • Nils@lemmy.ca
    link
    fedilink
    arrow-up
    4
    arrow-down
    1
    ·
    17 hours ago

    I do not understand people here defending misinformation/intolerance as a merit of discussion. The dichotomy of naive or complicity.

    People spreading misinformation and intolerance are not here for healthy arguments, you just need to check their history to see their dishonesty and ill temper.

    In the meanwhile, accounts like the one OP highlighted are just creating trouble for mods of other instances to solve.

    • Rentlar@lemmy.ca
      link
      fedilink
      arrow-up
      4
      arrow-down
      2
      ·
      edit-2
      16 hours ago

      The problem is, who is the arbiter of that? There are essentially 3 types of moderation styles here:

      Laissez-faire: Let people do whatever as long as it doesn’t actively hurt anyone. People can govern themselves and serious incidents are expected to be reported and dealt with. Some jerks will tiptoe around the rules but will eventually get caught. Lemm.ee, lemmy.ca and some others follow this.

      Casual enforcement of admin-philosophy: Most topics outside of politically contentious ones are not strictly monitored. Mods/admins will root out communities, comments and posts that actively go against the narrative, particularly on threads on political topics like Ukraine, Palestine, etc… Lemmy.world and lemmy.ml follow this.

      Strict enforcement of admin-philosophy: do not tolerate any potentially harmful statements (to that instance’s narrative or vibe). Any violation will be removed and repeated violations get you banned. This philosophy can be reasonable like Beehaw.org, which I think works very well for them and makes it a welcoming safe space, because there is no tolerance for bigotry and jerks. It can also be unreasonable like lemmygrad.ml, where dissent to the pro-Russian narrative is swiftly dealt with.

      Admins of other instances should ban users that go against their philosophy from reaching their servers, if they follow the latter two styles of moderation. That’s how it is with federation, sometimes different instances have conflicting philosophies (the vegan one for example). It’s up to each admin to decide whether a foreign Fediverse user belongs in their kingdom. The moderation style that lemmy.ca has lets it be a good neutral place to discuss various drama and lore from other servers.

      • Nils@lemmy.ca
        link
        fedilink
        arrow-up
        2
        ·
        9 hours ago

        The problem is, who is the arbiter of that?

        Intolerance is well-defined in many languages, and, so people do not confuse I am talking about milk intolerance, the hate crime is defined in many law codes across the globe, including Canada. There is no need for philosophical discussion of what is “intolerance”.

        There is no need for a linguistic expert to realize someone’s discourse is ill intentioned, when the semantics of “the victim deserves to suffer” is the same as the call to action.

        For countries that depend on common law, the account in question was already punished in other instances, creating precedent.

        The modus operandi of these kinds of accounts are also well-know and documented. And popularity contests should not be a tool to define what is right in an online platform where there is no real accountability. How many upvotes do you think a single worker in a troll farm can generate in a couple of minutes?

        We should not depend on admin humour for results (philosophies, as you suggest), but I agree that we should help when/where we can, their volunteer work is invaluable for the health of the instance.

        I think that the discussions worth having in these kinds of posts are about methods, checks and balance to prevent bad decisions from people in power, and that people will be fairly treated.

        Methods are many, and there are many examples out there.

        • would twitter like community notes solve some of these problems or create more? Would lemmy repo accept such PR?
        • the problem of twitter x Brazil: is it worth locking those accounts while an investigation is pending? One of them was instigating machete attacks in school/nursery. When would this lock be ok, or not ok?
        • how long should people complain/report before a something (an investigation, a lock, or a conclusion) happens. - The account we both mentioned not in this thread (but in this post) went on it for 2 months before being banned - they did not leave on their own. …
        • Rentlar@lemmy.ca
          link
          fedilink
          arrow-up
          1
          arrow-down
          1
          ·
          1 hour ago

          Sure, we should not tolerate intolerance, “No Bigotry” is rule #1 here so if you see that then please report it. Misinformation, though? That’s the main thing OP is talking about and they gave a few examples, they are propaganda but not intolerance.

  • Avid Amoeba@lemmy.ca
    link
    fedilink
    arrow-up
    8
    arrow-down
    4
    ·
    20 hours ago

    Unlike Xitter and Reddit where black box algorithms spread information to users’ feeds, Lemmy uses people’s vote to increase or decrease proliferation. It seems to me that the posts in that community aren’t going anywhere given how people have voted on them. The primary filter seems to work as expected. Maybe there isn’t need for another.

    • Deceptichum@quokk.auOP
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      5
      ·
      edit-2
      20 hours ago

      Having a less popular Nazi bar is still having a Nazi bar.

      And communities take time to grow, it’s only a month old. There isn’t any benefit in letting it fester until it’s a bigger threat.

      • Avid Amoeba@lemmy.ca
        link
        fedilink
        arrow-up
        7
        arrow-down
        3
        ·
        edit-2
        20 hours ago

        If the litmus is not having a Nazi bar, I don’t think that’ll ever happen unless we gate community creation. On the wider fediverse, it’ll never happen. I think it’ll always be about how unpopular it is and we should use that as the litmus. The scaleable approach is people’s votes and personally this is why I’m on Lemmy.

        • Deceptichum@quokk.auOP
          link
          fedilink
          English
          arrow-up
          8
          arrow-down
          4
          ·
          20 hours ago

          So you suggest never trying to do anything, because you can’t be sure it will 100% be gone forever?

          Think of it like a weed, when you see one you pull it out and move on with your day. Gardens require constant tending to.

          • Avid Amoeba@lemmy.ca
            link
            fedilink
            arrow-up
            6
            arrow-down
            2
            ·
            edit-2
            20 hours ago

            I’m not suggesting that. I’m suggesting that we’re already doing something and it seems effective. It seems to me the democratic process we have is already pulling the weed out.

            • Avid Amoeba@lemmy.ca
              link
              fedilink
              arrow-up
              2
              arrow-down
              3
              ·
              20 hours ago

              I won’t be mad if the admins delete that community, but it seems we’re already controlling how far it gets. If I saw high upvote rates which makes the misinfo spread wide, then I’d say we’re not doing a good enough job through the standard process and perhaps an additional action is needed.

          • Deceptichum@quokk.auOP
            link
            fedilink
            English
            arrow-up
            7
            arrow-down
            4
            ·
            20 hours ago

            Deleting a community is worse than Russia invading Ukraine?

            Huh good to know.

            • iii@mander.xyz
              link
              fedilink
              English
              arrow-up
              5
              arrow-down
              3
              ·
              20 hours ago

              If we examine the two possible actions: (1) banning the community, and (2) not banning the community. Then we can see that neither (1) nor (2) have influence on Russia invading Ukraine. That has already happened and is happening.

              Both actions, (1) and (2), are not worse than the invasion.

              • Deceptichum@quokk.auOP
                link
                fedilink
                English
                arrow-up
                8
                arrow-down
                1
                ·
                20 hours ago

                Russian propaganda has an impact on it.

                Deliberately allowing yourself to be a hub to spread it is ridiculous.

                • iii@mander.xyz
                  link
                  fedilink
                  English
                  arrow-up
                  3
                  arrow-down
                  4
                  ·
                  edit-2
                  10 hours ago

                  Russian propaganda has an impact on it.

                  I totally agree.

                  Deliberately allowing yourself to be a hub to spread it is ridiculous.

                  I think this argument is our main point of disagreement. In my opinion, censorship is a bad way to address misinformation. It creates parallel groups of thinking, each, from their point of view, growing more confident that they know all.

  • Rentlar@lemmy.ca
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    17 hours ago

    We have had multiple communities of that nature, all heavily downvoted. A previous example was Geopolitics where Russian and Conservative narratives were pushed daily. Eventually the creator, poster and community mod gave up, left and deleted it.

    You can always block communities or instances that you don’t want to see. Every instance has its own policy on what types of communities are allowed, and how strict they must align with admin values.

    I frequently disagree with Russian apologists, but pre-emptive restriction of viewpoints I don’t like doesn’t make for good discussion, even if in my eyes many of the arguments are on dubious ground. You see many complaints of lemmy.world and lemmy.ml admins enforcing certain policies and worldviews sitewide, which is fine for them to do, but not every server has to ascribe to that. Some servers restrict community creation to mods/admins, some like beehaw.org have a limited but curated set of communities. [email protected] is a hotbed of clown-take articles, doesn’t mean I think they should be banned.

    If you see posts with harmful misinformation, or harmful behaviour by the mod of such a community, please report it to the lemmy.ca admins. Demonstrating a pattern of harmful behaviour with evidence will get the mod and community banned.

    I vehemently disagree with an article posted there that having their experimental Mach-whatever missiles means that Russia and China’s going to get everything they want in a conflict, to me it’s a total bluff. But it was written by a Canadian so it would seem it belongs there, even if it is indeed a blatantly pro-Russian narrative.

    Perhaps as a solution going forward: new communities start as private communities for the instance only, and then upon admin review and approval can be federated to other instances?