hello,
bot or not???
hello,
bot or not???
ça sert à quoi le #nobot? y'a des bots qui spamment les gens au pif qui respectent ça?
@petrescatraian I noticed that.
will correct it!
by the way, you are the most advanced user;)
#nobot
Hrmmm I have #nobot in my profile and I *still* get those damn bots boosting me. Maybe it's case sensitive ;) or I made a mistake?
Interesante programa para proteger nuestras páginas web de scrappers indeseados.
#noAI #noIA #nobot #anubis
https://anubis.techaro.lol/docs/
J'ai changé ma description, parce que j'avais peur que le #nobot soit pris comme un instruction de ne pas faire de demande de suivi destinée aux entités qui s'identifient comme non organique. Vous êtes bienvenu-es, je vous aime!
@manna Ilmeisesti joku dataharvester / suodatustyökalu. JSONin lopussa oli jotain linkkejä projektiin, sielltä löysin "poistumistoiveen" (opt-out) ohjeen: https://github.com/awakari/.github/blob/master/OPT-OUT.md . #nobot riittänee ehkä profiilissa.
@eatyourglory @noondlyt It's a proposal to add an extension/field to the protocol to let people explicitly express consent or lack of consent for a few different types of uses of their data. Something that could very well be a thing in ActivityPub too, if it isn't already. Sort of like the #NoBot hashtags people put in the bio, but in a more organized format.
@MalthusJohn we allow bots on mas.to, but have very specific rules about how they can operate:
“All automated (bot) accounts must enable the bot flag in their account preferences, respect #nobot markers, and post as unlisted unless making less than one post per hour.”
So they’re easy to block and won’t flood your local timeline.
I’d say our moderation team is great, but, then, I’d be biased.
Seriously, though. I think we do a pretty good job, and we’d love to have you.
Hallo @printplayprotect Wieso bekomme ich eine Folgeanfrage von diesem Account? In meinem Profil steht explizit #nobot
There has been several times the request to make the Markov bot (like @Quackomatic ) talk to each other.
Two safety guards are against this:
1. The bot only talks to the account who mentioned it. All other accounts are ignored. This is so no one can annoy people by tagging the bot in a hell thread and I don't have to check for the #nobot flag
2. Answers to bots are blocked in the code at the moment.
This is just because I didn't want to program a routine to check how far in a conversation thread we are so it stops after, e.g. 20 notes.
The system would be like the reply system that was at work with @feli_bot@universe.feliver.se . This bot also mentioned all accounts who also was mentioned in the reply but only to a depth of 10 or 15 notes.
Poll time (Multiple selection possible):
Should I make a "Summoned MIKU" bot that posts the "OOO-EEE-OOO" video to every post mentioning Hatsune Miku (and ofc respects the #nobot flag)
And
Should I look into a bot that Markov-Chains Bible verses? (could be offensive to people who are Christians)
@AccidentalHaiku Your bot found and boosted my post that used no hashtags, meaning it had to scrape and process text. I also have had #NoIndex and #NoBot in my profile for years now. Does this bot intentionally disregard the filter flags or is this an oversight?
Please correct this to respect the flags some of us use to prevent this kind of data capture.
So... not a topic getting much interest, let me decide by myself.
Not a DoS as requests are coming every other second... not blocked by fail2ban as IP's are numerous and changing... not causing any trouble, I don't see what they can achieve by doing this as requests seem legitimate and reading random posts... but clearly not human nor social interactions, so undesirable. Maybe scraping content for AI training?
I will block the offending IP's this evening. This should not affect our other Singaporean users, who, let's be honest, are (understandably) not very many. Nor will it affect federation as user agent is voluntarily misleading.
I encourage other administrators to monitor this kind of case and to take action if necessary. Let's try and avoid scraping bots when we can.