originalucifer

joined 9 months ago
[–] [email protected] 1 points 6 months ago (1 children)

its running on AWS, US. you can see on 4/24 i finally solved for a long running bottleneck (disk writes), and it was smooth sailing until recently.

im not sure if anyone else has subscribed to any new communities, but ive been subbed to most of the big ones for awhile.

i don't see any saturation in the server network activity, connections, etc. my queues are not overloaded, and as i mentioned other instances are very performative.

maybe i just need to be patient and let it catch up. those graphs are awesome

 

I cant seem to get timely post federation to my instance (moist.catsweat.com) from lemmy.world. I might see a new post every few hours instead of every minute or so.

posts from other lemmy instances (lemmy.ca, lemmy.ml, dbzer0, etc) have no problem.

i see a ton of lemmy.world traffic in my nginx logs... it seems very specific to 'new posts'. voting/comments seem to come through no problem once the post shows up.

ive checked a few other instances similar to mine, and they are not having this issue. so it seems something specific to lemmy.world and my instance.

i do recall a similar issue several months back, and it involved a lemmy.world admin resetting some outbound queue.

any ideas?

 

im not seeing any traffic suddenly from lemmy.world to my instance at moist.catsweat.com

yes, its not a lemmy instance, but the 'bins are solidly interoperable and i was receiving inbounds from lemmy.world better than 2r/s before it just stopped.. usually a ton of POST to /f/Inbox

i have been blocking commercial bots left and right, but i checked all my logs and i dont see what i expect the lemmy.world ip address to be in any of them as blocked/error.

still successfully receiving inbound post from all other lemmy instances.

what could be the issue here?