We’re aware of ongoing federation issues for activities being sent to us by lemmy.ml.

We’re currently working on the issue, but we don’t have an ETA right now.

Cloudflare is reporting 520 - Origin Error when lemmy.ml is trying to send us activities, but the requests don’t seem to properly arrive on our proxy server. This is working fine for federation with all other instances so far, but we have seen a few more requests not related to activity sending that seem to occasionally report the same error.

Right now we’re about 1.25 days behind lemmy.ml.

You can still manually resolve posts in lemmy.ml communities or comments by lemmy.ml users in our communities to make them show up here without waiting for federation, but this obviously is not something that will replace regular federation.

We’ll update this post when there is any new information available.


Update 2024-11-19 17:19 UTC:

Federation is resumed and we’re down to less than 5 hours lag, the remainder should be caught up soon.

The root cause is still not identified unfortunately.


Update 2024-11-23 00:24 UTC:

We’ve explored several different approaches to identify and/or mitigate the issue, which included replacing our primary load balancer with a new VM, updating HAproxy from the latest version packaged in Ubuntu 24.04 LTS to the latest upstream version, finding and removing a configuration option that may have prevented logging of certain errors, but we still haven’t really made any progress other than ruling out various potential issues.

We’re currently waiting for lemmy.ml admins to be available to reset federation failures at a time when we can start capturing some traffic to get more insights on the traffic that is hitting our load balancer, as the problem seems to be either between Cloudflare and our load balancer, or within the load balancer itself. Due to real life time constraints, we weren’t able to find a suitable time this evening, we expect to be able to continue with this tomorrow during the day.

As of this update we’re about 2.37 days behind lemmy.ml.

We are still not aware of similar issues on other instances.


Update 2024-11-25 12:29 UTC:

We have identified the underlying issue, where a backport for a bugfix resulting in crashes in certain circumstances was accidentally reverted when another backport was applied. We have applied this patch again and we’re receiving activities from lemmy.ml again. It may take an hour or so to catch up, but this time we should reliably be getting there again. We’re currently 4.77 days behind.

We still don’t have an explanation why the logs were missing in HAproxy after going through Cloudflare, but this shouldn’t cause any further federation issues.


Update 2024-11-25 14:31 UTC:

Federation has fully caught up again.

  • @DarkCloud@lemmy.world
    link
    fedilink
    24
    edit-2
    1 month ago

    I’ve never had a positive interaction with Lemmy.ml. For me it serves as a quarantine space, and a set of pre-tagged users I don’t personally enjoy dealing with.

    …and I’m not particularly averse to Marxists sentiments either, but they’re certainly not good sales people, diplomats, or representative of their cause.

    Which is just part of their reputation now. Having a bad experience with a .ml user seems to be part of the lemmy experience. It’s kind of comical how consistent it seems.

    That said, I’m sure there’s good people on .ml.

    • @Pilferjinx@lemmy.world
      link
      fedilink
      -31 month ago

      Yeah, just avoid politics there and it’s fine. We all know about their zeal so it’s pointless to discuss it.