-
Notifications
You must be signed in to change notification settings - Fork 54
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Replication lag #864
Comments
we see in metrics: sometimes the lag can grow to a 5583808. |
Hi @tellienne are u able to check the cpu and memory usage of sink connector |
judging by k8s metrics CPU and memory usage has remained virtually unchanged since the first start. At the time of the snapshot, the connector uses 4GB of memory; after the snapshot, the use does not decrease. when replication starts to slow down there is no change in CPU and memory usage. the debezium_postgres_connector_metrics_TotalNumberOfEventsSeen metric shows that 92,000 events were transmitted in the last 7 hours. what other information can we provide? |
can you check the |
We checked it on the last lag.
|
Could this be causing the lag? are you running in if its multi threaded, you might want to increase the buffer flush time. |
we use default settings. As I understand it, by default it works in a multi-thread. |
added to config
But again we see a lag. In half an hour we see about 35,000 events and the lag is already 30 minutes, judging by the heartbeat table. There seems to be no effect. |
thanks for the ideas and your help :) The status endpoint only shows
without lsn UPD: we updated the connector to version 2.3.1 and the problem with offset in single-threaded mode was fixed |
It looks like switching to single.threaded solved the problem. Over the past hours, the clickhouse_sink_debezium_lag metric has not shown more than 120,000 ms. |
Hello!
Thank you for your work, we really like the connector-lightweight.
Recently we noticed that replication at some point begins to lag, although there are not many events in the database.
After the snapshot, replication works fine for a couple of hours, but then a lag begins and continues to grow.
We have a heartbeat table into which Debezium inserts a value every 15 seconds.
And it is clearly visible that the lag has begun.
Our connector is deployed in k8s, v2.3.0-lt.
On the k8s node at this moment there is enough CPU and RAM, in clickhouse and postgres too.
Do you have any idea why replication becomes so slow?
Also at this moment insertions into the altinity_sink_connector.replica_source_info table in the clickhouse become slow. Several seconds may pass between a row being deleted and a new one appearing in it.
The text was updated successfully, but these errors were encountered: