Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

TiCDC OOM(Kafka Sink ) #5163

Closed
Bowen-Tang opened this issue Apr 13, 2022 · 4 comments
Closed

TiCDC OOM(Kafka Sink ) #5163

Bowen-Tang opened this issue Apr 13, 2022 · 4 comments
Assignees
Labels
area/ticdc Issues or PRs related to TiCDC. severity/moderate type/bug The issue is confirmed as a bug.

Comments

@Bowen-Tang
Copy link

Bowen-Tang commented Apr 13, 2022

What did you do?

The operations on the application side is not yet identified. The default transaction size is 100MB, and the downstream is Kafka. Meanwhile, the changefeed is configured to synchronize single table.

The attachment contains golang heap and table schema.
cdcoom.zip

What did you expect to see?

No response

What did you see instead?

No matter the VM is configured with 32GB or 64 GB memory, CDC drains the memory entirely.

Versions of the cluster

V5.3.1

@Bowen-Tang Bowen-Tang added area/ticdc Issues or PRs related to TiCDC. type/bug The issue is confirmed as a bug. labels Apr 13, 2022
@zhaoxinyu
Copy link
Contributor

This issue is under investigation.

@CharlesCheung96
Copy link
Contributor

@Bowen-Tang Can you provide the monitoring data and detailed version information of cdc?

@Bowen-Tang
Copy link
Author

version: 5.3.1
@CharlesCheung96 Sorry,only this monitoring data:
9b227aded7f25335cc3cdd4e73e517c
1ec3197682f2b30df47451fa6c8e2fc
f1c40488039cbbd5d7e54289aa889cd
338f64a8fff3470467651904e04d3ca
a00d599df750f4c71feedecc6ad9940
e04ca426e44e7e7239f8e5b1196fd60
8e166eb84f52942a8249639db4869f3
d649084e763c3c9583963e986edcfc4
ce2da28bf57ab5da3bc483e42469718

@CharlesCheung96
Copy link
Contributor

image
@Bowen-Tang According to Puller's monitoring, the workload of this CDC node is relatively abnormal, which is far beyond the throughput capacity of the downstream under normal circumstances. Therefore, we highly suspect that inconsistent upstream and downstream throughput will cause data accumulation in memory, which will eventually lead to OOM.

This pr alleviates the OOM problem caused by upstream and downstream throughput to a certain extent, so we temporarily closed the issue. If similar problems occur again, feel free to reopen it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/ticdc Issues or PRs related to TiCDC. severity/moderate type/bug The issue is confirmed as a bug.
Projects
None yet
Development

No branches or pull requests

4 participants