-
Notifications
You must be signed in to change notification settings - Fork 726
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
PD can't redistribute the hot write regions among TiFlash nodes #3261
Comments
Does not like leader and followers in TiKV, all TiFlash learner peers are ready for reading. And TiDB chooses TiFlash peer by using round-robin. So I think that reporting Are there any key metrics for PD scheduling besides |
Will this be a fix in 4.0.x version? ZTO (one of our largest cluster user is trying TiFlash on 4.0). This issue is a troublesome one. |
Bug Report
What did you do?
I deployed a cluster with 1 TiDB + 1 PD + 1 TiKV + 2 TiFlash base on version v4.0.8.
The TiFlash and its proxy branch are "store_stats_4.0". These two branches fix the problem that the written bytes and written keys at the store level are not reported to PD.
What did you expect to see?
If the write-pressure between TiFlash nodes is not balanced, PD can generate some
move-hot-write-region
operators and make it balanced.What did you see instead?
By adding a sysbench workload on this cluster, I found that:
move-hot-write-region
between TiFlash store rarely happen.move-hot-write-region
between TiFlash stores.Another problem, maybe related or not:
I use the PD API:
/pd/api/v1/hotspot/regions/write
to check the stats of hot write regions. In the TiFlash node, the flow bytes by summing all regions is about 4 times to the TiFlash reported.What version of PD are you using (
pd-server -V
)?The text was updated successfully, but these errors were encountered: