Debug
简体中文 | English
Q: How to achieve higher QPS?
- A: When the QPS setting is too high, it can result in excessive server resource utilization, impacting business operations. To prevent accidental misconfiguration in production environments, kdoctor has added QPS checks in the webhook.
- If you wish to use a higher QPS, you can set the QPS limit using the parameter
--set feature.appHttpHealthyRequestMaxQPS=1000
,--set feature.netReachRequestMaxQPS=1000
,--set feature.netDnsRequestMaxQPS=1000
,You can also modify it through the configmap in kdoctor, And restart the relevant pods of kdoctor to reload the configmap.
Q: Why is my task unable to achieve the desired QPS ?
- A:There are several reasons why the expected QPS cannot be achieved, primarily categorized into the following reasons:
- The concurrency worker setting is too low. kdoctor can adjust the concurrency by setting the parameters
--set feature.netReachMaxConcurrency=50
,--set feature.appHttpHealthyMaxConcurrency=50
,--set feature.netDnsMaxConcurrency=50
. -
The kdoctor agent may have insufficient resource allocation. You can use the kdoctor aggregate report
kubectl get kdoctorreport
to check the CPU and memory usage of the task. This will help you determine if the resource allocation for the kdoctor agent is sufficient.shell ~kubectl get kdoctorreport test-task -oyaml ... SystemResource: MaxCPU: 52.951% MaxMemory: 120.00MB MeanCPU: 32.645% ...
-
Whether the kdoctor agent is concurrently executing other tasks and occupying resources can be determined by checking the QPS count of other tasks being executed simultaneously. You can use the kdoctor aggregate report
kubectl get kdoctorreport
to view the QPS count of other concurrently running tasks. Stagger the task execution time or isolate the task by defining agentSpec to specify the kdoctor agent to execute the task. Because QPS statistics are time-sensitive, they can be used together with the log as a reference. Before the task execution starts, the currently executing QPS will be output to the log。~kubectl logs kdoctor-agent-74rrp -n kdoctor |grep "Before the current task starts" {"level":"DEBUG","ts":"2023-11-07T10:01:02.821Z","agent":"agent.agentController.AppHttpHealthyReconciler.AppHttpHealthy.test-task.round1","caller":"pluginManager/agentTools.go:90","msg":"Before the current task starts, the total QPS of the tasks being executed is AppHttpHealth=100,NetReach=0,NetDNS=0","AppHttpHealthy":"test-task"}
~kubectl get kdoctorreport test-task -oyaml ... TotalRunningLoad: AppHttpHealthyQPS: 100 NetDnsQPS: 50 NetReachQPS: 0 ...
Q: Why does the Kdoctor agent experience OOM (Out of Memory) issues ?
- A: The Kdoctor agent serves as the default agent for executing tasks when no specific agent is specified in the task. Currently, the agent does not support features to reject or delay task execution based on the workload. Therefore, when the Kdoctor agent concurrently handles a large number of tasks, the high request volume combined with low memory limits can lead to memory overload and result in OOM (Out of Memory) errors. To mitigate this issue, you can stagger the usage of the Kdoctor agent based on the task workload, adjust the memory limits accordingly, or isolate tasks using specific agents that are better suited for the workload.