-
Notifications
You must be signed in to change notification settings - Fork 147
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
too much dns request on cf-deployment , via loggregator 106.x #401
Comments
We have created an issue in Pivotal Tracker to manage this: https://www.pivotaltracker.com/story/show/170456276 The labels on this github issue will be updated when the story is started. |
@jyriok This issue is related to a fix that we had made to actually get the bosh dns health check working for our doppler instances. Can you confirm that this high cpu still happens after you have updated to Bosh-DNS version 1.16? I am sure you are aware but in case you are not, Bosh-DNS is included in a runtime config in most environments so changing your version of CF-Deployment won't change it. |
@chentom88 hello ! thank for your answer. yes, i'm aware about issue fixed by bosh-dns 1.16 about using tls 1.2 instead 1.3 but i already use bosh-dns 1.16 :( so i thinks it's another issue. here my bosh deployment with releases i use
|
@jyriok and @tlwr, I believe the spike in CPU on both the |
Also been seeing this issue, currently running CF Deployment 12.15. We had to create a IPTABLES rules on the 'log-api' servers to BLOCK outbound port 53 traffic to our InfoBlox controllers as the SRC/TXT requests were burying them ... After enabling the IPTABLES rules, the CPU Load is still present. Commenting out the external DNS resolvers and leaving BOSH-DNS IP in /etc/resolv.conf did not fix the CPU load. |
This looks like it will be fixed in the next loggregator-release |
Need |
Hey @tlwr we also see this happen in our environment. |
I haven't tested any As I understand it, we want any release with this commit: 6d66cf7 EDIT I have deployed However deploying the From this, I can conclude that the latest |
This release: * https://github.com/cloudfoundry/bosh-dns-release/releases/tag/v1.17.0 Contains the following fixes: * stop redundant health check retries between poll intervals * stop redundant health check requests when querying a domain * Fix compilation error on windows due to broken symlinks * health server should shutdown cleanly on sigterm The following fixes: * stop redundant health check retries between poll intervals * stop redundant health check requests when querying a domain cause loggregator component trafficcontroller's very chatty dns healthcheck to be cached by bosh-dns, which will reduce the currently high cpu usage (there is another fix coming in a later loggregator release) See cloudfoundry/loggregator-release#401 Signed-off-by: Toby Lorne <toby.lornewelch-richards@digital.cabinet-office.gov.uk>
This release: * https://github.com/cloudfoundry/bosh-dns-release/releases/tag/v1.17.0 Contains the following fixes: * stop redundant health check retries between poll intervals * stop redundant health check requests when querying a domain * Fix compilation error on windows due to broken symlinks * health server should shutdown cleanly on sigterm The following fixes: * stop redundant health check retries between poll intervals * stop redundant health check requests when querying a domain cause loggregator component trafficcontroller's very chatty dns healthcheck to be cached by bosh-dns, which will reduce the currently high cpu usage (there is another fix coming in a later loggregator release) See cloudfoundry/loggregator-release#401 Signed-off-by: Toby Lorne <toby.lornewelch-richards@digital.cabinet-office.gov.uk>
@tlwr We still need to cut 106.3.7 with this fix. We will be doing so very soon. |
We have created an issue in Pivotal Tracker to manage this. Unfortunately, the Pivotal Tracker project is private so you may be unable to view the contents of the story. The labels on this github issue will be updated when the story is started. |
No, I shouldn't have referenced the version number explicitly - my bad The loggregator team(s) seem to have a continuous release process which automatically bumps relevant modules, without necessarily taking all the changes from develop As you can see from the difference between the latest release, and the develop branch, the commits which allegedly fix the issue, are still not in a public final release |
Ok, thanks. Do you have an estimate when this will be available? |
I think we can close this issue, as the release notes for the latest release suggest this is now fixed |
See cloudfoundry/loggregator-release#401 and https://github.com/cloudfoundry/loggregator-release/releases/tag/v106.3.8 Signed-off-by: Toby Lorne <toby.lornewelch-richards@digital.cabinet-office.gov.uk>
See cloudfoundry/loggregator-release#401 and https://github.com/cloudfoundry/loggregator-release/releases/tag/v106.3.8 Signed-off-by: Toby Lorne <toby.lornewelch-richards@digital.cabinet-office.gov.uk>
Confirming that 106.3.8 fixes this bug. |
Just to be clear here, there's two dns things here. 1) |
Bumps loggregator-release to 106.3.11 Resolves: cloudfoundry/loggregator-release#401 cloudfoundry/loggregator-release#405
Hello,
i use loggregator via cf-deployment , and recently, we have updated our cluster from cf-deployment 12.1.0 to 12.20.0 , so with loggregator 106.3.1. and we start to have dns issue, the log-api vm have huge load and send to much dns request (approx. 15k req/min).
Log :
[RequestLoggerHandler] 2019/12/27 14:30:15 INFO - handlers.DiscoveryHandler Request [33] [_grpclb._tcp.q-s0.doppler.default.cf.bosh.] 2 28000ns
i've try to upgrade loggregator to 106.3.2 (last one) but same issue.
i've rollbacked to a older loggregator release (105.6 but keep cf-deployment to 12.20.0) and issue disappear.
so, it's seem to have a bug on the 106.x loggregator series.
More info on the cloudfloundry slack :
https://cloudfoundry.slack.com/archives/C2U7KA7M4/p1576657695004400
thanks for your help :)
The text was updated successfully, but these errors were encountered: