-
Notifications
You must be signed in to change notification settings - Fork 66
OpenShift get watch is returning bogus datas #3240
Comments
if we can get a better description and breakdown than 'something' 'someone' 'some things' 'other things' - we can try and help triage the issue down. |
@kbsingh I am not sure what you are not understanding here, it's all in the past snippet, I have rephrased some of it if that's what you are asking? |
What version of oc tools are you using? |
@pbergene oc v3.6.0+c4dd4cf but that's happenning inside the loop of idler too |
Further to explain what i mean by hapenning from inside the loop of idler, idler loop is using oc watch to know when there is a new build event, this is what I get currently from the system, it works fine at first and then loop forever on this tenant The only added code is this diff : diff --git a/internal/condition/build_condition.go b/internal/condition/build_condition.go
index 64f7b97..f4444f5 100644
--- a/internal/condition/build_condition.go
+++ b/internal/condition/build_condition.go
@@ -28,6 +28,7 @@ func (c *BuildCondition) Eval(object interface{}) (bool, error) {
}
if u.HasActiveBuilds() {
+ fmt.Printf("DEBUG!!!: %s, %s, %s\n", u.ActiveBuild.Status.StartTimestamp, u.Name, u.ActiveBuild.Metadata.Name)
return false, nil
} I can't add anything in this loop to fix fabric8-services/fabric8-jenkins-idler#141 before I can get reliable data, |
We have done some discussion with @pbergene on this, According to the monitoring system (zabbix) there was some idling done on We only have some project looping forever not all of them have this issue. Since there is multiple workers on idler only one of them in this case would get stuck but others would still be able to properly works, which is what seems to happen. If other workers gets bugged on other bogus namespaces then we probably run in jenkins idler not functioning anymore. |
There are multiple builds that seem to be looping on 2a, but not for 2. Compiling a list on HK. |
@chmoul Isn't this just the sync plugin updating the model over and over for no reason? |
@aslakknutsen ah yes i forgot about that little bugger thanks, I am still not sure about this then :
I don't see difference between the different payloads that coming thru as well. fixing #2352 would anyway shut this up, |
removing the sev1 since this the |
I've done some troubleshooting on this. The following namespaces/builds are currently exhibiting this issue on starter-us-east-2a
Upon watching the builds with -ojson or -oyaml, I noticed that only the resourceVersion field kept changing. This field is updated whenever openshift updates an object. The number in the above builds is incrementing at at pace of ~60 per second. I then ran a watch in a terminal while I killed the jenkins instance from another and it stopped. The build object stopped being updated and thus being reported to my watch command. So it would appear that something in jenkins is sending tons of openshift API calls per second to update the builds, for some reason. I believe that finding and fixing what is doing this in Jenkins will help solve this issue. |
I also see this on the following builds running in starter-us-east-2
|
as identified this is caused by the sync plugin generating that bogus behaviour, i have create a new issue here to track this down and closing this. #3266 |
I am trying to debug an idler bug and have encountered a bug on prod-preview, when using
oc get watch
we seem to get a lot of build events that have already been captured which seems to loop for ever or at least for a long time.This buildname is called "app-test-1-sadsadsa-1" which seems to be located on 2a for user
ssamal-preview
without using idler and just via the CLI i am capturing the watch output :
oc get build -w --all-namespaces -o json > /tmp/a.json
i then wait for a while and I can see the same build object is showing up again and again :
This seems to go forever and i had to stop the capturing before this filled up my laptop disk space.
This is set as SEV1 because I need this fixed before I can fix this other SEV1 #2352
The text was updated successfully, but these errors were encountered: