id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
2692716889
🛑 FX blog (kty) is down In f50a034, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in 4239f75 after 34 minutes.
gharchive/issue
2024-11-26T00:55:39
2025-04-01T06:39:16.743951
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/42127", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2723466145
🛑 FX corretor (kty) is down In a84576e, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in 91bb9d8 after 11 minutes.
gharchive/issue
2024-12-06T16:29:49
2025-04-01T06:39:16.746277
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/44074", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2728393588
🛑 FX corretor (kty) is down In 2944063, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in b4f4465 after 10 minutes.
gharchive/issue
2024-12-09T22:13:57
2025-04-01T06:39:16.749164
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/44654", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2734195966
🛑 FX blog (kty) is down In 3c77710, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in ede7c81 after 7 minutes.
gharchive/issue
2024-12-11T22:56:40
2025-04-01T06:39:16.751746
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/45023", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1106125181
🛑 GIPI financial is down In 9cb7215, GIPI financial (https://financial.app.gipisistemas.com.br/public/check/health) was down: HTTP code: 404 Response time: 319 ms Resolved: GIPI financial is back up in 58e9035.
gharchive/issue
2022-01-17T17:49:26
2025-04-01T06:39:16.754148
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/710", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2367574381
🛑 FX corretor (kty) is down In 4aa7eb8, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in 10d9c19 after 9 minutes.
gharchive/issue
2024-06-22T05:18:19
2025-04-01T06:39:16.756636
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/9277", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
234190595
(node:8987) UnhandledPromiseRejectionWarning: Unhandled promise rejectio n (rejection id: 1): Error: No backlight device found This is the error that occurs when executing on linux platform. Please update the module to the latest node.js with catch() Error: No backlight device found Using the CLI or the regular API? Anyway, you can see why it fails in the error message. Please update the module to the latest node.js with catch() You should catch errors yourself. Try running ls /sys/class/backlight. Does it list anything? Nothing. But i installed backlight using apt-get install xbacklight What distro are you running? And I assume you can change brightness otherwise?
gharchive/issue
2017-06-07T12:21:49
2025-04-01T06:39:16.780386
{ "authors": [ "bharathvaj1995", "kevva" ], "repo": "kevva/brightness", "url": "https://github.com/kevva/brightness/issues/13", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
186712651
Decouple builtin plugins? Currently i just uses decompress-unzip, but it needs some more workarounds it because of the dest. I don't need the others. Yea they are pretty small deps, but... would be better. :) FYI you can directly use a plugin without decompress. Example: https://github.com/shinnn/get-zonetab/blob/ddf68e876dcc93cc891d3c289c45f97ea215d3ad/index.js#L15 Yea, that's what I'm doing now, but i should handle files manually. utils.unzip()(buf).then((files) => { // handle / decompress files }) maxogden/extract-zip seems to fit your use case.
gharchive/issue
2016-11-02T04:58:14
2025-04-01T06:39:16.783143
{ "authors": [ "shinnn", "tunnckoCore" ], "repo": "kevva/decompress", "url": "https://github.com/kevva/decompress/issues/44", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
212540055
Can't log into the Keybase app Keybase GUI Version: 1.0.18-20170223121400+af0e640 I tried to log in through the shell after resetting my keys and it returned 'ERROR SecretSyncer: no device found for ID=660e***'. I can log in in the browser, but not on the desktop. Thanks for the report, this should be fixed in our next release.
gharchive/issue
2017-03-07T19:52:17
2025-04-01T06:39:16.784607
{ "authors": [ "maxtaco", "thibaultdej" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/6160", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
259562025
Please Support Multiple Identities or Personas I have two personas online, my professional persona and my personal/family personas are kept separate online. In this real-life example, I will never associate my facebook with my github via way of keybase (i wont proof facebook and github on the same keybase account, because that then associates them). Are my concerns common, is anyone else concerned about the same thing? Can/Will keybase support multiple personas or identities? We do support this. You can keybase logout of one account and then log in to the second. We don't plan to make it much easier than it already is, because most people don't do this. I also would love the ability to be logged in to multiple accounts simultaneously. I understand Keybase is “single identity” driven, but I think that idea is misguided. The equivalent in GnuPG would be to say “ERROR: —genkeys: you can only generate one private key ring per instance of gnupg, please log out of current OS user and log into a new user to manage a second private keypair” You are just creating an artificial hurdle that won’t prevent people from doing it, and will only hurt the people who try jumping over that hurdle and trip and fall and break their face. People who want to do this will do it, so why not make it simpler for them? If the answer is “it’s not a priority” I totally understand and can put up with the inconvenience of switching between identities. But if the answer is “multiple email addresses are bad. multiple personas are bad, everyone should know about all your emails and PGP keys and social accounts” then I think that is misguided. The whole “discourage multiple accounts” thing makes sense for Facebook and twitter etc. but not for keybase imo. The logout/login process is not very user friendly because if people want to run multiple identities they usually want to be able to be logged into all of them at the same time which would mean the clients should be able to handle it. Running multiple instances on win/osx/linux would work, but multiple instances on a phone are problematic. @bradmurray I agree. Logging in/out is not representative of real-life, for example, any one of my acquaintances should be able to reach me at any time, despite whether they are professional acquaintances or otherwise. Just because i'm "at work" doesnt mean i shouldnt receive notifications from anyone else. Excellent software but honestly I'm surprised this wasn't a must-have in version 1. Part of managing one's privacy these days is managing one's identities. I know many people who would balk at using Keybase because they can't easily manage multiple identities. We already support this. On Wed, Feb 14, 2018 at 1:32 AM Wolf McNally notifications@github.com wrote: Excellent software but honestly I'm surprised this wasn't a must-have in version 1. Part of managing one's privacy these days is managing one's identities. I know many people who would balk at using Keybase because they can't easily manage multiple identities. — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/keybase/client/issues/8586#issuecomment-365509735, or mute the thread https://github.com/notifications/unsubscribe-auth/AA05_5nmKbxpy6Qo4gSBEbrGYXtP5js0ks5tUn35gaJpZM4Pfm9d . @maxtaco I just checked the iPhone app and logging in to multiple accounts simultaneously (the subject of this issue) is not yet supported. I need to log out and log back in with a different account. During which time I can no longer get notifications for the previous account. We already support this. How is this already supported? I run keybase on Android, Windows and OSX and on all of them I have to logout and log back in to change identities. I'm a new keybase user using it for multiple identities and I too would like to see the UX around being logged into multiple identities simultaneously become better. This is a sorely missing feature, imo. I'll add my vote for this one as well, and also point out that the Firefox container solution doesn't help with chat -- chat seems to be only supported with the native app, with the exception of the limited chat enabled by the browser extension. So, the browser can't really fix this issue. I have a similar use case to others here; I have a need to keep my work-related communication separate from personal accounts. I'm happy to see the new "exploding messages" feature; I hope this one might get some consideration now that that one is done. (Maybe also chat search via the GUI, but that's a different issue.) Hey, I just started to use keybase and the first thing I checked was if I can create separate accounts for private and business use cases. It would be great to have these two logged in in the app simultaneously. Jumping on this bandwagon as well. A lot of services nowadays do support multiple identities and/or personas logged in at the same time. It'd be awesome if keybase could do the same for those of us who keep separate online identities for separate areas of work. N+1'd. An important use case is the ability to make "bot" accounts to be members in a team (e.g. to auto-push DevOps secrets into a Keybase Git repo) +1 on this. Multiple identities should be supported. We don't plan to make it much easier than it already is, because most people don't do this. Many people who already have accounts in different places do have one for work and one for their personal identity. I know many, MANY people in this case, and looking at the list of people in this thread, there are indeed many of us. Please bring this to Keybase +1 I also have personal (where I am generally opinionated on things that matter to me) and my company account (fun but filtered) Conflicted if the "already supported" responses ought to be treated as depressing or hilarious. Dropping in to +1 here. Keybase looks fascinating but I have had my fingers burnt in the past by letting different areas of my life overlap (new rule: colleagues only get added on social media when one or other of us leaves!) Not doing that again! Want this as well. The logout/login process is not very user friendly because if people want to run multiple identities they usually want to be able to be logged into all of them at the same time which would mean the clients should be able to handle it. Running multiple instances on win/osx/linux would work, but multiple instances on a phone are problematic. If it is not feasible on phone, then don't try to implement this for phone. Do it for the desktop app. It is (obviously) a largely overlooked feature. +1000 For me, this is a dealbreaker - as much as I love keybase, until multiple personas will be properly supported, it will be just a toy app I'm checking out every now and then. +1 This is the main reason I'm not using keybase tbh. It's quite understandable that software developers want to discourage or prevent account creation spam. It can clog up the namespace, for one thing. It allows Sybil attacks in dicsussions, for another, so a person can pretend to be multiple people all arguing the same side (happens on Reddit and Twitter all the time). Making it too easy for a person to have many logins will also make it too easy for the person to take up many times the free allocation of cloud storage per person. So I definitely will not argue for multiple identities without limit. However, two is a bare minimum, because many of us find it desirable to keep our personal and professional lives separate. If carrying a single mobile device, you would want to have both identities logged in. Or you might have separate business and personal devices, and in that case, you can much more easily implement security procedures, so certain business-related files are never allowed to touch a personal device. You might need a third identity to say and do naughty things, not naughty enough to make you a criminal in the civilized world, but naughty enough to damage your personal or professional life if found out and, in some less civilized parts of the world, be stoned to death. You can probably imagine reasons for additional identities. I think about five or six would be a reasonable limit. Definitely a must have ! I just joined a workplace that uses Keybase and would love to have the ability to simultaneously sign into my work and personal accounts. I'd understand if it wasn't a priority though. We're close on this, either in this release or the next. In https://github.com/keybase/client/issues/2104#issuecomment-493079134 it's said that you can currently keybase login -s anotheraccount to switch account Is there any UI support planned for this? yes. however I currently can't get desktop notifications for any of the other logged in accounts. I have to switch and visually check for unread chats. Please add support for this! I honestly can't believe someone, never mind someone responsible for a product like keybase, would honestly think "most people don't have multiple identities on the internet" There's a subculture of people who seem to have melted their selves into the internet like butter on toast. People who do this will naturally interact with lots of others who have done the same to a greater or lesser extent so they come to believe it is normal. If a primary way you interact with other humans is online, then you might eventually arrive at the conclusion that the people posting prolifically online are representative of all humans. Because how else would you ever comes across anyone? It's a weird self perpetuating bubble. That said I am not sure if the devs/bosses here think this (though I haven't read all the other related issues yet so maybe they do) or if they are merely indifferent or uninterested. IMHO security and privacy/anonymity are intrinsically linked in every practical way. Like the others, I would also consider this tool much more useful if I didn't have to sit and think which portion of my life it is most suited for, which is my train of thought at the moment. Feature: is in current release version. Feature Request Issue: continues discussion as if it is not there. Feature: Feature: is in current release version. I am new to this whole github thing but why is the issue still marked as "open"? Also #2104 "Feature Request: Logging into multiple keybase accounts locally". In my limited experience, mostly lurking, devs/admins are extremely happy to close these things once they are resolved. Found this thread (and the many others which link here) trying to determine if it is possible or no. All I see is many people stating it's not possible. I can't see anything obvious on my phone about how to do this. "Add identities" lets me verify on about 800 mastodon servers. Because these threads get long, and after a while it's no longer clear what feature they were talking about to begin with. Also, Keybase client has 2.6k issues open... which tells you that this project is not focused much on closing / managing issues and is more focused on coding / reviewing code. The closed issues mostly tend to be bugs that were fixed, since they are much easier to put a binary value on completion. I think this issue is still open and people are still talking about this because currently, there is no Graphical User Interface (GUI) support for multiple identities or personas through the Keybase app. Here's my +1 for GUI support in addition to the Command Line Interface (CLI) implementation. @junderw Feature: is in current release version. The feature definitely doesn't exist in current version in an acceptable manner. +1. Especially with chat there's times where I'd want to use alternative identities. If Keybase is aiming for businesses to support Keybase, then it would be incredibly useful to have different accounts for work and private use. Note that these days twitter allows to simultaneously connect to up to 5 accounts/identities at the same time, right there through the normal official GUI. Just click on "More" then the blue "⊕" to add more accounts. (I never use that feature, but I'm glad it exists.) Also note that these days, gmail (or only G Suite?) supports multiple accounts natively: Kebase technically supports it, but having to re-enter your password every time is a massive pain. On Wed., Nov. 6, 2019, 7:19 a.m. Jean-Philippe Paradis, < notifications@github.com> wrote: Also note that these days, gmail (or only G Suite?) supports multiple accounts natively: [image: Multiple gmail accounts] https://camo.githubusercontent.com/5646d7d9503d7d5513b0623695dfd33968a03675/68747470733a2f2f7777772e68657873747265616d2e6e65742f746d702f676d61696c2d6d756c7469706c652d6163636f756e74732e706e67 — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/keybase/client/issues/8586?email_source=notifications&email_token=AIEGR7W5YENP5SGDKQTYRS3QSKY5DA5CNFSM4D36N5O2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEDGK3GY#issuecomment-550284699, or unsubscribe https://github.com/notifications/unsubscribe-auth/AIEGR7U5DEVHW4FWYV6FJFLQSKY5DANCNFSM4D36N5OQ . Being able to be logged in to multiple accounts simultaneously is distinctly different. I think this issue is still open and people are still talking about this because currently, there is no Graphical User Interface (GUI) support for multiple identities or personas through the Keybase app. Here's my +1 for GUI support in addition to the Command Line Interface (CLI) implementation. We just got our wish. Multiple identities or personas support through the Keybase app Graphical User Interface (GUI). That's in addition to the Command Line Interface (CLI) implementation. Yes! Sounds great, but as soon as I select "Log in as another user" it immediately logs me out and I have to enter my password again to log back in... This is now in the app in version 5.0 on all platforms. You should be able to switch users without entering your password every time. @Hexstream, if you have "Always stay logged in" checked in the Advanced Settings tab, that sounds like a bug. Please send feedback after reproducing and make another issue. Thanks. You are right, checking "Always stay logged in" solves the problem, but this behavior is not intuitive at all. I created a new issue. Thanks so much for fixing this issue! I use this feature on a daily basis, and it works great. I have two suggestions to make it even more awesome: https://github.com/keybase/client/issues/23850 to show notifications for all accounts configured as "always stay logged in". https://github.com/keybase/client/issues/23851 to allow switching between "always stay logged in" accounts with just a keyboard shortcut, or a single click, and remember what tab was last open for each account.
gharchive/issue
2017-09-21T16:46:35
2025-04-01T06:39:16.821384
{ "authors": [ "AngryAnt", "Avamander", "Hexstream", "IpsmLorem", "ItsExtra", "Mikaela", "QuentinBellus", "TheKangaroo", "bradmurray", "cbluth", "cjb", "dabura667", "dlumma", "domq", "fortran77", "getraf", "grilldos", "heronhaye", "hexagile", "jcausey-astate", "jfinstrom", "junderw", "mahmad-nutrien", "maxtaco", "mvdan", "neunenak", "rsoltanzadeh", "srd424", "supakeen", "tstrijdhorst", "w3bb", "wolfmcnally" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/8586", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
310127146
fix jump to new conversation of kbfs chats @keybase/react-hackers this fixes the jump to new convo flow I renamed the payload param and flow didn't catch it because the payload types weren't exact. i don't remember why we undid that (it used to be) but it really should go back i think
gharchive/pull-request
2018-03-30T17:49:36
2025-04-01T06:39:16.823953
{ "authors": [ "chrisnojima" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/pull/11157", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
535371730
fix loading team on create Messy because team get takes a name Spawn a getDetails on team created because the component will not have the teamname Store entry teamNameToID on team created so we have it as soon as possible cc @keybase/y2ksquad We can easily expose an RPC that takes an ID (if we don’t have one already) @maxtaco changing the gui to use the team ID here is part of y2k-974
gharchive/pull-request
2019-12-09T22:17:54
2025-04-01T06:39:16.826098
{ "authors": [ "buoyad", "maxtaco" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/pull/21496", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
318328642
[Mac OS] Old block DBs not being cleaned up diskCacheBlocks.leveldb admin$ date Fri 27 Apr 2018 18:36:24 diskCacheBlocks.leveldb admin$ pwd /Users/admin/Library/Application Support/Keybase/kbfs_block_cache/v1/diskCacheBlocks.leveldb diskCacheBlocks.leveldb admin$ ls -l total 10540072 drwxr-xr-x 117 admin staff 3.9K 27 Apr 15:36 . drwxr-xr-x 5 admin staff 170B 30 Sep 2017 .. -rw-r--r-- 1 admin staff 50M 20 Feb 23:11 009400.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:11 009401.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:20 009459.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:20 009460.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:39 009608.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:39 009609.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:39 009610.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:39 009611.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:47 009668.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:47 009669.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:53 009726.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:55 009751.ldb -rw-r--r-- 1 admin staff 50M 20 Feb 23:55 009752.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:00 009804.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:15 009981.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:15 009982.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:15 009983.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:15 009984.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:15 009985.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:20 010043.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:20 010044.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:25 010098.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:25 010099.ldb -rw-r--r-- 1 admin staff 50M 21 Feb 00:25 010100.ldb -rw-r--r-- 1 admin staff 17M 21 Feb 00:25 010101.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 10:27 010303.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 10:40 010451.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 10:40 010452.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 10:40 010453.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 10:40 010454.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 10:40 010455.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 10:40 010456.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 10:40 010457.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 10:40 010458.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 10:51 010510.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:16 010752.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:16 010753.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:16 010754.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:16 010755.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:16 010756.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:16 010757.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:16 010758.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:16 010759.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:16 010760.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:16 010761.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:35 010991.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:35 010992.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:35 010993.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:35 010994.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:35 010995.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:35 010996.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:35 010997.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:35 010998.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:35 010999.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:35 011000.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:44 011104.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:44 011105.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:44 011106.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:44 011107.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:47 011153.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:47 011154.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:47 011155.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:51 011213.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:51 011214.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:52 011251.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 11:55 011299.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:00 011421.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:00 011422.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:00 011423.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:00 011424.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:00 011425.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:00 011426.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:00 011427.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:04 011473.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:04 011474.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:04 011475.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:08 011533.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:08 011534.ldb -rw-r--r-- 1 admin staff 50M 24 Feb 12:08 011535.ldb -rw-r--r-- 1 admin staff 44M 24 Feb 12:08 011536.ldb -rw-r--r-- 1 admin staff 50M 11 Mar 14:21 011632.ldb -rw-r--r-- 1 admin staff 50M 11 Mar 14:21 011633.ldb -rw-r--r-- 1 admin staff 50M 11 Mar 14:21 011634.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:29 011852.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:29 011853.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:29 011854.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:29 011855.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:29 011856.ldb -rw-r--r-- 1 admin staff 10M 26 Apr 10:29 011857.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:29 011858.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:29 011859.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:29 011860.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:29 011861.ldb -rw-r--r-- 1 admin staff 12M 26 Apr 10:29 011862.ldb -rw-r--r-- 1 admin staff 7.5M 26 Apr 10:29 011863.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:30 011864.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:30 011865.ldb -rw-r--r-- 1 admin staff 35M 26 Apr 10:30 011866.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:30 011867.ldb -rw-r--r-- 1 admin staff 45M 26 Apr 10:30 011868.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:30 011869.ldb -rw-r--r-- 1 admin staff 19M 26 Apr 10:30 011870.ldb -rw-r--r-- 1 admin staff 47M 26 Apr 10:30 011871.ldb -rw-r--r-- 1 admin staff 11M 26 Apr 10:30 011872.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:30 011873.ldb -rw-r--r-- 1 admin staff 50M 26 Apr 10:30 011874.ldb -rw-r--r-- 1 admin staff 43M 26 Apr 10:30 011875.ldb -rw-r--r-- 1 admin staff 0B 27 Apr 09:30 011887.log -rw-r--r-- 1 admin staff 32M 27 Apr 15:36 011889.ldb -rw-r--r-- 1 admin staff 46K 27 Apr 15:36 011890.ldb -rw-r--r-- 1 admin staff 16B 27 Apr 09:30 CURRENT -rw-r--r-- 1 admin staff 0B 30 Sep 2017 LOCK -rw-r--r-- 1 admin staff 368K 27 Apr 15:37 LOG -rw-r--r-- 1 admin staff 1.1M 21 Feb 00:28 LOG.old -rw-r--r-- 1 admin staff 10K 27 Apr 15:36 MANIFEST-011888 Old .ldb's not being cleaned up. I made a backup, removed the old ones, relaunched keybase and everything still running normally. As a sidenote, why are these DBs so big when I only have 17mb of files in my KBFS? (currently, had more before) If there's any extra debugging information I can provide, just let me know! @kklash: as explained here under the "Local disk usage policy" section, we keep up to 20 GB of blocks, depending on the free space of your drive. If you use more of the drive, we will shrink that cache automatically if KBFS is running. We cache all blocks you've even seen, including from team and public folders, and past blocks from old versions of the files in some cases.
gharchive/issue
2018-04-27T08:54:42
2025-04-01T06:39:16.830010
{ "authors": [ "kklash", "strib" ], "repo": "keybase/kbfs", "url": "https://github.com/keybase/kbfs/issues/1558", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
182457338
Update kbfstool URL it has been changed in https://github.com/keybase/kbfs/commit/3cf9045546540bb72435bb48cd53471773a46c76 but URL didn't updated. can be related to #309 Thanks!
gharchive/pull-request
2016-10-12T07:47:11
2025-04-01T06:39:16.831689
{ "authors": [ "oconnor663", "sijad" ], "repo": "keybase/kbfs", "url": "https://github.com/keybase/kbfs/pull/419", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
185832218
cr: speed up while journaling by avoiding extraneous block uploads and downloads This PR enables a journaled TLF to use only local information during conflict resolution, and completely avoids creating a conflict branch on the servers. Getting there required quite a few changes (let me know if you'd prefer them broken out into separate PRs): tlfJournal pauses flushing as soon as it notices a conflict. conflictResolver figures out which blocks don't need to be uploaded, and passes that to MDOps.ResolveBranch so the journal can ignore them. We now write the resolution MD to the journal, rather than directly to the server (due to the ResolveBranch semantics). Since the journal might have uploaded some blocks before discovering a conflict, BlockServer now has an IsFlushed method that the conflictResolver can use to tell which blocks have already been flushed. blockJournal now has a way to avoid removing the data for flushed blocks from disk, until the next successful MD flush. This helps us avoid re-downloading blocks in the case where the resolution MD itself hits a conflict during flushing, and we have to re-do CR all over again. (This is possible now because we are writing the resolution MD to the journal, and won't know if there's a conflict until later.) conflictResolver needed several tweaks to work correctly when the unmerged branch contains a resolutionOp, which wasn't possible before. RootMetadata.deepCopy needs to also copy the cached block changes -- otherwise when the journal converts a branch and replaces the cached MD entries, it could wind up forgetting the cached block change pointer. This fixes a bug in master that we just didn't have test coverage for, but now we do. Ok fixed up, ready for another look please. Known unrelated CI failures. Merging.
gharchive/pull-request
2016-10-28T03:44:06
2025-04-01T06:39:16.836158
{ "authors": [ "strib" ], "repo": "keybase/kbfs", "url": "https://github.com/keybase/kbfs/pull/473", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
221991159
Switch on major versions when reading Otherwise, we may panic when reading a header from an unknown minor version. Move some non-common functions out to better places. Coverage increased (+0.04%) to 86.244% when pulling 72453ddb0f9d49ab06adb214e119187bb313b94d on akalin/fix-switch into 5ff37ca6c69573e0786615153f260a45b266476b on master. Hmm. I should probably write a test for this. I'll prep one tomorrow.
gharchive/pull-request
2017-04-16T05:12:02
2025-04-01T06:39:16.838675
{ "authors": [ "akalin-keybase", "coveralls" ], "repo": "keybase/saltpack", "url": "https://github.com/keybase/saltpack/pull/28", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1596719277
roadmap: update for 2023 This updates the roadmap with the major changes planned for Keylime in 2023. Please have a look and see, if the goals are defined well enough and if I forgot something to add. Codecov Report Merging #1323 (cc93e8a) into master (8ddfd12) will decrease coverage by 0.49%. The diff coverage is n/a. :mega: This organization is not using Codecov’s GitHub App Integration. We recommend you install it so Codecov can continue to function properly for your repositories. Learn more Additional details and impacted files Flag Coverage Δ packit-e2e 74.15% <ø> (-0.49%) :arrow_down: testsuite 11.07% <ø> (ø) unittests 1.92% <ø> (ø) Flags with carried forward coverage won't be shown. Click here to find out more. Impacted Files Coverage Δ keylime/keylime_agent.py 72.55% <0.00%> (-0.58%) :arrow_down: Just to make clear the key realization on the parenthetical comment on the 4th bullet on @THS-on answer. It seems clear to me there is the need for separating the "ascertaining the integrity" of the attestation artifacts (i.e., tpm2_checkquote, the replaying of both MB and IMA log and matching of PCRs) and the subsequent validation against policy.
gharchive/pull-request
2023-02-23T11:52:19
2025-04-01T06:39:16.900387
{ "authors": [ "THS-on", "codecov-commenter", "maugustosilva" ], "repo": "keylime/keylime", "url": "https://github.com/keylime/keylime/pull/1323", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
156535078
Installation failed with "yo keystone" Expected behavior I was following the "Get Started" page, "Commands". I wanted to install the latest version of keystone and startup new project. Actual behavior After the keystone-generator installed, tried to do "yo keystone" and the installation has started but it does not generate the project. I got the following error: screen shot 2016-05-24 at 10 21 49 Thanks for helping me out! Issue was fixed with #195, the Gruntfile is out of date on the 0.3.16 with npm install @mxstbr
gharchive/issue
2016-05-24T15:17:22
2025-04-01T06:39:16.949687
{ "authors": [ "morenoh149", "pll33" ], "repo": "keystonejs/generator-keystone", "url": "https://github.com/keystonejs/generator-keystone/issues/200", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
929998322
(WIP) build: restructure project Changes [x] Setup Dependabot [x] Use standard instead of xo [x] Unify Licenses [x] Setup npm metadata [x] Setup editorconfig Should homepage be changed to whatever site we deploy docs to?
gharchive/pull-request
2021-06-25T09:20:40
2025-04-01T06:39:16.956905
{ "authors": [ "Jytesh", "Kikobeats" ], "repo": "keyvhq/keyv", "url": "https://github.com/keyvhq/keyv/pull/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2425674358
🛑 Helsinki Regional Transport Authority is down In 629e789, Helsinki Regional Transport Authority (https://www.hsl.fi/) was down: HTTP code: 502 Response time: 27501 ms Resolved: Helsinki Regional Transport Authority is back up in 17c0c16 after 11 minutes.
gharchive/issue
2024-07-23T16:55:08
2025-04-01T06:39:16.959619
{ "authors": [ "keywordnew" ], "repo": "keywordnew/helsinki-watchtower", "url": "https://github.com/keywordnew/helsinki-watchtower/issues/200", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2413277227
Included RML Strategy, fixed some mistakes, and updated sources to respect RML-IO @anaigmo: I have removed "join" from the comment in the ontology to make rml:Strategy and rml:strategy implicitly specific to RML-CC. @frmichel Can you check? Hi @chrdebru, sorry I've been away from the last discussions. I'm just confused about commit https://github.com/kg-construct/rml-cc/pull/49/commits/29216b27393ed1b1ab2a83e8fd94cd09e1b394a1, I though the plan was to move rml:Strategy to core, but it is reintegrated into CC here. Why is that? @frmichel the explanation is here https://github.com/kg-construct/rml-core/issues/51 (last comment from Ben) Ok got it, thx @dachafra, sorry for missing that. no worries! :-) Just one tiny remark, the rendered.html mentions 10 January 2024 as the spec date. This is on purpose or should be updated? Apart from that I've good, you can merge the PR whenever you like.
gharchive/pull-request
2024-07-17T10:47:51
2025-04-01T06:39:16.963016
{ "authors": [ "chrdebru", "dachafra", "frmichel" ], "repo": "kg-construct/rml-cc", "url": "https://github.com/kg-construct/rml-cc/pull/49", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
2101722153
Join specification when logical source is the same Let's say we have two triple maps that refer to the same logical source (and with same, we really mean same URI, not "same because the descriptions lead to the semantically same logical source"). Sample source (CSV) id,parent_id 1,2 2,1 Base mapping (YARRRML) prefixes ex: http://example.com# sources: test: [data.csv] mappings: test1: s: ex:$(id) po: p: ex:parent o: mapping: test2 test2: s: ex:$(parent_id) We have following use cases that are underspecified in de spec the spec currently says If the logical source of the child triples map and the logical source of the parent triples map of a referencing object map are not identical, then the referencing object map must have at least one join condition. If a join condition is specified AND the logical source is not the same: common case, execute join condition between each iteration pair If a join condition is specified AND the logical source is the same: same as above If no join condition is specified AND the logical source is not the same: do a full join (i.e., take all iterations into account) example output: ex:1 ex:parent ex:2, ex:1 ex:parent ex:1, ex:2 ex:parent ex:2, ex:2 ex:parent ex:1 If no join condition is specified AND the logical source is the same: don't do a full join, but take the current iteration into account example output: ex:1 ex:parent ex:2, ex:2 ex:parent ex:1 this last one is the edge case, but allows to 'join per iteration'. Question is: should we make this edge case explicit, or should there be a different way to tackle this edge case? To continue the discussion of this issue, and considering that there is already a spec written, I would suggest making a PR @elsdvlee so the rest can review it and provide comments! See https://github.com/kg-construct/rml-core/pull/78 awesome! Please assign @andimou @pmaria @bjdmeest @DylanVanAssche as potential reviewers My view on defining equality of logical sources: Object equality in programming languages is used as the basis for many things. For example comparison in different data structures for uniqueness and hashing. (Think dictionaries, sets etc.) I strongly believe we should be able to leverage this for logical sources. I think source and logical source equality is something that is very useful to have when building RML processors. Therefor, I would propose to come up with a definition of equality which can be implemented as such. My proposal would be to define a logical source or source to be equal to another logical source or source if the RML-defined properties of the description of both are equal. RML-defined: Those properties that are defined by a specification to have behavior that influences the behavior of an RML processor. These properties MUST be listed for the rml:LogicalSource specification. These properties MUST be listed for any rml:Source description. Doing so will allow RML processors to map these descriptions to standard object equality mechanisms in their respective programming languages to best leverage the language's abilities.
gharchive/issue
2023-01-18T12:50:53
2025-04-01T06:39:16.972887
{ "authors": [ "bjdmeest", "dachafra", "elsdvlee", "pmaria" ], "repo": "kg-construct/rml-core", "url": "https://github.com/kg-construct/rml-core/issues/74", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1967471304
🛑 P is down In 00a3866, P ($P) was down: HTTP code: 0 Response time: 0 ms Resolved: P is back up in 5880ac1 after 8 minutes.
gharchive/issue
2023-10-30T04:24:51
2025-04-01T06:39:16.978651
{ "authors": [ "khadanja" ], "repo": "khadanja/upptime", "url": "https://github.com/khadanja/upptime/issues/161", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2060916233
🛑 SF is down In 227f417, SF ($SF) was down: HTTP code: 530 Response time: 3063 ms Resolved: SF is back up in 83259c4 after 8 minutes.
gharchive/issue
2023-12-31T02:38:34
2025-04-01T06:39:16.980666
{ "authors": [ "khadanja" ], "repo": "khadanja/upptime", "url": "https://github.com/khadanja/upptime/issues/326", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
229966649
Fatal error with latest update Hi! Thanks for your image, I'm happy using it! Unfortunately, with the latest update, I'm not able to run postgres in docker-compose.yml The error is FATAL: could not create lock file "/run/postgresql/.s.PGSQL.5432.lock": No such file or directory Your image part from docker-compose.yml postgres: restart: always image: kiasaki/alpine-postgres:latest environment: - POSTGRES_PASSWORD=test - POSTGRES_DB=test_db - POSTGRES_USER=test_user volumes: - ./persistant/pgdata:/var/lib/postgresql/data I've switched to 9.5 tag and everything works fine. Sorry looks like the same as #32 You can close the issue if you decide that it's a duplicate. Thank you!
gharchive/issue
2017-05-19T12:56:03
2025-04-01T06:39:17.016576
{ "authors": [ "khorolets" ], "repo": "kiasaki/docker-alpine-postgres", "url": "https://github.com/kiasaki/docker-alpine-postgres/issues/33", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
189364456
Telepathy does not work with Fortune enchantment The telepathy enchantment disables fortune from allowing ore blocks to yield more than 1 of their resource, whereas the fortune enchantment without telepathy would allow ores to yield 2-3. Is there any possible fix to the telepathy enchantment to allow more than 1 resource to be sent to the inventory when fortune is being used on the same tool? Ill be fixing it in the next update. #Badbones69totherescue This has been fixed and will be in the next update.
gharchive/issue
2016-11-15T11:15:14
2025-04-01T06:39:17.018193
{ "authors": [ "Scirrio", "Yukun99", "kicjow" ], "repo": "kicjow/Crazy-Enchantments", "url": "https://github.com/kicjow/Crazy-Enchantments/issues/110", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
408310364
Logged in drawer updates What ❓ making avatar 48 by 48 making name bigger Story 📖 Trello before and after 👀 killing it! One tweak: could you please increase the distance from the top of the screen to the header avatar? Between top-of-header-avatar and system bar: 16dp (or 18 if you want to use our grid) Between top-of-header-avatar and top-of-screen: 40dp (or 42 if you want to use our grid) That space is 16dp, my avi has a white background so it's hard to see Improve navigation and filtering on Android app
gharchive/pull-request
2019-02-08T20:24:19
2025-04-01T06:39:17.027636
{ "authors": [ "dannyalright", "eoji" ], "repo": "kickstarter/android-oss", "url": "https://github.com/kickstarter/android-oss/pull/471", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
922880477
[Ignore] CI Failure Validation Just drafting PR to verify if a CI failure is related to changes in Cartfile or general environment issue. SwiftFormat found issues: File Rules Kickstarter-iOS/Library/OptimizelyResultTypeTests.swift:13:1 warning: (spaceInsideComments) Add leading and/or trailing space inside comments. Generated by :no_entry_sign: Danger SwiftFormat found issues: File Rules Kickstarter-iOS/AppDelegate.swift:10:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:11:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:12:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:13:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:14:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:15:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:16:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:17:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:18:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:19:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:20:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:21:1 warning: (sortedImports) Sort import statements alphabetically. Kickstarter-iOS/AppDelegate.swift:22:1 warning: (sortedImports) Sort import statements alphabetically. Generated by :no_entry_sign: Danger No longer needed to check ci validation
gharchive/pull-request
2021-06-16T17:20:33
2025-04-01T06:39:17.035892
{ "authors": [ "msadoon", "nativeksr" ], "repo": "kickstarter/ios-oss", "url": "https://github.com/kickstarter/ios-oss/pull/1521", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
327861802
Password reveal functionality What Show/hide password on login. 👁 See 👀 Native: Password reveal option to make entering passwords easier
gharchive/pull-request
2018-05-30T19:09:57
2025-04-01T06:39:17.038781
{ "authors": [ "Scollaco", "ifbarrera" ], "repo": "kickstarter/ios-oss", "url": "https://github.com/kickstarter/ios-oss/pull/368", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1342621943
AX6S不定时自动重启 ` _____ _____________________________ __________ /_ _ __ __ __ \ _ _ __ _ | /| / /_ / __/ / // / // / / / / / |/ |/ / / / /_ _/ ./_/// //____/|_/ // _/ /_/ _________________________________________ OpenWrt 08.02.2022 by Kiddin' _________________________________________ 系统负载: 0.33 0.25 0.21 运行时间: 21 min 内存已用: 21 % of 227MB IP 地址: 192.168.0.1 系统存储: 17% of 78.7M CPU 信息: ARMv8 Processor x 2 [root@OpenWrt:02:28 PM ~] # cat /sys/fs/pstore/* Oops#1 Part1 <6>[ 9.048398] tun: Universal TUN/TAP device driver, 1.6 <6>[ 9.074886] cryptodev: driver 1.12 loaded. <6>[ 9.081134] Loading modules backported from Linux version v5.15.58-0-g7d8048d4e064 <6>[ 9.088728] Backport generated by backports.git v5.15.58-1-0-g42a95ce7 <6>[ 9.116066] xt_time: kernel timezone is -0000 <7>[ 9.172139] ieee80211 phy0: Selected rate control algorithm 'minstrel_ht' <7>[ 9.183164] mt7915e 0000:01:00.0: assign IRQ: got 142 <7>[ 9.188314] pci 0000:00:00.0: enabling bus mastering <6>[ 9.193291] mt7915e 0000:01:00.0: enabling device (0000 -> 0002) <7>[ 9.199395] mt7915e 0000:01:00.0: enabling bus mastering <6>[ 9.261106] mt7622-wmac 18000000.wmac: HW/SW Version: 0x8a108a10, Build Time: 20190801210006a <6>[ 9.261106] <6>[ 9.366162] mt7622-wmac 18000000.wmac: N9 Firmware Version: reserved, Build Time: 20220630094834 <6>[ 9.408501] mt7915e 0000:01:00.0: HW/SW Version: 0x8a108a10, Build Time: 20211222184017a <6>[ 9.408501] <6>[ 9.549747] mt7915e 0000:01:00.0: WM Firmware Version: ____000000, Build Time: 20211222184052 <6>[ 9.614454] mt7915e 0000:01:00.0: WA Firmware Version: DEV_000000, Build Time: 20211222184111 <6>[ 9.735732] PPP generic driver version 2.4.2 <6>[ 9.740941] PPP MPPE Compression module registered <6>[ 9.746777] NET: Registered protocol family 24 <14>[ 9.753146] kmodloader: done loading kernel modules from /etc/modules.d/* <6>[ 10.355906] zram0: detected capacity change from 0 to 78643200 <6>[ 10.426265] Adding 76796k swap on /dev/zram0. Priority:100 extents:1 across:76796k SSDsc <6>[ 12.700070] mtk_soc_eth 1b100000.ethernet eth0: Link is Down <6>[ 12.703505] mtk_soc_eth 1b100000.ethernet eth0: configuring for fixed/2500base-x link mode <6>[ 12.703677] mtk_soc_eth 1b100000.ethernet eth0: Link is Up - 2.5Gbps/Full - flow control rx/tx <6>[ 12.703695] IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready <6>[ 12.707013] mt7530 mdio-bus:00 lan1: configuring for phy/gmii link mode <6>[ 12.707758] 8021q: adding VLAN 0 to HW filter on device lan1 <6>[ 12.712393] br-lan: port 1(lan1) entered blocking state <6>[ 12.712404] br-lan: port 1(lan1) entered disabled state <6>[ 12.716849] device lan1 entered promiscuous mode <6>[ 12.716857] device eth0 entered promiscuous mode <6>[ 12.726779] mt7530 mdio-bus:00 lan2: configuring for phy/gmii link mode <6>[ 12.727334] 8021q: adding VLAN 0 to HW filter on device lan2 <6>[ 12.731970] br-lan: port 2(lan2) entered blocking state <6>[ 12.731981] br-lan: port 2(lan2) entered disabled state <6>[ 12.732684] device lan2 entered promiscuous mode <6>[ 12.741533] mt7530 mdio-bus:00 lan3: configuring for phy/gmii link mode <6>[ 12.743953] 8021q: adding VLAN 0 to HW filter on device lan3 <6>[ 12.748836] mt7530 mdio-bus:00 lan3: Link is Up - 1Gbps/Full - flow control rx/tx <6>[ 12.752498] br-lan: port 3(lan3) entered blocking state <6>[ 12.752509] br-lan: port 3(lan3) entered disabled state <6>[ 12.753896] device lan3 entered promiscuous mode <6>[ 12.756669] br-lan: port 3(lan3) entered blocking state <6>[ 12.756681] br-lan: port 3(lan3) entered forwarding state <6>[ 12.756768] IPv6: ADDRCONF(NETDEV_CHANGE): br-lan: link becomes ready <6>[ 12.769739] mt7530 mdio-bus:00 wan: configuring for phy/gmii link mode <6>[ 12.770260] 8021q: adding VLAN 0 to HW filter on device wan <6>[ 12.774823] mt7530 mdio-bus:00 wan: Link is Up - 1Gbps/Full - flow control off <6>[ 12.777093] IPv6: ADDRCONF(NETDEV_CHANGE): wan: link becomes ready <6>[ 13.159944] pppoe-wan: renamed from ppp0 <6>[ 13.433189] mt7530 mdio-bus:00 lan2: Link is Up - 100Mbps/Full - flow control rx/tx <6>[ 14.286329] mt7530 mdio-bus:00 lan2: Link is Down <6>[ 14.294018] mt7530 mdio-bus:00 lan1: Link is Up - 100Mbps/Full - flow control rx/tx <6>[ 15.002148] br-lan: port 1(lan1) entered blocking state <6>[ 15.002168] br-lan: port 1(lan1) entered forwarding state <6>[ 15.259329] br-lan: port 4(wlan0) entered blocking state <6>[ 15.259342] br-lan: port 4(wlan0) entered disabled state <6>[ 15.259568] device wlan0 entered promiscuous mode <6>[ 15.259640] br-lan: port 4(wlan0) entered blocking state <6>[ 15.259647] br-lan: port 4(wlan0) entered forwarding state <6>[ 15.918171] mt7530 mdio-bus:00 lan2: Link is Up - 100Mbps/Full - flow control rx/tx <6>[ 15.918229] br-lan: port 2(lan2) entered blocking state <6>[ 15.918243] br-lan: port 2(lan2) entered forwarding state <6>[ 16.006841] br-lan: port 4(wlan0) entered disabled state <6>[ 16.640989] br-lan: port 5(wlan1) entered blocking state <6>[ 16.641001] br-lan: port 5(wlan1) entered disabled state <6>[ 16.641170] device wlan1 entered promiscuous mode <6>[ 16.641231] br-lan: port 5(wlan1) entered blocking state <6>[ 16.641237] br-lan: port 5(wlan1) entered forwarding state <6>[ 17.038752] br-lan: port 5(wlan1) entered disabled state <6>[ 17.940844] IPv6: ADDRCONF(NETDEV_CHANGE): wlan1: link becomes ready <6>[ 17.941026] br-lan: port 5(wlan1) entered blocking state <6>[ 17.941035] br-lan: port 5(wlan1) entered forwarding state <6>[ 22.031440] IPv6: ADDRCONF(NETDEV_CHANGE): wlan0: link becomes ready <6>[ 22.031584] br-lan: port 4(wlan0) entered blocking state <6>[ 22.031593] br-lan: port 4(wlan0) entered forwarding state <1>[10901.799290] Unable to handle kernel paging request at virtual address dead000000000110 <1>[10901.807213] Mem abort info: <1>[10901.810041] ESR = 0x96000004 <1>[10901.813095] EC = 0x25: DABT (current EL), IL = 32 bits <1>[10901.818401] SET = 0, FnV = 0 <1>[10901.821489] EA = 0, S1PTW = 0 <1>[10901.824646] Data abort info: <1>[10901.827516] ISV = 0, ISS = 0x00000004 <1>[10901.831378] CM = 0, WnR = 0 <1>[10901.834364] [dead000000000110] address between user and kernel address ranges <0>[10901.841554] Internal error: Oops: 96000004 [#1] SMP <7>[10901.846425] Modules linked in: pppoe ppp_async pppox ppp_mppe ppp_generic mt7915e mt7615e mt7615_common mt76_connac_lib mt76 mac80211 ipt_REJECT cfg80211 xt_time xt_tcpudp xt_state xt_quota xt_pkttype xt_owner xt_nat xt_multiport xt_mark xt_mac xt_limit xt_conntrack xt_comment xt_cgroup xt_addrtype xt_TPROXY xt_TCPMSS xt_REDIRECT xt_MASQUERADE xt_LOG xt_FLOWOFFLOAD xt_CT tcp_bbr slhc nf_tproxy_ipv6 nf_tproxy_ipv4 nf_reject_ipv4 nf_log_ipv6 nf_log_ipv4 nf_log_common nf_flow_table nf_conntrack_netlink macvlan iptable_nat iptable_mangle iptable_filter ip_tables hwmon crc_ccitt compat cryptodev xt_set ip_set_list_set ip_set_hash_netportnet ip_set_hash_netport ip_set_hash_netnet ip_set_hash_netiface ip_set_hash_net ip_set_hash_mac ip_set_hash_ipportnet ip_set_hash_ipportip ip_set_hash_ipport ip_set_hash_ipmark ip_set_hash_ip ip_set_bitmap_port ip_set_bitmap_ipmac ip_set_bitmap_ip ip_set nfnetlink ip6table_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 ip6t_NPT ip6table_mangle <7>[10901.846575] ip6table_filter ip6_tables ip6t_REJECT x_tables nf_reject_ipv6 tun zram zsmalloc crypto_user algif_skcipher algif_rng algif_hash algif_aead af_alg sha1_generic seqiv authenc arc4 leds_gpio gpio_button_hotplug <7>[10901.952790] CPU: 1 PID: 4271 Comm: kworker/u4:3 Not tainted 5.10.134 #0 <7>[10901.959392] Hardware name: Xiaomi Redmi Router AX6S (DT) <7>[10901.964712] Workqueue: nf_ft_offload_stats nf_flow_table_offload_setup [nf_flow_table] <7>[10901.972622] pstate: 20000005 (nzCv daif -PAN -UAO -TCO BTYPE=--) <7>[10901.978622] pc : nf_flow_offload_ip_hook+0x7c4/0x138c [nf_flow_table] <7>[10901.985056] lr : nf_flow_offload_ip_hook+0x7d8/0x138c [nf_flow_table] <7>[10901.991485] sp : ffffffc014a5bc20 <7>[10901.994789] x29: ffffffc014a5bc20 x28: 0000000000000000 <7>[10902.000093] x27: ffffff80038a4248 x26: 0000000000000000 <7>[10902.005397] x25: ffffff8001b71a00 x24: ffffffc008ab9660 <7>[10902.010702] x23: ffffffc014a5bd40 x22: 0000000000000002 <7>[10902.016007] x21: ffffffc008ab9650 x20: 0000000000000000 <7>[10902.021311] x19: dead0000000000f0 x18: 0000000000000000 <7>[10902.026615] x17: 0000000000000015 x16: ffffffc0108185d0 <7>[10902.031919] x15: 000000000000003d x14: 0000000000000000 <7>[10902.037224] x13: ffffffc008ab9398 x12: ffffffffffffffff <7>[10902.042528] x11: 0000000000000008 x10: ffffff8000005100 <7>[10902.047832] x9 : ffffff8003afff7c x8 : 0000000000000000 <7>[10902.053136] x7 : ffffffffffffffff x6 : 0089e70200000000 <7>[10902.058441] x5 : 0000000000000002 x4 : 0000000000000000 <7>[10902.063744] x3 : ffffff800018ad00 x2 : 0000000000000000 <7>[10902.069049] x1 : dead000000000100 x0 : 0000000000000001 <7>[10902.074354] Call trace: <7>[10902.076795] nf_flow_offload_ip_hook+0x7c4/0x138c [nf_flow_table] <7>[10902.082880] nf_flow_table_offload_setup+0x478/0x670 [nf_flow_table] <7>[10902.089230] process_one_work+0x1d4/0x370 <7>[10902.093231] worker_thread+0x178/0x4e0 <7>[10902.096971] kthread+0x120/0x124 <7>[10902.100192] ret_from_fork+0x10/0x18 <0>[10902.103761] Code: d1004013 eb0002bf 54000580 52800014 (f9401263) <4>[10902.109846] ---[ end trace 35a8132a39490859 ]--- Panic#2 Part1 <6>[ 9.116066] xt_time: kernel timezone is -0000 <7>[ 9.172139] ieee80211 phy0: Selected rate control algorithm 'minstrel_ht' <7>[ 9.183164] mt7915e 0000:01:00.0: assign IRQ: got 142 <7>[ 9.188314] pci 0000:00:00.0: enabling bus mastering <6>[ 9.193291] mt7915e 0000:01:00.0: enabling device (0000 -> 0002) <7>[ 9.199395] mt7915e 0000:01:00.0: enabling bus mastering <6>[ 9.261106] mt7622-wmac 18000000.wmac: HW/SW Version: 0x8a108a10, Build Time: 20190801210006a <6>[ 9.261106] <6>[ 9.366162] mt7622-wmac 18000000.wmac: N9 Firmware Version: reserved, Build Time: 20220630094834 <6>[ 9.408501] mt7915e 0000:01:00.0: HW/SW Version: 0x8a108a10, Build Time: 20211222184017a <6>[ 9.408501] <6>[ 9.549747] mt7915e 0000:01:00.0: WM Firmware Version: ____000000, Build Time: 20211222184052 <6>[ 9.614454] mt7915e 0000:01:00.0: WA Firmware Version: DEV_000000, Build Time: 20211222184111 <6>[ 9.735732] PPP generic driver version 2.4.2 <6>[ 9.740941] PPP MPPE Compression module registered <6>[ 9.746777] NET: Registered protocol family 24 <14>[ 9.753146] kmodloader: done loading kernel modules from /etc/modules.d/* <6>[ 10.355906] zram0: detected capacity change from 0 to 78643200 <6>[ 10.426265] Adding 76796k swap on /dev/zram0. Priority:100 extents:1 across:76796k SSDsc <6>[ 12.700070] mtk_soc_eth 1b100000.ethernet eth0: Link is Down <6>[ 12.703505] mtk_soc_eth 1b100000.ethernet eth0: configuring for fixed/2500base-x link mode <6>[ 12.703677] mtk_soc_eth 1b100000.ethernet eth0: Link is Up - 2.5Gbps/Full - flow control rx/tx <6>[ 12.703695] IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready <6>[ 12.707013] mt7530 mdio-bus:00 lan1: configuring for phy/gmii link mode <6>[ 12.707758] 8021q: adding VLAN 0 to HW filter on device lan1 <6>[ 12.712393] br-lan: port 1(lan1) entered blocking state <6>[ 12.712404] br-lan: port 1(lan1) entered disabled state <6>[ 12.716849] device lan1 entered promiscuous mode <6>[ 12.716857] device eth0 entered promiscuous mode <6>[ 12.726779] mt7530 mdio-bus:00 lan2: configuring for phy/gmii link mode <6>[ 12.727334] 8021q: adding VLAN 0 to HW filter on device lan2 <6>[ 12.731970] br-lan: port 2(lan2) entered blocking state <6>[ 12.731981] br-lan: port 2(lan2) entered disabled state <6>[ 12.732684] device lan2 entered promiscuous mode <6>[ 12.741533] mt7530 mdio-bus:00 lan3: configuring for phy/gmii link mode <6>[ 12.743953] 8021q: adding VLAN 0 to HW filter on device lan3 <6>[ 12.748836] mt7530 mdio-bus:00 lan3: Link is Up - 1Gbps/Full - flow control rx/tx <6>[ 12.752498] br-lan: port 3(lan3) entered blocking state <6>[ 12.752509] br-lan: port 3(lan3) entered disabled state <6>[ 12.753896] device lan3 entered promiscuous mode <6>[ 12.756669] br-lan: port 3(lan3) entered blocking state <6>[ 12.756681] br-lan: port 3(lan3) entered forwarding state <6>[ 12.756768] IPv6: ADDRCONF(NETDEV_CHANGE): br-lan: link becomes ready <6>[ 12.769739] mt7530 mdio-bus:00 wan: configuring for phy/gmii link mode <6>[ 12.770260] 8021q: adding VLAN 0 to HW filter on device wan <6>[ 12.774823] mt7530 mdio-bus:00 wan: Link is Up - 1Gbps/Full - flow control off <6>[ 12.777093] IPv6: ADDRCONF(NETDEV_CHANGE): wan: link becomes ready <6>[ 13.159944] pppoe-wan: renamed from ppp0 <6>[ 13.433189] mt7530 mdio-bus:00 lan2: Link is Up - 100Mbps/Full - flow control rx/tx <6>[ 14.286329] mt7530 mdio-bus:00 lan2: Link is Down <6>[ 14.294018] mt7530 mdio-bus:00 lan1: Link is Up - 100Mbps/Full - flow control rx/tx <6>[ 15.002148] br-lan: port 1(lan1) entered blocking state <6>[ 15.002168] br-lan: port 1(lan1) entered forwarding state <6>[ 15.259329] br-lan: port 4(wlan0) entered blocking state <6>[ 15.259342] br-lan: port 4(wlan0) entered disabled state <6>[ 15.259568] device wlan0 entered promiscuous mode <6>[ 15.259640] br-lan: port 4(wlan0) entered blocking state <6>[ 15.259647] br-lan: port 4(wlan0) entered forwarding state <6>[ 15.918171] mt7530 mdio-bus:00 lan2: Link is Up - 100Mbps/Full - flow control rx/tx <6>[ 15.918229] br-lan: port 2(lan2) entered blocking state <6>[ 15.918243] br-lan: port 2(lan2) entered forwarding state <6>[ 16.006841] br-lan: port 4(wlan0) entered disabled state <6>[ 16.640989] br-lan: port 5(wlan1) entered blocking state <6>[ 16.641001] br-lan: port 5(wlan1) entered disabled state <6>[ 16.641170] device wlan1 entered promiscuous mode <6>[ 16.641231] br-lan: port 5(wlan1) entered blocking state <6>[ 16.641237] br-lan: port 5(wlan1) entered forwarding state <6>[ 17.038752] br-lan: port 5(wlan1) entered disabled state <6>[ 17.940844] IPv6: ADDRCONF(NETDEV_CHANGE): wlan1: link becomes ready <6>[ 17.941026] br-lan: port 5(wlan1) entered blocking state <6>[ 17.941035] br-lan: port 5(wlan1) entered forwarding state <6>[ 22.031440] IPv6: ADDRCONF(NETDEV_CHANGE): wlan0: link becomes ready <6>[ 22.031584] br-lan: port 4(wlan0) entered blocking state <6>[ 22.031593] br-lan: port 4(wlan0) entered forwarding state <1>[10901.799290] Unable to handle kernel paging request at virtual address dead000000000110 <1>[10901.807213] Mem abort info: <1>[10901.810041] ESR = 0x96000004 <1>[10901.813095] EC = 0x25: DABT (current EL), IL = 32 bits <1>[10901.818401] SET = 0, FnV = 0 <1>[10901.821489] EA = 0, S1PTW = 0 <1>[10901.824646] Data abort info: <1>[10901.827516] ISV = 0, ISS = 0x00000004 <1>[10901.831378] CM = 0, WnR = 0 <1>[10901.834364] [dead000000000110] address between user and kernel address ranges <0>[10901.841554] Internal error: Oops: 96000004 [#1] SMP <7>[10901.846425] Modules linked in: pppoe ppp_async pppox ppp_mppe ppp_generic mt7915e mt7615e mt7615_common mt76_connac_lib mt76 mac80211 ipt_REJECT cfg80211 xt_time xt_tcpudp xt_state xt_quota xt_pkttype xt_owner xt_nat xt_multiport xt_mark xt_mac xt_limit xt_conntrack xt_comment xt_cgroup xt_addrtype xt_TPROXY xt_TCPMSS xt_REDIRECT xt_MASQUERADE xt_LOG xt_FLOWOFFLOAD xt_CT tcp_bbr slhc nf_tproxy_ipv6 nf_tproxy_ipv4 nf_reject_ipv4 nf_log_ipv6 nf_log_ipv4 nf_log_common nf_flow_table nf_conntrack_netlink macvlan iptable_nat iptable_mangle iptable_filter ip_tables hwmon crc_ccitt compat cryptodev xt_set ip_set_list_set ip_set_hash_netportnet ip_set_hash_netport ip_set_hash_netnet ip_set_hash_netiface ip_set_hash_net ip_set_hash_mac ip_set_hash_ipportnet ip_set_hash_ipportip ip_set_hash_ipport ip_set_hash_ipmark ip_set_hash_ip ip_set_bitmap_port ip_set_bitmap_ipmac ip_set_bitmap_ip ip_set nfnetlink ip6table_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 ip6t_NPT ip6table_mangle <7>[10901.846575] ip6table_filter ip6_tables ip6t_REJECT x_tables nf_reject_ipv6 tun zram zsmalloc crypto_user algif_skcipher algif_rng algif_hash algif_aead af_alg sha1_generic seqiv authenc arc4 leds_gpio gpio_button_hotplug <7>[10901.952790] CPU: 1 PID: 4271 Comm: kworker/u4:3 Not tainted 5.10.134 #0 <7>[10901.959392] Hardware name: Xiaomi Redmi Router AX6S (DT) <7>[10901.964712] Workqueue: nf_ft_offload_stats nf_flow_table_offload_setup [nf_flow_table] <7>[10901.972622] pstate: 20000005 (nzCv daif -PAN -UAO -TCO BTYPE=--) <7>[10901.978622] pc : nf_flow_offload_ip_hook+0x7c4/0x138c [nf_flow_table] <7>[10901.985056] lr : nf_flow_offload_ip_hook+0x7d8/0x138c [nf_flow_table] <7>[10901.991485] sp : ffffffc014a5bc20 <7>[10901.994789] x29: ffffffc014a5bc20 x28: 0000000000000000 <7>[10902.000093] x27: ffffff80038a4248 x26: 0000000000000000 <7>[10902.005397] x25: ffffff8001b71a00 x24: ffffffc008ab9660 <7>[10902.010702] x23: ffffffc014a5bd40 x22: 0000000000000002 <7>[10902.016007] x21: ffffffc008ab9650 x20: 0000000000000000 <7>[10902.021311] x19: dead0000000000f0 x18: 0000000000000000 <7>[10902.026615] x17: 0000000000000015 x16: ffffffc0108185d0 <7>[10902.031919] x15: 000000000000003d x14: 0000000000000000 <7>[10902.037224] x13: ffffffc008ab9398 x12: ffffffffffffffff <7>[10902.042528] x11: 0000000000000008 x10: ffffff8000005100 <7>[10902.047832] x9 : ffffff8003afff7c x8 : 0000000000000000 <7>[10902.053136] x7 : ffffffffffffffff x6 : 0089e70200000000 <7>[10902.058441] x5 : 0000000000000002 x4 : 0000000000000000 <7>[10902.063744] x3 : ffffff800018ad00 x2 : 0000000000000000 <7>[10902.069049] x1 : dead000000000100 x0 : 0000000000000001 <7>[10902.074354] Call trace: <7>[10902.076795] nf_flow_offload_ip_hook+0x7c4/0x138c [nf_flow_table] <7>[10902.082880] nf_flow_table_offload_setup+0x478/0x670 [nf_flow_table] <7>[10902.089230] process_one_work+0x1d4/0x370 <7>[10902.093231] worker_thread+0x178/0x4e0 <7>[10902.096971] kthread+0x120/0x124 <7>[10902.100192] ret_from_fork+0x10/0x18 <0>[10902.103761] Code: d1004013 eb0002bf 54000580 52800014 (f9401263) <4>[10902.109846] ---[ end trace 35a8132a39490859 ]--- <0>[10902.119936] Kernel panic - not syncing: Oops: Fatal exception <2>[10902.125675] SMP: stopping secondary CPUs <0>[10902.129591] Kernel Offset: disabled <0>[10902.133073] CPU features: 0x0000002,04002004 <0>[10902.137333] Memory Limit: none ` 如果安装了Watchcat 模块(服务菜单里),可以尝试把 默认的 8.8.8.8那个ip改成你能ping通的国内地址,就不会导致误判重启了,或者用不到的话你可以关闭这个模块,希望对你有帮助 如果安装了Watchcat模块(菜单地址里),可以试一下8.8.8.8那个ip成你ping通的国内模块,就不会导致判别重启了,或者这个用不到的话你可以关闭,希望对你有帮助 谢谢你提供的方法,我的 模块只安装了openchash upnp bypass 和自带的系统在线更新 同样ax6s,模块只装了 系统在线更新、bypass 和 upnp,也是不定时重启,每天大概重启3、4次。 尝试下关闭网络-防火墙-路由/NAT 分载?我之前也是没装啥也反复重启,关了就稳定多了. 我的 网络-防火墙 页面里面没有路由/Nat 分载 这一项啊..
gharchive/issue
2022-08-18T06:32:11
2025-04-01T06:39:17.106123
{ "authors": [ "ERxhao", "SRGOD", "digiw", "dreamflandre" ], "repo": "kiddin9/OpenWrt_x86-r2s-r4s-r5s-N1", "url": "https://github.com/kiddin9/OpenWrt_x86-r2s-r4s-r5s-N1/issues/889", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1110099930
KOGITO-6606 KOGITO-6446 Produce event listener before starting Data I… …ndex The (build) kogito-runtimes check has failed. Please check the logs. The (build) kogito-apps check has failed. Please check the logs. The (build) kogito-examples check has failed. Please check the logs. The (build) kogito-apps check has failed. Please check the logs. Jenkins rerun kogito-examples tests Jenkins rerun kogito-apps tests The (build) kogito-examples check has failed. Please check the logs. Jenkins rerun kogito-examples tests
gharchive/pull-request
2022-01-21T06:20:06
2025-04-01T06:39:17.130269
{ "authors": [ "cristianonicolai", "kie-ci", "nmirasch" ], "repo": "kiegroup/kogito-runtimes", "url": "https://github.com/kiegroup/kogito-runtimes/pull/1916", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1522681709
[KOGITO-8432] - CVE-2022-25857 - Upgrade SnakeYAML on Spring Boot related components See: https://issues.redhat.com/browse/KOGITO-8432 Small update on our SB libraries to fix the affected components. See: https://github.com/spring-projects/spring-boot/issues/32221 for more info. Signed-off-by: Ricardo Zanini zanini@redhat.com Many thanks for submitting your Pull Request :heart:! Please make sure that your PR meets the following requirements: [x] You have read the contributors guide [x] Your code is properly formatted according to this configuration [x] Pull Request title is properly formatted: KOGITO-XYZ Subject [x] Pull Request title contains the target branch if not targeting main: [0.9.x] KOGITO-XYZ Subject [x] Pull Request contains link to the JIRA issue [x] Pull Request contains link to any dependent or related Pull Request [x] Pull Request contains description of the issue [x] Pull Request does not include fixes for issues other than the main ticket How to replicate CI configuration locally? Build Chain tool does "simple" maven build(s), the builds are just Maven commands, but because the repositories relates and depends on each other and any change in API or class method could affect several of those repositories there is a need to use build-chain tool to handle cross repository builds and be sure that we always use latest version of the code for each repository. build-chain tool is a build tool which can be used on command line locally or in Github Actions workflow(s), in case you need to change multiple repositories and send multiple dependent pull requests related with a change you can easily reproduce the same build by executing it on Github hosted environment or locally in your development environment. See local execution details to get more information about it. How to retest this PR or trigger a specific build: for pull request checks Please add comment: Jenkins retest this for a specific pull request check Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] tests for quarkus branch checks Run checks against Quarkus current used branch Please add comment: Jenkins run quarkus-branch for a quarkus branch specific check Run checks against Quarkus current used branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-branch for quarkus main checks Run checks against Quarkus main branch Please add comment: Jenkins run quarkus-main for a specific quarkus main check Run checks against Quarkus main branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-main for quarkus lts checks Run checks against Quarkus lts branch Please add comment: Jenkins run quarkus-lts for a specific quarkus lts check Run checks against Quarkus lts branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-lts for native checks Run native checks Please add comment: Jenkins run native for a specific native check Run native checks Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] native for mandrel checks Run native checks against Mandrel image Please add comment: Jenkins run mandrel for a specific mandrel check Run native checks against Mandrel image Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] mandrel for mandrel lts checks Run native checks against Mandrel image and quarkus lts branch Please add comment: Jenkins run mandrel-lts for a specific mandrel lts check Run native checks against Mandrel image and quarkus lts branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] mandrel-lts Full Kogito testing (with cloud images and operator BDD testing) Please add comment: Jenkins run BDD This check should be used only if a big change is done as it takes time to run, need resources and one full BDD tests check can be done at a time ... How to backport a pull request to a different branch? In order to automatically create a backporting pull request please add one or more labels having the following format backport-<branch-name>, where <branch-name> is the name of the branch where the pull request must be backported to (e.g., backport-7.67.x to backport the original PR to the 7.67.x branch). NOTE: backporting is an action aiming to move a change (usually a commit) from a branch (usually the main one) to another one, which is generally referring to a still maintained release branch. Keeping it simple: it is about to move a specific change or a set of them from one branch to another. Once the original pull request is successfully merged, the automated action will create one backporting pull request per each label (with the previous format) that has been added. If something goes wrong, the author will be notified and at this point a manual backporting is needed. NOTE: this automated backporting is triggered whenever a pull request on main branch is labeled or closed, but both conditions must be satisfied to get the new PR created. (tests) - kogito-runtimes job #1657 was: UNSTABLE Possible explanation: This should be test failures Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/1657/display/redirect Test results: PASSED: 3147 FAILED: 1 Those are the test failures: org.kie.kogito.addons.quarkus.kubernetes.KubernetesAddOnTest.(?) java.lang.RuntimeException: java.lang.RuntimeException: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor#setupKubernetesDevService threw an exception: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:142) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 24 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 25 moreCaused by: com.github.dockerjava.api.exception.BadRequestException: Status 400: {"message":"failed to create shim task: OCI runtime create failed: runc create failed: unable to create new parent process: namespace path: lstat /proc/6964/ns/net: no such file or directory: unknown"} at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.execute(DefaultInvocationBuilder.java:237) at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.post(DefaultInvocationBuilder.java:102) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:31) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:13) at org.testcontainers.shaded.com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21) at org.testcontainers.shaded.com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35) at org.testcontainers.shaded.com.github.dockerjava.core.command.StartContainerCmdImpl.exec(StartContainerCmdImpl.java:43) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:441) ... 27 more at io.quarkus.test.QuarkusUnitTest.beforeAll(QuarkusUnitTest.java:689) at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$invokeBeforeAllCallbacks$12(ClassBasedTestDescriptor.java:395) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.invokeBeforeAllCallbacks(ClassBasedTestDescriptor.java:395) at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.before(ClassBasedTestDescriptor.java:211) at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.before(ClassBasedTestDescriptor.java:84) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:148) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95) at java.base/java.util.ArrayList.forEach(ArrayList.java:1541) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:35) at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57) at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:54) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:55) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:102) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:54) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86) at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86) at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:188) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:128) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)Caused by: java.lang.RuntimeException: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor#setupKubernetesDevService threw an exception: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:142) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 24 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 25 moreCaused by: com.github.dockerjava.api.exception.BadRequestException: Status 400: {"message":"failed to create shim task: OCI runtime create failed: runc create failed: unable to create new parent process: namespace path: lstat /proc/6964/ns/net: no such file or directory: unknown"} at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.execute(DefaultInvocationBuilder.java:237) at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.post(DefaultInvocationBuilder.java:102) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:31) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:13) at org.testcontainers.shaded.com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21) at org.testcontainers.shaded.com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35) at org.testcontainers.shaded.com.github.dockerjava.core.command.StartContainerCmdImpl.exec(StartContainerCmdImpl.java:43) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:441) ... 27 more at io.quarkus.runner.bootstrap.AugmentActionImpl.runAugment(AugmentActionImpl.java:335) at io.quarkus.runner.bootstrap.AugmentActionImpl.createInitialRuntimeApplication(AugmentActionImpl.java:252) at io.quarkus.test.QuarkusUnitTest.beforeAll(QuarkusUnitTest.java:635) ... 43 moreCaused by: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor#setupKubernetesDevService threw an exception: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:142) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 24 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 25 moreCaused by: com.github.dockerjava.api.exception.BadRequestException: Status 400: {"message":"failed to create shim task: OCI runtime create failed: runc create failed: unable to create new parent process: namespace path: lstat /proc/6964/ns/net: no such file or directory: unknown"} at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.execute(DefaultInvocationBuilder.java:237) at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.post(DefaultInvocationBuilder.java:102) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:31) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:13) at org.testcontainers.shaded.com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21) at org.testcontainers.shaded.com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35) at org.testcontainers.shaded.com.github.dockerjava.core.command.StartContainerCmdImpl.exec(StartContainerCmdImpl.java:43) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:441) ... 27 more at io.quarkus.builder.Execution.run(Execution.java:123) at io.quarkus.builder.BuildExecutionBuilder.execute(BuildExecutionBuilder.java:79) at io.quarkus.deployment.QuarkusAugmentor.run(QuarkusAugmentor.java:160) at io.quarkus.runner.bootstrap.AugmentActionImpl.runAugment(AugmentActionImpl.java:331) ... 45 moreCaused by: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:142) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 24 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 25 moreCaused by: com.github.dockerjava.api.exception.BadRequestException: Status 400: {"message":"failed to create shim task: OCI runtime create failed: runc create failed: unable to create new parent process: namespace path: lstat /proc/6964/ns/net: no such file or directory: unknown"} at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.execute(DefaultInvocationBuilder.java:237) at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.post(DefaultInvocationBuilder.java:102) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:31) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:13) at org.testcontainers.shaded.com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21) at org.testcontainers.shaded.com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35) at org.testcontainers.shaded.com.github.dockerjava.core.command.StartContainerCmdImpl.exec(StartContainerCmdImpl.java:43) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:441) ... 27 more jenkins rerun kogito-runtimes tests (tests) - kogito-runtimes job #1659 was: UNSTABLE Possible explanation: This should be test failures Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/1659/display/redirect Test results: PASSED: 3145 FAILED: 2 Those are the test failures: ilove.quark.us.GreetTest.testEnglish java.lang.RuntimeException: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor#setupKubernetesDevService threw an exception: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.shaded.org.awaitility.core.ConditionTimeoutException: Lambda expression in com.dajudge.kindcontainer.ApiServerContainer: expected the predicate to return but it returned for input of within 10 seconds. at org.testcontainers.shaded.org.awaitility.core.ConditionAwaiter.await(ConditionAwaiter.java:167) at org.testcontainers.shaded.org.awaitility.core.AbstractHamcrestCondition.await(AbstractHamcrestCondition.java:86) at org.testcontainers.shaded.org.awaitility.core.ConditionFactory.until(ConditionFactory.java:985) at org.testcontainers.shaded.org.awaitility.core.ConditionFactory.until(ConditionFactory.java:691) at org.testcontainers.shaded.org.awaitility.core.ConditionFactory.until(ConditionFactory.java:708) at com.dajudge.kindcontainer.ApiServerContainer.waitForDefaultNamespace(ApiServerContainer.java:150) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:144) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.FutureTask.get(FutureTask.java:204) at org.testcontainers.shaded.org.awaitility.core.Uninterruptibles.getUninterruptibly(Uninterruptibles.java:101) at org.testcontainers.shaded.org.awaitility.core.Uninterruptibles.getUninterruptibly(Uninterruptibles.java:81) at org.testcontainers.shaded.org.awaitility.core.ConditionAwaiter.await(ConditionAwaiter.java:103) ... 28 more org.kie.kogito.quarkus.ServerlessWorkflowCodestartTest.buildAllProjectsForLocalUse org.opentest4j.AssertionFailedError: [Run project return status is zero] expected: 0 but was: 1 at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at io.quarkus.devtools.testing.codestarts.QuarkusCodestartTest.buildProject(QuarkusCodestartTest.java:158) at io.quarkus.devtools.testing.codestarts.QuarkusCodestartTest.buildAllProjects(QuarkusCodestartTest.java:145) at org.kie.kogito.quarkus.ServerlessWorkflowCodestartTest.buildAllProjectsForLocalUse(ServerlessWorkflowCodestartTest.java:45) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.platform.commons.util.ReflectionUtils.invokeMethod(ReflectionUtils.java:727) at org.junit.jupiter.engine.execution.MethodInvocation.proceed(MethodInvocation.java:60) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$ValidatingInvocation.proceed(InvocationInterceptorChain.java:131) at org.junit.jupiter.engine.extension.TimeoutExtension.intercept(TimeoutExtension.java:156) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestableMethod(TimeoutExtension.java:147) at org.junit.jupiter.engine.extension.TimeoutExtension.interceptTestMethod(TimeoutExtension.java:86) at org.junit.jupiter.engine.execution.InterceptingExecutableInvoker$ReflectiveInterceptorCall.lambda$ofVoidMethod$0(InterceptingExecutableInvoker.java:103) at org.junit.jupiter.engine.execution.InterceptingExecutableInvoker.lambda$invoke$0(InterceptingExecutableInvoker.java:93) at org.junit.jupiter.engine.execution.InvocationInterceptorChain$InterceptedInvocation.proceed(InvocationInterceptorChain.java:106) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.proceed(InvocationInterceptorChain.java:64) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.chainAndInvoke(InvocationInterceptorChain.java:45) at org.junit.jupiter.engine.execution.InvocationInterceptorChain.invoke(InvocationInterceptorChain.java:37) at org.junit.jupiter.engine.execution.InterceptingExecutableInvoker.invoke(InterceptingExecutableInvoker.java:92) at org.junit.jupiter.engine.execution.InterceptingExecutableInvoker.invoke(InterceptingExecutableInvoker.java:86) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.lambda$invokeTestMethod$7(TestMethodTestDescriptor.java:217) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.invokeTestMethod(TestMethodTestDescriptor.java:213) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:138) at org.junit.jupiter.engine.descriptor.TestMethodTestDescriptor.execute(TestMethodTestDescriptor.java:68) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:151) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95) at java.base/java.util.ArrayList.forEach(ArrayList.java:1541) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95) at java.base/java.util.ArrayList.forEach(ArrayList.java:1541) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:35) at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57) at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:54) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:55) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:102) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:54) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86) at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86) at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:188) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:128) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) jenkins rerun kogito-runtimes tests (tests) - kogito-runtimes job #1661 was: ABORTED Possible explanation: Most probably a timeout, please review Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/1661/display/redirect Test results: PASSED: 3064 FAILED: 0 Those are the test failures: none See console log: Console Logs [2023-01-06T22:06:19.691Z] [INFO] | | | +- org.rnorth.duct-tape:duct-tape:jar:1.0.8:compile[2023-01-06T22:06:19.691Z] [INFO] | | | | \- org.jetbrains:annotations:jar:17.0.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | | \- com.github.docker-java:docker-java-transport-zerodep:jar:3.2.13:compile[2023-01-06T22:06:19.691Z] [INFO] | | | +- com.github.docker-java:docker-java-transport:jar:3.2.13:compile[2023-01-06T22:06:19.691Z] [INFO] | | | \- net.java.dev.jna:jna:jar:5.8.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | \- org.bouncycastle:bcpkix-jdk15on:jar:1.70:runtime[2023-01-06T22:06:19.691Z] [INFO] | | +- org.bouncycastle:bcprov-jdk15on:jar:1.70:runtime[2023-01-06T22:06:19.691Z] [INFO] | | \- org.bouncycastle:bcutil-jdk15on:jar:1.70:runtime[2023-01-06T22:06:19.691Z] [INFO] | \- io.quarkus:quarkus-jackson-deployment:jar:2.15.0.Final:compile[2023-01-06T22:06:19.691Z] [INFO] | \- io.quarkus:quarkus-jackson-spi:jar:2.15.0.Final:compile[2023-01-06T22:06:19.691Z] [INFO] +- io.quarkus:quarkus-openshift-client-deployment:jar:2.15.0.Final:compile[2023-01-06T22:06:19.691Z] [INFO] | \- io.quarkus:quarkus-openshift-client:jar:2.15.0.Final:compile[2023-01-06T22:06:19.691Z] [INFO] | \- io.fabric8:openshift-client:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | +- io.fabric8:openshift-client-api:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-clusterautoscaling:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-operator:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-operatorhub:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-machine:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-whereabouts:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-monitoring:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-storageversionmigrator:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-tuned:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-console:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-config:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-machineconfig:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-miscellaneous:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.fabric8:openshift-model-hive:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | | \- io.fabric8:openshift-model-installer:jar:6.2.0:compile[2023-01-06T22:06:19.691Z] [INFO] | \- com.github.mifmif:generex:jar:1.0.2:compile[2023-01-06T22:06:19.691Z] [INFO] | \- dk.brics.automaton:automaton:jar:1.11-8:compile[2023-01-06T22:06:19.691Z] [INFO] +- io.quarkus:quarkus-cache-deployment:jar:2.15.0.Final:compile[2023-01-06T22:06:19.691Z] [INFO] | +- io.quarkus:quarkus-cache:jar:2.15.0.Final:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.quarkus:quarkus-caffeine:jar:2.15.0.Final:compile[2023-01-06T22:06:19.691Z] [INFO] | | | \- com.github.ben-manes.caffeine:caffeine:jar:3.1.1:compile[2023-01-06T22:06:19.691Z] [INFO] | | | \- com.google.errorprone:error_prone_annotations:jar:2.16:compile[2023-01-06T22:06:19.691Z] [INFO] | | \- io.quarkus:quarkus-mutiny:jar:2.15.0.Final:compile[2023-01-06T22:06:19.691Z] [INFO] | | +- io.smallrye.reactive:mutiny:jar:1.8.0:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- org.reactivestreams:reactive-streams:jar:1.0.3:compile[2023-01-06T22:06:19.692Z] [INFO] | | | \- io.smallrye.common:smallrye-common-annotation:jar:1.13.2:compile[2023-01-06T22:06:19.692Z] [INFO] | | +- io.quarkus:quarkus-smallrye-context-propagation:jar:2.15.0.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | \- io.smallrye:smallrye-context-propagation:jar:1.2.2:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.smallrye:smallrye-context-propagation-api:jar:1.2.2:compile[2023-01-06T22:06:19.692Z] [INFO] | | | \- io.smallrye:smallrye-context-propagation-storage:jar:1.2.2:compile[2023-01-06T22:06:19.692Z] [INFO] | | \- io.smallrye.reactive:mutiny-smallrye-context-propagation:jar:1.8.0:compile[2023-01-06T22:06:19.692Z] [INFO] | +- io.quarkus:quarkus-arc-deployment:jar:2.15.0.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | +- io.quarkus:quarkus-smallrye-context-propagation-spi:jar:2.15.0.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | +- io.quarkus:quarkus-vertx-http-dev-console-spi:jar:2.15.0.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.quarkus:quarkus-vertx-http-dev-console-runtime-spi:jar:2.15.0.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | \- io.vertx:vertx-web:jar:4.3.5:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.vertx:vertx-web-common:jar:4.3.5:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.vertx:vertx-auth-common:jar:4.3.5:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.vertx:vertx-bridge-common:jar:4.3.5:compile[2023-01-06T22:06:19.692Z] [INFO] | | | \- io.vertx:vertx-core:jar:4.3.5:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.netty:netty-common:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.netty:netty-buffer:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.netty:netty-transport:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.netty:netty-handler:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | | +- io.netty:netty-transport-native-unix-common:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | | \- io.netty:netty-codec:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.netty:netty-handler-proxy:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | | \- io.netty:netty-codec-socks:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.netty:netty-codec-http:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.netty:netty-codec-http2:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | +- io.netty:netty-resolver:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | \- io.netty:netty-resolver-dns:jar:4.1.85.Final:compile[2023-01-06T22:06:19.692Z] [INFO] | | | \- io.netty:netty-codec-dns:jar:4.1.85.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | | \- io.quarkus.arc:arc-processor:jar:2.15.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | +- io.quarkus:quarkus-caffeine-deployment:jar:2.15.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | \- io.quarkus:quarkus-mutiny-deployment:jar:2.15.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | \- io.quarkus:quarkus-smallrye-context-propagation-deployment:jar:2.15.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] +- org.kie.kogito:kogito-addons-quarkus-common-deployment:jar:2.0.0-SNAPSHOT:compile[2023-01-06T22:06:19.693Z] [INFO] | \- io.quarkus:quarkus-core-deployment:jar:2.15.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | +- org.aesh:readline:jar:2.2:compile[2023-01-06T22:06:19.693Z] [INFO] | | \- org.fusesource.jansi:jansi:jar:1.18:compile[2023-01-06T22:06:19.693Z] [INFO] | +- org.aesh:aesh:jar:2.6:compile[2023-01-06T22:06:19.693Z] [INFO] | +- org.apache.commons:commons-lang3:jar:3.12.0:compile[2023-01-06T22:06:19.693Z] [INFO] | +- org.wildfly.common:wildfly-common:jar:1.5.4.Final-format-001:compile[2023-01-06T22:06:19.693Z] [INFO] | +- io.quarkus.gizmo:gizmo:jar:1.4.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | | \- org.ow2.asm:asm-util:jar:9.4:compile[2023-01-06T22:06:19.693Z] [INFO] | | \- org.ow2.asm:asm-analysis:jar:9.4:compile[2023-01-06T22:06:19.693Z] [INFO] | +- io.smallrye:jandex:jar:3.0.5:compile[2023-01-06T22:06:19.693Z] [INFO] | +- org.ow2.asm:asm:jar:9.4:compile[2023-01-06T22:06:19.693Z] [INFO] | +- org.ow2.asm:asm-commons:jar:9.4:compile[2023-01-06T22:06:19.693Z] [INFO] | | \- org.ow2.asm:asm-tree:jar:9.4:compile[2023-01-06T22:06:19.693Z] [INFO] | +- io.quarkus:quarkus-development-mode-spi:jar:2.15.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | +- io.quarkus:quarkus-class-change-agent:jar:2.15.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | +- io.quarkus:quarkus-bootstrap-core:jar:2.15.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | | +- io.quarkus:quarkus-bootstrap-app-model:jar:2.15.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | | \- io.smallrye.common:smallrye-common-io:jar:1.13.2:compile[2023-01-06T22:06:19.693Z] [INFO] | +- org.eclipse.sisu:org.eclipse.sisu.inject:jar:0.3.5:runtime[2023-01-06T22:06:19.693Z] [INFO] | +- io.quarkus:quarkus-builder:jar:2.15.0.Final:compile[2023-01-06T22:06:19.693Z] [INFO] | +- org.graalvm.sdk:graal-sdk:jar:22.3.0:compile[2023-01-06T22:06:19.693Z] [INFO] | +- org.junit.platform:junit-platform-launcher:jar:1.9.1:compile[2023-01-06T22:06:19.694Z] [INFO] | \- org.junit.jupiter:junit-jupiter:jar:5.9.1:test[2023-01-06T22:06:19.694Z] [INFO] +- org.kie.kogito:kogito-addons-quarkus-kubernetes:jar:2.0.0-SNAPSHOT:compile[2023-01-06T22:06:19.694Z] [INFO] | +- io.fabric8:knative-client:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:knative-model:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | | \- io.fabric8:kubernetes-model-core:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | | \- io.fabric8:kubernetes-model-common:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | \- io.fabric8:kubernetes-client-api:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-gatewayapi:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-rbac:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-admissionregistration:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-apps:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-autoscaling:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-apiextensions:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-batch:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-certificates:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-coordination:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-discovery:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-events:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-extensions:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-flowcontrol:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-networking:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-metrics:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-policy:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-scheduling:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-storageclass:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-model-node:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- com.fasterxml.jackson.dataformat:jackson-dataformat-yaml:jar:2.14.1:compile[2023-01-06T22:06:19.694Z] [INFO] | | \- com.fasterxml.jackson.core:jackson-core:jar:2.14.1:compile[2023-01-06T22:06:19.694Z] [INFO] | +- io.fabric8:kubernetes-client:jar:6.2.0:compile[2023-01-06T22:06:19.694Z] [INFO] | | +- io.fabric8:kubernetes-httpclient-okhttp:jar:6.2.0:runtime[2023-01-06T22:06:19.694Z] [INFO] | | | +- com.squareup.okhttp3:okhttp:jar:3.14.9:runtime[2023-01-06T22:06:19.694Z] [INFO] | | | | \- com.squareup.okio:okio:jar:1.17.2:runtime[2023-01-06T22:06:19.694Z] [INFO] | | | \- com.squareup.okhttp3:logging-interceptor:jar:3.14.9:runtime[2023-01-06T22:06:19.694Z] [INFO] | | \- io.fabric8:zjsonpatch:jar:0.3.0:compile[2023-01-06T22:06:19.694Z] [INFO] | \- org.kie.kogito:kogito-addons-kubernetes:jar:2.0.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | +- org.kie.kogito:kogito-api:jar:2.0.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | +- org.kie:kie-api:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | +- com.fasterxml.jackson.core:jackson-annotations:jar:2.14.1:compile[2023-01-06T22:06:19.695Z] [INFO] | | +- io.cloudevents:cloudevents-api:jar:2.3.0:compile[2023-01-06T22:06:19.695Z] [INFO] | | +- org.drools:drools-ruleunits-api:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | | \- org.kie:kie-internal:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | +- org.kie:kie-dmn-api:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | | +- org.kie:efesto-common-api:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | | | +- org.kie:kie-memory-compiler:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | | | | \- org.drools:drools-util:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | | | \- org.assertj:assertj-core:jar:3.22.0:test[2023-01-06T22:06:19.695Z] [INFO] | | | \- org.kie:kie-dmn-model:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | +- org.kie.kogito:kogito-timer:jar:2.0.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | \- org.kie.kogito:kogito-pmml-api-dependencies:pom:2.0.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | +- org.kie:kie-pmml-evaluator-api:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | | \- org.kie:kie-pmml-api:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | | +- org.kie:efesto-runtime-manager-api:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | | +- org.apache.commons:commons-math3:jar:3.4.1:compile[2023-01-06T22:06:19.695Z] [INFO] | | | \- org.kie:efesto-compilation-manager-api:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | \- org.kie:kie-pmml-commons:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | +- org.kie:efesto-compilation-manager-core:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | | \- org.kie:efesto-common-core:jar:8.33.0-SNAPSHOT:compile[2023-01-06T22:06:19.695Z] [INFO] | | \- org.apache.commons:commons-text:jar:1.10.0:compile[2023-01-06T22:06:19.695Z] [INFO] | \- com.fasterxml.jackson.core:jackson-databind:jar:2.14.1:compile[2023-01-06T22:06:19.695Z] [INFO] \- io.quarkus:quarkus-junit5-internal:jar:2.15.0.Final:test[2023-01-06T22:06:19.695Z] [INFO] +- org.jboss.shrinkwrap:shrinkwrap-depchain:pom:1.2.6:test[2023-01-06T22:06:19.695Z] [INFO] | +- org.jboss.shrinkwrap:shrinkwrap-api:jar:1.2.6:test[2023-01-06T22:06:19.695Z] [INFO] | +- org.jboss.shrinkwrap:shrinkwrap-spi:jar:1.2.6:test[2023-01-06T22:06:19.695Z] [INFO] | \- org.jboss.shrinkwrap:shrinkwrap-impl-base:jar:1.2.6:test[2023-01-06T22:06:19.695Z] [INFO] +- org.junit.jupiter:junit-jupiter-api:jar:5.9.1:test[2023-01-06T22:06:19.695Z] [INFO] | +- org.opentest4j:opentest4j:jar:1.2.0:test[2023-01-06T22:06:19.695Z] [INFO] | +- org.junit.platform:junit-platform-commons:jar:1.9.1:test[2023-01-06T22:06:19.695Z] [INFO] | \- org.apiguardian:apiguardian-api:jar:1.1.2:compile[2023-01-06T22:06:19.695Z] [INFO] +- org.junit.jupiter:junit-jupiter-params:jar:5.9.1:test[2023-01-06T22:06:19.695Z] [INFO] +- org.junit.jupiter:junit-jupiter-engine:jar:5.9.1:test[2023-01-06T22:06:19.695Z] [INFO] | \- org.junit.platform:junit-platform-engine:jar:1.9.1:test[2023-01-06T22:06:19.695Z] [INFO] +- io.quarkus:quarkus-core:jar:2.15.0.Final:compile[2023-01-06T22:06:19.695Z] [INFO] | +- jakarta.inject:jakarta.inject-api:jar:1.0:compile[2023-01-06T22:06:19.695Z] [INFO] | +- io.quarkus:quarkus-ide-launcher:jar:2.15.0.Final:compile[2023-01-06T22:06:19.695Z] [INFO] | +- io.smallrye.config:smallrye-config:jar:2.13.0:compile[2023-01-06T22:06:19.695Z] [INFO] | | \- io.smallrye.config:smallrye-config-core:jar:2.13.0:compile[2023-01-06T22:06:19.695Z] [INFO] | | \- io.smallrye.common:smallrye-common-expression:jar:1.13.2:compile[2023-01-06T22:06:19.695Z] [INFO] | | \- io.smallrye.common:smallrye-common-function:jar:1.13.2:compile[2023-01-06T22:06:19.695Z] [INFO] | +- org.jboss.logging:jboss-logging:jar:3.5.0.Final:compile[2023-01-06T22:06:19.695Z] [INFO] | +- org.jboss.logmanager:jboss-logmanager-embedded:jar:1.0.11:compile[2023-01-06T22:06:19.696Z] [INFO] | +- org.jboss.logging:jboss-logging-annotations:jar:2.2.1.Final:compile[2023-01-06T22:06:19.696Z] [INFO] | +- org.jboss.threads:jboss-threads:jar:3.4.3.Final:compile[2023-01-06T22:06:19.696Z] [INFO] | +- org.slf4j:slf4j-api:jar:1.7.36:compile[2023-01-06T22:06:19.696Z] [INFO] | +- org.jboss.slf4j:slf4j-jboss-logmanager:jar:1.2.0.Final:compile[2023-01-06T22:06:19.696Z] [INFO] | +- io.quarkus:quarkus-bootstrap-runner:jar:2.15.0.Final:compile[2023-01-06T22:06:19.696Z] [INFO] | | \- io.github.crac:org-crac:jar:0.1.3:compile[2023-01-06T22:06:19.696Z] [INFO] | \- io.quarkus:quarkus-fs-util:jar:0.0.9:compile[2023-01-06T22:06:19.696Z] [INFO] +- io.quarkus:quarkus-test-common:jar:2.15.0.Final:test[2023-01-06T22:06:19.696Z] [INFO] | +- io.quarkus:quarkus-bootstrap-maven-resolver:jar:2.15.0.Final:test[2023-01-06T22:06:19.696Z] [INFO] | | +- org.apache.maven:maven-embedder:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven:maven-settings:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven:maven-core:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | | +- org.apache.maven:maven-artifact:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | | \- org.codehaus.plexus:plexus-component-annotations:jar:1.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven:maven-plugin-api:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven:maven-model:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven:maven-model-builder:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven:maven-builder-support:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-api:jar:1.6.3:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-util:jar:1.6.3:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven.shared:maven-shared-utils:jar:3.3.4:test[2023-01-06T22:06:19.696Z] [INFO] | | | | \- commons-io:commons-io:jar:2.11.0:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- com.google.inject:guice:jar:no_aop:4.2.2:test[2023-01-06T22:06:19.696Z] [INFO] | | | | \- com.google.guava:guava:jar:31.1-jre:test[2023-01-06T22:06:19.696Z] [INFO] | | | | \- com.google.guava:failureaccess:jar:1.0.1:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.codehaus.plexus:plexus-classworlds:jar:2.6.0:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.codehaus.plexus:plexus-sec-dispatcher:jar:2.0:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.codehaus.plexus:plexus-cipher:jar:2.0:test[2023-01-06T22:06:19.696Z] [INFO] | | | \- commons-cli:commons-cli:jar:1.4:test[2023-01-06T22:06:19.696Z] [INFO] | | +- org.eclipse.sisu:org.eclipse.sisu.plexus:jar:0.3.5:test[2023-01-06T22:06:19.696Z] [INFO] | | +- org.apache.maven:maven-settings-builder:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | \- org.codehaus.plexus:plexus-interpolation:jar:1.26:test[2023-01-06T22:06:19.696Z] [INFO] | | +- org.apache.maven:maven-resolver-provider:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven:maven-repository-metadata:jar:3.8.6:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-spi:jar:1.6.3:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-impl:jar:1.6.3:test[2023-01-06T22:06:19.696Z] [INFO] | | | \- org.codehaus.plexus:plexus-utils:jar:3.3.0:test[2023-01-06T22:06:19.696Z] [INFO] | | +- org.apache.maven.resolver:maven-resolver-connector-basic:jar:1.6.3:test[2023-01-06T22:06:19.696Z] [INFO] | | +- org.apache.maven.resolver:maven-resolver-transport-wagon:jar:1.6.3:test[2023-01-06T22:06:19.696Z] [INFO] | | +- org.apache.maven.wagon:wagon-http:jar:3.5.1:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.maven.wagon:wagon-http-shared:jar:3.5.1:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.httpcomponents:httpclient:jar:4.5.13:test[2023-01-06T22:06:19.696Z] [INFO] | | | | \- commons-codec:commons-codec:jar:1.15:test[2023-01-06T22:06:19.696Z] [INFO] | | | +- org.apache.httpcomponents:httpcore:jar:4.4.15:test[2023-01-06T22:06:19.696Z] [INFO] | | | \- org.apache.maven.wagon:wagon-provider-api:jar:3.5.1:test[2023-01-06T22:06:19.696Z] [INFO] | | \- org.apache.maven.wagon:wagon-file:jar:3.5.1:test[2023-01-06T22:06:19.696Z] [INFO] | +- io.quarkus:quarkus-bootstrap-gradle-resolver:jar:2.15.0.Final:test[2023-01-06T22:06:19.696Z] [INFO] | \- org.jboss.logging:commons-logging-jboss-logging:jar:1.0.0.Final:runtime[2023-01-06T22:06:19.696Z] [INFO] +- jakarta.enterprise:jakarta.enterprise.cdi-api:jar:2.0.2:compile[2023-01-06T22:06:19.696Z] [INFO] | +- jakarta.el:jakarta.el-api:jar:3.0.3:compile[2023-01-06T22:06:19.696Z] [INFO] | \- jakarta.interceptor:jakarta.interceptor-api:jar:1.2.5:compile[2023-01-06T22:06:19.696Z] [INFO] +- io.quarkus:quarkus-devtools-utilities:jar:2.15.0.Final:compile[2023-01-06T22:06:19.696Z] [INFO] \- org.awaitility:awaitility:jar:4.2.0:test[2023-01-06T22:06:19.696Z] [INFO] \- org.hamcrest:hamcrest:jar:2.1:test[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] ----< org.kie.kogito:kogito-quarkus-serverless-workflow-deployment >----[2023-01-06T22:06:19.696Z] [INFO] Building Kogito :: Quarkus Workflows Extension :: Deployment 2.0.0-SNAPSHOT [308/314][2023-01-06T22:06:19.696Z] [INFO] --------------------------------[ jar ]---------------------------------[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- maven-clean-plugin:3.1.0:clean (default-clean) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- maven-enforcer-plugin:3.0.0-M2:enforce (enforce-versions) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- maven-enforcer-plugin:3.0.0-M2:enforce (enforce-java-version) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- maven-enforcer-plugin:3.0.0-M2:enforce (enforce-maven-version) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- formatter-maven-plugin:2.13.0:validate (default) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] Using 'UTF-8' encoding to format source files.[2023-01-06T22:06:19.696Z] [INFO] Number of files to be formatted: 17[2023-01-06T22:06:19.696Z] [INFO] Successfully formatted: 0 file(s)[2023-01-06T22:06:19.696Z] [INFO] Fail to format: 0 file(s)[2023-01-06T22:06:19.696Z] [INFO] Skipped: 17 file(s)[2023-01-06T22:06:19.696Z] [INFO] Read only skipped: 0 file(s)[2023-01-06T22:06:19.696Z] [INFO] Approximate time taken: 0s[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- maven-checkstyle-plugin:3.1.1:check (default) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] You have 0 Checkstyle violations.[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- buildnumber-maven-plugin:1.4:create (get-scm-revision) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- jacoco-maven-plugin:0.8.5:prepare-agent (jacoco-prepare-agent) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] jacoco.agent.argLine set to -javaagent:/home/jenkins/.m2/repository/org/jacoco/org.jacoco.agent/0.8.5/org.jacoco.agent-0.8.5-runtime.jar=destfile=/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito_runtimes/kogito-runtimes/target/jacoco.exec,append=true,excludes=*Lexer:org.kie.kogito.codegen.data.*[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- impsort-maven-plugin:1.5.0:check (default) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] Using compiler compliance level: JAVA_11[2023-01-06T22:06:19.696Z] [INFO] Total Files Processed: 17 in 00:00.046[2023-01-06T22:06:19.696Z] [INFO] Already Sorted: 17[2023-01-06T22:06:19.696Z] [INFO] Needed Sorting: 0[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- maven-resources-plugin:3.1.0:resources (default-resources) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] Using 'UTF-8' encoding to copy filtered resources.[2023-01-06T22:06:19.696Z] [INFO] Copying 6 resources[2023-01-06T22:06:19.696Z] [INFO] [2023-01-06T22:06:19.696Z] [INFO] --- maven-compiler-plugin:3.8.1:compile (default-compile) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-06T22:06:19.696Z] [INFO] Changes detected - recompiling the module![2023-01-06T22:06:19.697Z] [INFO] Compiling 17 source files to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito_runtimes/kogito-runtimes/quarkus/extensions/kogito-quarkus-serverless-workflow-extension/kogito-quarkus-serverless-workflow-deployment/target/classes[2023-01-06T22:06:19.697Z] ::error title= [kiegroup/kogito-runtimes] [Execution ERROR]::mvn -f kogito-runtimes/pom.xml clean install -Dvalidate-formatting -s /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes@tmp/config5955749182657767749tmp -Dmaven.wagon.http.ssl.insecure=true -Dmaven.test.failure.ignore=true -nsu -ntp -fae -e -Dhttp.keepAlive=false -Dmaven.wagon.http.pool=false -Dmaven.wagon.httpconnectionManager.ttlSeconds=120 -Dmaven.wagon.http.retryHandler.count=3 -T 1C -Prun-code-coverage dependency:tree -Dorg.slf4j.simpleLogger.log.org.apache.maven.cli.transfer.Slf4jMavenTransferListener=warn -B[2023-01-06T22:06:19.697Z] [INFO] Archive artifact won't be executed[2023-01-06T22:06:19.697Z] ::group::[Pull Request Flow] Execution Summary...[2023-01-06T22:06:19.697Z] [INFO] ----------------------------------------------[2023-01-06T22:06:19.697Z] [INFO] [kiegroup/drools]. Execution Result: ok. Time: 2m 38.5s (158519 ms)[2023-01-06T22:06:19.697Z] [INFO] [kiegroup/kogito-runtimes]. Execution Result: error. Time: 2h 56m 34.4s (10594423 ms)[2023-01-06T22:06:19.697Z] [INFO] ----------------------------------------------[2023-01-06T22:06:19.697Z] ::endgroup::[2023-01-06T22:06:19.701Z] script returned exit code 143[Pipeline] }[Pipeline] // withCredentials[Pipeline] }[2023-01-06T22:06:19.771Z] Deleting 1 temporary files[Pipeline] // configFileProvider[Pipeline] }[Pipeline] // scriptPost stage[Pipeline] junit[2023-01-06T22:06:19.873Z] Recording test results[2023-01-06T22:06:23.967Z] [Checks API] No suitable checks publisher found.[Pipeline] archiveArtifacts[2023-01-06T22:06:23.986Z] Archiving artifacts[Pipeline] script[Pipeline] {[Pipeline] sh[2023-01-06T22:06:25.346Z] + rm -rf console.log[Pipeline] sh[2023-01-06T22:06:25.634Z] + wget --no-check-certificate -qO - https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/1661/consoleText[2023-01-06T22:06:25.634Z] + tail -n 300 Jenkins rerun kogito-runtimes tests @ricardozanini could you please cherry-pick this into 1.13.x and 1.13.x-blue, thanks you can use backport labels to generate backporting PRs automatically ;) (tests) - kogito-runtimes job #1663 was: ABORTED Possible explanation: Most probably a timeout, please review Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/1663/display/redirect Test results: PASSED: 3064 FAILED: 0 Those are the test failures: none See console log: Console Logs [2023-01-09T09:35:59.157Z] [INFO] | | | +- org.rnorth.duct-tape:duct-tape:jar:1.0.8:compile[2023-01-09T09:35:59.157Z] [INFO] | | | | \- org.jetbrains:annotations:jar:17.0.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | | \- com.github.docker-java:docker-java-transport-zerodep:jar:3.2.13:compile[2023-01-09T09:35:59.157Z] [INFO] | | | +- com.github.docker-java:docker-java-transport:jar:3.2.13:compile[2023-01-09T09:35:59.157Z] [INFO] | | | \- net.java.dev.jna:jna:jar:5.8.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | \- org.bouncycastle:bcpkix-jdk15on:jar:1.70:runtime[2023-01-09T09:35:59.157Z] [INFO] | | +- org.bouncycastle:bcprov-jdk15on:jar:1.70:runtime[2023-01-09T09:35:59.157Z] [INFO] | | \- org.bouncycastle:bcutil-jdk15on:jar:1.70:runtime[2023-01-09T09:35:59.157Z] [INFO] | \- io.quarkus:quarkus-jackson-deployment:jar:2.15.0.Final:compile[2023-01-09T09:35:59.157Z] [INFO] | \- io.quarkus:quarkus-jackson-spi:jar:2.15.0.Final:compile[2023-01-09T09:35:59.157Z] [INFO] +- io.quarkus:quarkus-openshift-client-deployment:jar:2.15.0.Final:compile[2023-01-09T09:35:59.157Z] [INFO] | \- io.quarkus:quarkus-openshift-client:jar:2.15.0.Final:compile[2023-01-09T09:35:59.157Z] [INFO] | \- io.fabric8:openshift-client:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | +- io.fabric8:openshift-client-api:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model-clusterautoscaling:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model-operator:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model-operatorhub:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model-machine:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model-whereabouts:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model-monitoring:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model-storageversionmigrator:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model-tuned:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model-console:jar:6.2.0:compile[2023-01-09T09:35:59.157Z] [INFO] | | +- io.fabric8:openshift-model-config:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:openshift-model-machineconfig:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:openshift-model-miscellaneous:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:openshift-model-hive:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- io.fabric8:openshift-model-installer:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | \- com.github.mifmif:generex:jar:1.0.2:compile[2023-01-09T09:35:59.158Z] [INFO] | \- dk.brics.automaton:automaton:jar:1.11-8:compile[2023-01-09T09:35:59.158Z] [INFO] +- io.quarkus:quarkus-cache-deployment:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | +- io.quarkus:quarkus-cache:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.quarkus:quarkus-caffeine:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- com.github.ben-manes.caffeine:caffeine:jar:3.1.1:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- com.google.errorprone:error_prone_annotations:jar:2.16:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- io.quarkus:quarkus-mutiny:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.smallrye.reactive:mutiny:jar:1.8.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- org.reactivestreams:reactive-streams:jar:1.0.3:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- io.smallrye.common:smallrye-common-annotation:jar:1.13.2:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.quarkus:quarkus-smallrye-context-propagation:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- io.smallrye:smallrye-context-propagation:jar:1.2.2:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.smallrye:smallrye-context-propagation-api:jar:1.2.2:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- io.smallrye:smallrye-context-propagation-storage:jar:1.2.2:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- io.smallrye.reactive:mutiny-smallrye-context-propagation:jar:1.8.0:compile[2023-01-09T09:35:59.158Z] [INFO] | +- io.quarkus:quarkus-arc-deployment:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.quarkus:quarkus-smallrye-context-propagation-spi:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.quarkus:quarkus-vertx-http-dev-console-spi:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.quarkus:quarkus-vertx-http-dev-console-runtime-spi:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- io.vertx:vertx-web:jar:4.3.5:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.vertx:vertx-web-common:jar:4.3.5:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.vertx:vertx-auth-common:jar:4.3.5:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.vertx:vertx-bridge-common:jar:4.3.5:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- io.vertx:vertx-core:jar:4.3.5:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.netty:netty-common:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.netty:netty-buffer:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.netty:netty-transport:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.netty:netty-handler:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | | +- io.netty:netty-transport-native-unix-common:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | | \- io.netty:netty-codec:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.netty:netty-handler-proxy:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | | \- io.netty:netty-codec-socks:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.netty:netty-codec-http:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.netty:netty-codec-http2:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- io.netty:netty-resolver:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- io.netty:netty-resolver-dns:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- io.netty:netty-codec-dns:jar:4.1.85.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- io.quarkus.arc:arc-processor:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | +- io.quarkus:quarkus-caffeine-deployment:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | \- io.quarkus:quarkus-mutiny-deployment:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | \- io.quarkus:quarkus-smallrye-context-propagation-deployment:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] +- org.kie.kogito:kogito-addons-quarkus-common-deployment:jar:2.0.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | \- io.quarkus:quarkus-core-deployment:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | +- org.aesh:readline:jar:2.2:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- org.fusesource.jansi:jansi:jar:1.18:compile[2023-01-09T09:35:59.158Z] [INFO] | +- org.aesh:aesh:jar:2.6:compile[2023-01-09T09:35:59.158Z] [INFO] | +- org.apache.commons:commons-lang3:jar:3.12.0:compile[2023-01-09T09:35:59.158Z] [INFO] | +- org.wildfly.common:wildfly-common:jar:1.5.4.Final-format-001:compile[2023-01-09T09:35:59.158Z] [INFO] | +- io.quarkus.gizmo:gizmo:jar:1.4.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- org.ow2.asm:asm-util:jar:9.4:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- org.ow2.asm:asm-analysis:jar:9.4:compile[2023-01-09T09:35:59.158Z] [INFO] | +- io.smallrye:jandex:jar:3.0.5:compile[2023-01-09T09:35:59.158Z] [INFO] | +- org.ow2.asm:asm:jar:9.4:compile[2023-01-09T09:35:59.158Z] [INFO] | +- org.ow2.asm:asm-commons:jar:9.4:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- org.ow2.asm:asm-tree:jar:9.4:compile[2023-01-09T09:35:59.158Z] [INFO] | +- io.quarkus:quarkus-development-mode-spi:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | +- io.quarkus:quarkus-class-change-agent:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | +- io.quarkus:quarkus-bootstrap-core:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.quarkus:quarkus-bootstrap-app-model:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- io.smallrye.common:smallrye-common-io:jar:1.13.2:compile[2023-01-09T09:35:59.158Z] [INFO] | +- org.eclipse.sisu:org.eclipse.sisu.inject:jar:0.3.5:runtime[2023-01-09T09:35:59.158Z] [INFO] | +- io.quarkus:quarkus-builder:jar:2.15.0.Final:compile[2023-01-09T09:35:59.158Z] [INFO] | +- org.graalvm.sdk:graal-sdk:jar:22.3.0:compile[2023-01-09T09:35:59.158Z] [INFO] | +- org.junit.platform:junit-platform-launcher:jar:1.9.1:compile[2023-01-09T09:35:59.158Z] [INFO] | \- org.junit.jupiter:junit-jupiter:jar:5.9.1:test[2023-01-09T09:35:59.158Z] [INFO] +- org.kie.kogito:kogito-addons-quarkus-kubernetes:jar:2.0.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | +- io.fabric8:knative-client:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:knative-model:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- io.fabric8:kubernetes-model-core:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- io.fabric8:kubernetes-model-common:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- io.fabric8:kubernetes-client-api:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-gatewayapi:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-rbac:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-admissionregistration:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-apps:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-autoscaling:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-apiextensions:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-batch:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-certificates:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-coordination:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-discovery:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-events:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-extensions:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-flowcontrol:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-networking:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-metrics:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-policy:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-scheduling:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-storageclass:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-model-node:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- com.fasterxml.jackson.dataformat:jackson-dataformat-yaml:jar:2.14.1:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- com.fasterxml.jackson.core:jackson-core:jar:2.14.1:compile[2023-01-09T09:35:59.158Z] [INFO] | +- io.fabric8:kubernetes-client:jar:6.2.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.fabric8:kubernetes-httpclient-okhttp:jar:6.2.0:runtime[2023-01-09T09:35:59.158Z] [INFO] | | | +- com.squareup.okhttp3:okhttp:jar:3.14.9:runtime[2023-01-09T09:35:59.158Z] [INFO] | | | | \- com.squareup.okio:okio:jar:1.17.2:runtime[2023-01-09T09:35:59.158Z] [INFO] | | | \- com.squareup.okhttp3:logging-interceptor:jar:3.14.9:runtime[2023-01-09T09:35:59.158Z] [INFO] | | \- io.fabric8:zjsonpatch:jar:0.3.0:compile[2023-01-09T09:35:59.158Z] [INFO] | \- org.kie.kogito:kogito-addons-kubernetes:jar:2.0.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | +- org.kie.kogito:kogito-api:jar:2.0.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- org.kie:kie-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- com.fasterxml.jackson.core:jackson-annotations:jar:2.14.1:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- io.cloudevents:cloudevents-api:jar:2.3.0:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- org.drools:drools-ruleunits-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- org.kie:kie-internal:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- org.kie:kie-dmn-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- org.kie:efesto-common-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | | | +- org.kie:kie-memory-compiler:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | | | | \- org.drools:drools-util:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | | | \- org.assertj:assertj-core:jar:3.22.0:test[2023-01-09T09:35:59.158Z] [INFO] | | | \- org.kie:kie-dmn-model:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- org.kie.kogito:kogito-timer:jar:2.0.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- org.kie.kogito:kogito-pmml-api-dependencies:pom:2.0.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- org.kie:kie-pmml-evaluator-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- org.kie:kie-pmml-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- org.kie:efesto-runtime-manager-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | | +- org.apache.commons:commons-math3:jar:3.4.1:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- org.kie:efesto-compilation-manager-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- org.kie:kie-pmml-commons:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | +- org.kie:efesto-compilation-manager-core:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | | \- org.kie:efesto-common-core:jar:8.33.0-SNAPSHOT:compile[2023-01-09T09:35:59.158Z] [INFO] | | \- org.apache.commons:commons-text:jar:1.10.0:compile[2023-01-09T09:35:59.158Z] [INFO] | \- com.fasterxml.jackson.core:jackson-databind:jar:2.14.1:compile[2023-01-09T09:35:59.158Z] [INFO] \- io.quarkus:quarkus-junit5-internal:jar:2.15.0.Final:test[2023-01-09T09:35:59.158Z] [INFO] +- org.jboss.shrinkwrap:shrinkwrap-depchain:pom:1.2.6:test[2023-01-09T09:35:59.158Z] [INFO] | +- org.jboss.shrinkwrap:shrinkwrap-api:jar:1.2.6:test[2023-01-09T09:35:59.158Z] [INFO] | +- org.jboss.shrinkwrap:shrinkwrap-spi:jar:1.2.6:test[2023-01-09T09:35:59.158Z] [INFO] | \- org.jboss.shrinkwrap:shrinkwrap-impl-base:jar:1.2.6:test[2023-01-09T09:35:59.158Z] [INFO] +- org.junit.jupiter:junit-jupiter-api:jar:5.9.1:test[2023-01-09T09:35:59.158Z] [INFO] | +- org.opentest4j:opentest4j:jar:1.2.0:test[2023-01-09T09:35:59.158Z] [INFO] | +- org.junit.platform:junit-platform-commons:jar:1.9.1:test[2023-01-09T09:35:59.159Z] [INFO] | \- org.apiguardian:apiguardian-api:jar:1.1.2:compile[2023-01-09T09:35:59.159Z] [INFO] +- org.junit.jupiter:junit-jupiter-params:jar:5.9.1:test[2023-01-09T09:35:59.159Z] [INFO] +- org.junit.jupiter:junit-jupiter-engine:jar:5.9.1:test[2023-01-09T09:35:59.159Z] [INFO] | \- org.junit.platform:junit-platform-engine:jar:1.9.1:test[2023-01-09T09:35:59.159Z] [INFO] +- io.quarkus:quarkus-core:jar:2.15.0.Final:compile[2023-01-09T09:35:59.159Z] [INFO] | +- jakarta.inject:jakarta.inject-api:jar:1.0:compile[2023-01-09T09:35:59.159Z] [INFO] | +- io.quarkus:quarkus-ide-launcher:jar:2.15.0.Final:compile[2023-01-09T09:35:59.159Z] [INFO] | +- io.smallrye.config:smallrye-config:jar:2.13.0:compile[2023-01-09T09:35:59.159Z] [INFO] | | \- io.smallrye.config:smallrye-config-core:jar:2.13.0:compile[2023-01-09T09:35:59.159Z] [INFO] | | \- io.smallrye.common:smallrye-common-expression:jar:1.13.2:compile[2023-01-09T09:35:59.159Z] [INFO] | | \- io.smallrye.common:smallrye-common-function:jar:1.13.2:compile[2023-01-09T09:35:59.159Z] [INFO] | +- org.jboss.logging:jboss-logging:jar:3.5.0.Final:compile[2023-01-09T09:35:59.159Z] [INFO] | +- org.jboss.logmanager:jboss-logmanager-embedded:jar:1.0.11:compile[2023-01-09T09:35:59.159Z] [INFO] | +- org.jboss.logging:jboss-logging-annotations:jar:2.2.1.Final:compile[2023-01-09T09:35:59.159Z] [INFO] | +- org.jboss.threads:jboss-threads:jar:3.4.3.Final:compile[2023-01-09T09:35:59.159Z] [INFO] | +- org.slf4j:slf4j-api:jar:1.7.36:compile[2023-01-09T09:35:59.159Z] [INFO] | +- org.jboss.slf4j:slf4j-jboss-logmanager:jar:1.2.0.Final:compile[2023-01-09T09:35:59.159Z] [INFO] | +- io.quarkus:quarkus-bootstrap-runner:jar:2.15.0.Final:compile[2023-01-09T09:35:59.159Z] [INFO] | | \- io.github.crac:org-crac:jar:0.1.3:compile[2023-01-09T09:35:59.159Z] [INFO] | \- io.quarkus:quarkus-fs-util:jar:0.0.9:compile[2023-01-09T09:35:59.159Z] [INFO] +- io.quarkus:quarkus-test-common:jar:2.15.0.Final:test[2023-01-09T09:35:59.159Z] [INFO] | +- io.quarkus:quarkus-bootstrap-maven-resolver:jar:2.15.0.Final:test[2023-01-09T09:35:59.159Z] [INFO] | | +- org.apache.maven:maven-embedder:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven:maven-settings:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven:maven-core:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | | +- org.apache.maven:maven-artifact:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | | \- org.codehaus.plexus:plexus-component-annotations:jar:1.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven:maven-plugin-api:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven:maven-model:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven:maven-model-builder:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven:maven-builder-support:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-api:jar:1.6.3:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-util:jar:1.6.3:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven.shared:maven-shared-utils:jar:3.3.4:test[2023-01-09T09:35:59.159Z] [INFO] | | | | \- commons-io:commons-io:jar:2.11.0:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- com.google.inject:guice:jar:no_aop:4.2.2:test[2023-01-09T09:35:59.159Z] [INFO] | | | | \- com.google.guava:guava:jar:31.1-jre:test[2023-01-09T09:35:59.159Z] [INFO] | | | | \- com.google.guava:failureaccess:jar:1.0.1:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.codehaus.plexus:plexus-classworlds:jar:2.6.0:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.codehaus.plexus:plexus-sec-dispatcher:jar:2.0:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.codehaus.plexus:plexus-cipher:jar:2.0:test[2023-01-09T09:35:59.159Z] [INFO] | | | \- commons-cli:commons-cli:jar:1.4:test[2023-01-09T09:35:59.159Z] [INFO] | | +- org.eclipse.sisu:org.eclipse.sisu.plexus:jar:0.3.5:test[2023-01-09T09:35:59.159Z] [INFO] | | +- org.apache.maven:maven-settings-builder:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | \- org.codehaus.plexus:plexus-interpolation:jar:1.26:test[2023-01-09T09:35:59.159Z] [INFO] | | +- org.apache.maven:maven-resolver-provider:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven:maven-repository-metadata:jar:3.8.6:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-spi:jar:1.6.3:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-impl:jar:1.6.3:test[2023-01-09T09:35:59.159Z] [INFO] | | | \- org.codehaus.plexus:plexus-utils:jar:3.3.0:test[2023-01-09T09:35:59.159Z] [INFO] | | +- org.apache.maven.resolver:maven-resolver-connector-basic:jar:1.6.3:test[2023-01-09T09:35:59.159Z] [INFO] | | +- org.apache.maven.resolver:maven-resolver-transport-wagon:jar:1.6.3:test[2023-01-09T09:35:59.159Z] [INFO] | | +- org.apache.maven.wagon:wagon-http:jar:3.5.1:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.maven.wagon:wagon-http-shared:jar:3.5.1:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.httpcomponents:httpclient:jar:4.5.13:test[2023-01-09T09:35:59.159Z] [INFO] | | | | \- commons-codec:commons-codec:jar:1.15:test[2023-01-09T09:35:59.159Z] [INFO] | | | +- org.apache.httpcomponents:httpcore:jar:4.4.15:test[2023-01-09T09:35:59.159Z] [INFO] | | | \- org.apache.maven.wagon:wagon-provider-api:jar:3.5.1:test[2023-01-09T09:35:59.159Z] [INFO] | | \- org.apache.maven.wagon:wagon-file:jar:3.5.1:test[2023-01-09T09:35:59.159Z] [INFO] | +- io.quarkus:quarkus-bootstrap-gradle-resolver:jar:2.15.0.Final:test[2023-01-09T09:35:59.159Z] [INFO] | \- org.jboss.logging:commons-logging-jboss-logging:jar:1.0.0.Final:runtime[2023-01-09T09:35:59.159Z] [INFO] +- jakarta.enterprise:jakarta.enterprise.cdi-api:jar:2.0.2:compile[2023-01-09T09:35:59.159Z] [INFO] | +- jakarta.el:jakarta.el-api:jar:3.0.3:compile[2023-01-09T09:35:59.159Z] [INFO] | \- jakarta.interceptor:jakarta.interceptor-api:jar:1.2.5:compile[2023-01-09T09:35:59.159Z] [INFO] +- io.quarkus:quarkus-devtools-utilities:jar:2.15.0.Final:compile[2023-01-09T09:35:59.159Z] [INFO] \- org.awaitility:awaitility:jar:4.2.0:test[2023-01-09T09:35:59.159Z] [INFO] \- org.hamcrest:hamcrest:jar:2.1:test[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] ----< org.kie.kogito:kogito-quarkus-serverless-workflow-deployment >----[2023-01-09T09:35:59.159Z] [INFO] Building Kogito :: Quarkus Workflows Extension :: Deployment 2.0.0-SNAPSHOT [308/314][2023-01-09T09:35:59.159Z] [INFO] --------------------------------[ jar ]---------------------------------[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- maven-clean-plugin:3.1.0:clean (default-clean) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- maven-enforcer-plugin:3.0.0-M2:enforce (enforce-versions) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- maven-enforcer-plugin:3.0.0-M2:enforce (enforce-java-version) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- maven-enforcer-plugin:3.0.0-M2:enforce (enforce-maven-version) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- formatter-maven-plugin:2.13.0:validate (default) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] Using 'UTF-8' encoding to format source files.[2023-01-09T09:35:59.159Z] [INFO] Number of files to be formatted: 17[2023-01-09T09:35:59.159Z] [INFO] Successfully formatted: 0 file(s)[2023-01-09T09:35:59.159Z] [INFO] Fail to format: 0 file(s)[2023-01-09T09:35:59.159Z] [INFO] Skipped: 17 file(s)[2023-01-09T09:35:59.159Z] [INFO] Read only skipped: 0 file(s)[2023-01-09T09:35:59.159Z] [INFO] Approximate time taken: 0s[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- maven-checkstyle-plugin:3.1.1:check (default) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] You have 0 Checkstyle violations.[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- buildnumber-maven-plugin:1.4:create (get-scm-revision) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- jacoco-maven-plugin:0.8.5:prepare-agent (jacoco-prepare-agent) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] jacoco.agent.argLine set to -javaagent:/home/jenkins/.m2/repository/org/jacoco/org.jacoco.agent/0.8.5/org.jacoco.agent-0.8.5-runtime.jar=destfile=/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito_runtimes/kogito-runtimes/target/jacoco.exec,append=true,excludes=*Lexer:org.kie.kogito.codegen.data.*[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- impsort-maven-plugin:1.5.0:check (default) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] Using compiler compliance level: JAVA_11[2023-01-09T09:35:59.159Z] [INFO] Total Files Processed: 17 in 00:00.043[2023-01-09T09:35:59.159Z] [INFO] Already Sorted: 17[2023-01-09T09:35:59.159Z] [INFO] Needed Sorting: 0[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- maven-resources-plugin:3.1.0:resources (default-resources) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] Using 'UTF-8' encoding to copy filtered resources.[2023-01-09T09:35:59.159Z] [INFO] Copying 6 resources[2023-01-09T09:35:59.159Z] [INFO] [2023-01-09T09:35:59.159Z] [INFO] --- maven-compiler-plugin:3.8.1:compile (default-compile) @ kogito-quarkus-serverless-workflow-deployment ---[2023-01-09T09:35:59.159Z] [INFO] Changes detected - recompiling the module![2023-01-09T09:35:59.159Z] [INFO] Compiling 17 source files to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito_runtimes/kogito-runtimes/quarkus/extensions/kogito-quarkus-serverless-workflow-extension/kogito-quarkus-serverless-workflow-deployment/target/classes[2023-01-09T09:35:59.159Z] ::error title= [kiegroup/kogito-runtimes] [Execution ERROR]::mvn -f kogito-runtimes/pom.xml clean install -Dvalidate-formatting -s /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes@tmp/config11957976266486756553tmp -Dmaven.wagon.http.ssl.insecure=true -Dmaven.test.failure.ignore=true -nsu -ntp -fae -e -Dhttp.keepAlive=false -Dmaven.wagon.http.pool=false -Dmaven.wagon.httpconnectionManager.ttlSeconds=120 -Dmaven.wagon.http.retryHandler.count=3 -T 1C -Prun-code-coverage dependency:tree -Dorg.slf4j.simpleLogger.log.org.apache.maven.cli.transfer.Slf4jMavenTransferListener=warn -B[2023-01-09T09:35:59.159Z] [INFO] Archive artifact won't be executed[2023-01-09T09:35:59.159Z] ::group::[Pull Request Flow] Execution Summary...[2023-01-09T09:35:59.159Z] [INFO] ----------------------------------------------[2023-01-09T09:35:59.159Z] [INFO] [kiegroup/drools]. Execution Result: ok. Time: 2m 47.2s (167287 ms)[2023-01-09T09:35:59.159Z] [INFO] [kiegroup/kogito-runtimes]. Execution Result: error. Time: 2h 56m 26.8s (10586850 ms)[2023-01-09T09:35:59.159Z] [INFO] ----------------------------------------------[2023-01-09T09:35:59.159Z] ::endgroup::[2023-01-09T09:35:59.163Z] script returned exit code 143[Pipeline] }[Pipeline] // withCredentials[Pipeline] }[2023-01-09T09:35:59.229Z] Deleting 1 temporary files[Pipeline] // configFileProvider[Pipeline] }[Pipeline] // scriptPost stage[Pipeline] junit[2023-01-09T09:35:59.328Z] Recording test results[2023-01-09T09:36:04.746Z] [Checks API] No suitable checks publisher found.[Pipeline] archiveArtifacts[2023-01-09T09:36:04.763Z] Archiving artifacts[Pipeline] script[Pipeline] {[Pipeline] sh[2023-01-09T09:36:06.197Z] + rm -rf console.log[Pipeline] sh[2023-01-09T09:36:06.488Z] + wget --no-check-certificate -qO - https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/1663/consoleText[2023-01-09T09:36:06.488Z] + tail -n 300 Jenkins rerun kogito-runtimes tests (tests) - kogito-runtimes job #1664 was: ABORTED Possible explanation: Most probably a timeout, please review Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/1664/display/redirect Test results: PASSED: 3064 FAILED: 0 Those are the test failures: none See console log: Console Logs [2023-01-09T15:16:06.278Z] [INFO] --- maven-failsafe-plugin:3.0.0-M5:integration-test (integration-test) @ kogito-addons-quarkus-kubernetes-deployment ---[2023-01-09T15:16:06.278Z] [INFO] [2023-01-09T15:16:06.278Z] [INFO] --- maven-failsafe-plugin:3.0.0-M5:verify (integration-test) @ kogito-addons-quarkus-kubernetes-deployment ---[2023-01-09T15:16:06.278Z] [INFO] Failsafe report directory: /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito_runtimes/kogito-runtimes/quarkus/addons/kubernetes/deployment/target/failsafe-reports[2023-01-09T15:16:06.278Z] [INFO] [2023-01-09T15:16:06.278Z] [INFO] --- maven-install-plugin:2.5.2:install (default-install) @ kogito-addons-quarkus-kubernetes-deployment ---[2023-01-09T15:16:06.278Z] [INFO] Installing /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito_runtimes/kogito-runtimes/quarkus/addons/kubernetes/deployment/target/kogito-addons-quarkus-kubernetes-deployment-2.0.0-SNAPSHOT.jar to /home/jenkins/.m2/repository/org/kie/kogito/kogito-addons-quarkus-kubernetes-deployment/2.0.0-SNAPSHOT/kogito-addons-quarkus-kubernetes-deployment-2.0.0-SNAPSHOT.jar[2023-01-09T15:16:06.278Z] [INFO] Installing /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito_runtimes/kogito-runtimes/quarkus/addons/kubernetes/deployment/pom.xml to /home/jenkins/.m2/repository/org/kie/kogito/kogito-addons-quarkus-kubernetes-deployment/2.0.0-SNAPSHOT/kogito-addons-quarkus-kubernetes-deployment-2.0.0-SNAPSHOT.pom[2023-01-09T15:16:06.278Z] [INFO] Installing /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito_runtimes/kogito-runtimes/quarkus/addons/kubernetes/deployment/target/kogito-addons-quarkus-kubernetes-deployment-2.0.0-SNAPSHOT-tests.jar to /home/jenkins/.m2/repository/org/kie/kogito/kogito-addons-quarkus-kubernetes-deployment/2.0.0-SNAPSHOT/kogito-addons-quarkus-kubernetes-deployment-2.0.0-SNAPSHOT-tests.jar[2023-01-09T15:16:06.278Z] [INFO] Installing /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito_runtimes/kogito-runtimes/quarkus/addons/kubernetes/deployment/target/kogito-addons-quarkus-kubernetes-deployment-2.0.0-SNAPSHOT-sources.jar to /home/jenkins/.m2/repository/org/kie/kogito/kogito-addons-quarkus-kubernetes-deployment/2.0.0-SNAPSHOT/kogito-addons-quarkus-kubernetes-deployment-2.0.0-SNAPSHOT-sources.jar[2023-01-09T15:16:06.278Z] [INFO] Installing /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito_runtimes/kogito-runtimes/quarkus/addons/kubernetes/deployment/target/kogito-addons-quarkus-kubernetes-deployment-2.0.0-SNAPSHOT-test-sources.jar to /home/jenkins/.m2/repository/org/kie/kogito/kogito-addons-quarkus-kubernetes-deployment/2.0.0-SNAPSHOT/kogito-addons-quarkus-kubernetes-deployment-2.0.0-SNAPSHOT-test-sources.jar[2023-01-09T15:16:06.278Z] [INFO] [2023-01-09T15:16:06.278Z] [INFO] --- maven-dependency-plugin:2.8:tree (default-cli) @ kogito-addons-quarkus-kubernetes-deployment ---[2023-01-09T15:16:06.278Z] [INFO] org.kie.kogito:kogito-addons-quarkus-kubernetes-deployment:jar:2.0.0-SNAPSHOT[2023-01-09T15:16:06.278Z] [INFO] +- io.quarkus:quarkus-kubernetes-client-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | +- io.quarkus:quarkus-kubernetes-client:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.quarkus:quarkus-arc:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.quarkus.arc:arc:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | | \- jakarta.transaction:jakarta.transaction-api:jar:1.3.3:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- org.eclipse.microprofile.context-propagation:microprofile-context-propagation-api:jar:1.2:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.quarkus:quarkus-kubernetes-client-internal:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.quarkus:quarkus-jackson:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- com.fasterxml.jackson.datatype:jackson-datatype-jsr310:jar:2.14.1:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- com.fasterxml.jackson.datatype:jackson-datatype-jdk8:jar:2.14.1:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- com.fasterxml.jackson.module:jackson-module-parameter-names:jar:2.14.1:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- org.apache.commons:commons-compress:jar:1.22:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- jakarta.annotation:jakarta.annotation-api:jar:1.3.5:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- org.jboss.spec.javax.xml.bind:jboss-jaxb-api_2.3_spec:jar:2.0.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | \- io.smallrye.config:smallrye-config-source-yaml:jar:2.13.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- org.yaml:snakeyaml:jar:1.33:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.smallrye.config:smallrye-config-common:jar:2.13.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- org.eclipse.microprofile.config:microprofile-config-api:jar:2.0.1:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- io.smallrye.common:smallrye-common-classloader:jar:1.13.2:compile[2023-01-09T15:16:06.278Z] [INFO] | | \- io.smallrye.common:smallrye-common-constraint:jar:1.13.2:compile[2023-01-09T15:16:06.278Z] [INFO] | +- io.quarkus:quarkus-kubernetes-client-internal-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.quarkus:quarkus-kubernetes-client-spi:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.quarkus:quarkus-kubernetes-spi:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.quarkus:quarkus-devservices-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- io.quarkus:quarkus-devservices-common:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- com.github.docker-java:docker-java-api:jar:3.2.13:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- io.quarkus:quarkus-junit4-mock:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | \- com.dajudge.kindcontainer:kindcontainer:jar:1.3.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- org.testcontainers:testcontainers:jar:1.17.6:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- org.rnorth.duct-tape:duct-tape:jar:1.0.8:compile[2023-01-09T15:16:06.278Z] [INFO] | | | | \- org.jetbrains:annotations:jar:17.0.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- com.github.docker-java:docker-java-transport-zerodep:jar:3.2.13:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- com.github.docker-java:docker-java-transport:jar:3.2.13:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- net.java.dev.jna:jna:jar:5.8.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | \- org.bouncycastle:bcpkix-jdk15on:jar:1.70:runtime[2023-01-09T15:16:06.278Z] [INFO] | | +- org.bouncycastle:bcprov-jdk15on:jar:1.70:runtime[2023-01-09T15:16:06.278Z] [INFO] | | \- org.bouncycastle:bcutil-jdk15on:jar:1.70:runtime[2023-01-09T15:16:06.278Z] [INFO] | \- io.quarkus:quarkus-jackson-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | \- io.quarkus:quarkus-jackson-spi:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] +- io.quarkus:quarkus-openshift-client-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | \- io.quarkus:quarkus-openshift-client:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | \- io.fabric8:openshift-client:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | +- io.fabric8:openshift-client-api:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-clusterautoscaling:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-operator:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-operatorhub:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-machine:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-whereabouts:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-monitoring:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-storageversionmigrator:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-tuned:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-console:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-config:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-machineconfig:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-miscellaneous:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.fabric8:openshift-model-hive:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | \- io.fabric8:openshift-model-installer:jar:6.2.0:compile[2023-01-09T15:16:06.278Z] [INFO] | \- com.github.mifmif:generex:jar:1.0.2:compile[2023-01-09T15:16:06.278Z] [INFO] | \- dk.brics.automaton:automaton:jar:1.11-8:compile[2023-01-09T15:16:06.278Z] [INFO] +- io.quarkus:quarkus-cache-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | +- io.quarkus:quarkus-cache:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.quarkus:quarkus-caffeine:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- com.github.ben-manes.caffeine:caffeine:jar:3.1.1:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- com.google.errorprone:error_prone_annotations:jar:2.16:compile[2023-01-09T15:16:06.278Z] [INFO] | | \- io.quarkus:quarkus-mutiny:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.smallrye.reactive:mutiny:jar:1.8.0:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- org.reactivestreams:reactive-streams:jar:1.0.3:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- io.smallrye.common:smallrye-common-annotation:jar:1.13.2:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.quarkus:quarkus-smallrye-context-propagation:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- io.smallrye:smallrye-context-propagation:jar:1.2.2:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.smallrye:smallrye-context-propagation-api:jar:1.2.2:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- io.smallrye:smallrye-context-propagation-storage:jar:1.2.2:compile[2023-01-09T15:16:06.278Z] [INFO] | | \- io.smallrye.reactive:mutiny-smallrye-context-propagation:jar:1.8.0:compile[2023-01-09T15:16:06.278Z] [INFO] | +- io.quarkus:quarkus-arc-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.quarkus:quarkus-smallrye-context-propagation-spi:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | +- io.quarkus:quarkus-vertx-http-dev-console-spi:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.quarkus:quarkus-vertx-http-dev-console-runtime-spi:jar:2.15.0.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- io.vertx:vertx-web:jar:4.3.5:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.vertx:vertx-web-common:jar:4.3.5:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.vertx:vertx-auth-common:jar:4.3.5:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.vertx:vertx-bridge-common:jar:4.3.5:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- io.vertx:vertx-core:jar:4.3.5:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.netty:netty-common:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.netty:netty-buffer:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.netty:netty-transport:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.netty:netty-handler:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | | +- io.netty:netty-transport-native-unix-common:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | | \- io.netty:netty-codec:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.netty:netty-handler-proxy:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | | \- io.netty:netty-codec-socks:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.netty:netty-codec-http:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.netty:netty-codec-http2:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | +- io.netty:netty-resolver:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- io.netty:netty-resolver-dns:jar:4.1.85.Final:compile[2023-01-09T15:16:06.278Z] [INFO] | | | \- io.netty:netty-codec-dns:jar:4.1.85.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- io.quarkus.arc:arc-processor:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.quarkus:quarkus-caffeine-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | \- io.quarkus:quarkus-mutiny-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | \- io.quarkus:quarkus-smallrye-context-propagation-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] +- org.kie.kogito:kogito-addons-quarkus-common-deployment:jar:2.0.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | \- io.quarkus:quarkus-core-deployment:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.aesh:readline:jar:2.2:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- org.fusesource.jansi:jansi:jar:1.18:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.aesh:aesh:jar:2.6:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.apache.commons:commons-lang3:jar:3.12.0:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.wildfly.common:wildfly-common:jar:1.5.4.Final-format-001:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.quarkus.gizmo:gizmo:jar:1.4.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- org.ow2.asm:asm-util:jar:9.4:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- org.ow2.asm:asm-analysis:jar:9.4:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.smallrye:jandex:jar:3.0.5:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.ow2.asm:asm:jar:9.4:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.ow2.asm:asm-commons:jar:9.4:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- org.ow2.asm:asm-tree:jar:9.4:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.quarkus:quarkus-development-mode-spi:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.quarkus:quarkus-class-change-agent:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.quarkus:quarkus-bootstrap-core:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.quarkus:quarkus-bootstrap-app-model:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- io.smallrye.common:smallrye-common-io:jar:1.13.2:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.eclipse.sisu:org.eclipse.sisu.inject:jar:0.3.5:runtime[2023-01-09T15:16:06.279Z] [INFO] | +- io.quarkus:quarkus-builder:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.graalvm.sdk:graal-sdk:jar:22.3.0:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.junit.platform:junit-platform-launcher:jar:1.9.1:compile[2023-01-09T15:16:06.279Z] [INFO] | \- org.junit.jupiter:junit-jupiter:jar:5.9.1:test[2023-01-09T15:16:06.279Z] [INFO] +- org.kie.kogito:kogito-addons-quarkus-kubernetes:jar:2.0.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.fabric8:knative-client:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:knative-model:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | | \- io.fabric8:kubernetes-model-core:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | | \- io.fabric8:kubernetes-model-common:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- io.fabric8:kubernetes-client-api:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-gatewayapi:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-rbac:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-admissionregistration:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-apps:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-autoscaling:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-apiextensions:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-batch:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-certificates:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-coordination:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-discovery:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-events:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-extensions:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-flowcontrol:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-networking:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-metrics:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-policy:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-scheduling:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-storageclass:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-model-node:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- com.fasterxml.jackson.dataformat:jackson-dataformat-yaml:jar:2.14.1:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- com.fasterxml.jackson.core:jackson-core:jar:2.14.1:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.fabric8:kubernetes-client:jar:6.2.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.fabric8:kubernetes-httpclient-okhttp:jar:6.2.0:runtime[2023-01-09T15:16:06.279Z] [INFO] | | | +- com.squareup.okhttp3:okhttp:jar:3.14.9:runtime[2023-01-09T15:16:06.279Z] [INFO] | | | | \- com.squareup.okio:okio:jar:1.17.2:runtime[2023-01-09T15:16:06.279Z] [INFO] | | | \- com.squareup.okhttp3:logging-interceptor:jar:3.14.9:runtime[2023-01-09T15:16:06.279Z] [INFO] | | \- io.fabric8:zjsonpatch:jar:0.3.0:compile[2023-01-09T15:16:06.279Z] [INFO] | \- org.kie.kogito:kogito-addons-kubernetes:jar:2.0.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.kie.kogito:kogito-api:jar:2.0.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- org.kie:kie-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- com.fasterxml.jackson.core:jackson-annotations:jar:2.14.1:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- io.cloudevents:cloudevents-api:jar:2.3.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- org.drools:drools-ruleunits-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | | \- org.kie:kie-internal:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- org.kie:kie-dmn-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.kie:efesto-common-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | | | +- org.kie:kie-memory-compiler:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | | | | \- org.drools:drools-util:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | | | \- org.assertj:assertj-core:jar:3.22.0:test[2023-01-09T15:16:06.279Z] [INFO] | | | \- org.kie:kie-dmn-model:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- org.kie.kogito:kogito-timer:jar:2.0.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- org.kie.kogito:kogito-pmml-api-dependencies:pom:2.0.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- org.kie:kie-pmml-evaluator-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | | \- org.kie:kie-pmml-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.kie:efesto-runtime-manager-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.apache.commons:commons-math3:jar:3.4.1:compile[2023-01-09T15:16:06.279Z] [INFO] | | | \- org.kie:efesto-compilation-manager-api:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- org.kie:kie-pmml-commons:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | +- org.kie:efesto-compilation-manager-core:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | | \- org.kie:efesto-common-core:jar:8.33.0-SNAPSHOT:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- org.apache.commons:commons-text:jar:1.10.0:compile[2023-01-09T15:16:06.279Z] [INFO] | \- com.fasterxml.jackson.core:jackson-databind:jar:2.14.1:compile[2023-01-09T15:16:06.279Z] [INFO] \- io.quarkus:quarkus-junit5-internal:jar:2.15.0.Final:test[2023-01-09T15:16:06.279Z] [INFO] +- org.jboss.shrinkwrap:shrinkwrap-depchain:pom:1.2.6:test[2023-01-09T15:16:06.279Z] [INFO] | +- org.jboss.shrinkwrap:shrinkwrap-api:jar:1.2.6:test[2023-01-09T15:16:06.279Z] [INFO] | +- org.jboss.shrinkwrap:shrinkwrap-spi:jar:1.2.6:test[2023-01-09T15:16:06.279Z] [INFO] | \- org.jboss.shrinkwrap:shrinkwrap-impl-base:jar:1.2.6:test[2023-01-09T15:16:06.279Z] [INFO] +- org.junit.jupiter:junit-jupiter-api:jar:5.9.1:test[2023-01-09T15:16:06.279Z] [INFO] | +- org.opentest4j:opentest4j:jar:1.2.0:test[2023-01-09T15:16:06.279Z] [INFO] | +- org.junit.platform:junit-platform-commons:jar:1.9.1:test[2023-01-09T15:16:06.279Z] [INFO] | \- org.apiguardian:apiguardian-api:jar:1.1.2:compile[2023-01-09T15:16:06.279Z] [INFO] +- org.junit.jupiter:junit-jupiter-params:jar:5.9.1:test[2023-01-09T15:16:06.279Z] [INFO] +- org.junit.jupiter:junit-jupiter-engine:jar:5.9.1:test[2023-01-09T15:16:06.279Z] [INFO] | \- org.junit.platform:junit-platform-engine:jar:1.9.1:test[2023-01-09T15:16:06.279Z] [INFO] +- io.quarkus:quarkus-core:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- jakarta.inject:jakarta.inject-api:jar:1.0:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.quarkus:quarkus-ide-launcher:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.smallrye.config:smallrye-config:jar:2.13.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- io.smallrye.config:smallrye-config-core:jar:2.13.0:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- io.smallrye.common:smallrye-common-expression:jar:1.13.2:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- io.smallrye.common:smallrye-common-function:jar:1.13.2:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.jboss.logging:jboss-logging:jar:3.5.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.jboss.logmanager:jboss-logmanager-embedded:jar:1.0.11:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.jboss.logging:jboss-logging-annotations:jar:2.2.1.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.jboss.threads:jboss-threads:jar:3.4.3.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.slf4j:slf4j-api:jar:1.7.36:compile[2023-01-09T15:16:06.279Z] [INFO] | +- org.jboss.slf4j:slf4j-jboss-logmanager:jar:1.2.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | +- io.quarkus:quarkus-bootstrap-runner:jar:2.15.0.Final:compile[2023-01-09T15:16:06.279Z] [INFO] | | \- io.github.crac:org-crac:jar:0.1.3:compile[2023-01-09T15:16:06.279Z] [INFO] | \- io.quarkus:quarkus-fs-util:jar:0.0.9:compile[2023-01-09T15:16:06.279Z] [INFO] +- io.quarkus:quarkus-test-common:jar:2.15.0.Final:test[2023-01-09T15:16:06.279Z] [INFO] | +- io.quarkus:quarkus-bootstrap-maven-resolver:jar:2.15.0.Final:test[2023-01-09T15:16:06.279Z] [INFO] | | +- org.apache.maven:maven-embedder:jar:3.8.6:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.apache.maven:maven-settings:jar:3.8.6:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.apache.maven:maven-core:jar:3.8.6:test[2023-01-09T15:16:06.279Z] [INFO] | | | | +- org.apache.maven:maven-artifact:jar:3.8.6:test[2023-01-09T15:16:06.279Z] [INFO] | | | | \- org.codehaus.plexus:plexus-component-annotations:jar:1.6:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.apache.maven:maven-plugin-api:jar:3.8.6:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.apache.maven:maven-model:jar:3.8.6:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.apache.maven:maven-model-builder:jar:3.8.6:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.apache.maven:maven-builder-support:jar:3.8.6:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-api:jar:1.6.3:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-util:jar:1.6.3:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.apache.maven.shared:maven-shared-utils:jar:3.3.4:test[2023-01-09T15:16:06.279Z] [INFO] | | | | \- commons-io:commons-io:jar:2.11.0:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- com.google.inject:guice:jar:no_aop:4.2.2:test[2023-01-09T15:16:06.279Z] [INFO] | | | | \- com.google.guava:guava:jar:31.1-jre:test[2023-01-09T15:16:06.279Z] [INFO] | | | | \- com.google.guava:failureaccess:jar:1.0.1:test[2023-01-09T15:16:06.279Z] [INFO] | | | +- org.codehaus.plexus:plexus-classworlds:jar:2.6.0:test[2023-01-09T15:16:06.280Z] [INFO] | | | +- org.codehaus.plexus:plexus-sec-dispatcher:jar:2.0:test[2023-01-09T15:16:06.280Z] [INFO] | | | +- org.codehaus.plexus:plexus-cipher:jar:2.0:test[2023-01-09T15:16:06.280Z] [INFO] | | | \- commons-cli:commons-cli:jar:1.4:test[2023-01-09T15:16:06.280Z] [INFO] | | +- org.eclipse.sisu:org.eclipse.sisu.plexus:jar:0.3.5:test[2023-01-09T15:16:06.280Z] [INFO] | | +- org.apache.maven:maven-settings-builder:jar:3.8.6:test[2023-01-09T15:16:06.280Z] [INFO] | | | \- org.codehaus.plexus:plexus-interpolation:jar:1.26:test[2023-01-09T15:16:06.280Z] [INFO] | | +- org.apache.maven:maven-resolver-provider:jar:3.8.6:test[2023-01-09T15:16:06.280Z] [INFO] | | | +- org.apache.maven:maven-repository-metadata:jar:3.8.6:test[2023-01-09T15:16:06.280Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-spi:jar:1.6.3:test[2023-01-09T15:16:06.280Z] [INFO] | | | +- org.apache.maven.resolver:maven-resolver-impl:jar:1.6.3:test[2023-01-09T15:16:06.280Z] [INFO] | | | \- org.codehaus.plexus:plexus-utils:jar:3.3.0:test[2023-01-09T15:16:06.280Z] [INFO] | | +- org.apache.maven.resolver:maven-resolver-connector-basic:jar:1.6.3:test[2023-01-09T15:16:06.280Z] [INFO] | | +- org.apache.maven.resolver:maven-resolver-transport-wagon:jar:1.6.3:test[2023-01-09T15:16:06.280Z] [INFO] | | +- org.apache.maven.wagon:wagon-http:jar:3.5.1:test[2023-01-09T15:16:06.280Z] [INFO] | | | +- org.apache.maven.wagon:wagon-http-shared:jar:3.5.1:test[2023-01-09T15:16:06.280Z] [INFO] | | | +- org.apache.httpcomponents:httpclient:jar:4.5.13:test[2023-01-09T15:16:06.280Z] [INFO] | | | | \- commons-codec:commons-codec:jar:1.15:test[2023-01-09T15:16:06.280Z] [INFO] | | | +- org.apache.httpcomponents:httpcore:jar:4.4.15:test[2023-01-09T15:16:06.280Z] [INFO] | | | \- org.apache.maven.wagon:wagon-provider-api:jar:3.5.1:test[2023-01-09T15:16:06.280Z] [INFO] | | \- org.apache.maven.wagon:wagon-file:jar:3.5.1:test[2023-01-09T15:16:06.280Z] [INFO] | +- io.quarkus:quarkus-bootstrap-gradle-resolver:jar:2.15.0.Final:test[2023-01-09T15:16:06.280Z] [INFO] | \- org.jboss.logging:commons-logging-jboss-logging:jar:1.0.0.Final:runtime[2023-01-09T15:16:06.280Z] [INFO] +- jakarta.enterprise:jakarta.enterprise.cdi-api:jar:2.0.2:compile[2023-01-09T15:16:06.280Z] [INFO] | +- jakarta.el:jakarta.el-api:jar:3.0.3:compile[2023-01-09T15:16:06.280Z] [INFO] | \- jakarta.interceptor:jakarta.interceptor-api:jar:1.2.5:compile[2023-01-09T15:16:06.280Z] [INFO] +- io.quarkus:quarkus-devtools-utilities:jar:2.15.0.Final:compile[2023-01-09T15:16:06.280Z] [INFO] \- org.awaitility:awaitility:jar:4.2.0:test[2023-01-09T15:16:06.280Z] [INFO] \- org.hamcrest:hamcrest:jar:2.1:test[2023-01-09T15:16:06.280Z] [INFO] [2023-01-09T15:16:06.280Z] [INFO] ----< org.kie.kogito:kogito-quarkus-serverless-workflow-deployment >----[2023-01-09T15:16:06.280Z] [INFO] Building Kogito :: Quarkus Workflows Extension :: Deployment 2.0.0-SNAPSHOT [308/314][2023-01-09T15:16:06.280Z] [INFO] --------------------------------[ jar ]---------------------------------[2023-01-09T15:16:06.280Z] ::error title= [kiegroup/kogito-runtimes] [Execution ERROR]::mvn -f kogito-runtimes/pom.xml clean install -Dvalidate-formatting -s /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes@tmp/config11146782411654787205tmp -Dmaven.wagon.http.ssl.insecure=true -Dmaven.test.failure.ignore=true -nsu -ntp -fae -e -Dhttp.keepAlive=false -Dmaven.wagon.http.pool=false -Dmaven.wagon.httpconnectionManager.ttlSeconds=120 -Dmaven.wagon.http.retryHandler.count=3 -T 1C -Prun-code-coverage dependency:tree -Dorg.slf4j.simpleLogger.log.org.apache.maven.cli.transfer.Slf4jMavenTransferListener=warn -B[2023-01-09T15:16:06.280Z] [INFO] Archive artifact won't be executed[2023-01-09T15:16:06.280Z] ::group::[Pull Request Flow] Execution Summary...[2023-01-09T15:16:06.280Z] [INFO] ----------------------------------------------[2023-01-09T15:16:06.284Z] script returned exit code 143[Pipeline] }[Pipeline] // withCredentials[Pipeline] }[2023-01-09T15:16:06.351Z] Deleting 1 temporary files[Pipeline] // configFileProvider[Pipeline] }[Pipeline] // scriptPost stage[Pipeline] junit[2023-01-09T15:16:06.457Z] Recording test results[2023-01-09T15:16:11.541Z] [Checks API] No suitable checks publisher found.[Pipeline] archiveArtifacts[2023-01-09T15:16:11.561Z] Archiving artifacts[Pipeline] script[Pipeline] {[Pipeline] sh[2023-01-09T15:16:12.987Z] + rm -rf console.log[Pipeline] sh[2023-01-09T15:16:13.277Z] + wget --no-check-certificate -qO - https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/1664/consoleText[2023-01-09T15:16:13.277Z] + tail -n 300 @cristianonicolai the tests failures are not related. jenkins rerun kogito-runtimes tests (tests) - kogito-runtimes job #1665 was: UNSTABLE Possible explanation: This should be test failures Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/1665/display/redirect Test results: PASSED: 3147 FAILED: 1 Those are the test failures: org.kie.kogito.addons.quarkus.kubernetes.KubernetesAddOnTest.(?) java.lang.RuntimeException: java.lang.RuntimeException: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor#setupKubernetesDevService threw an exception: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:142) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 24 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 25 moreCaused by: com.github.dockerjava.api.exception.ConflictException: Status 409: {"message":"cannot join network of a non running container: d48a1afba5b37a19da4b976e2a2ee4d00ef9a33135563032b0b875a2eb5d4483"} at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.execute(DefaultInvocationBuilder.java:245) at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.post(DefaultInvocationBuilder.java:102) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:31) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:13) at org.testcontainers.shaded.com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21) at org.testcontainers.shaded.com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35) at org.testcontainers.shaded.com.github.dockerjava.core.command.StartContainerCmdImpl.exec(StartContainerCmdImpl.java:43) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:441) ... 27 more at io.quarkus.test.QuarkusUnitTest.beforeAll(QuarkusUnitTest.java:689) at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.lambda$invokeBeforeAllCallbacks$12(ClassBasedTestDescriptor.java:395) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.invokeBeforeAllCallbacks(ClassBasedTestDescriptor.java:395) at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.before(ClassBasedTestDescriptor.java:211) at org.junit.jupiter.engine.descriptor.ClassBasedTestDescriptor.before(ClassBasedTestDescriptor.java:84) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:148) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95) at java.base/java.util.ArrayList.forEach(ArrayList.java:1541) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.invokeAll(SameThreadHierarchicalTestExecutorService.java:41) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$6(NodeTestTask.java:155) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$8(NodeTestTask.java:141) at org.junit.platform.engine.support.hierarchical.Node.around(Node.java:137) at org.junit.platform.engine.support.hierarchical.NodeTestTask.lambda$executeRecursively$9(NodeTestTask.java:139) at org.junit.platform.engine.support.hierarchical.ThrowableCollector.execute(ThrowableCollector.java:73) at org.junit.platform.engine.support.hierarchical.NodeTestTask.executeRecursively(NodeTestTask.java:138) at org.junit.platform.engine.support.hierarchical.NodeTestTask.execute(NodeTestTask.java:95) at org.junit.platform.engine.support.hierarchical.SameThreadHierarchicalTestExecutorService.submit(SameThreadHierarchicalTestExecutorService.java:35) at org.junit.platform.engine.support.hierarchical.HierarchicalTestExecutor.execute(HierarchicalTestExecutor.java:57) at org.junit.platform.engine.support.hierarchical.HierarchicalTestEngine.execute(HierarchicalTestEngine.java:54) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:55) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:102) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:54) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86) at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86) at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:188) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:128) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)Caused by: java.lang.RuntimeException: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor#setupKubernetesDevService threw an exception: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:142) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 24 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 25 moreCaused by: com.github.dockerjava.api.exception.ConflictException: Status 409: {"message":"cannot join network of a non running container: d48a1afba5b37a19da4b976e2a2ee4d00ef9a33135563032b0b875a2eb5d4483"} at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.execute(DefaultInvocationBuilder.java:245) at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.post(DefaultInvocationBuilder.java:102) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:31) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:13) at org.testcontainers.shaded.com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21) at org.testcontainers.shaded.com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35) at org.testcontainers.shaded.com.github.dockerjava.core.command.StartContainerCmdImpl.exec(StartContainerCmdImpl.java:43) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:441) ... 27 more at io.quarkus.runner.bootstrap.AugmentActionImpl.runAugment(AugmentActionImpl.java:335) at io.quarkus.runner.bootstrap.AugmentActionImpl.createInitialRuntimeApplication(AugmentActionImpl.java:252) at io.quarkus.test.QuarkusUnitTest.beforeAll(QuarkusUnitTest.java:635) ... 43 moreCaused by: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor#setupKubernetesDevService threw an exception: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:142) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 24 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 25 moreCaused by: com.github.dockerjava.api.exception.ConflictException: Status 409: {"message":"cannot join network of a non running container: d48a1afba5b37a19da4b976e2a2ee4d00ef9a33135563032b0b875a2eb5d4483"} at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.execute(DefaultInvocationBuilder.java:245) at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.post(DefaultInvocationBuilder.java:102) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:31) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:13) at org.testcontainers.shaded.com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21) at org.testcontainers.shaded.com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35) at org.testcontainers.shaded.com.github.dockerjava.core.command.StartContainerCmdImpl.exec(StartContainerCmdImpl.java:43) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:441) ... 27 more at io.quarkus.builder.Execution.run(Execution.java:123) at io.quarkus.builder.BuildExecutionBuilder.execute(BuildExecutionBuilder.java:79) at io.quarkus.deployment.QuarkusAugmentor.run(QuarkusAugmentor.java:160) at io.quarkus.runner.bootstrap.AugmentActionImpl.runAugment(AugmentActionImpl.java:331) ... 45 moreCaused by: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:142) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 24 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 25 moreCaused by: com.github.dockerjava.api.exception.ConflictException: Status 409: {"message":"cannot join network of a non running container: d48a1afba5b37a19da4b976e2a2ee4d00ef9a33135563032b0b875a2eb5d4483"} at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.execute(DefaultInvocationBuilder.java:245) at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.post(DefaultInvocationBuilder.java:102) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:31) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:13) at org.testcontainers.shaded.com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21) at org.testcontainers.shaded.com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35) at org.testcontainers.shaded.com.github.dockerjava.core.command.StartContainerCmdImpl.exec(StartContainerCmdImpl.java:43) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:441) ... 27 more Jenkins rerun kogito-runtimes tests
gharchive/pull-request
2023-01-06T14:34:17
2025-04-01T06:39:17.260563
{ "authors": [ "cristianonicolai", "kie-ci3", "mareknovotny", "radtriste", "ricardozanini" ], "repo": "kiegroup/kogito-runtimes", "url": "https://github.com/kiegroup/kogito-runtimes/pull/2721", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1652287705
[kie-issues-156] Use specific conversion method for numbers Fixes https://github.com/kiegroup/kie-issues/issues/156. When the convert method is executed on IBM JDK, the order of iteration of methods is different. For Integer, the original implementation picked a wrong method (Integer.getInteger()), which caused a wrong behaviour and test fail. This PR uses valueOf() method specifically for numbers. Adds more tests for various number types. How to replicate CI configuration locally? Build Chain tool does "simple" maven build(s), the builds are just Maven commands, but because the repositories relates and depends on each other and any change in API or class method could affect several of those repositories there is a need to use build-chain tool to handle cross repository builds and be sure that we always use latest version of the code for each repository. build-chain tool is a build tool which can be used on command line locally or in Github Actions workflow(s), in case you need to change multiple repositories and send multiple dependent pull requests related with a change you can easily reproduce the same build by executing it on Github hosted environment or locally in your development environment. See local execution details to get more information about it. How to retest this PR or trigger a specific build: for pull request checks Please add comment: Jenkins retest this for a specific pull request check Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] tests for quarkus branch checks Run checks against Quarkus current used branch Please add comment: Jenkins run quarkus-branch for a quarkus branch specific check Run checks against Quarkus current used branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-branch for quarkus main checks Run checks against Quarkus main branch Please add comment: Jenkins run quarkus-main for a specific quarkus main check Run checks against Quarkus main branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-main for quarkus lts checks Run checks against Quarkus lts branch Please add comment: Jenkins run quarkus-lts for a specific quarkus lts check Run checks against Quarkus lts branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-lts for native checks Run native checks Please add comment: Jenkins run native for a specific native check Run native checks Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] native for native lts checks Run native checks against quarkus lts branch Please add comment: Jenkins run native-lts for a specific native lts check Run native checks against quarkus lts branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] native-lts Full Kogito testing (with cloud images and operator BDD testing) Please add comment: Jenkins run BDD This check should be used only if a big change is done as it takes time to run, need resources and one full BDD tests check can be done at a time ... How to backport a pull request to a different branch? In order to automatically create a backporting pull request please add one or more labels having the following format backport-<branch-name>, where <branch-name> is the name of the branch where the pull request must be backported to (e.g., backport-7.67.x to backport the original PR to the 7.67.x branch). NOTE: backporting is an action aiming to move a change (usually a commit) from a branch (usually the main one) to another one, which is generally referring to a still maintained release branch. Keeping it simple: it is about to move a specific change or a set of them from one branch to another. Once the original pull request is successfully merged, the automated action will create one backporting pull request per each label (with the previous format) that has been added. If something goes wrong, the author will be notified and at this point a manual backporting is needed. NOTE: this automated backporting is triggered whenever a pull request on main branch is labeled or closed, but both conditions must be satisfied to get the new PR created. @kiegroup/gatekeepers Could you please merge this, squashing commits? Thanks! I see @baldimir asked the review to @fjtirado too. @baldimir Do you want to wait or is it fine for you to proceed? hi @danielezonca Let's wait for @fjtirado @danielezonca ^^
gharchive/pull-request
2023-04-03T15:23:17
2025-04-01T06:39:17.284993
{ "authors": [ "baldimir", "danielezonca", "gitgabrio" ], "repo": "kiegroup/kogito-runtimes", "url": "https://github.com/kiegroup/kogito-runtimes/pull/2919", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1854958942
KOGITO-9145 Create an implementation of the Service Discovery catalog based on MicroProfile Config Jira: https://issues.redhat.com/browse/KOGITO-9145 @VaniHaripriya's PR is lacking of integration tests and she won't be able to finish it on time. So, I'm opening this one with my commits on top of hers. [x] You have read the contributors guide [x] Your code is properly formatted according to this configuration [x] Pull Request title is properly formatted: KOGITO-XYZ Subject [x] Pull Request title contains the target branch if not targeting main: [0.9.x] KOGITO-XYZ Subject [x] Pull Request contains link to the JIRA issue [x] Pull Request contains link to any dependent or related Pull Request [x] Pull Request contains description of the issue [x] Pull Request does not include fixes for issues other than the main ticket How to replicate CI configuration locally? Build Chain tool does "simple" maven build(s), the builds are just Maven commands, but because the repositories relates and depends on each other and any change in API or class method could affect several of those repositories there is a need to use build-chain tool to handle cross repository builds and be sure that we always use latest version of the code for each repository. build-chain tool is a build tool which can be used on command line locally or in Github Actions workflow(s), in case you need to change multiple repositories and send multiple dependent pull requests related with a change you can easily reproduce the same build by executing it on Github hosted environment or locally in your development environment. See local execution details to get more information about it. How to retest this PR or trigger a specific build: for pull request checks Please add comment: Jenkins retest this for a specific pull request check Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] tests for quarkus branch checks Run checks against Quarkus current used branch Please add comment: Jenkins run quarkus-branch for a quarkus branch specific check Run checks against Quarkus current used branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-branch for quarkus main checks Run checks against Quarkus main branch Please add comment: Jenkins run quarkus-main for a specific quarkus main check Run checks against Quarkus main branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-main for quarkus lts checks Run checks against Quarkus lts branch Please add comment: Jenkins run quarkus-lts for a specific quarkus lts check Run checks against Quarkus lts branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] quarkus-lts for native checks Run native checks Please add comment: Jenkins run native for a specific native check Run native checks Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] native for native lts checks Run native checks against quarkus lts branch Please add comment: Jenkins run native-lts for a specific native lts check Run native checks against quarkus lts branch Please add comment: Jenkins (re)run [kogito-runtimes|kogito-apps|kogito-examples] native-lts Full Kogito testing (with cloud images and operator BDD testing) Please add comment: Jenkins run BDD This check should be used only if a big change is done as it takes time to run, need resources and one full BDD tests check can be done at a time ... How to backport a pull request to a different branch? In order to automatically create a backporting pull request please add one or more labels having the following format backport-<branch-name>, where <branch-name> is the name of the branch where the pull request must be backported to (e.g., backport-7.67.x to backport the original PR to the 7.67.x branch). NOTE: backporting is an action aiming to move a change (usually a commit) from a branch (usually the main one) to another one, which is generally referring to a still maintained release branch. Keeping it simple: it is about to move a specific change or a set of them from one branch to another. Once the original pull request is successfully merged, the automated action will create one backporting pull request per each label (with the previous format) that has been added. If something goes wrong, the author will be notified and at this point a manual backporting is needed. NOTE: this automated backporting is triggered whenever a pull request on main branch is labeled or closed, but both conditions must be satisfied to get the new PR created. Quarkus-3 PR check is failing ... what to do ? The Quarkus 3 check is applying patches from the `.ci/environments/quarkus-3/patches`. The first patch, called 0001_before_sh.patch, is generated from Openrewrite .ci/environments/quarkus-3/quarkus3.yml recipe. The patch is created to speed up the check. But it may be that some changes in the PR broke this patch. No panic, there is an easy way to regenerate it. You just need to comment on the PR: jenkins rewrite quarkus-3 and it should, after some minutes (~20/30min) apply a commit on the PR with the patch regenerated. Other patches were generated manually. If any of it fails, you will need to manually update it... and push your changes. jenkins retest this (tests) - kogito-runtimes job #3029 was: UNSTABLE Possible explanation: This should be test failures Reproducer export BUILD_MVN_OPTS_CURRENT=-T 1C -Dvalidate-formatting -Prun-code-coverage build-chain build cross_pr -f 'https://raw.githubusercontent.com/kiegroup/kogito-pipelines/main/.ci/pull-request-config.yaml' -o 'bc' -p kiegroup/kogito-runtimes -u https://github.com/kiegroup/kogito-runtimes/pull/3179 --skipParallelCheckout NOTE: To install the build-chain tool, please refer to https://github.com/kiegroup/github-action-build-chain#local-execution Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3029/display/redirect Test results: PASSED: 3356 FAILED: 1 Those are the test failures: org.kie.kogito.addons.quarkus.microprofile.config.service.catalog.it.MicroProfileConfigServiceAddonIT.executeWithEmptyParameters java.lang.RuntimeException: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor#setupKubernetesDevService threw an exception: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:142) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 24 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 25 moreCaused by: com.github.dockerjava.api.exception.ConflictException: Status 409: {"message":"cannot join network of a non running container: 3a08247341cc6dd467de5426b224f627decfe06f12291db85c14424c14e7459f"} at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.execute(DefaultInvocationBuilder.java:245) at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.post(DefaultInvocationBuilder.java:102) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:31) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:13) at org.testcontainers.shaded.com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21) at org.testcontainers.shaded.com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35) at org.testcontainers.shaded.com.github.dockerjava.core.command.StartContainerCmdImpl.exec(StartContainerCmdImpl.java:43) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:441) ... 27 more jenkins rerun kogito-runtimes tests (tests) - kogito-runtimes job #3030 was: ABORTED Possible explanation: Most probably a timeout, please review Reproducer export BUILD_MVN_OPTS_CURRENT=-T 1C -Dvalidate-formatting -Prun-code-coverage build-chain build cross_pr -f 'https://raw.githubusercontent.com/kiegroup/kogito-pipelines/main/.ci/pull-request-config.yaml' -o 'bc' -p kiegroup/kogito-runtimes -u https://github.com/kiegroup/kogito-runtimes/pull/3179 --skipParallelCheckout NOTE: To install the build-chain tool, please refer to https://github.com/kiegroup/github-action-build-chain#local-execution Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3030/display/redirect Test results: PASSED: 3356 FAILED: 0 Those are the test failures: none See console log: Console Logs [2023-08-17T17:05:11.655Z] [ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:3.1.2:verify (integration-test) on project kogito-addons-quarkus-microprofile-config-service-catalog-integration-tests: [2023-08-17T17:05:11.655Z] [ERROR] [2023-08-17T17:05:11.655Z] [ERROR] Please refer to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/failsafe-reports for the individual test results.[2023-08-17T17:05:11.655Z] [ERROR] Please refer to dump files (if any exist) [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.[2023-08-17T17:05:11.655Z] [ERROR] The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T17:05:11.655Z] [ERROR] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817102247202_396.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T10-10-17_516-jvmRun3' 'surefire-20230817102247202_394tmp' 'surefire_36-20230817102247202_395tmp'[2023-08-17T17:05:11.655Z] [ERROR] Error occurred in starting fork, check output in log[2023-08-17T17:05:11.655Z] [ERROR] Process Exit Code: 143[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T17:05:11.655Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T17:05:11.655Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.655Z] [ERROR] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T17:05:11.655Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.655Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T17:05:11.655Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T17:05:11.655Z] [ERROR] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T17:05:11.655Z] [ERROR] [2023-08-17T17:05:11.655Z] [ERROR] org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T17:05:11.656Z] [ERROR] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817102247202_396.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T10-10-17_516-jvmRun3' 'surefire-20230817102247202_394tmp' 'surefire_36-20230817102247202_395tmp'[2023-08-17T17:05:11.656Z] [ERROR] Error occurred in starting fork, check output in log[2023-08-17T17:05:11.656Z] [ERROR] Process Exit Code: 143[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T17:05:11.656Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T17:05:11.656Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.656Z] [ERROR] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T17:05:11.656Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.656Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T17:05:11.656Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T17:05:11.656Z] [ERROR] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T17:05:11.656Z] [ERROR] -> [Help 1][2023-08-17T17:05:11.656Z] org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:3.1.2:verify (integration-test) on project kogito-addons-quarkus-microprofile-config-service-catalog-integration-tests: [2023-08-17T17:05:11.656Z] [2023-08-17T17:05:11.656Z] Please refer to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/failsafe-reports for the individual test results.[2023-08-17T17:05:11.656Z] Please refer to dump files (if any exist) [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.[2023-08-17T17:05:11.656Z] The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T17:05:11.656Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817102247202_396.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T10-10-17_516-jvmRun3' 'surefire-20230817102247202_394tmp' 'surefire_36-20230817102247202_395tmp'[2023-08-17T17:05:11.656Z] Error occurred in starting fork, check output in log[2023-08-17T17:05:11.656Z] Process Exit Code: 143[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T17:05:11.656Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.656Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T17:05:11.656Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.656Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T17:05:11.656Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T17:05:11.656Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T17:05:11.656Z] [2023-08-17T17:05:11.656Z] org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T17:05:11.656Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817102247202_396.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T10-10-17_516-jvmRun3' 'surefire-20230817102247202_394tmp' 'surefire_36-20230817102247202_395tmp'[2023-08-17T17:05:11.656Z] Error occurred in starting fork, check output in log[2023-08-17T17:05:11.656Z] Process Exit Code: 143[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T17:05:11.656Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.656Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T17:05:11.656Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.656Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T17:05:11.656Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T17:05:11.656Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T17:05:11.656Z] [2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2 (MojoExecutor.java:375)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:351)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:215)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:171)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:163)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:210)[2023-08-17T17:05:11.656Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:195)[2023-08-17T17:05:11.656Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T17:05:11.656Z] at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:515)[2023-08-17T17:05:11.656Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T17:05:11.656Z] at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1128)[2023-08-17T17:05:11.656Z] at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:628)[2023-08-17T17:05:11.656Z] at java.lang.Thread.run (Thread.java:829)[2023-08-17T17:05:11.656Z] Caused by: org.apache.maven.plugin.MojoExecutionException: [2023-08-17T17:05:11.656Z] [2023-08-17T17:05:11.656Z] Please refer to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/failsafe-reports for the individual test results.[2023-08-17T17:05:11.656Z] Please refer to dump files (if any exist) [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.[2023-08-17T17:05:11.656Z] The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T17:05:11.656Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817102247202_396.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T10-10-17_516-jvmRun3' 'surefire-20230817102247202_394tmp' 'surefire_36-20230817102247202_395tmp'[2023-08-17T17:05:11.656Z] Error occurred in starting fork, check output in log[2023-08-17T17:05:11.656Z] Process Exit Code: 143[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T17:05:11.656Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T17:05:11.657Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T17:05:11.657Z] [2023-08-17T17:05:11.657Z] org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T17:05:11.657Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817102247202_396.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T10-10-17_516-jvmRun3' 'surefire-20230817102247202_394tmp' 'surefire_36-20230817102247202_395tmp'[2023-08-17T17:05:11.657Z] Error occurred in starting fork, check output in log[2023-08-17T17:05:11.657Z] Process Exit Code: 143[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T17:05:11.657Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T17:05:11.657Z] [2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.SurefireHelper.reportExecution (SurefireHelper.java:154)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.failsafe.VerifyMojo.execute (VerifyMojo.java:190)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2 (MojoExecutor.java:370)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:351)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:215)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:171)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:163)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:210)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:195)[2023-08-17T17:05:11.657Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T17:05:11.657Z] at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:515)[2023-08-17T17:05:11.657Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T17:05:11.657Z] at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1128)[2023-08-17T17:05:11.657Z] at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:628)[2023-08-17T17:05:11.657Z] at java.lang.Thread.run (Thread.java:829)[2023-08-17T17:05:11.657Z] Caused by: org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T17:05:11.657Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817102247202_396.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T10-10-17_516-jvmRun3' 'surefire-20230817102247202_394tmp' 'surefire_36-20230817102247202_395tmp'[2023-08-17T17:05:11.657Z] Error occurred in starting fork, check output in log[2023-08-17T17:05:11.657Z] Process Exit Code: 143[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T17:05:11.657Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T17:05:11.657Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T17:05:11.657Z] [2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.failsafe.VerifyMojo.getBooterForkException (VerifyMojo.java:198)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.failsafe.VerifyMojo.execute (VerifyMojo.java:190)[2023-08-17T17:05:11.657Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2 (MojoExecutor.java:370)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:351)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:215)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:171)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:163)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:210)[2023-08-17T17:05:11.657Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:195)[2023-08-17T17:05:11.657Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T17:05:11.657Z] at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:515)[2023-08-17T17:05:11.657Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T17:05:11.657Z] at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1128)[2023-08-17T17:05:11.657Z] at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:628)[2023-08-17T17:05:11.657Z] at java.lang.Thread.run (Thread.java:829)[2023-08-17T17:05:11.657Z] [ERROR] [2023-08-17T17:05:11.657Z] [ERROR] Re-run Maven using the -X switch to enable full debug logging.[2023-08-17T17:05:11.657Z] [ERROR] [2023-08-17T17:05:11.657Z] [ERROR] For more information about the errors and possible solutions, please read the following articles:[2023-08-17T17:05:11.657Z] [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[2023-08-17T17:05:11.657Z] [ERROR] [2023-08-17T17:05:11.657Z] [ERROR] After correcting the problems, you can resume the build with the command[2023-08-17T17:05:11.658Z] [ERROR] mvn -rf :kogito-addons-quarkus-microprofile-config-service-catalog-integration-tests[2023-08-17T17:05:11.658Z] [INFO] kiegroup/kogito-runtimes failed. Won't execute remaining commands and projects[2023-08-17T17:05:11.658Z] [INFO] Execution summary for kiegroup/kogito-runtimes[2023-08-17T17:05:11.658Z] # [BEFORE] [kiegroup/kogito-runtimes] export INTEGRATION_BRANCH=[2023-08-17T17:05:11.658Z] [INFO] OK [Executed in 0.179566 ms][2023-08-17T17:05:11.658Z] [2023-08-17T17:05:11.658Z] # [BEFORE] [kiegroup/kogito-runtimes] bash -c "if [ ! -z '' ] && [ -f .ci/environments/update.sh ]; then .ci/environments/update.sh ; fi"[2023-08-17T17:05:11.658Z] [INFO] OK [Executed in 7.996456 ms][2023-08-17T17:05:11.658Z] [2023-08-17T17:05:11.658Z] # [COMMANDS] [kiegroup/kogito-runtimes] export MVN_CMD=`bash -c "if [ '' = 'true' ]; then printf 'deploy '; else printf 'install'; fi"`[2023-08-17T17:05:11.658Z] [INFO] OK [Executed in 6.844716 ms][2023-08-17T17:05:11.658Z] [2023-08-17T17:05:11.658Z] # [COMMANDS] [kiegroup/kogito-runtimes] mvn dependency:tree clean -Dfull install -s /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes@tmp/config1758373660637564tmp -Dmaven.wagon.http.ssl.insecure=true -Dmaven.test.failure.ignore=true -nsu -ntp -fae -e -Dhttp.keepAlive=false -Dmaven.wagon.http.pool=false -Dmaven.wagon.httpconnectionManager.ttlSeconds=120 -Dmaven.wagon.http.retryHandler.count=3 -T 1C -Dvalidate-formatting -Prun-code-coverage -Dorg.slf4j.simpleLogger.log.org.apache.maven.cli.transfer.Slf4jMavenTransferListener=warn -B[2023-08-17T17:05:11.658Z] [INFO] NOT OK [Executed in 10509862.525863 ms][2023-08-17T17:05:11.658Z] [ERROR] The process '/opt/tools/apache-maven-3.8.7/bin/mvn' failed with exit code 143[2023-08-17T17:05:11.658Z] [2023-08-17T17:05:11.658Z] [INFO] [AFTER] No commands were found for kiegroup/kogito-runtimes[2023-08-17T17:05:11.658Z] [2023-08-17T17:05:11.658Z] # Uploading artifacts[2023-08-17T17:05:11.658Z] [INFO] Will not upload any artifacts in CLI environment[2023-08-17T17:05:11.658Z] [2023-08-17T17:05:11.658Z] [ERROR] Failed to execute commands for kiegroup/kogito-runtimes[2023-08-17T17:05:11.658Z] [ERROR] Failed to execute mvn dependency:tree clean -Dfull install -s /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes@tmp/config1758373660637564tmp -Dmaven.wagon.http.ssl.insecure=true -Dmaven.test.failure.ignore=true -nsu -ntp -fae -e -Dhttp.keepAlive=false -Dmaven.wagon.http.pool=false -Dmaven.wagon.httpconnectionManager.ttlSeconds=120 -Dmaven.wagon.http.retryHandler.count=3 -T 1C -Dvalidate-formatting -Prun-code-coverage -Dorg.slf4j.simpleLogger.log.org.apache.maven.cli.transfer.Slf4jMavenTransferListener=warn -B :[2023-08-17T17:05:11.658Z] [ERROR] The process '/opt/tools/apache-maven-3.8.7/bin/mvn' failed with exit code 143[2023-08-17T17:05:11.666Z] script returned exit code 1[Pipeline] }[Pipeline] // withCredentials[Pipeline] }[2023-08-17T17:05:11.733Z] Deleting 1 temporary files[Pipeline] // configFileProvider[Pipeline] }[Pipeline] // scriptPost stage[Pipeline] script[Pipeline] {[Pipeline] sh[2023-08-17T17:05:12.138Z] + find . -type d -name node_modules -exec rm -rf '{}' ';'[Pipeline] junit[2023-08-17T17:05:20.228Z] Recording test results[2023-08-17T17:05:27.281Z] [Checks API] No suitable checks publisher found.[Pipeline] archiveArtifacts[2023-08-17T17:05:27.299Z] Archiving artifacts[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh[2023-08-17T17:05:28.782Z] + rm -rf console.log[Pipeline] sh[2023-08-17T17:05:29.074Z] + wget --no-check-certificate -qO - https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3030/consoleText[2023-08-17T17:05:29.074Z] + tail -n 300 jenkins rerun kogito-runtimes tests (tests) - kogito-runtimes job #3032 was: ABORTED Possible explanation: Most probably a timeout, please review Reproducer export BUILD_MVN_OPTS_CURRENT=-T 1C -Dvalidate-formatting -Prun-code-coverage build-chain build cross_pr -f 'https://raw.githubusercontent.com/kiegroup/kogito-pipelines/main/.ci/pull-request-config.yaml' -o 'bc' -p kiegroup/kogito-runtimes -u https://github.com/kiegroup/kogito-runtimes/pull/3179 --skipParallelCheckout NOTE: To install the build-chain tool, please refer to https://github.com/kiegroup/github-action-build-chain#local-execution Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3032/display/redirect Test results: PASSED: 3356 FAILED: 0 Those are the test failures: none See console log: Console Logs [2023-08-17T20:24:33.195Z] [ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:3.1.2:verify (integration-test) on project kogito-addons-quarkus-microprofile-config-service-catalog-integration-tests: [2023-08-17T20:24:33.195Z] [ERROR] [2023-08-17T20:24:33.195Z] [ERROR] Please refer to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/failsafe-reports for the individual test results.[2023-08-17T20:24:33.195Z] [ERROR] Please refer to dump files (if any exist) [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.[2023-08-17T20:24:33.195Z] [ERROR] The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T20:24:33.195Z] [ERROR] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817134247864_388.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T13-29-46_362-jvmRun1' 'surefire-20230817134247864_386tmp' 'surefire_35-20230817134247864_387tmp'[2023-08-17T20:24:33.195Z] [ERROR] Error occurred in starting fork, check output in log[2023-08-17T20:24:33.195Z] [ERROR] Process Exit Code: 143[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T20:24:33.195Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T20:24:33.195Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.195Z] [ERROR] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T20:24:33.195Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.195Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T20:24:33.195Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T20:24:33.195Z] [ERROR] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T20:24:33.195Z] [ERROR] [2023-08-17T20:24:33.195Z] [ERROR] org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T20:24:33.196Z] [ERROR] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817134247864_388.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T13-29-46_362-jvmRun1' 'surefire-20230817134247864_386tmp' 'surefire_35-20230817134247864_387tmp'[2023-08-17T20:24:33.196Z] [ERROR] Error occurred in starting fork, check output in log[2023-08-17T20:24:33.196Z] [ERROR] Process Exit Code: 143[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T20:24:33.196Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T20:24:33.196Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.196Z] [ERROR] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T20:24:33.196Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.196Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T20:24:33.196Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T20:24:33.196Z] [ERROR] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T20:24:33.196Z] [ERROR] -> [Help 1][2023-08-17T20:24:33.196Z] org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:3.1.2:verify (integration-test) on project kogito-addons-quarkus-microprofile-config-service-catalog-integration-tests: [2023-08-17T20:24:33.196Z] [2023-08-17T20:24:33.196Z] Please refer to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/failsafe-reports for the individual test results.[2023-08-17T20:24:33.196Z] Please refer to dump files (if any exist) [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.[2023-08-17T20:24:33.196Z] The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T20:24:33.196Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817134247864_388.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T13-29-46_362-jvmRun1' 'surefire-20230817134247864_386tmp' 'surefire_35-20230817134247864_387tmp'[2023-08-17T20:24:33.196Z] Error occurred in starting fork, check output in log[2023-08-17T20:24:33.196Z] Process Exit Code: 143[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T20:24:33.196Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.196Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T20:24:33.196Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.196Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T20:24:33.196Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T20:24:33.196Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T20:24:33.196Z] [2023-08-17T20:24:33.196Z] org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T20:24:33.196Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817134247864_388.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T13-29-46_362-jvmRun1' 'surefire-20230817134247864_386tmp' 'surefire_35-20230817134247864_387tmp'[2023-08-17T20:24:33.196Z] Error occurred in starting fork, check output in log[2023-08-17T20:24:33.196Z] Process Exit Code: 143[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T20:24:33.196Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T20:24:33.196Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.196Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T20:24:33.196Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.196Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T20:24:33.196Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T20:24:33.196Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T20:24:33.196Z] [2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2 (MojoExecutor.java:375)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:351)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:215)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:171)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:163)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:210)[2023-08-17T20:24:33.196Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:195)[2023-08-17T20:24:33.196Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T20:24:33.196Z] at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:515)[2023-08-17T20:24:33.196Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T20:24:33.196Z] at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1128)[2023-08-17T20:24:33.196Z] at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:628)[2023-08-17T20:24:33.196Z] at java.lang.Thread.run (Thread.java:829)[2023-08-17T20:24:33.196Z] Caused by: org.apache.maven.plugin.MojoExecutionException: [2023-08-17T20:24:33.196Z] [2023-08-17T20:24:33.196Z] Please refer to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/failsafe-reports for the individual test results.[2023-08-17T20:24:33.196Z] Please refer to dump files (if any exist) [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.[2023-08-17T20:24:33.196Z] The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T20:24:33.196Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817134247864_388.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T13-29-46_362-jvmRun1' 'surefire-20230817134247864_386tmp' 'surefire_35-20230817134247864_387tmp'[2023-08-17T20:24:33.197Z] Error occurred in starting fork, check output in log[2023-08-17T20:24:33.197Z] Process Exit Code: 143[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T20:24:33.197Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T20:24:33.197Z] [2023-08-17T20:24:33.197Z] org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T20:24:33.197Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817134247864_388.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T13-29-46_362-jvmRun1' 'surefire-20230817134247864_386tmp' 'surefire_35-20230817134247864_387tmp'[2023-08-17T20:24:33.197Z] Error occurred in starting fork, check output in log[2023-08-17T20:24:33.197Z] Process Exit Code: 143[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T20:24:33.197Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T20:24:33.197Z] [2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.SurefireHelper.reportExecution (SurefireHelper.java:154)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.failsafe.VerifyMojo.execute (VerifyMojo.java:190)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2 (MojoExecutor.java:370)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:351)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:215)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:171)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:163)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:210)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:195)[2023-08-17T20:24:33.197Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T20:24:33.197Z] at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:515)[2023-08-17T20:24:33.197Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T20:24:33.197Z] at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1128)[2023-08-17T20:24:33.197Z] at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:628)[2023-08-17T20:24:33.197Z] at java.lang.Thread.run (Thread.java:829)[2023-08-17T20:24:33.197Z] Caused by: org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T20:24:33.197Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817134247864_388.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T13-29-46_362-jvmRun1' 'surefire-20230817134247864_386tmp' 'surefire_35-20230817134247864_387tmp'[2023-08-17T20:24:33.197Z] Error occurred in starting fork, check output in log[2023-08-17T20:24:33.197Z] Process Exit Code: 143[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T20:24:33.197Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T20:24:33.197Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T20:24:33.197Z] [2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.failsafe.VerifyMojo.getBooterForkException (VerifyMojo.java:198)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.failsafe.VerifyMojo.execute (VerifyMojo.java:190)[2023-08-17T20:24:33.197Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2 (MojoExecutor.java:370)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:351)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:215)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:171)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:163)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:210)[2023-08-17T20:24:33.197Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:195)[2023-08-17T20:24:33.197Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T20:24:33.197Z] at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:515)[2023-08-17T20:24:33.197Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T20:24:33.197Z] at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1128)[2023-08-17T20:24:33.197Z] at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:628)[2023-08-17T20:24:33.197Z] at java.lang.Thread.run (Thread.java:829)[2023-08-17T20:24:33.197Z] [ERROR] [2023-08-17T20:24:33.197Z] [ERROR] Re-run Maven using the -X switch to enable full debug logging.[2023-08-17T20:24:33.197Z] [ERROR] [2023-08-17T20:24:33.197Z] [ERROR] For more information about the errors and possible solutions, please read the following articles:[2023-08-17T20:24:33.197Z] [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[2023-08-17T20:24:33.197Z] [ERROR] [2023-08-17T20:24:33.197Z] [ERROR] After correcting the problems, you can resume the build with the command[2023-08-17T20:24:33.197Z] [ERROR] mvn -rf :kogito-addons-quarkus-microprofile-config-service-catalog-integration-tests[2023-08-17T20:24:33.198Z] [INFO] kiegroup/kogito-runtimes failed. Won't execute remaining commands and projects[2023-08-17T20:24:33.198Z] [INFO] Execution summary for kiegroup/kogito-runtimes[2023-08-17T20:24:33.198Z] # [BEFORE] [kiegroup/kogito-runtimes] export INTEGRATION_BRANCH=[2023-08-17T20:24:33.198Z] [INFO] OK [Executed in 0.154279 ms][2023-08-17T20:24:33.198Z] [2023-08-17T20:24:33.198Z] # [BEFORE] [kiegroup/kogito-runtimes] bash -c "if [ ! -z '' ] && [ -f .ci/environments/update.sh ]; then .ci/environments/update.sh ; fi"[2023-08-17T20:24:33.198Z] [INFO] OK [Executed in 5.541521 ms][2023-08-17T20:24:33.198Z] [2023-08-17T20:24:33.198Z] # [COMMANDS] [kiegroup/kogito-runtimes] export MVN_CMD=`bash -c "if [ '' = 'true' ]; then printf 'deploy '; else printf 'install'; fi"`[2023-08-17T20:24:33.198Z] [INFO] OK [Executed in 4.813595 ms][2023-08-17T20:24:33.198Z] [2023-08-17T20:24:33.198Z] # [COMMANDS] [kiegroup/kogito-runtimes] mvn dependency:tree clean -Dfull install -s /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes@tmp/config11369415331975608143tmp -Dmaven.wagon.http.ssl.insecure=true -Dmaven.test.failure.ignore=true -nsu -ntp -fae -e -Dhttp.keepAlive=false -Dmaven.wagon.http.pool=false -Dmaven.wagon.httpconnectionManager.ttlSeconds=120 -Dmaven.wagon.http.retryHandler.count=3 -T 1C -Dvalidate-formatting -Prun-code-coverage -Dorg.slf4j.simpleLogger.log.org.apache.maven.cli.transfer.Slf4jMavenTransferListener=warn -B[2023-08-17T20:24:33.198Z] [INFO] NOT OK [Executed in 10509874.394501 ms][2023-08-17T20:24:33.198Z] [ERROR] The process '/opt/tools/apache-maven-3.8.7/bin/mvn' failed with exit code 143[2023-08-17T20:24:33.198Z] [2023-08-17T20:24:33.198Z] [INFO] [AFTER] No commands were found for kiegroup/kogito-runtimes[2023-08-17T20:24:33.198Z] [2023-08-17T20:24:33.198Z] # Uploading artifacts[2023-08-17T20:24:33.198Z] [INFO] Will not upload any artifacts in CLI environment[2023-08-17T20:24:33.198Z] [2023-08-17T20:24:33.198Z] [ERROR] Failed to execute commands for kiegroup/kogito-runtimes[2023-08-17T20:24:33.198Z] [ERROR] Failed to execute mvn dependency:tree clean -Dfull install -s /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes@tmp/config11369415331975608143tmp -Dmaven.wagon.http.ssl.insecure=true -Dmaven.test.failure.ignore=true -nsu -ntp -fae -e -Dhttp.keepAlive=false -Dmaven.wagon.http.pool=false -Dmaven.wagon.httpconnectionManager.ttlSeconds=120 -Dmaven.wagon.http.retryHandler.count=3 -T 1C -Dvalidate-formatting -Prun-code-coverage -Dorg.slf4j.simpleLogger.log.org.apache.maven.cli.transfer.Slf4jMavenTransferListener=warn -B :[2023-08-17T20:24:33.198Z] [ERROR] The process '/opt/tools/apache-maven-3.8.7/bin/mvn' failed with exit code 143[2023-08-17T20:24:33.458Z] script returned exit code 1[Pipeline] }[Pipeline] // withCredentials[Pipeline] }[2023-08-17T20:24:33.637Z] Deleting 1 temporary files[Pipeline] // configFileProvider[Pipeline] }[Pipeline] // scriptPost stage[Pipeline] script[Pipeline] {[Pipeline] sh[2023-08-17T20:24:34.117Z] + find . -type d -name node_modules -exec rm -rf '{}' ';'[Pipeline] junit[2023-08-17T20:24:38.285Z] Recording test results[2023-08-17T20:24:42.928Z] [Checks API] No suitable checks publisher found.[Pipeline] archiveArtifacts[2023-08-17T20:24:42.948Z] Archiving artifacts[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh[2023-08-17T20:24:44.446Z] + rm -rf console.log[Pipeline] sh[2023-08-17T20:24:44.742Z] + wget --no-check-certificate -qO - https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3032/consoleText[2023-08-17T20:24:44.742Z] + tail -n 300 (tests) - kogito-runtimes job #3034 was: ABORTED Possible explanation: Most probably a timeout, please review Reproducer export BUILD_MVN_OPTS_CURRENT=-T 1C -Dvalidate-formatting -Prun-code-coverage build-chain build cross_pr -f 'https://raw.githubusercontent.com/kiegroup/kogito-pipelines/main/.ci/pull-request-config.yaml' -o 'bc' -p kiegroup/kogito-runtimes -u https://github.com/kiegroup/kogito-runtimes/pull/3179 --skipParallelCheckout NOTE: To install the build-chain tool, please refer to https://github.com/kiegroup/github-action-build-chain#local-execution Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3034/display/redirect Test results: PASSED: 3356 FAILED: 0 Those are the test failures: none See console log: Console Logs [2023-08-17T22:22:42.727Z] [ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:3.1.2:verify (integration-test) on project kogito-addons-quarkus-microprofile-config-service-catalog-integration-tests: [2023-08-17T22:22:42.727Z] [ERROR] [2023-08-17T22:22:42.727Z] [ERROR] Please refer to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/failsafe-reports for the individual test results.[2023-08-17T22:22:42.727Z] [ERROR] Please refer to dump files (if any exist) [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.[2023-08-17T22:22:42.727Z] [ERROR] The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T22:22:42.728Z] [ERROR] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817154129763_383.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T15-28-12_486-jvmRun4' 'surefire-20230817154129763_381tmp' 'surefire_34-20230817154129763_382tmp'[2023-08-17T22:22:42.728Z] [ERROR] Error occurred in starting fork, check output in log[2023-08-17T22:22:42.728Z] [ERROR] Process Exit Code: 143[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T22:22:42.728Z] [ERROR] [2023-08-17T22:22:42.728Z] [ERROR] org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T22:22:42.728Z] [ERROR] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817154129763_383.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T15-28-12_486-jvmRun4' 'surefire-20230817154129763_381tmp' 'surefire_34-20230817154129763_382tmp'[2023-08-17T22:22:42.728Z] [ERROR] Error occurred in starting fork, check output in log[2023-08-17T22:22:42.728Z] [ERROR] Process Exit Code: 143[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T22:22:42.728Z] [ERROR] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T22:22:42.728Z] [ERROR] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T22:22:42.728Z] [ERROR] -> [Help 1][2023-08-17T22:22:42.728Z] org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:3.1.2:verify (integration-test) on project kogito-addons-quarkus-microprofile-config-service-catalog-integration-tests: [2023-08-17T22:22:42.728Z] [2023-08-17T22:22:42.728Z] Please refer to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/failsafe-reports for the individual test results.[2023-08-17T22:22:42.728Z] Please refer to dump files (if any exist) [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.[2023-08-17T22:22:42.728Z] The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T22:22:42.728Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817154129763_383.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T15-28-12_486-jvmRun4' 'surefire-20230817154129763_381tmp' 'surefire_34-20230817154129763_382tmp'[2023-08-17T22:22:42.728Z] Error occurred in starting fork, check output in log[2023-08-17T22:22:42.728Z] Process Exit Code: 143[2023-08-17T22:22:42.728Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T22:22:42.728Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T22:22:42.728Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T22:22:42.728Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T22:22:42.728Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T22:22:42.728Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T22:22:42.728Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T22:22:42.728Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T22:22:42.728Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T22:22:42.728Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T22:22:42.728Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T22:22:42.728Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T22:22:42.728Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T22:22:42.728Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T22:22:42.728Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T22:22:42.728Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.728Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T22:22:42.728Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.728Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T22:22:42.728Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T22:22:42.728Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T22:22:42.728Z] [2023-08-17T22:22:42.728Z] org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T22:22:42.728Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817154129763_383.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T15-28-12_486-jvmRun4' 'surefire-20230817154129763_381tmp' 'surefire_34-20230817154129763_382tmp'[2023-08-17T22:22:42.728Z] Error occurred in starting fork, check output in log[2023-08-17T22:22:42.728Z] Process Exit Code: 143[2023-08-17T22:22:42.728Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T22:22:42.728Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T22:22:42.728Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T22:22:42.729Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T22:22:42.729Z] [2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2 (MojoExecutor.java:375)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:351)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:215)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:171)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:163)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:210)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:195)[2023-08-17T22:22:42.729Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T22:22:42.729Z] at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:515)[2023-08-17T22:22:42.729Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T22:22:42.729Z] at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1128)[2023-08-17T22:22:42.729Z] at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:628)[2023-08-17T22:22:42.729Z] at java.lang.Thread.run (Thread.java:829)[2023-08-17T22:22:42.729Z] Caused by: org.apache.maven.plugin.MojoExecutionException: [2023-08-17T22:22:42.729Z] [2023-08-17T22:22:42.729Z] Please refer to /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/failsafe-reports for the individual test results.[2023-08-17T22:22:42.729Z] Please refer to dump files (if any exist) [date].dump, [date]-jvmRun[N].dump and [date].dumpstream.[2023-08-17T22:22:42.729Z] The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T22:22:42.729Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817154129763_383.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T15-28-12_486-jvmRun4' 'surefire-20230817154129763_381tmp' 'surefire_34-20230817154129763_382tmp'[2023-08-17T22:22:42.729Z] Error occurred in starting fork, check output in log[2023-08-17T22:22:42.729Z] Process Exit Code: 143[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T22:22:42.729Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T22:22:42.729Z] [2023-08-17T22:22:42.729Z] org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T22:22:42.729Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817154129763_383.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T15-28-12_486-jvmRun4' 'surefire-20230817154129763_381tmp' 'surefire_34-20230817154129763_382tmp'[2023-08-17T22:22:42.729Z] Error occurred in starting fork, check output in log[2023-08-17T22:22:42.729Z] Process Exit Code: 143[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T22:22:42.729Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T22:22:42.729Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T22:22:42.729Z] [2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.surefire.SurefireHelper.reportExecution (SurefireHelper.java:154)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.failsafe.VerifyMojo.execute (VerifyMojo.java:190)[2023-08-17T22:22:42.729Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2 (MojoExecutor.java:370)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:351)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:215)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:171)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:163)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:210)[2023-08-17T22:22:42.729Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:195)[2023-08-17T22:22:42.729Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T22:22:42.729Z] at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:515)[2023-08-17T22:22:42.729Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T22:22:42.730Z] at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1128)[2023-08-17T22:22:42.730Z] at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:628)[2023-08-17T22:22:42.730Z] at java.lang.Thread.run (Thread.java:829)[2023-08-17T22:22:42.730Z] Caused by: org.apache.maven.surefire.booter.SurefireBooterForkException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called?[2023-08-17T22:22:42.730Z] Command was /bin/sh -c cd '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests' && '/usr/lib/jvm/java-11-openjdk-11.0.20.0.8-2.el8.x86_64/bin/java' '-Xms2048m' '-Xmx4g' '-XX:-TieredCompilation' '-XX:TieredStopAtLevel=1' '-Xverify:none' '-jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire/surefirebooter-20230817154129763_383.jar' '/home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes/bc/kiegroup_kogito-runtimes/quarkus/addons/microprofile-config-service-catalog/integration-tests/target/surefire' '2023-08-17T15-28-12_486-jvmRun4' 'surefire-20230817154129763_381tmp' 'surefire_34-20230817154129763_382tmp'[2023-08-17T22:22:42.730Z] Error occurred in starting fork, check output in log[2023-08-17T22:22:42.730Z] Process Exit Code: 143[2023-08-17T22:22:42.730Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.fork(ForkStarter.java:643)[2023-08-17T22:22:42.730Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:285)[2023-08-17T22:22:42.730Z] at org.apache.maven.plugin.surefire.booterclient.ForkStarter.run(ForkStarter.java:250)[2023-08-17T22:22:42.730Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeProvider(AbstractSurefireMojo.java:1203)[2023-08-17T22:22:42.730Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.executeAfterPreconditionsChecked(AbstractSurefireMojo.java:1055)[2023-08-17T22:22:42.730Z] at org.apache.maven.plugin.surefire.AbstractSurefireMojo.execute(AbstractSurefireMojo.java:871)[2023-08-17T22:22:42.730Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:137)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2(MojoExecutor.java:370)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute(MojoExecutor.java:351)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:215)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:171)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:163)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:117)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:210)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call(MultiThreadedBuilder.java:195)[2023-08-17T22:22:42.730Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.730Z] at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)[2023-08-17T22:22:42.730Z] at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)[2023-08-17T22:22:42.730Z] at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)[2023-08-17T22:22:42.730Z] at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)[2023-08-17T22:22:42.730Z] at java.base/java.lang.Thread.run(Thread.java:829)[2023-08-17T22:22:42.730Z] [2023-08-17T22:22:42.730Z] at org.apache.maven.plugin.failsafe.VerifyMojo.getBooterForkException (VerifyMojo.java:198)[2023-08-17T22:22:42.730Z] at org.apache.maven.plugin.failsafe.VerifyMojo.execute (VerifyMojo.java:190)[2023-08-17T22:22:42.730Z] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo (DefaultBuildPluginManager.java:137)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute2 (MojoExecutor.java:370)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.MojoExecutor.doExecute (MojoExecutor.java:351)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:215)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:171)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.MojoExecutor.execute (MojoExecutor.java:163)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject (LifecycleModuleBuilder.java:117)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:210)[2023-08-17T22:22:42.730Z] at org.apache.maven.lifecycle.internal.builder.multithreaded.MultiThreadedBuilder$1.call (MultiThreadedBuilder.java:195)[2023-08-17T22:22:42.730Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T22:22:42.730Z] at java.util.concurrent.Executors$RunnableAdapter.call (Executors.java:515)[2023-08-17T22:22:42.730Z] at java.util.concurrent.FutureTask.run (FutureTask.java:264)[2023-08-17T22:22:42.730Z] at java.util.concurrent.ThreadPoolExecutor.runWorker (ThreadPoolExecutor.java:1128)[2023-08-17T22:22:42.730Z] at java.util.concurrent.ThreadPoolExecutor$Worker.run (ThreadPoolExecutor.java:628)[2023-08-17T22:22:42.730Z] at java.lang.Thread.run (Thread.java:829)[2023-08-17T22:22:42.730Z] [ERROR] [2023-08-17T22:22:42.730Z] [ERROR] Re-run Maven using the -X switch to enable full debug logging.[2023-08-17T22:22:42.730Z] [ERROR] [2023-08-17T22:22:42.730Z] [ERROR] For more information about the errors and possible solutions, please read the following articles:[2023-08-17T22:22:42.730Z] [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException[2023-08-17T22:22:42.730Z] [ERROR] [2023-08-17T22:22:42.730Z] [ERROR] After correcting the problems, you can resume the build with the command[2023-08-17T22:22:42.730Z] [ERROR] mvn -rf :kogito-addons-quarkus-microprofile-config-service-catalog-integration-tests[2023-08-17T22:22:42.730Z] [INFO] kiegroup/kogito-runtimes failed. Won't execute remaining commands and projects[2023-08-17T22:22:42.730Z] [INFO] Execution summary for kiegroup/kogito-runtimes[2023-08-17T22:22:42.730Z] # [BEFORE] [kiegroup/kogito-runtimes] export INTEGRATION_BRANCH=[2023-08-17T22:22:42.730Z] [INFO] OK [Executed in 0.17045 ms][2023-08-17T22:22:42.730Z] [2023-08-17T22:22:42.730Z] # [BEFORE] [kiegroup/kogito-runtimes] bash -c "if [ ! -z '' ] && [ -f .ci/environments/update.sh ]; then .ci/environments/update.sh ; fi"[2023-08-17T22:22:42.730Z] [INFO] OK [Executed in 7.033159 ms][2023-08-17T22:22:42.730Z] [2023-08-17T22:22:42.730Z] # [COMMANDS] [kiegroup/kogito-runtimes] export MVN_CMD=`bash -c "if [ '' = 'true' ]; then printf 'deploy '; else printf 'install'; fi"`[2023-08-17T22:22:42.730Z] [INFO] OK [Executed in 6.895191 ms][2023-08-17T22:22:42.730Z] [2023-08-17T22:22:42.730Z] # [COMMANDS] [kiegroup/kogito-runtimes] mvn dependency:tree clean -Dfull install -s /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes@tmp/config269414667400062756tmp -Dmaven.wagon.http.ssl.insecure=true -Dmaven.test.failure.ignore=true -nsu -ntp -fae -e -Dhttp.keepAlive=false -Dmaven.wagon.http.pool=false -Dmaven.wagon.httpconnectionManager.ttlSeconds=120 -Dmaven.wagon.http.retryHandler.count=3 -T 1C -Dvalidate-formatting -Prun-code-coverage -Dorg.slf4j.simpleLogger.log.org.apache.maven.cli.transfer.Slf4jMavenTransferListener=warn -B[2023-08-17T22:22:42.730Z] [INFO] NOT OK [Executed in 10487094.659419 ms][2023-08-17T22:22:42.730Z] [ERROR] The process '/opt/tools/apache-maven-3.8.7/bin/mvn' failed with exit code 143[2023-08-17T22:22:42.730Z] [2023-08-17T22:22:42.730Z] [INFO] [AFTER] No commands were found for kiegroup/kogito-runtimes[2023-08-17T22:22:42.730Z] [2023-08-17T22:22:42.730Z] # Uploading artifacts[2023-08-17T22:22:42.730Z] [INFO] Will not upload any artifacts in CLI environment[2023-08-17T22:22:42.730Z] [2023-08-17T22:22:42.730Z] [ERROR] Failed to execute commands for kiegroup/kogito-runtimes[2023-08-17T22:22:42.730Z] [ERROR] Failed to execute mvn dependency:tree clean -Dfull install -s /home/jenkins/workspace/KIE/kogito/main/pullrequest/kogito-runtimes.tests.kogito-runtimes@tmp/config269414667400062756tmp -Dmaven.wagon.http.ssl.insecure=true -Dmaven.test.failure.ignore=true -nsu -ntp -fae -e -Dhttp.keepAlive=false -Dmaven.wagon.http.pool=false -Dmaven.wagon.httpconnectionManager.ttlSeconds=120 -Dmaven.wagon.http.retryHandler.count=3 -T 1C -Dvalidate-formatting -Prun-code-coverage -Dorg.slf4j.simpleLogger.log.org.apache.maven.cli.transfer.Slf4jMavenTransferListener=warn -B :[2023-08-17T22:22:42.730Z] [ERROR] The process '/opt/tools/apache-maven-3.8.7/bin/mvn' failed with exit code 143[2023-08-17T22:22:42.735Z] script returned exit code 1[Pipeline] }[Pipeline] // withCredentials[Pipeline] }[2023-08-17T22:22:42.813Z] Deleting 1 temporary files[Pipeline] // configFileProvider[Pipeline] }[Pipeline] // scriptPost stage[Pipeline] script[Pipeline] {[Pipeline] sh[2023-08-17T22:22:43.223Z] + find . -type d -name node_modules -exec rm -rf '{}' ';'[Pipeline] junit[2023-08-17T22:22:48.473Z] Recording test results[2023-08-17T22:22:53.438Z] [Checks API] No suitable checks publisher found.[Pipeline] archiveArtifacts[2023-08-17T22:22:53.495Z] Archiving artifacts[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh[2023-08-17T22:22:55.119Z] + rm -rf console.log[Pipeline] sh[2023-08-17T22:22:55.407Z] + wget --no-check-certificate -qO - https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3034/consoleText[2023-08-17T22:22:55.407Z] + tail -n 300 jenkins rerun kogito-runtimes tests (tests) - kogito-runtimes job #3036 was: UNSTABLE Possible explanation: This should be test failures Reproducer export BUILD_MVN_OPTS_CURRENT=-T 1C -Dvalidate-formatting -Prun-code-coverage build-chain build cross_pr -f 'https://raw.githubusercontent.com/kiegroup/kogito-pipelines/main/.ci/pull-request-config.yaml' -o 'bc' -p kiegroup/kogito-runtimes -u https://github.com/kiegroup/kogito-runtimes/pull/3179 --skipParallelCheckout NOTE: To install the build-chain tool, please refer to https://github.com/kiegroup/github-action-build-chain#local-execution Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3036/display/redirect Test results: PASSED: 3356 FAILED: 1 Those are the test failures: org.kie.kogito.addons.quarkus.microprofile.config.service.catalog.it.MicroProfileConfigServiceAddonIT.executeWithEmptyParameters java.lang.RuntimeException: io.quarkus.builder.BuildException: Build failure: Build failed due to errors [error]: Build step io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor#setupKubernetesDevService threw an exception: java.lang.RuntimeException: org.testcontainers.containers.ContainerLaunchException: Container startup failed at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:119) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at io.quarkus.deployment.ExtensionLoader$3.execute(ExtensionLoader.java:909) at io.quarkus.builder.BuildContext.run(BuildContext.java:281) at org.jboss.threads.ContextHandler$1.runWith(ContextHandler.java:18) at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2449) at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1478) at java.base/java.lang.Thread.run(Thread.java:829) at org.jboss.threads.JBossThread.run(JBossThread.java:501)Caused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.lambda$startKubernetes$7(DevServicesKubernetesProcessor.java:227) at java.base/java.util.Optional.orElseGet(Optional.java:369) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.startKubernetes(DevServicesKubernetesProcessor.java:241) at io.quarkus.kubernetes.client.deployment.DevServicesKubernetesProcessor.setupKubernetesDevService(DevServicesKubernetesProcessor.java:109) ... 11 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 16 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 17 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Container startup failed at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:349) at org.testcontainers.containers.GenericContainer.start(GenericContainer.java:322) at com.dajudge.kindcontainer.ApiServerContainer.containerIsStarting(ApiServerContainer.java:142) at org.testcontainers.containers.GenericContainer.containerIsStarting(GenericContainer.java:692) at com.dajudge.kindcontainer.KubernetesContainer.containerIsStarting(KubernetesContainer.java:151) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:474) ... 19 moreCaused by: org.rnorth.ducttape.RetryCountExceededException: Retry limit hit with exception at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:88) at org.testcontainers.containers.GenericContainer.doStart(GenericContainer.java:334) ... 24 moreCaused by: org.testcontainers.containers.ContainerLaunchException: Could not create/start container at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:542) at org.testcontainers.containers.GenericContainer.lambda$doStart$0(GenericContainer.java:344) at org.rnorth.ducttape.unreliables.Unreliables.retryUntilSuccess(Unreliables.java:81) ... 25 moreCaused by: com.github.dockerjava.api.exception.ConflictException: Status 409: {"message":"cannot join network of a non running container: 8a347daa4f468501b5e7e199011ba9fb4e45bcba8a7f07efb42215e74526b09f"} at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.execute(DefaultInvocationBuilder.java:245) at org.testcontainers.shaded.com.github.dockerjava.core.DefaultInvocationBuilder.post(DefaultInvocationBuilder.java:102) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:31) at org.testcontainers.shaded.com.github.dockerjava.core.exec.StartContainerCmdExec.execute(StartContainerCmdExec.java:13) at org.testcontainers.shaded.com.github.dockerjava.core.exec.AbstrSyncDockerCmdExec.exec(AbstrSyncDockerCmdExec.java:21) at org.testcontainers.shaded.com.github.dockerjava.core.command.AbstrDockerCmd.exec(AbstrDockerCmd.java:35) at org.testcontainers.shaded.com.github.dockerjava.core.command.StartContainerCmdImpl.exec(StartContainerCmdImpl.java:43) at org.testcontainers.containers.GenericContainer.tryStart(GenericContainer.java:441) ... 27 more (tests) - kogito-runtimes job #3037 was: UNSTABLE Possible explanation: This should be test failures Reproducer export BUILD_MVN_OPTS_CURRENT=-T 1C -Dvalidate-formatting -Prun-code-coverage build-chain build cross_pr -f 'https://raw.githubusercontent.com/kiegroup/kogito-pipelines/main/.ci/pull-request-config.yaml' -o 'bc' -p kiegroup/kogito-runtimes -u https://github.com/kiegroup/kogito-runtimes/pull/3179 --skipParallelCheckout NOTE: To install the build-chain tool, please refer to https://github.com/kiegroup/github-action-build-chain#local-execution Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3037/display/redirect Test results: PASSED: 3356 FAILED: 1 Those are the test failures: org.kie.kogito.persistence.kafka.KafkaProcessInstancesIT.testFindByIdReadMode expected: 5 but was: 0 Jenkins rerun kogito-runtimes tests (tests) - kogito-runtimes job #3038 was: UNSTABLE Possible explanation: This should be test failures Reproducer export BUILD_MVN_OPTS_CURRENT=-T 1C -Dvalidate-formatting -Prun-code-coverage build-chain build cross_pr -f 'https://raw.githubusercontent.com/kiegroup/kogito-pipelines/main/.ci/pull-request-config.yaml' -o 'bc' -p kiegroup/kogito-runtimes -u https://github.com/kiegroup/kogito-runtimes/pull/3179 --skipParallelCheckout NOTE: To install the build-chain tool, please refer to https://github.com/kiegroup/github-action-build-chain#local-execution Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3038/display/redirect Test results: PASSED: 3356 FAILED: 1 Those are the test failures: org.jbpm.process.TimerTest.testTimer expected: 4 but was: 5 Jenkins rerun kogito-runtimes tests (tests) - kogito-runtimes job #3039 was: UNSTABLE Possible explanation: This should be test failures Reproducer export BUILD_MVN_OPTS_CURRENT=-T 1C -Dvalidate-formatting -Prun-code-coverage build-chain build cross_pr -f 'https://raw.githubusercontent.com/kiegroup/kogito-pipelines/main/.ci/pull-request-config.yaml' -o 'bc' -p kiegroup/kogito-runtimes -u https://github.com/kiegroup/kogito-runtimes/pull/3179 --skipParallelCheckout NOTE: To install the build-chain tool, please refer to https://github.com/kiegroup/github-action-build-chain#local-execution Please look here: https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/job/KIE/job/kogito/job/main/job/pullrequest/job/kogito-runtimes.tests.kogito-runtimes/3039/display/redirect Test results: PASSED: 3355 FAILED: 2 Those are the test failures: org.kie.kogito.persistence.kafka.KafkaProcessInstancesIT.testFindByIdReadMode expected: 5 but was: 0 org.kie.kogito.it.KafkaPersistenceIT.testMultipleEmbeddedInstance 1 expectation failed.Expected status code <404> but was <200>. Jenkins rerun kogito-runtimes tests All green @ricardozanini.
gharchive/pull-request
2023-08-17T13:00:51
2025-04-01T06:39:17.406883
{ "authors": [ "hbelmiro", "kie-ci3" ], "repo": "kiegroup/kogito-runtimes", "url": "https://github.com/kiegroup/kogito-runtimes/pull/3179", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2137959334
ECG Data Nice job on this app! I wanted to ask if you have plans to incorporate the ECG characteristics in the future, as I see the UUIDs are called out in the .py file. Thanks for any info! @zanovis you'll see now I'm connecting to the sensors with a python package I've written that also has support for ECG https://github.com/kieranabrennan/blehrm/blob/master/blehrm/clients/polar_h10.py
gharchive/issue
2024-02-16T06:29:35
2025-04-01T06:39:17.438942
{ "authors": [ "kieranabrennan", "zanovis" ], "repo": "kieranabrennan/every-breath-you-take", "url": "https://github.com/kieranabrennan/every-breath-you-take/issues/21", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2631004400
Slower execution when compared with Pytorch 2.1.2 + CU 11.8 + xformers 0.0.23.post1 I just updated my environment to 2.5.0 + CU 12.4 + xformers 0.0.28.post2, and it seems that it is slower and if I have a few images to loads, it will stop for a while before processing next image (at Florence2Run node). In the past, processing 10 images took about 40 secs now it needs 100 secs. Are there anything I can do to speed up? Issue gone after updating to Pytorch 2.5.1 + CU 12.4 (xformers 0.0.28.post3)
gharchive/issue
2024-11-03T06:20:25
2025-04-01T06:39:17.448821
{ "authors": [ "Orenji-Tangerine" ], "repo": "kijai/ComfyUI-Florence2", "url": "https://github.com/kijai/ComfyUI-Florence2/issues/92", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
150935779
Fix verify_signature in Python 3 We were not properly accounting for the fact that the hashing functions in Python 3 take bytestrings. Note that in Python 3, the body of the HTML request must be a bytestring when passed into the verify_signature function, this is consistent with how Flask and other HTTP libraries work. Fixes #14 @mmailhot :+1:
gharchive/pull-request
2016-04-25T17:47:29
2025-04-01T06:39:17.451037
{ "authors": [ "cgurnik", "mmailhot" ], "repo": "kikinteractive/kik-python", "url": "https://github.com/kikinteractive/kik-python/pull/16", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1692176185
Failed to start DB cause it is stopped - ionic angular there is error showing on ionic angular with the latest version not able to connect the Database. by doing let isStopped = false; in nativeIonicBackend.ts file error is resolve let isStopped = true; database is not started yet ` async initialize() { if (isStopped) throw new Error("Failed to start DB- cause it is stopped"); db = await SQLite.create({ name: dbName, location: "default", }); if (isStopped) { await db.close(); } }, ` is isStopped is always true so its not creating database I faced the same issue just now, using tauri-backend. Seems like nothing is setting it to false 🤷 i have made a service in angular ` import { Injectable } from '@angular/core'; import { SQLite, SQLiteObject } from '@awesome-cordova-plugins/sqlite/ngx'; import { buildAsyncQueryRunner, getTime, IDbBackend, initJobsState, IPrimitiveValue, IQuery, IQueryResult, IRunRes, ITransactionOpts, } from "@kikko-land/kikko"; @Injectable({ providedIn: 'root' }) export class SqliteService { constructor( public sqlite:SQLite,) { } async ionicBackend(dbName:string) { let sqlDB = await this.sqlite.create({ name:`${dbName}.db`, location: 'default', }); // console.log('db',JSON.stringify(sqlDB)); return () => { let isStopped = false; let db: SQLiteObject | undefined = undefined; const jobsState = initJobsState(); const runQueries = async ( queries: | { type: "usual"; values: IQuery[] } | { type: "prepared"; query: IQuery; preparedValues: IPrimitiveValue[][]; } ): Promise<IRunRes[]> => { if (!db) { throw new Error(`Failed to run queries, db not initialized`); } const res: IRunRes[] = []; const queriesToRun = queries.type === "usual" ? queries.values : queries.preparedValues.map( (v): IQuery => ({ text: queries.query.text, values: v, }) ); for (const q of queriesToRun) { const startTime = getTime(); // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment const execResult = await (async () => { try { // eslint-disable-next-line @typescript-eslint/no-unsafe-return return await db.executeSql(q.text, q.values); } catch (e) { if (e instanceof Error) { e.message = `Error while executing query: ${q.text} - ${e.message}`; } throw e; } })(); // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment, @typescript-eslint/no-unsafe-member-access const rows: IQueryResult = new Array(execResult.rows.length); // eslint-disable-next-line @typescript-eslint/no-unsafe-member-access for (let i = 0; i < execResult.rows.length; i++) { // eslint-disable-next-line @typescript-eslint/no-unsafe-assignment, @typescript-eslint/no-unsafe-call, @typescript-eslint/no-unsafe-member-access rows[i] = execResult.rows.item(i); } const end = getTime(); res.push({ rows: rows, performance: { execTime: end - startTime, }, }); } return res; }; const queryRunner = buildAsyncQueryRunner({ async execPrepared(query: IQuery, preparedValues: IPrimitiveValue[][]) { return await runQueries({ type: "prepared", query, preparedValues }); }, async execUsualBatch(queriesToRun: IQuery[]): Promise<IRunRes[]> { return await runQueries({ type: "usual", values: queriesToRun }); }, async rollback() { if (!db) { throw new Error(`Failed to run queries, db not initialized`); } await db.executeSql("ROLLBACK"); }, }); return { async initialize() { if (isStopped) throw new Error("Failed to start DB - cause it is stopped"); db = sqlDB; if (isStopped) { await db.close(); } }, async execQueries( q: | { type: "usual"; values: IQuery[] } | { type: "prepared"; query: IQuery; preparedValues: IPrimitiveValue[][]; }, transactionOpts?: ITransactionOpts ) { const startedAt = getTime(); const res = await queryRunner.run(jobsState, q, transactionOpts); const endAt = getTime(); return { ...res, performance: { ...res.performance, totalTime: endAt - startedAt, }, }; }, async stop() { isStopped = true; if (db) { await db.close(); } }, }; }; } } ` and in start function ` let ionicBackend = await this.sqliteKikko.ionicBackend(this.dbName); this.sqliteDb = ionicBackend().initialize(); console.log("sqlite", this.sqliteDb); console.log("db_start"); await initDbClient({ dbName: this.dbName, dbBackend: ionicBackend , plugins: [] }).then(async (dbConnect)=>{ this.kikkoDb = dbConnect ; }).catch(e => console.log("db",e)); ` hope it help
gharchive/issue
2023-05-02T10:34:12
2025-04-01T06:39:17.460266
{ "authors": [ "pioneersingh321", "valstu" ], "repo": "kikko-land/kikko", "url": "https://github.com/kikko-land/kikko/issues/161", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1617231083
Deep flat-file convention causing wrong route added If we add something like this - a+ - b+ - index.tsx This would generate route for a+/b+ and a/b at the same time. Odd. The + are removed and treated like you used . a+/b+/index.tsx should map to a.b.index.tsx Can you show me the remix.config? The content of remix.config const { flatRoutes } = require("remix-flat-routes"); /** @type {import('@remix-run/dev').AppConfig} */ module.exports = { devServerBroadcastDelay: 1000, ignoredRouteFiles: ["**/.*"], server: "./server.js", serverBuildPath: "functions/[[path]].js", serverConditions: ["worker"], serverDependenciesToBundle: "all", serverMainFields: ["browser", "module", "main"], serverMinify: true, serverModuleFormat: "esm", serverPlatform: "neutral", future: { unstable_tailwind: true, unstable_postcss: true, v2_routeConvention: true, }, routes: (defineRoutes) => { return flatRoutes("routes", defineRoutes); }, }; My actual routes looks like this <Route path="admin+/users+" file="routes/admin+/users+/index.tsx" /> <Route path="admin" file="routes/admin+/_layout.tsx"> <Route path="users/edit/:id?" file="routes/admin+/users+/edit.($id).tsx" /> <Route path="users/" index file="routes/admin+/users+/index.tsx" /> <Route path="users/new" file="routes/admin+/users+/new.tsx" /> </Route> Ah, you can't use v2_routeConvention and remix-flat-routes at the same time. v2 doesn't support the + convention. I see. I thought the routes function would take over the route generation. Thanks for the explanation. You are right! Change that fix the issue
gharchive/issue
2023-03-09T13:15:57
2025-04-01T06:39:17.466525
{ "authors": [ "kiliman", "zhuhaow" ], "repo": "kiliman/remix-flat-routes", "url": "https://github.com/kiliman/remix-flat-routes/issues/43", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1878608041
Index was out of range error Hello Getting this error every time I open the program, clicking the "Apply" button or finishing quick setting. As the result, settings are not applying. Pic 2 is coming after the first one (it's saying application is shutting down) but nothing happens. By the way russian translate is bad and breaking the interface, how can I change it to english? hello It seems like the app settings are messed up. Do the following to reset and try again. Additional -> Reset config change language Additional -> Advanced settings -> App language Resetting config helped, thanks. Its weird, because it was my 1st launch and already with broken config. And that "Config file" submenu in russian looks like file setup/configuration, and the buttons are "Setting", "Save settings" and "reload". It seems that a problem occurred during the initialization process. Thank you for reporting the worng text! I'll fix it
gharchive/issue
2023-09-02T10:52:24
2025-04-01T06:39:17.471053
{ "authors": [ "Dmitryqq", "killkimno" ], "repo": "killkimno/MORT", "url": "https://github.com/killkimno/MORT/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2502675703
Updating stuff for deployment outside initial plant Redo : install script update script correct bugs at startup Update globals variable and their use in email and other stuff.
gharchive/issue
2024-09-03T11:52:41
2025-04-01T06:39:17.473768
{ "authors": [ "kiloutyg" ], "repo": "kiloutyg/efnc", "url": "https://github.com/kiloutyg/efnc/issues/100", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1110490075
Documentation of setup for BioBERT and neleval First of all, thanks for sharing this repo! However, I have a bit of trouble reproducing the results, maybe you can help. Running the training script python run_e2e_span.py --data_dir data/BC5CDR/processed_data --model_type bert --output_dir o utput --do_train --n_gpu 1 --model_name_or_path ./biobert_v1.1_pubmed yields different errors. If I use the BioBERT checkpoint from here https://github.com/dmis-lab/biobert: I get the error: OSError: file ./biobert_v1.1_pubmed/config.json not found If, however, I use this checkpoint https://huggingface.co/dmis-lab/biobert-v1.1/tree/main I'm getting OSError: Error no file named ['pytorch_model-1000000.bin', 'tf_model.h5', 'model.ckpt.index'] found in directory ./biobert-v1.1/ or from_tf set to False neleval The Readme states that evaluation is run with ./neleval/BC5CDR-AllSpan/gold.csv. How do I get this file? Is it some produced as part of running the e2e_span script? Thanks in advance! Hi, Thanks for your interest in our paper. If you are using https://github.com/dmis-lab/biobert, please rename the bert_config.json file in ./biobert_v1.1_pubmed to config.json. Alternatively, you can specify the path to the config file using the --config_name argument in the training script. If you are using https://huggingface.co/dmis-lab/biobert-v1.1/tree/main, please rename pytorch_model.bin to pytorch_model-1000000.bin. gold.csv will be created when you do inference on the trained model. Use the --do_eval argument to do so. Then copy the gold.csv file to a directory named BC5CDR-AllSpan. I hope this is helpful. Thank you so much! Regarding 1. I tried the second option (using the HuggingFace model) and actually progressed a bit further. However, now I am getting an error 01/25/2022 12:12:17 - INFO - __main__ - Loading features from cached file data/BC5CDR/processed_data/cached_train_biobert-v1.1 Traceback (most recent call last): File "run_e2e_span.py", line 1071, in <module> main() File "run_e2e_span.py", line 1021, in main global_step, tr_loss = train(args, model, tokenizer) File "run_e2e_span.py", line 76, in train train_dataset, _, _ = load_and_cache_examples(args, tokenizer) File "run_e2e_span.py", line 725, in load_and_cache_examples all_mention_start_indices = torch.tensor([f.mention_start_indices for f in features], dtype=torch.long) ValueError: expected sequence of length 16 at dim 1 (got 9) Do you have any hints what might be the problem? use --overwrite_output_dir, --overwrite_cache, and --use_hard_and_random_negative during training. During inference use --use_all_candidates
gharchive/issue
2022-01-21T13:47:54
2025-04-01T06:39:17.529691
{ "authors": [ "kingsaint", "phlobo" ], "repo": "kingsaint/BioMedical-EL", "url": "https://github.com/kingsaint/BioMedical-EL/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2197903816
fix(vite.config): mdi icon convert camelCase to kebab-case When the previous solution encounters mdiBattery80, it will be converted to mdi-battery-8-0. When using mdi-battery-80, an error that the icon cannot be found will be prompted. This solution first converts uppercase letters and then matches one or more numbers. There are some weird names here, such as mdiBattery10Bluetooth. : ) Thank you very much!
gharchive/pull-request
2024-03-20T15:56:49
2025-04-01T06:39:17.531539
{ "authors": [ "kingyue737", "xj63" ], "repo": "kingyue737/vitify-next", "url": "https://github.com/kingyue737/vitify-next/pull/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2142851072
terminal: pasting causes cursor to go to end of line E.g. imagine I have a line foo baz and I have my cursor on the space and bar in the clipboard. Pasting will properly modify the line, but will move my cursor to the end (i.e. to the z in baz), rather than, as expected, the end of the paste (the space in bar ). FYI @tadad @dr-frmr (not sure who owns this part of terminal). fixed in develop!
gharchive/issue
2024-02-19T17:20:53
2025-04-01T06:39:17.541160
{ "authors": [ "dr-frmr", "hosted-fornet", "nick1udwig" ], "repo": "kinode-dao/kinode", "url": "https://github.com/kinode-dao/kinode/issues/259", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
672253572
Some improvements for network policy advisor Align "--namespace" and "--namespaces" flags Add --output option for report comman @alban I think these changes are still relevant.
gharchive/pull-request
2020-08-03T18:10:48
2025-04-01T06:39:17.544550
{ "authors": [ "mauriciovasquezbernal" ], "repo": "kinvolk/inspektor-gadget", "url": "https://github.com/kinvolk/inspektor-gadget/pull/126", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
255237321
Support overlayfs on btrfs after moby's new release Currently we don't support btrfs as underlying device's filesystem, mainly because moby doesn't allow it. Linux kernel 4.7 or newer already supports overlayfs on btrfs, but moby started to support since https://github.com/moby/moby/pull/33281 (2017-05-19). The commit has not been included in any release so far. Even the most recent release 17.05.0-ce doesn't have it. So we need to wait for a while until moby released a new version with it, as well as Container Linux includes the new version. Then we can allow overlayfs on btrfs. /cc @iaguis fixed by #193 Right. Thanks!
gharchive/issue
2017-09-05T10:53:57
2025-04-01T06:39:17.546756
{ "authors": [ "dongsupark", "jonboulle" ], "repo": "kinvolk/kube-spawn", "url": "https://github.com/kinvolk/kube-spawn/issues/130", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
244623298
index.php hi thank for sharing good source but i am running this code if am facing index.php problem each and every open page from admin. http://localhost/shopping/admin/publish Problem http://localhost/shopping/index.php/admin/publish right @vikaskumarmca This is not a problem. You can load admin/publish with index.php before and without :).
gharchive/issue
2017-07-21T10:03:59
2025-04-01T06:39:17.562037
{ "authors": [ "kirilkirkov", "vikaskumarmca" ], "repo": "kirilkirkov/Shopping-Cart-Solution-CodeIgniter", "url": "https://github.com/kirilkirkov/Shopping-Cart-Solution-CodeIgniter/issues/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
202809391
DropperSelectedRow never called i followed the instructions to the word and it didn't work then i tried some other variations and it didn't work as well no need for code because i'm doing what you wrote the framework looks great! i really hope to use it thanks :) Hey! I created a new project and followed my instructions again, but I wasn't able to replicate the issue you are having. Here is my View controller: class ViewController: UIViewController { let dropper = Dropper(width: 75, height: 200) @IBOutlet weak var dropButton: UIButton! override func viewDidLoad() { super.viewDidLoad() } override func didReceiveMemoryWarning() { super.didReceiveMemoryWarning() // Dispose of any resources that can be recreated. } @IBAction func buttonSelected() { if dropper.status == .hidden { dropper.items = ["Item 1", "Item 2", "Item 3", "Item 4"] // Item displayed dropper.theme = Dropper.Themes.white dropper.delegate = self dropper.cornerRadius = 3 dropper.showWithAnimation(0.15, options: Dropper.Alignment.center, button: dropButton) } else { dropper.hideWithAnimation(0.1) } } } extension ViewController: DropperDelegate { func DropperSelectedRow(_ path: IndexPath, contents: String) { print(path) print(contents) } }
gharchive/issue
2017-01-24T12:49:28
2025-04-01T06:39:17.564105
{ "authors": [ "kirkbyo", "ocnur" ], "repo": "kirkbyo/Dropper", "url": "https://github.com/kirkbyo/Dropper/issues/9", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1127678924
Transport auto doesn't run value through sanitize_callback Using 4.0.20 I have a text field where users can put in a number or a CSS value like 30px, 2rem, 2em, etc. If an integer is added, my sanitize_callback adds px at the end. Using transport => auto just uses the integer value in the CSS. This may be a product of how transport auto works in JS only, but I wanted to confirm. Hi @JiveDig , thanks for confirming this. From what I understand, the postMessage CSS output will output the value based on: how the control handle the value and then displaying it to the markup via content_template method in PHP or via JS part of the control. what the control prints into the JS object (via to_json() method in PHP) how the JS part of the control handles that object and how the JS part of the control handles the value returned from the customizer In your use-case, let's say you enter 11 to the text input. Your custom sanitize_callback will add px to it before it's saved to the database. But the control still seeing the value as 11 instead of 11px. In this case, you would need a custom JS to handle this in order to make the CSS output generated by postMessage works as expected. You might already know, the filter name is kirkiPostMessageStylesOutput. You might also already have the script for that. So, this is just an example to do that: (function () { /** * Check if the provided value is a numeric. * * @see https://stackoverflow.com/questions/175739/built-in-way-in-javascript-to-check-if-a-string-is-a-valid-number#answer-175787 * * @param {string|number} str The provided value. * @return bool */ function isNumeric(str) { // Number is a numeric. if ("number" === typeof str) return true; // We only process strings. if ("string" !== typeof str) return false; // Use type coercion to parse the entirety of the string (`parseFloat` alone does not do this) and ensure strings of whitespace fail. return !isNaN(str) && !isNaN(parseFloat(str)); } /** * Function to hook into `kirkiPostMessageStylesOutput` filter. * * @param {string} styles The styles to be filtered. * @param {string|Object|int} value The control's value. * @param {Object} output The control's output argument. * @param {string} controlType The control type. * * @return {string} The filtered styles. */ function stylesOutput(styles, value, output, controlType) { // These checks are just example :). if ("kirki-generic" !== controlType) return styles; if (!isNumeric(value)) return styles; styles += output.element + "{" + output.property + ": " + value + "px" + ";\ }"; return styles; } // Hook the function to the `kirkiPostMessageStylesOutput` filter. wp.hooks.addFilter("kirkiPostMessageStylesOutput", "kirki", stylesOutput); })(); We also do custom JS in some controls such as in control-react-colorful, field-typography, and field-dimensions to make the CSS output of postMessage output the styles in the expected format. There, people can check where to enqueue, and what are the dependencies.
gharchive/issue
2022-02-08T19:38:21
2025-04-01T06:39:17.569375
{ "authors": [ "JiveDig", "contactjavas" ], "repo": "kirki-framework/kirki", "url": "https://github.com/kirki-framework/kirki/issues/2453", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2221379046
feat(s3): autoDeleteObjects log group allows retention period and removal policy definitions Issue #24815 Closes #24815 Reason for this change S3 bucket autoDeleteObjects leaves behind a log group for each bucket that uses the feature. This results in a lot of cruft, especially in test accounts, which should be configurable by the bucket owner. The account limit for log groups is 10,000 and I've got test accounts that have hit this limit several times. Description of changes Creates a log group rather than relying on the underlying custom-resource to create it automatically (a side effect of using CfnResource for AWS::Lambda::Function) Sets a default retention period of 90 days on the log group (I picked a number) Sets a default removal policy of delete on the log group (I don't think anyone wants these after they delete a bucket) Denies the custom-resource Lambda role permission to create a log group (prevents log group recreation on delete) Adds log group name as an optional to the interface of the custom-resource. This is plumbed into the loggingConfig and results in an undefined entry if not provided. Description of how you validated changes Unit tests in addition to some simple functional tests. When making a bucket with autoDeleteObjects enabled I wanted to confirm that the log group for the lambda was, in fact, gone after I deleted the stack. This is how I found that I needed to modify the permission of the Lambda role to deny log group creation. I also confirmed that the custom-resources which do not provide a log group name still produce a log group and logs within. Also, over 100 snapshot tests (RIP me). Checklist [x] My code adheres to the CONTRIBUTING GUIDE and DESIGN GUIDELINES By submitting this pull request, I confirm that my contribution is made under the terms of the Apache-2.0 license There are about 10 snapshot tests which are failing that I'm unable to resolve on my own and I could use some help in running them. I believe that a few of them are because I'm using an internal AWS account so hopefully they just need to be run from someone's not-quite-so-special account. Oh, I definitely did something wrong on opening this PR I have no idea what though, changing to a draft. Opened this backwards. Closing. Actual PR: https://github.com/aws/aws-cdk/pull/29698
gharchive/pull-request
2024-04-02T20:22:17
2025-04-01T06:39:17.576686
{ "authors": [ "kishiel" ], "repo": "kishiel/eks-ipv6", "url": "https://github.com/kishiel/eks-ipv6/pull/2", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1243399711
Open source NVIDIA drivers It would be nice to update the repository and add the nvidia-open package 😃 (sway with native wlroots works for me without artifacts) The open source kernel drivers only support GPUs from the 3xxx series onwards and I only have a 1660, so I won't be able to test it. Nvm, it does support 16xx and 20xx series aswell. Fixed and added an option to use proprietary module Got it, let's hope for better support for the open source driver.
gharchive/issue
2022-05-20T16:47:43
2025-04-01T06:39:17.579297
{ "authors": [ "git-bruh", "nvidiaLinuxUser" ], "repo": "kiss-community/grepo", "url": "https://github.com/kiss-community/grepo/issues/31", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
368166810
Removed deprecated class PrioQueueForInts and its tests The PrioQueueForInts class is deprecated and there is no NetworKit class using it. This is uncontroversial. I will just merge it.
gharchive/pull-request
2018-10-09T11:36:38
2025-04-01T06:39:17.581587
{ "authors": [ "avdgrinten", "eug93" ], "repo": "kit-parco/networkit", "url": "https://github.com/kit-parco/networkit/pull/243", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
257814293
runbeforeSendingCallbacks errors on undefined each This may be entirely to do with some oddity of the project where I'm trying to switch to Zttp (Lumen v5.1.7, due to requirements back when it was built and lack of time/resources to rework it in 5.5 or full Laravel), but the changes in runBeforeSendingCallbacks in commit 7b6dddc8c824671a460de8a093acc92a40f4ffe8 throw an error for me on any request, including a simple get with no parameters: $response = Zttp::get('https://github.com'); PHP error: Undefined property: Illuminate\Support\Collection::$each in [...]/vendor/kitetail/zttp/src/Zttp.php on line 178 Using kitetail/zttp v0.3.0 and illuminate/support v5.1.41. Illuminate\Support\Collection definitely has the each function, but I'm probably in over my head as to what I'm doing wrong. If I switch back to array_reduce and [] instead of collect(), there are no errors and the call behaves as expected. Any idea what I'm running up against? Again, fully expecting it to be something on my end (and I'm not expecting help with that) with the older Lumen/Illuminate versions in play rather than a Zttp issue, but I figured I could check. I bet this is because we use the collect helper function and Lumen already has it defined so we are getting an old version of the Collection class. Not at my computer right now but I'll double check this when I get home 👍🏻 So the issue is unfortunately somewhat complex, I've opened an issue on the package we use with more details: https://github.com/tightenco/collect/issues/54 I'm going to rename this issue to match the root cause and will think if there's a good interim solution. I ran into this on a pretty blank API project that had laravel/tinker in its dependencies. Tinker apparently installed illuminate/support v 5.2.x when depending on the latest tinker.
gharchive/issue
2017-09-14T18:29:41
2025-04-01T06:39:17.587564
{ "authors": [ "10bass", "adamwathan", "dakira" ], "repo": "kitetail/zttp", "url": "https://github.com/kitetail/zttp/issues/46", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2139783822
🛑 Quasar RPC is down In d482ad7, Quasar RPC (https://quasar-rpc.kitkat.zone) was down: HTTP code: 502 Response time: 426 ms Resolved: Quasar RPC is back up in a8b9bef after 13 minutes.
gharchive/issue
2024-02-17T03:10:24
2025-04-01T06:39:17.591205
{ "authors": [ "kitkatfund" ], "repo": "kitkatfund/kitkatfund.github.io", "url": "https://github.com/kitkatfund/kitkatfund.github.io/issues/277", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1313059111
🛑 Vinos Divertidos is down In 025448c, Vinos Divertidos (https://vinosdivertidos.es) was down: HTTP code: 500 Response time: 7877 ms Resolved: Vinos Divertidos is back up in 6d0252c.
gharchive/issue
2022-07-21T10:46:17
2025-04-01T06:39:17.593550
{ "authors": [ "kitos9112" ], "repo": "kitos9112/uptime", "url": "https://github.com/kitos9112/uptime/issues/196", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
55094535
Error after inserting non-ASCII char from Cliboard to TextInput widget Xubuntu 14.10, Python 2.7.8. Inserting non-ASCII char "Д" by Control+V into the TextInput widget [INFO ] [Logger ] Record log in /root/.kivy/logs/kivy_15-01-22_0.txt [INFO ] Kivy v1.9.0-dev [INFO ] [Python ] v2.7.8 (default, Oct 20 2014, 15:05:19) [GCC 4.9.1] [INFO ] [Factory ] 173 symbols loaded [INFO ] [Image ] Providers: img_tex, img_dds, img_gif, img_pygame, img_pil (img_ffpyplayer ignored) [INFO ] [Window ] Provider: pygame(['window_egl_rpi'] ignored) [INFO ] [GL ] OpenGL version <3.0 Mesa 10.3.2> [INFO ] [GL ] OpenGL vendor <Intel Open Source Technology Center> [INFO ] [GL ] OpenGL renderer <Mesa DRI Intel(R) Sandybridge Mobile > [INFO ] [GL ] OpenGL parsed version: 3, 0 [INFO ] [GL ] Shading version <1.30> [INFO ] [GL ] Texture max size <8192> [INFO ] [GL ] Texture max units <16> [INFO ] [Window ] virtual keyboard not allowed, single mode, not docked [INFO ] [Text ] Provider: pygame [INFO ] [Video ] Provider: pygst [INFO ] [OSC ] using <multiprocessing> for socket [INFO ] [ProbeSysfs ] device match: /dev/input/event5 [INFO ] [MTD ] Read event from </dev/input/event5> [INFO ] [Base ] Start application main loop [INFO ] [MTD ] </dev/input/event5> range position X is 1212 - 5756 [INFO ] [MTD ] </dev/input/event5> range position Y is 996 - 4876 [INFO ] [MTD ] </dev/input/event5> range touch major is 0 - 0 [INFO ] [MTD ] </dev/input/event5> range touch minor is 0 - 0 [INFO ] [MTD ] </dev/input/event5> range pressure is 0 - 255 [INFO ] [MTD ] </dev/input/event5> axes invertion: X is 0, Y is 0 [INFO ] [GL ] NPOT texture support is available [INFO ] [Clipboard ] Provider: pygame(['clipboard_dbusklipper'] ignored) [INFO ] [Base ] Leaving application in progress... Traceback (most recent call last): File "/home/johndoe/kivy-master/examples/RST_Editor/main.py", line 61, in <module> Editor().run() File "/usr/lib/python2.7/dist-packages/kivy/app.py", line 824, in run runTouchApp() File "/usr/lib/python2.7/dist-packages/kivy/base.py", line 484, in runTouchApp EventLoop.window.mainloop() File "/usr/lib/python2.7/dist-packages/kivy/core/window/window_pygame.py", line 381, in mainloop self._mainloop() File "/usr/lib/python2.7/dist-packages/kivy/core/window/window_pygame.py", line 348, in _mainloop self.modifiers): File "_event.pyx", line 697, in kivy._event.EventDispatcher.dispatch (kivy/_event.c:6788) File "_event.pyx", line 1159, in kivy._event.EventObservers.dispatch (kivy/_event.c:11470) File "_event.pyx", line 1083, in kivy._event.EventObservers._dispatch (kivy/_event.c:11066) File "/usr/lib/python2.7/dist-packages/kivy/core/window/__init__.py", line 149, in _on_window_key_down return self.dispatch('on_key_down', keycode, text, modifiers) File "_event.pyx", line 697, in kivy._event.EventDispatcher.dispatch (kivy/_event.c:6788) File "_event.pyx", line 1159, in kivy._event.EventObservers.dispatch (kivy/_event.c:11470) File "_event.pyx", line 1083, in kivy._event.EventObservers._dispatch (kivy/_event.c:11066) File "/usr/lib/python2.7/dist-packages/kivy/uix/textinput.py", line 2017, in keyboard_on_key_down self.paste() File "/usr/lib/python2.7/dist-packages/kivy/uix/textinput.py", line 1388, in paste data = Clipboard.paste() File "/usr/lib/python2.7/dist-packages/kivy/core/clipboard/__init__.py", line 88, in paste return self._paste() File "/usr/lib/python2.7/dist-packages/kivy/core/clipboard/__init__.py", line 106, in _paste data = self.get(mime_type) File "/usr/lib/python2.7/dist-packages/kivy/core/clipboard/clipboard_pygame.py", line 35, in get text = text.encode('utf-8') UnicodeDecodeError: 'ascii' codec can't decode byte 0xd0 in position 0: ordinal not in range(128) Want to back this issue? Place a bounty on it! We accept bounties via Bountysource. I have the same error. It looks like pygame.scrap.get() returns string, and the following text.encode('utf-8') works correctly with unicode string. The following patch fixed the issue for me: diff --git a/kivy/core/clipboard/clipboard_pygame.py b/kivy/core/clipboard/clipboard_pygame.py index 6f053ec..8dfbac3 100644 --- a/kivy/core/clipboard/clipboard_pygame.py +++ b/kivy/core/clipboard/clipboard_pygame.py @@ -31,8 +31,8 @@ class ClipboardPygame(ClipboardBase): def get(self, mimetype='text/plain'): self.init() text = pygame.scrap.get(mimetype) - if PY2: - text = text.encode('utf-8') + # if PY2: + # text = text.encode('utf-8') return text def put(self, data, mimetype='text/plain'): @hey-sancho It's strange, but there is no any answer from kivy developers... I cannot seem to reproduce it on windows with master. Perhaps if it's already bytes we should not encode it again. closed via 33bfc526add488571b2dadea7771579a9cbb9042
gharchive/issue
2015-01-21T23:44:09
2025-04-01T06:39:17.663059
{ "authors": [ "akshayaurora", "hey-sancho", "macropas", "matham" ], "repo": "kivy/kivy", "url": "https://github.com/kivy/kivy/issues/2876", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
100805627
Multisamples causing GL error I created 2 .exe packages for windows according to the documentation. I used stable and dev version kivy. And I have a problem - stable version detects opengl version 2.1 and dev version - 1.1 log files stable [INFO ] Logger: Record log in C:\Users\maxim_000\.kivy\logs\kivy_15-08-13_6.txt [WARNING ] [Config ] Upgrading configuration in progress. [WARNING ] [Config ] Older configuration version detected (14 instead of 13) [INFO ] Kivy: v1.9.0 [INFO ] Python: v2.7.9 (default, Dec 10 2014, 12:24:55) [MSC v.1500 32 bit (Intel)] [INFO ] Factory: 173 symbols loaded [INFO ] Image: Providers: img_tex, img_dds, img_gif, img_sdl2, img_pil (img_ffpyplayer ignored) [INFO ] Text: Provider: sdl2 [INFO ] OSC: using <thread> for socket [INFO ] Window: Provider: sdl2 [INFO ] GL: GLEW initialization succeeded [INFO ] GL: OpenGL version <2.1.0 - Build 8.15.10.1872> [INFO ] GL: OpenGL vendor <Intel> [INFO ] GL: OpenGL renderer <Mobile Intel(R) 4 Series Express Chipset Family> [INFO ] GL: OpenGL parsed version: 2, 1 [INFO ] GL: Shading version <1.20 - Intel Build 8.15.10.1872> [INFO ] GL: Texture max size <4096> [INFO ] GL: Texture max units <16> [INFO ] Shader: fragment shader: <No errors.> [INFO ] Shader: vertex shader: <No errors.> [INFO ] Shader: program: <No errors.> [INFO ] Window: auto add sdl2 input provider [INFO ] Window: virtual keyboard not allowed, single mode, not docked [INFO ] GL: NPOT texture support is available [INFO ] Base: Start application main loop [INFO ] Base: Leaving application in progress... dev [INFO ] Logger: Record log in C:\Users\maxim_000\.kivy\logs\kivy_15-08-13_5.txt [INFO ] Kivy: v1.9.1-dev [INFO ] Python: v2.7.9 (default, Dec 10 2014, 12:24:55) [MSC v.1500 32 bit (Intel)] [INFO ] Factory: 177 symbols loaded [INFO ] Image: Providers: img_tex, img_dds, img_gif, img_sdl2, img_pil (img_ffpyplayer ignored) [INFO ] Text: Provider: sdl2 [INFO ] OSC: using <thread> for socket [INFO ] Window: Provider: sdl2 [INFO ] GL: GLEW initialization succeeded [INFO ] GL: OpenGL version <1.1.0> [INFO ] GL: OpenGL vendor <Microsoft Corporation> [INFO ] GL: OpenGL renderer <GDI Generic> [INFO ] GL: OpenGL parsed version: 1, 1 [CRITICAL ] GL: Minimum required OpenGL version (2.0) NOT found! OpenGL version detected: 1.1 Version: 1.1.0 Vendor: Microsoft Corporation Renderer: GDI Generic Try upgrading your graphics drivers and/or your graphics hardware in case of problems. The application will leave now. I have the same issue with 1.9.1-dev : [INFO ] [Logger ] Record log in C:\Documents and Settings\gfj138\.kivy\logs\kivy_15-08-25_22.txt [INFO ] [Kivy ] v1.9.1-dev [INFO ] [Python ] v2.7.10 |Continuum Analytics, Inc.| (default, May 28 2015, 17:02:00) [MSC v.1500 32 bit (Intel)] [INFO ] [Factory ] 177 symbols loaded [DEBUG ] [Cache ] register <kv.lang> with limit=None, timeout=None [DEBUG ] [Cache ] register <kv.image> with limit=None, timeout=60 [DEBUG ] [Cache ] register <kv.atlas> with limit=None, timeout=None [INFO ] [Image ] Providers: img_tex, img_dds, img_gif, img_sdl2, img_pil (img_ffpyplayer ignored) [DEBUG ] [Cache ] register <kv.texture> with limit=1000, timeout=60 [DEBUG ] [Cache ] register <kv.shader> with limit=1000, timeout=3600 [INFO ] [Text ] Provider: sdl2 [INFO ] [OSC ] using <thread> for socket [WARNING ] [Input ] WM_Touch/WM_Pen not supported by your version of Windows [INFO ] [Window ] Provider: sdl2 [INFO ] [GL ] GLEW initialization succeeded GL: glGenFramebuffers is NULL, try to detect an extension GL: available extensions: GL_WIN_swap_hint GL_EXT_bgra GL_EXT_paletted_texture GL: No framebuffers extension is supported GL: Any call to Fbo will crash ! [INFO ] [GL ] OpenGL version <1.1.0> [INFO ] [GL ] OpenGL vendor <Microsoft Corporation> [INFO ] [GL ] OpenGL renderer <GDI Generic> [INFO ] [GL ] OpenGL parsed version: 1, 1 [CRITICAL ] [GL ] Minimum required OpenGL version (2.0) NOT found! and 1.9.0 [INFO ] [Kivy ] v1.9.0 [INFO ] [Python ] v2.7.10 |Continuum Analytics, Inc.| (default, May 28 2015, 17:02:00) [MSC v.1500 32 bit (Intel)] [INFO ] [Factory ] 173 symbols loaded [INFO ] [Image ] Providers: img_tex, img_dds, img_gif, img_sdl2, img_pil (img_ffpyplayer ignored) [INFO ] [Text ] Provider: sdl2 [INFO ] [OSC ] using <thread> for socket [WARNING ] [Input ] WM_Touch/WM_Pen not supported by your version of Windows [INFO ] [Window ] Provider: sdl2 [INFO ] [GL ] GLEW initialization succeeded [INFO ] [GL ] OpenGL version <3.1.0 - Build 6.14.10.5398> [INFO ] [GL ] OpenGL vendor <Intel> [INFO ] [GL ] OpenGL renderer <Intel(R) HD Graphics 4000> [INFO ] [GL ] OpenGL parsed version: 3, 1 [INFO ] [GL ] Shading version <1.40 - Intel Build 6.14.10.5398> [INFO ] [GL ] Texture max size <8192> [INFO ] [GL ] Texture max units <16> [INFO ] [Shader ] fragment shader: <No errors.> [INFO ] [Shader ] vertex shader: <No errors.> [INFO ] [Shader ] program: <No errors.> [INFO ] [Window ] auto add sdl2 input provider libpng warning: iCCP: known incorrect sRGB profile [INFO ] [Window ] virtual keyboard not allowed, single mode, not docked As I understand it - for determining the version of opengl responsible kivy.graphics. I'm assuming commit history here https://github.com/kivy/kivy/commits/master/kivy/graphics?page=1 3eb5b4844105fe83349a59612f8850e73622c26c found that it works, and since f15283ffd99b226e6cfe9c99d18d1010176aea21 error exists. So either this error because of a commit or error corrections for some is the above-mentioned package. Error in this commit fd54e811f9c9413d22a2486920ed8d89ae84fc11 What is precisely the error ? Did you succeed in correcting this error ? I dont succeed( Error on these lines - https://github.com/kivy/kivy/commit/fd54e811f9c9413d22a2486920ed8d89ae84fc11#diff-b91d9923ff01305495e47990f1ac7951R85 SDL_GL_SetAttribute(SDL_GL_MULTISAMPLEBUFFERS, ...) SDL_GL_SetAttribute(SDL_GL_MULTISAMPLESAMPLES, ...) If you set any values to these attributes - error exists. Why - i dont know. I'm looking for a solution, but so far without results. I do not get exactly what you are trying to do and what error you are fighting against... Do you have compilation errors ? Googling the issue, I found a thread related to a game using OpenGL where a user had the same issue (other games detecting correct OpenGL driver with OpenGL >= 2.0, but this game only using the default Microsoft driver with OpenGL 1.1.0) => see https://betaguide.wz2100.net/viewtopic.php?f=4&t=11314&sid=a20074f87518a1a6a6d39bb97cabd18b&start=0 . At the end, the game detected the correct driver when he user changed the driver acceleration mode from quality to performance. Can this help us in solving the issue ? OK, maybe it will help, I'll read it and think about it. There is one way that circumvents the problem but not solve it. Just remove the 124 and 125 lines in https://github.com/kivy/kivy/blob/master/kivy/core/window/_window_sdl2.pyx. sdementen Do you have compilation errors ? No, i didn`t have them. At the end, the game detected the correct driver when he user changed the driver acceleration mode from quality to performance. Perhaps a similar thing can be done by working with the registry or installing a SDL_GL_SetAttribute. Commenting lines 124+125 did not solve the issue. However, by skipping entirely the new code related to multisamples, ie change line 116 into if multisamples > 0 and False: it did work. Thank you for your help ! This is a show stopper bug we need to either find a solution or revert this before 1.9.1 This version works for me https://github.com/kivy/kivy/commit/c394fa891d34fbc37224e2c5b713140d6dc216a2 .About better fix the problem - for support SDL_GL_MULTISAMPLEBUFFERS requires the GL_ARB_multisample extension(according this sdl.beuc.net/sdl.wiki/SDL_GLattr) , and i don`t have it. Need to add a check for this extension, and in his absence - set in config multisamples to 0. I'll try to do it Perhaps there is another solution, the information here link_to_SO_1 link_to_SO_2 link_to_SO_3 can help to find it, or make sure that it is not (I had not understand due to poor knowledge of English and even less knowledge in SDL and opengl) I dont know how to fix it. I know that for the code you need to write, but I dont know where it should be LaTeX-Python confirmed on irc that it seems to happen with integrated GPUs, he uses this processor and was running a packaged app. His code runs without problems on machines with a dedicated graphics card. @Kazun3500 i'm unsure if that commit is the real trigger, LaTeX tried with multisamples 0 (which stops the relevant part of the patch from executing) and the crash happened the same way. Could you check if running the source code shows the same issue, or does it crash only with a packaged app? What is your processor model and driver version? @dessant running the source code shows the same issue. And for me setting multisamples to 0 helps to avoid crashes. Processor version - Celeron(R) Dual-Core CPU T3500 @ 2.10GHz × 2, driver version - i will say a little bit later. And - i have integrated GRU. driver version -6.3.9600.16384, driver provider - Microsoft. Now I try to find a driver from Intel. driver version -6.3.9600.16384, driver provider - Microsoft. I can`t find driver from intel. And i checked again - setting multisamples to 0 helps to avoid crashes, and depends on multisamples count it shows different lists avaliable gl extensions with 0 GL: glGenFramebuffers is NULL, try to detect an extension GL: available extensions: GL_EXT_blend_minmax GL_EXT_blend_subtract GL_EXT_blend _color GL_EXT_abgr GL_EXT_texture3D GL_EXT_clip_volume_hint GL_EXT_compiled_vert ex_array GL_SGIS_texture_edge_clamp GL_SGIS_generate_mipmap GL_EXT_draw_range_el ements GL_SGIS_texture_lod GL_EXT_rescale_normal GL_EXT_packed_pixels GL_EXT_sep arate_specular_color GL_ARB_multitexture GL_EXT_texture_env_combine GL_EXT_bgra GL_EXT_blend_func_separate GL_EXT_secondary_color GL_EXT_fog_coord GL_EXT_textur e_env_add GL_ARB_texture_cube_map GL_ARB_transpose_matrix GL_ARB_texture_env_add GL_IBM_texture_mirrored_repeat GL_EXT_multi_draw_arrays GL_NV_blend_square GL_A RB_texture_compression GL_3DFX_texture_compression_FXT1 GL_EXT_texture_filter_an isotropic GL_ARB_texture_border_clamp GL_ARB_point_parameters GL_ARB_texture_env _combine GL_ARB_texture_env_dot3 GL_ARB_texture_env_crossbar GL_EXT_texture_comp ression_s3tc GL_ARB_shadow GL_ARB_window_pos GL_EXT_shadow_funcs GL_EXT_stencil_ wrap GL_ARB_vertex_program GL_EXT_texture_rectangle GL_ARB_fragment_program GL_E XT_stencil_two_side GL_ATI_separate_stencil GL_ARB_vertex_buffer_object GL_EXT_t exture_lod_bias GL_ARB_occlusion_query GL_ARB_fragment_shader GL_ARB_shader_obje cts GL_ARB_shading_language_100 GL_ARB_texture_non_power_of_two GL_ARB_vertex_sh ader GL_NV_texgen_reflection GL_ARB_point_sprite GL_EXT_blend_equation_separate GL_ARB_depth_texture GL_ARB_texture_rectangle GL_ARB_draw_buffers GL_ARB_pixel_b uffer_object GL_WIN_swap_hint GL_EXT_framebuffer_object GL_EXT_texture_sRGB GL_A RB_color_buffer_float GL_ARB_half_float_pixel GL_ARB_texture_float GL_NV_conditi onal_render GL_EXT_texture_swizzle GL: EXT_framebuffer_object is supported with 2 GL: glGenFramebuffers is NULL, try to detect an extension GL: available extensions: GL_WIN_swap_hint GL_EXT_bgra GL_EXT_paletted_texture GL: No framebuffers extension is supported GL: Any call to Fbo will crash ! I confirm that i have this problem with a Windows 10 machine and with Virtualbox Windows machines. Config.set('graphics', 'multisamples', '0') solve the problem. @lupin3rd - can you give more details where exactly (file, line, kivy version) you made the change so I can try this on a win7pro 32bit machine? Just to share an update about this, a work to have ANGLE available on Windows (ANGLE is a project that translate OpenGL instructions to DirectX instructions, used by Firefox, Chrome, etc.) My personnal wish is to have ANGLE by default on windows, and prevent theses kind of issues. ANGLE require DirectX 9, which is the default version installed on Windows 10, so no graphics drivers installation are required! I tested my app on lots of different machines, Windows 10 and Windows 7. I have no problems with the latest stable 1.9.1 or the latest 1.9.2-dev on these machines. But none of them works on my tablet samsung slate 700t with Windows 10. The workaround with multisamples does not work either. I think I'm having the same issue as @jegger, works fine on Windows 8.1 on one of my machines, but when testing my packaged app on Windows 10 I get the 1.1 error: I did some further testing: On a Windows 10 (64bit) it does not work (as described in my last comment) On a Windows 7 (32bit, in a virtual-machine) it does work Could it be related to the architecture (32/64bit?) What do you have @steinnes ? I created via Pyinstaller3.3-dev (and kivy latest master) a: single file executable: Does work on Win10 and Win7 folder based executable: Does show OpenGL error on Win10, but not on Win7 @jegger, there have been substantial graphics changes since the last stable release, you might want to test master too. @dessant Today I tried it with the latest kivy build (windows wheel https://kivy.org/downloads/appveyor/kivy/Kivy-1.9.2.dev0-cp27-cp27m-win32.whl) I built with pyinstaller (also latest master branch) a folder based executable on windows 10 (virtual machine 64bit). This runs on the virtual machine itself but runs into this error on the other physical windows10 64bit machine. When building a pyinstaller single file executable it works on both windows 10 installations (virtual an physical). I don't get this... Try setting the environment to KIVY_GL_BACKEND=sdl2 and see if that fixes it. That needs to be set before kivy is imported. You can set it in environ before the first kivy import. On Jan 18, 2017 11:16 AM, "Dominique Burnand" notifications@github.com wrote: @dessant https://github.com/dessant Today I tried it with the latest kivy build (windows wheel https://kivy.org/downloads/ appveyor/kivy/Kivy-1.9.2.dev0-cp27-cp27m-win32.whl) I built with pyinstaller (also latest master branch) a folder based executable on windows 10 (virtual machine 64bit). This runs on the virtual machine itself but runs into this error on the other physical windows10 64bit machine. When building a pyinstaller single file executable it works on both windows 10 installations (virtual an physical). I don't get this... — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/kivy/kivy/issues/3576#issuecomment-273521033, or mute the thread https://github.com/notifications/unsubscribe-auth/ABkW_iRWtsJUmrmsCumyZef905gklotKks5rTjrYgaJpZM4FrD4a . @matham Thanks for your input. I tried setting the variable by: os.environ["KIVY_GL_BACKEND"] = "sdl2" before any kivy imports in my main file. I can confirm that the flag is recognized when I run the script on the machine on which there is no error. The log shows: "Backend used glew" without the variable beeing set and "Backend used sdl2" with the env-variable set to sdl2. But this does not seems to change anything: Here is the full log when the error is happening: http://codepad.org/QnEvh6Gl I think you'll have to debug it on your machine since none of us can reproduce it. The place to look at is https://github.com/kivy/kivy/blob/master/kivy/core/gl/init.py and https://github.com/kivy/kivy/blob/master/kivy/graphics/opengl_utils.pyx#L240. I'd also try to add the referenced _kivy_opengl_required_func to sys (e.g. sys._kivy_opengl_required_func = lambda *largs: 1) which will make kivy ignore that the version is too low. Maybe kivy will still work or it could crash. I get this error too. I tried all the below (separately an together) but I still get the same error. Config.set('graphics', 'multisamples', '0') sys._kivy_opengl_required_func = lambda *largs: 1 os.environ["KIVY_GL_BACKEND"] = "sdl2" [INFO ] [Window ] Provider: sdl2 [INFO ] [GL ] GLEW initialization succeeded [INFO ] [GL ] OpenGL version <b'1.1.0'> [INFO ] [GL ] OpenGL vendor <b'Microsoft Corporation'> [INFO ] [GL ] OpenGL renderer <b'GDI Generic'> [INFO ] [GL ] OpenGL parsed version: 1, 1 [CRITICAL ] [GL ] Minimum required OpenGL version (2.0) NOT found! OpenGL version detected: 1.1 Version: b'1.1.0' Vendor: b'Microsoft Corporation' Renderer: b'GDI Generic' Try upgrading your graphics drivers and/or your graphics hardware in case of problems. The application will leave now. GL: glGenFramebuffers is NULL, try to detect an extension GL: available extensions: GL_WIN_swap_hint GL_EXT_bgra GL_EXT_paletted_texture GL: No framebuffers extension is supported GL: Any call to Fbo will crash ! Process finished with exit code 1 Any other ideas? Same problem here guys!! Running windows 10 64bit. confirmed i have opengl 3.2 but still detecting opengl 1.1 Just a mention, the same OpenGL 1.1 error is available on Appveyor, so if this gets solved, it might be a nice test case. kivy_17-02-24_3.txt I am also getting this misidentifying error. I tried running the /share/kivy-examples/demo/showcase/main.py file and got that error. It worked after I edited the main.py file and added the following two lines just after the rest of the import statements: from kivy import Config Config.set('graphics', 'multisamples', '0') Currently running on: Manufacturer: Acer Computer Model: Asprie Z3-715 Intel i5 6400T CPU @ 2.2GHz 8GB RAM 64-bit Windows 10 Integrated Graphics: Intel HD 530 with Shared Memory Dedicated Graphics: Nvidia Geforce 940M with 2GB RAM The issue is with Windows 10 support for interested graphics, had this problem with a Java library, changed to the last SDK version before windows 10 was released and the issue was fixed @encloinc Can you explain what you mean? What error did you get in the java library? Also, you changed the sdk version of what? When compiling Angle? Btw, @KeyWeeUsr The problem with appveyor being 1.1 will likely never be fixed since they literally don't have a graphics card. So that's the version windows is limited to I believe. But I could be wrong, since they do seem to have DirectX support. @matham I think they really have DirectX, so if we can force kivy to use angle on appveyor, it might test this (or the 4971 part at least). There seems to be a mishmash of issues in this thread. Let's use this thread for cases where gl detection issues are solved by Config.set('graphics', 'multisamples', '0'), regardless of the gl backend being used. If that is not the case, please open a new issue, describing the error you get. I was using JDK v#121 for the library https://libgdx.badlogicgames.com/ , when I switched the JDK version to an older one that was not built around Windows 10, (I switched to v51), the opengl error on java went away. @KeyWeeUsr Yes, they do have DirectX installed, but that is different from the graphics driver. That link is just for the software package, it doesn't mean the graphics driver would also work. I should have my head examined. I already enabled the tests on appveyor using angle with 3.5+ as you can see here: https://ci.appveyor.com/project/KivyOrg/kivy/build/1.0.719/job/j7k7c123ojwt0ba9#L1080. So the tests already work there with angle. It could be other people who have problems is because they are lacking some dll. Appveyor comes preinstalled with the direct3d sdk so it has all the dlls. I can probably test on appveyor to see what other dlls it depends on Reproduced on a PC I don't really have a stable access to, but these are the specs from dxdiag. I'll try angle if there will be a chance to do so: Win 10 Education 64bit Build 14393 Intel Core2 Quad Q9400 (6M Cache, 2.66 GHz, 1333 MHz FSB) Intel GMA 4500M I think it's this machine I had no errors such as: GL: glGenFramebuffers is NULL, try to detect an extension GL: available extensions: GL_WIN_swap_hint GL_EXT_bgra GL_EXT_paletted_texture GL: No framebuffers extension is supported GL: Any call to Fbo will crash ! just a simple log without any issues + OpenGL 1.1 error at the end, although the CPU should support the required OpenGL version (or at least he and he say so). @KeyWeeUsr, is it fixed by Config.set('graphics', 'multisamples', '0')? If not, please open a separate issue. @dessant I'm not quite sure if it is an issue as I didn't really find anything OpenGL > 1.1 in there and even WebGL checker reports no WebGL2 support (so probably just no DLL available). Anyway, I checked today even with disabled multisamples. It did nothing, but that was rather expected and using angle fixed the error, therefore I guess it's just that - a missing DLL. As @dessant mentioned, this thread has become full of too many issues and suggestions. @jeeger's (#5071) new ticket summarizes the remaining issues. Please follow any remaining discussion there. Note: I have a work laptop which reproduces the issues as discussed there. I can run any requested tests or make this machine available remotely upon request. In the meantime, I will continue investigating and discussing on that thread. Thanks Sorry, you're thinking of @jegger ☺️ Am 29. März 2017 12:07:29 MESZ schrieb Richard Larkin notifications@github.com: As @dessant mentioned, this thread has become full of too many issues and suggestions. @jeeger's (#5071) new ticket summarizes the remaining issues. Please follow any remaining discussion there. Note: I have a work laptop which reproduces the issues as discussed there. I can run any requested tests or make this machine available remotely upon request. In the meantime, I will continue investigating and discussing on that thread. Thanks -- You are receiving this because you were mentioned. Reply to this email directly or view it on GitHub: https://github.com/kivy/kivy/issues/3576#issuecomment-290044678 -- Diese Nachricht wurde von meinem Mobiltelefon gesendet. Comments not related to multisamples have been deleted, let's reopen this. Kivy 1.10.0 was released! So I installed it right away to check. Alas the problem stays the same. I still need to add from kivy import Config Config.set('graphics', 'multisamples', '0') to the kivy showcase "share\kivy-examples\demo\showcase\main.py" to get it to run. The showcase itself runs, apart from the scatter demo, which just does nothing. Tried on Win7 Pro 32bit, Python 3.6.1 32 bit, Intel HD on board graphics. (Core i5), driver 8.15.10.2281 (not updatable) Did anyone try it with the sdl2 backend (KIVY_GL_BACKEND=sdl2) or the angle backend for py3.5+ (pip install kivy.deps.angle and KIVY_GL_BACKEND=angle_sdl2)? @matham can you give beginner friendly instructions on how to activate KIVY_GL_BACKEND=angle_sdl2 assuming I have performed pip install kivy.deps.angle on python 3.6.1 and am now about to start share\kivy-examples\demo\showcase\main.py? Hmm, that is very strange, I don't see how that error could occur with the last release. @mborus, dozens of posts have been deleted from this thread to bring it to a usable state, please do not go off-topic, and debug your angle issues on irc or the mailing list. I also have same experience with python 3.6 but not with python 2.7. My apps running well with python 2.7 but have false detected about opengl version and can't find any solution yet. Every solution I try including angle and multisamples didn't work. the issue seems to be persisting still in my machine windows 10 I have also tried the above solutions to see the issue... import os os.environ['KIVY_GL_BACKEND'] = 'angle_sdl2' import kivy Adding this to the top of my script is the work around I'm using on windows 10, kivy 1.10.0, python 3.6.2. I want to clarify: from kivy import Config Config.set('graphics', 'multisamples', '0') Does not resolve my error. RESTART: C:/Users/AppData/Local/Programs/Python/Python36-32/kivy_label.py [INFO ] [Logger ] Record log in C:\Users\Corie.kivy\logs\kivy_17-09-05_15.txt [INFO ] [Kivy ] v1.10.0 [INFO ] [Python ] v3.6.2 (v3.6.2:5fd33b5, Jul 8 2017, 04:14:34) [MSC v.1900 32 bit (Intel)] [INFO ] [Factory ] 194 symbols loaded [INFO ] [Image ] Providers: img_tex, img_dds, img_sdl2, img_gif (img_pil, img_ffpyplayer ignored) [INFO ] [Text ] Provider: sdl2 [INFO ] [OSC ] using for socket [INFO ] [Window ] Provider: sdl2 [INFO ] [GL ] Using the "OpenGL" graphics system [INFO ] [GL ] GLEW initialization succeeded [INFO ] [GL ] No framebuffers extension is supported [INFO ] [GL ] Backend used [INFO ] [GL ] OpenGL version <b'1.1.0'> [INFO ] [GL ] OpenGL vendor <b'Microsoft Corporation'> [INFO ] [GL ] OpenGL renderer <b'GDI Generic'> [INFO ] [GL ] OpenGL parsed version: 1, 1 [CRITICAL] [GL ] Minimum required OpenGL version (2.0) NOT found! OpenGL version detected: 1.1 Version: b'1.1.0' Vendor: b'Microsoft Corporation' Renderer: b'GDI Generic' Try upgrading your graphics drivers and/or your graphics hardware in case of problems. The application will leave now. However, import os os.environ['KIVY_GL_BACKEND'] = 'angle_sdl2' does. I'm not sure if this is the same or related issue. Should we apply the: Config.set('graphics', 'multisamples', '0') On windows machines only? Will there be a performance impact on other platforms if applied globally, such as on Android, iOS or OSX? It's only a problem on windows AFAIK. Thanks. What I meant was, by setting this config option regardless of the platform is running on, will it have any detrimental performance impact? I don't think so. From the docs, Sets the MultiSample Anti-Aliasing (MSAA) level. Increasing this value results in smoother graphics but at the cost of processing time. which seems to imply that a lower level will only increase performance. But it may look worse. I'd just do import platform from kivy.config import Config if platform.system() == 'Windows: Config.set('graphics', 'multisamples', '0')' Perfect. Thank you. This worked for me import os os.environ['KIVY_GL_BACKEND'] = 'angle_sdl2' from kivy import Config Config.set('graphics', 'multisamples', '0') What would be ideal is to allow the default multisample value, but pre-emptively set it to zero only if: platform is windows compatibility issue is detected Basically, how do we test ahead of time before Kivy crashes out? I'd like to have the optimal graphics for the 99% of our software installations out there. I don't think that is possible. You need to run the app twice, once to see if there's an issue, at which point you need to terminate python because gl is already initialized and then you have to run it again with the "fix". So we have no way to detect the issue and then adjust the setting. You can make it available as a config option that your user changes if they have issues. Not work for me: Pc config : Intel GPU with updated driver: Source test: ` import kivy kivy.require('1.10.0') import os os.environ['KIVY_GL_BACKEND'] = 'angle_sdl2' from kivy import Config Config.set('graphics', 'multisamples', '0') from kivy.app import App from kivy.uix.button import Button class TestApp(App): def build(self): return Button(text='Hello World') TestApp().run()> ` Occours: Python 3.6.2 (v3.6.2:5fd33b5, Jul 8 2017, 04:14:34) [MSC v.1900 32 bit (Intel)] on Walmir-not, Standard [INFO ] [Logger ] Record log in C:\Users\Walmir.kivy\logs\kivy_17-09-29_7.txt [INFO ] [Kivy ] v1.10.0 [INFO ] [Python ] v3.6.2 (v3.6.2:5fd33b5, Jul 8 2017, 04:14:34) [MSC v.1900 32 bit (Intel)] [INFO ] [Factory ] 194 symbols loaded [INFO ] [Image ] Providers: img_tex, img_dds, img_sdl2, img_gif (img_pil, img_ffpyplayer ignored) [INFO ] [Text ] Provider: sdl2 [INFO ] [OSC ] using for socket [INFO ] [Window ] Provider: sdl2 [INFO ] [Window ] Activate GLES2/ANGLE context [CRITICAL] [Window ] Unable to find any valuable Window provider. sdl2 - RuntimeError: b'Could not initialize EGL' File "C:\Python36\lib\site-packages\kivy\core_init_.py", line 67, in core_select_lib cls = cls() File "C:\Python36\lib\site-packages\kivy\core\window\window_sdl2.py", line 140, in init super(WindowSDL, self).init() File "C:\Python36\lib\site-packages\kivy\core\window_init_.py", line 899, in init self.create_window() File "C:\Python36\lib\site-packages\kivy\core\window\window_sdl2.py", line 269, in create_window self.fullscreen, resizable, state) File "kivy\core\window_window_sdl2.pyx", line 142, in kivy.core.window._window_sdl2._WindowSDL2Storage.setup_window (kivy\core/window_window_sdl2.c:2782) File "kivy\core\window_window_sdl2.pyx", line 57, in kivy.core.window._window_sdl2._WindowSDL2Storage.die (kivy\core/window_window_sdl2.c:1872) [CRITICAL] [App ] Unable to get a Window, abort. Exception ignored in: 'kivy.properties.dpi2px' Traceback (most recent call last): File "C:\Python36\lib\site-packages\kivy\utils.py", line 496, in get retval = self.func(inst) File "C:\Python36\lib\site-packages\kivy\metrics.py", line 174, in dpi EventLoop.ensure_window() File "C:\Python36\lib\site-packages\kivy\base.py", line 127, in ensure_window sys.exit(1) SystemExit: 1 [CRITICAL] [App ] Unable to get a Window, abort. Please, any suggestions for a lost beginner? Tks. The issue is still happening on Windows 7 x64 (2011 computer, but still...). I have an Intel HD Graphics Arrandale. The way to fix it is to place: from kivy import Config Config.set('graphics', 'multisamples', '0') Before any kivy import. Another way to fix just on the current machine is to modify the %HOMEPATH%\.kivy\config.ini to change multisamples = 2 into multisamples = 0. I do have a compliant openGL on my Windows 10 laptop: But even with those addition in my code import os os.environ['KIVY_GL_BACKEND'] = 'angle_sdl2' from kivy import Config Config.set('graphics', 'multisamples', '0') I get the 1.1 version error ... The full log: [INFO ] [Logger ] Record log in C:\Users\Jean-Pierre\.kivy\logs\kivy_18-04-21_14.txt [INFO ] [Kivy ] v1.10.0 [INFO ] [Python ] v3.6.4 |Anaconda, Inc.| (default, Jan 16 2018, 10:22:32) [MSC v.1900 64 bit (AMD64)] [INFO ] [Factory ] 194 symbols loaded [INFO ] [Image ] Providers: img_tex, img_dds, img_sdl2, img_pil, img_gif (img_ffpyplayer ignored) [INFO ] [OSC ] using <thread> for socket [INFO ] [Window ] Provider: sdl2 [INFO ] [GL ] Using the "OpenGL" graphics system [INFO ] [GL ] GLEW initialization succeeded [INFO ] [GL ] No framebuffers extension is supported [INFO ] [GL ] Backend used <glew> [INFO ] [GL ] OpenGL version <b'1.1.0'> [INFO ] [GL ] OpenGL vendor <b'Microsoft Corporation'> [INFO ] [GL ] OpenGL renderer <b'GDI Generic'> [INFO ] [GL ] OpenGL parsed version: 1, 1 [CRITICAL] [GL ] Minimum required OpenGL version (2.0) NOT found! OpenGL version detected: 1.1 Version: b'1.1.0' Vendor: b'Microsoft Corporation' Renderer: b'GDI Generic' Try upgrading your graphics drivers and/or your graphics hardware in case of problems. The application will leave now. This is really very problematic ! @Archanciel [INFO ] [Python ] v3.6.4 |Anaconda, Inc.| (default, Jan 16 2018, 10:22:32) [MSC v.1900 64 bit (AMD64)] Anaconda overrides a lot of libraries with their own versions. Please try again with a CPython from python.org Thanks for the suggestion, dolang. I did uninstall Python Anaconda 3.6.4 and installed CPython 3.6.5 instead, but this did not solved the problem. But I finally found the solution thanks to a contribution on an Intel graphical card forum. The idea is to use Shims to solve a driver compatibility problem. See the end of the thread here for a description of the solution. Thank you for contributing your solution back. It might be able to help someone else with the same problem. @gutcheschiro ,Thanks It worked for me. import os os.environ['KIVY_GL_BACKEND'] = 'angle_sdl2' import kivy I'm using on windows 10, kivy 1.10.0, python 3.6.4 Is this is still present on 1.10.1? If so, would it make sense to alter the order of https://github.com/kivy/kivy/blob/b4d3e7d0db67c9ad23dff8c48c4a275bafb9d76b/kivy/graphics/cgl.pyx#L59 so sdl2 is picked first (at least on windows)? Is this is still present on 1.10.1? If so, would it make sense to alter the order of https://github.com/kivy/kivy/blob/b4d3e7d0db67c9ad23dff8c48c4a275bafb9d76b/kivy/graphics/cgl.pyx#L59 so sdl2 is picked first (at least on windows)? In Kivy 2.2.1 that code picks sdl2 or gl. Does that mean this can be closed?
gharchive/issue
2015-08-13T15:58:48
2025-04-01T06:39:17.747161
{ "authors": [ "Archanciel", "Julian-O", "Kazun3500", "Kerang", "KeyWeeUsr", "Naveenkariyappa", "Walpa", "Zen-CODE", "akshayaurora", "amitmarathe", "brentpicasso", "codypace68", "dessant", "dolang", "encloinc", "gutcheschiro", "jeeger", "jegger", "jimanvlad", "lrq3000", "lupin3rd", "matham", "maxcrow", "mborus", "schitzN", "sdementen", "steinnes", "sushihuye", "tito", "tshirtman" ], "repo": "kivy/kivy", "url": "https://github.com/kivy/kivy/issues/3576", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
156167301
Rare should equal failure with dictionary comparison [Remastered]. Well, as I tell before dictionaries equal doesn't work well. I would like to narrow down the area for inspection, but I could only give you a clue: git: yourkarma/jwt branch: master ( latest release, for example, 2.0.2 ) JWTSpec.m describe encoding -> context claims set -> it decode claims set and verify it correctly. steps to reproduce: download latest release. open JWT.xcworkspace cleanup everything. run tests. do 4 - 3 - 4 - 4 - 4 - 3 - 4 or whatever order you'd like. catch results as failure of example mentioned above. Thanks for the remastered version! 😁 So I guess the problem is still occurring? Is it possible to reduce your test case at all? That is, does a dictionary comparison fail stochastically when not using code from the JWT project? I ask because Kiwi uses -[NSObject isEqual:] for its equality matcher -- there isn't any special code path when comparing NSDictionary. For two dictionaries to only sometimes not be equal, I'd be forced to suspect either: All Kiwi equality comparisons fail stochastically, or... ...Apple's -[NSDictionary isEqual:] sometimes returns different values. I think the more likely explanation is that something in the JWTSpec_encoding_claims_set_decode_claims_set_and_verify_it_correctly test has a race condition, which causes the two dictionaries to not be equal sometimes. Of course, it's possible there's some bug I'm overlooking in Kiwi. Reducing the test case would help find that, too! @modocache hey, how could I isolate this test? Is it possible to move it to single spec file? I think no, because kiwi doesn't allow to run single test :( Single test project as an option? Hey @lolgear! Yeah, I think a new Xcode project with just this test, and just the source code it's testing, will probably help isolate the problem better. @modocache check master branch: https://github.com/lolgear/JWT Inspection -> SherlockHolmes project. ( In spec you could find steps to reproduce issue ) @modocache any update? Not yet, sorry! Will try to find a spare minute during a weekend soon. Or if you have some spare time before I do, you could take a look at the Kiwi internals to try and figure it out!
gharchive/issue
2016-05-22T19:09:24
2025-04-01T06:39:17.797884
{ "authors": [ "lolgear", "modocache" ], "repo": "kiwi-bdd/Kiwi", "url": "https://github.com/kiwi-bdd/Kiwi/issues/682", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
641271742
feat: support for multiple examples Implements #589 Updates: add specs/openapi/examples to hold the logic for retrieving examples from endpoints add support for OpenAPI examples Tests: add unit tests for retrieving examples from "examples" add tests to make sure the static_parameter objects are created correctly Note: Implementation allows the "example" and "examples" keywords to be used at the same time. We get all examples from "examples", and if "example" is used also, we create one more strategy using the "example" values. One concern I have is that, when a request body has multiple mediaTypes, we retrieve examples from the "first" mediaType. However, because the API def is a dictionary, we don't know that calling next(iter(media_types_dict)) actually gives the first mediaType. The important thing is that we get examples from the same mediaType that is used in specs/openapi/serialization. But I don't know that we can rely on getting the same mediaType because different dictionaries may have different hashing functions. Great job! Re: dictionaries - I assume, that there might be corner-cases on non-CPython implementation of Python 3.6, but not sure how important it is to handle as we only declare CPython compatibility in pyproject.toml. Also, I don't know if PyPy has the same implementation detail as CPython for dicts on 3.6 (it might be, because that dicts impl came originally from PyPy) - otherwise, I am not sure if we need to support other implementations, that don't guarantee insertion ordering for dicts. I.e. I'd like to know what would be the affected space before taking action on this regard.
gharchive/pull-request
2020-06-18T14:50:19
2025-04-01T06:39:17.806059
{ "authors": [ "Stranger6667", "barrett-schonefeld" ], "repo": "kiwicom/schemathesis", "url": "https://github.com/kiwicom/schemathesis/pull/620", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1784588852
🛑 Thainess is down In 1844861, Thainess (https://thainess.net) was down: HTTP code: 0 Response time: 0 ms Resolved: Thainess is back up in 0db381b.
gharchive/issue
2023-07-02T11:16:37
2025-04-01T06:39:17.834846
{ "authors": [ "kizzna" ], "repo": "kizzna/uptime", "url": "https://github.com/kizzna/uptime/issues/405", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1221899697
confusing example on helper - color for prompt only needed - no matching brackets, completions etc. Hello, I have seen the example file but i find it quite confusing. I do not need bracket matching, file completions or hinting, just plain colored prompt with some ANSI string (i got e.g. from colored crate, or manually written). Can I add highlight_prompt functionality, without all the "underlated" traits (completion, hints, validations, ...)? There are some references to colored_prompt in the example code, but helper has extra methods & traits that seem completely unrelated to the prompt higlighting, and it's not clear whether they can/should be replaced with some placeholder code to keep other functionality intact. There is also an example with only an highlighter: https://github.com/kkawakam/rustyline/blob/master/examples/read_password.rs#L30-L32 And all Highlighter trait methods have a default implementation that does nothing (no highlighting) so you just have to overwrite highlight_prompt impl. perfect, thank you for direction! :)
gharchive/issue
2022-04-30T19:05:21
2025-04-01T06:39:17.854381
{ "authors": [ "RustyJoeM", "gwenn" ], "repo": "kkawakam/rustyline", "url": "https://github.com/kkawakam/rustyline/issues/628", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
683364932
重构 Icon 组件 Tasking [x] 解决逻辑报错 [x] 使用 composition api 重构逻辑 [x] 更新使用文档 Comment 我即将重构这个组件。doing 我即将重构这个组件。doing 其实 已经重构完了的emmmm 文档页demo显示有问题,因为少了一个div所以应用不到.demo-icon .source > div > i 选择器样式,其实所有组件demo都有这个问题
gharchive/issue
2020-08-21T07:35:17
2025-04-01T06:39:17.856870
{ "authors": [ "ImJustAMan", "calcyu", "dengshenkk" ], "repo": "kkbjs/element3", "url": "https://github.com/kkbjs/element3/issues/42", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
657585025
discussion: improve project layout and etc... now youtube have more code and functions than before. To improve readability and maintainability, we would like to reorganize project layout, some plans are under discussion: project layout: change to cmd/pkg(internal)/etc reorganize errors move some functions from youtube to own package Thanks for any suggestion in advance. closed by v2
gharchive/issue
2020-07-15T19:07:18
2025-04-01T06:39:17.858697
{ "authors": [ "Julian-Chu" ], "repo": "kkdai/youtube", "url": "https://github.com/kkdai/youtube/issues/73", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2523958839
[Snyk] Security upgrade node from 18.13.0 to 18.20.4 Snyk has created this PR to fix 2 vulnerabilities in the dockerfile dependencies of this project. Keeping your Docker base image up-to-date means you’ll benefit from security fixes in the latest version of your chosen image. Snyk changed the following file(s): Dockerfile We recommend upgrading to node:18.20.4, as this image has only 184 known vulnerabilities. To do this, merge this pull request, then verify your application still works as expected. Vulnerabilities that will be fixed with an upgrade: Issue Score Out-of-bounds Write SNYK-DEBIAN11-GLIBC-5927133   829   Out-of-bounds Write SNYK-DEBIAN11-LIBWEBP-5893094   829   Out-of-bounds Write SNYK-DEBIAN11-LIBWEBP-5893094   829   Out-of-bounds Write SNYK-DEBIAN11-LIBWEBP-5893094   829   Out-of-bounds Write SNYK-DEBIAN11-LIBWEBP-5893094   829   [!IMPORTANT] Check the changes in this PR to ensure they won't cause issues with your project. Max score is 1000. Note that the real score may have changed since the PR was raised. This PR was automatically created by Snyk using the credentials of a real user. Note: You are seeing this because you or someone else with access to this repository has authorized Snyk to open fix PRs. For more information: 🧐 View latest project report 📜 Customise PR templates 🛠 Adjust project settings 📚 Read about Snyk's upgrade logic Learn how to fix vulnerabilities with free interactive lessons: 🦉 Learn about vulnerability in an interactive lesson of Snyk Learn. Opened a fix PR for one issue as per the document
gharchive/pull-request
2024-09-13T06:13:20
2025-04-01T06:39:17.873622
{ "authors": [ "kkebaara" ], "repo": "kkebaara/nodejs-goof", "url": "https://github.com/kkebaara/nodejs-goof/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1786121552
bug: Error when get_caller_address() is called within a view call Bug Report Current behavior: In a view call, CALLER and ORIGIN opcodes return 0. Expected behavior: The should behave the same in view and execute call. Related code: https://github.com/kkrt-labs/kakarot/blob/e5556e30b7560b8d647899b6dc27bf4b20857f38/src/kakarot/library.cairo#L292 https://github.com/kkrt-labs/kakarot/blob/e5556e30b7560b8d647899b6dc27bf4b20857f38/src/kakarot/instructions/environmental_information.cairo#L129 https://github.com/kkrt-labs/kakarot/blob/e5556e30b7560b8d647899b6dc27bf4b20857f38/src/kakarot/instructions/environmental_information.cairo#L159 Other information: See JSON-RPC spec and evm.codes How to reproduce fork: https://github.com/danilowhk/kakarot-opcode-test add "@view execute" function on Kakarot adapt: .env (kakarot contract and class_hash) View function code: @view func execute{ syscall_ptr: felt*, pedersen_ptr: HashBuiltin*, range_check_ptr, bitwise_ptr: BitwiseBuiltin* }( starknet_contract_address: felt, evm_contract_address: felt, bytecode_len: felt, bytecode: felt*, calldata_len: felt, calldata: felt*, value: felt, gas_limit: felt, gas_price: felt, ) -> ( stack_accesses_len: felt, stack_accesses: felt*, stack_len: felt, memory_accesses_len: felt, memory_accesses: felt*, memory_bytes_len: felt, starknet_contract_address: felt, evm_contract_address: felt, return_data_len: felt, return_data: felt*, gas_used: felt, ) { return Kakarot.execute( starknet_contract_address, evm_contract_address, bytecode_len, bytecode, calldata_len, calldata, value, gas_limit, gas_price, ); } Ok, want to make sure I am understanding the problem space here-- our tests in python are calling these opcodes in the exact same way, so we are debugging a difference in VM issue and not a logic issue? really important context here. this comment is around our integration tests in python, anyone can give context for it? https://github.com/kkrt-labs/kakarot/blame/e273a58716f6a9a6f787b3b3817cbf1ad35035df/tests/integration/test_kakarot.py#L73 What is the status of this issue? This is still an issue: In this contract: pragma solidity ^0.8.0; contract TestContract { // This function will return the address of the caller function getCallerAddress() public view returns (address) { return msg.sender; } }``` When trying to call getCallerAddress() from Remix, there is an error: ```2023-07-26 11:55:36 [2023-07-26T09:55:36Z WARN katana_core::backend] Call error: VirtualMachineExecutionErrorWithTrace { trace: "Error in the called contract (0x07a20c8450211766ecde6cb15882f330381531a29da72648a893940a41728c8b):\nError at pc=0:37:\nGot an exception while executing a hint: Custom Hint Error: Requested contract address ContractAddress(PatriciaKey(StarkFelt(\"0x0000000000000000000000000000000000000000000000000000000000000000\"))) is not deployed.\nCairo traceback (most recent call last):\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:17986)\nUnknown location (pc=0:17364)\nUnknown location (pc=0:7532)\nUnknown location (pc=0:1737)\n", source: CairoRunError(VmException(VmException { pc: 37, inst_location: None, inner_exc: Hint((0, CustomHint("Requested contract address ContractAddress(PatriciaKey(StarkFelt(\"0x0000000000000000000000000000000000000000000000000000000000000000\"))) is not deployed."))), error_attr_value: None, traceback: Some("Cairo traceback (most recent call last):\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:18084)\nUnknown location (pc=0:17986)\nUnknown location (pc=0:17364)\nUnknown location (pc=0:7532)\nUnknown location (pc=0:1737)\n") })) }``` It seems this is not an issue when calling directly from forge: The test for this is the address sender_address = counter.getCallerAddress(); // SPDX-License-Identifier: UNLICENSED pragma solidity ^0.8.13; import "forge-std/Script.sol"; import "kakarot/PlainOpcodes/GetCaller.sol"; contract GetCallerScript is Script { GetCaller public getCaller; function run() external { uint256 deployerPrivateKey = vm.envUint("EVM_PRIVATE_KEY"); vm.startBroadcast(deployerPrivateKey); getCaller = new GetCaller(); address sender_address = getCaller.getCallerAddress(); console.logAddress(sender_address); require( sender_address == address(0xf39Fd6e51aad88F6F4ce6aB8827279cffFb92266), "Address should be 0xf39Fd6e51aad88F6F4ce6aB8827279cffFb92266" ); vm.stopBroadcast(); } } GetCaller Contract: // SPDX-License-Identifier: MIT pragma solidity >=0.8.0; contract GetCaller { address private caller; // This function will return the address of the caller function getCallerAddress() public view returns (address) { return msg.sender; } } What is the status of this? Is it linked to get_caller_address issue? What's the status of this? Tried it locally, it works for latest Kakarot commit. Closing this issue
gharchive/issue
2023-07-03T12:50:46
2025-04-01T06:39:17.885530
{ "authors": [ "Eikix", "danilowhk", "jobez" ], "repo": "kkrt-labs/kakarot", "url": "https://github.com/kkrt-labs/kakarot/issues/617", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2518217021
Usage double actual use Hiya, love your work on this and appreciate your efforts. This is more of an FYI unless there's some clarity others had. I was setting up my own glow and having trouble adjusting the sensitivity on the photodiode. It seemed unless I got it perfect it would either not register pulses or double count them. My meter is outside in a not very sealed box so perhaps that doesn't help. I just added the following options to my config: internal_filter: 200ms internal_filter_mode: PULSE I have had the 200ms filter in for a while but adding the filter mode seems to have solved the doublt counting issue. I'm not entirely across why, I think it might be something to do with the daylight but I don't really understand why it would be. I have exactly same issue. All the values are doubled to what the actual usage is. I have checked all the settings should be ok.
gharchive/issue
2024-09-11T00:38:28
2025-04-01T06:39:17.888094
{ "authors": [ "Dale-Muccignat", "osro" ], "repo": "klaasnicolaas/home-assistant-glow", "url": "https://github.com/klaasnicolaas/home-assistant-glow/issues/592", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
513510356
Kill or restart? Sorry if this is obvious, but does this kill and not start the worker back up, or will it do a restart. Thanks Hi @randy-girard It doesn't restart, you may have to do it manually or have a monitor tool like upstart or monit to restart Sidekiq automatically. @randy-girard if you are using Heroku, restarts are automated as well :) I don't think we could handle a restart since the process is kind of killing itself: we are generating a child thread from sidekiq process, and it is this thread that sends SIGTERM, and then dies as well... Thanks yall.
gharchive/issue
2019-10-28T19:28:21
2025-04-01T06:39:17.893228
{ "authors": [ "BuonOmo", "msxavi", "randy-girard" ], "repo": "klaxit/sidekiq-worker-killer", "url": "https://github.com/klaxit/sidekiq-worker-killer/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1173185569
Added compatibility contents with tx constructor Proposed changes From caver-js v1.8.1-rc.4, to support multiple Caver instances, transaction instance creation through the constructor for each transaction type is not supported. Also i checked the example repo, and there was no places where use constructor to create tx. In this PR, i've added something to what's described above. v1.8.1-rc.4 is not released yet, so i wanna merge this after release. Types of changes Please put an x in the boxes related to your change. [x] Minor Issues and Typos [ ] Major Content Contribution [ ] Others Checklist Put an x in the boxes that apply. You can also fill these out after creating the PR. If you're unsure about any of them, don't hesitate to reach out. We're here to help! This is simply a reminder of what we are going to look for before merging your code. [x] I have read the CONTRIBUTING GUIDELINES [x] I have signed the CLA [x] I have added necessary documentation (if appropriate) [ ] Any dependent changes have been merged and published in downstream modules Related issues Please leave the issue numbers or links related to this PR here. Further comments If this is a relatively large content contribution, kick off the discussion by explaining why you would suggest the content contribution, etc... Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it. @terrikwak Please take a look :) @terrikwak @kjhman21 I think i don't have permission to merge. Can you merge this for me? @dcground @neoofklaytn Please take a look this :) Thank you
gharchive/pull-request
2022-03-18T04:48:51
2025-04-01T06:39:17.913041
{ "authors": [ "CLAassistant", "jimni1222" ], "repo": "klaytn/klaytn-docs", "url": "https://github.com/klaytn/klaytn-docs/pull/400", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1305474624
Create CLA.yml Proposed changes Create CLA pipeline on Github Actions Types of changes Please put an x in the boxes related to your change. [ ] Bugfix [ ] New feature or enhancement [x] Others Checklist Put an x in the boxes that apply. You can also fill these out after creating the PR. If you're unsure about any of them, don't hesitate to ask. We're here to help! This is simply a reminder of what we are going to look for before merging your code. [x] I have read the CONTRIBUTING GUIDELINES doc [x] I have signed the CLA [ ] Lint and unit tests pass locally with my changes ($ make test) [ ] I have added tests that prove my fix is effective or that my feature works [ ] I have added necessary documentation (if appropriate) [ ] Any dependent changes have been merged and published in downstream modules Related issues Further comments Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
gharchive/pull-request
2022-07-15T01:47:48
2025-04-01T06:39:17.919104
{ "authors": [ "CLAassistant", "iv0rish" ], "repo": "klaytn/klaytn-integration-tests", "url": "https://github.com/klaytn/klaytn-integration-tests/pull/12", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2318440952
update the workload proposl What type of PR is this? update the workload proposl What this PR does / why we need it: Which issue(s) this PR fixes: Fixes # Special notes for your reviewer: Does this PR introduce a user-facing change?: Codecov Report All modified and coverable lines are covered by tests :white_check_mark: :exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality. Flag Coverage Δ unittests 31.67% <ø> (?) Flags with carried forward coverage won't be shown. Click here to find out more.
gharchive/pull-request
2024-05-27T07:27:17
2025-04-01T06:39:17.966095
{ "authors": [ "codecov-commenter", "kwb0523" ], "repo": "kmesh-net/kmesh", "url": "https://github.com/kmesh-net/kmesh/pull/369", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
279464654
Added album name support. I have added Album Name support to SpotMenu, of course with a preference to configure it. By default, it is set to true. I also updated the Xcode Workspace project format to be Xcode 8 and higher compatible only, since we don't target any macOS versions that still use Xcode 3.2. I have tested this a bunch, and so far it works really well aside from the fact that I haven't figured out a way to hide the album name if it's the exact same as the song title unfortunately. looks good to me!
gharchive/pull-request
2017-12-05T17:35:42
2025-04-01T06:39:17.967870
{ "authors": [ "KamranMackey", "kmikiy" ], "repo": "kmikiy/SpotMenu", "url": "https://github.com/kmikiy/SpotMenu/pull/86", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1000414041
How do I find keycodes for special German / Danish / Nordic keys to use in defsrc? I'm trying to set up using a laptop with a Danish keyboard. It's a ThinkPad, and I have been using the X220 DE template by @slotThe to get started defining my defsrc block. The template quickly doesn't match, as my keyboard has ½ left of 1, so I replace that just with ^ as in the X220 template, to get started. Alas: kmonad: Parse error at 24:3: | 24 | ^ 1 2 3 4 5 6 7 8 9 0 kp+ grv bspc | ^ unexpected '^' expecting ')' or keycode Figuring that maybe I can't just paste, I try replacing ^ with +: kmonad: Parse error at 24:3: | 24 | + 1 2 3 4 5 6 7 8 9 0 kp+ grv bspc | ^ unexpected '+' expecting ')' or keycode OK. So I look up the keycodes in Keycode.hs and see that maybe I should use kp+ instead of just +. Succés! Now that doesn't error. Alas, my first Danish character does: kmonad: Parse error at 25:58: | 25 | tab q w e r t y u i o p å kp+ ret | ^ unexpected 'å' expecting ')' or keycode OK. I cant find anything relevant in Keycodes.hs, so to move on, I replace it with what is in the X220 template, namely ß -- eszett, the german double s. Alas, to no avail: unexpected 'ß' expecting ')' or keycode Now I'm confused. I can't seem to find anything in Keycodes.hs. Others have seemingly had it working earlier. I'm at a loss. How do I figure out what keycodes I should use to define my defsrc block? I apologize if I have missed something obvious. I'm on the Ubuntu-related Pop!_OS 21.04 running Wayland. The template quickly doesn't match, as my keyboard has ½ left of 1 The defsrc block is not really meant as an accurate representation of the layout that you want, but more of your physical keyboard as a whole. As a rule of thumb, always specify a generic US keyboard layout there. It is also rather restricted in terms of what it accepts, hence symbols that wouldn't phase the parser in a deflayer block are giving you some troubles here. @slotThe is spot on. The ink on your keys doesn't match the events that get sent to the kernel. The easiest solution would be to run evtest on your keyboard and inspect the events that get sent to the kernel. Those have to line up with your (defsrc ...) definition, and then the rest should work. Thank you! Defining a US layout was straightforward, and I have now verified its correctness with evtest. Mapping the US layout back to Danish, I run into some unforeseen and undesirable behavior. Following the "Special characters" section in this suggested wiki-etry, I've added the following to else empty ~/.XCompose include "%L" <Multi_key> <a> <o> : "å" <Multi_key> <a> <e> : "æ" <Multi_key> <o> <e> : "ø" <Multi_key> <1> <2> : "½" <Multi_key> <1> <0> : "´" <Multi_key> <1> <p> : "¨" <Multi_key> <1> <a> : "'" and use the following config.kbd: (defcfg input (device-file "/dev/input/by-path/platform-i8042-serio-0-event-kbd") output (uinput-sink "KMonad: X1C9" "/usr/bin/sleep 1 && /usr/bin/setxkbmap -option compose:ralt") cmp-seq ralt ) (defalias å #(ralt a o) æ #(ralt a e) ø #(ralt o e) ½ #(ralt 1 2) ´ #(ralt 1 0) ¨ #(ralt 1 p) ' #(ralt 1 a) ) (defsrc esc f1 f2 f3 f4 f5 f6 f7 f8 f9 f10 f11 f12 home end ins del grv 1 2 3 4 5 6 7 8 9 0 - = bspc tab q w e r t y u i o p [ ] ret caps a s d f g h j k l ; ' \ lsft 102d z x c v b n m , . / rsft wkup lctl lmet lalt spc ralt cmps rctl back up fwd left down rght ) (deflayer default esc f1 f2 f3 f4 f5 f6 f7 f8 f9 f10 f11 f12 home end ins del @½ 1 2 3 4 5 6 7 8 9 0 + @´ bspc tab q w e r t y u i o p @å @¨ ret caps a s d f g h j k l @æ @ø @' lsft 102d z x c v b n m , . - rsft wkup lctl lmet lalt spc ralt cmps rctl back up fwd left down rght ) That works, partially, but with a lot of issues. The keys work, but in different ways, each key acts unexpected in combination with ´Shift´ and ´AltGr´. Am I missing some step? No, that is how special characters work in Linux, we use compose sequences to emit rapid macros that tell the OS what to encode. I.e. we emit something like altgr -> shifted-' -> e to emit an e-umlaut. This is an alright method of emitting special characters if you don't need them too much. The other option that might work better if you have to use special characters a lot is to simply emit the raw keycodes that your keyboard would have (i.e. if you test evtest on æ and it says that internally it's coded as ;, you can encode your keyboard to just emit a ; (maybe document it in the comments of your keymap). Then you can let the internationalization settings of your OS deal with translating it into a æ. Do you understand what I'm describing? I believe I do, and thank you for it! As option 1 doesn't seem to work ideally for me (e.g., I do need my Shift + ø to be Ø, not Œ ), I tried option 2. Option 2 works under X, but not under Wayland, when I boot with no remappings active and my input source set to Danish, run kmonad with a config file where deflayer default is identical with defscr from my comment above, with nothing in my .XCompose. The KMonad name did seem to hint to somebody's preferences :) I'm sorry if I'm being slow here, but what you are suggesting is that I as a first step get my hands on an external, non-Danish keyboard and check whether that acts as desired with my input source set to Danish, under Wayland? And if not, then troubleshoot that, learn a lesson, and apply the same to KMonad? I'm sorry if I'm being slow here, but what you are suggesting is that I as a first step get my hands on an external, non-Danish keyboard and check whether that acts as desired with my input source set to Danish, under Wayland? Oops: didn't explain my suggestion correctly. What I was suggesting was treating your KMonad remapping entirely as US-english, but keeping in the back of your head that your OS is going to be remapping some US keys to Danish keys. So don't try to get KMonad to emit special characters, like Ø (which it will try to do using compose sequences), instead just get KMonad to emit the US character which your OS will interpret as Ø. That was all the shifting behavior etc. should just work out of the box. Oh: and as an addendum, how you could do that under X: After KMonad is launched, call an setxkbmap dk command, setting the (OS) keyboard map to something. Whenever you plug in a new keyboard, that command needs to be called, and when KMonad starts it 'plugs in' a simulated keyboard, so you'll have to call that after you start KMonad. There is a post-init setting in the defcfg section that will let you do this automatically, but experiment by hand first. You can also switch to a vt (with ctrl-alt-f2) and use the evtest utility. It will give you the linux keynames directly, which you can use in your config as-is (just remove the prefix and convert to lowercase). This seems fixed so I'm closing for now Sorry for commenting on an old, closed issue, but I wondered how exactly I could achieve the behavior described in: What I was suggesting was treating your KMonad remapping entirely as US-english, but keeping in the back of your head that your OS is going to be remapping some US keys to Danish keys. So don't try to get KMonad to emit special characters, like Ø (which it will try to do using compose sequences), instead just get KMonad to emit the US character which your OS will interpret as Ø. That was all the shifting behavior etc. should just work out of the box. Because this is exactly what I'm looking for. Basically, I want to be able to insert foreign symbols (German umlauts in my case) while a modifier key is pressed. I want to be able to get upper- and lowercase characters, which doesn't seem to work if I use the symbols directly in my kmonad mappings.
gharchive/issue
2021-09-19T20:40:07
2025-04-01T06:39:17.988278
{ "authors": [ "Artefact2", "david-janssen", "humanplayer2", "slotThe", "tuxflo" ], "repo": "kmonad/kmonad", "url": "https://github.com/kmonad/kmonad/issues/360", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
836982016
Configurable Privacy Options While discussing this project + chat privacy in general, we had a small discussion around chat privacy and it will be useful for this project to have configurable privacy options so that this can serve varied groups which have different privacy needs. Posting some initial thoughts on this Option to show / hide timestamp - the template can show / hide timestamp based on this. Link Mode :- Option to pick only links / media and scrub off any comments. This lets the group archive only resources while skipping the chatter. This lets community have freedom to chat, prevents chatty noise on archives, while preserving value. This can be True / False Include / Exclude from archives based on select hashtags - If LinkMode is True, a set of hashtags can be added as 'Include' hashtags so those important messages get archived, wihle leaving out remaining chats. If LinkMode is False, a set of hashtags can be added to 'Exclude' hashtag (like #DontArchive #KeepThisPrivate), so that the group can still have private non-archiving conversations even while on Full Archive mode. Thoughts? "This is a privacy risk, lets stay as is" Just setting some context, this was discussed in Foss United as well before publishing the archive. Like @knadh mentioned, for an already public Telegram group, the chats are available publicly anyway. About privacy, like @mr-karan pointed out, by definition, there is no privacy in a public Telegram group. Anyone can join and read/export/copy/re-publish messages at any point. Option to pick only links / media and scrub off any comments This is too niche an option to include in the global config, but can be easily achieved externally by doing a DELETE from messages WHERE media_id is NULL on the .sqlite file. Include / Exclude from archives based on select hashtags Again, too specific to include in the config but can be easily achieved by querying the .sqlite file. These can be done with the SQLite CLI.
gharchive/issue
2021-03-21T04:33:30
2025-04-01T06:39:18.008394
{ "authors": [ "knadh", "mr-karan", "srikanthlogic" ], "repo": "knadh/tg-archive", "url": "https://github.com/knadh/tg-archive/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2002719173
Make SECURITY.md consistent We're missing these across a lot of Knative repos, this is copied/improved from the 3 that existed. /approve /lgtm
gharchive/pull-request
2023-11-20T17:56:18
2025-04-01T06:39:18.009838
{ "authors": [ "davidhadas", "evankanderson" ], "repo": "knative-extensions/kn-plugin-source-kamelet", "url": "https://github.com/knative-extensions/kn-plugin-source-kamelet/pull/239", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
744180676
e2e test framework poc needs more flexibility So none of the networking conformance tests currently have levels. Only serving has some. I think for serving a runtime conformance test will be structured as Create a Service Make a get request to get some info about the running environment Run the levelled assertions against the returned info Repeat step 3 until you have no more assertions A problem I see is the info from step 2 needs to go to step 3. With the framework as is this is going to be cumbersome - especially if each levelled assertion has unique logic. Originally posted by @dprotaso in https://github.com/knative-sandbox/reconciler-test/pull/30#discussion_r520992411 /assign @n3wscott Another example would be the API operations for different types https://github.com/knative/docs/blob/master/docs/serving/spec/knative-api-specification-1.0.md#service Do you have an example test with code so I can understand? If I was using https://github.com/knative-sandbox/reconciler-test/blob/master/pkg/test/context.go I'd write it as follows func TestService_APIOperations(t *test.T) { var s v1.Service = someTestService() t.Must("Create", func(t *test.T) { t.ServingClient.Create(...) ... }) t.Must("Update", func(t *test.T) { t.ServingClient.Update(...) ... }) t.Must("Get", func(t *test.T) { t.ServingClient.Get(...) ... }) t.Should("Patch", func(t *test.T) { t.ServingClient.Patch(...) ... }) t.Must("Delete", func(t *test.T) { t.ServingClient.Delete(...) ... }) } I would assume that looks something like: func StableAPIOpsFeature() *feature.Feature { f := new(feature.Feature) f.Stable("Service"). Must("Create", func(ctx context.Context, t *test.T) { servingclient.Get(ctx).Create(...) ... })). Must("Update", func(t *test.T) { servingclient.Get(ctx).Update(...) ... }). Must("Get", func(t *test.T) { servingclient.Get(ctx).Get(...) ... }). Should("Patch", func(t *test.T) { servingclient.Get(ctx).Patch(...) ... }). Must("Delete", func(t *test.T) { servingclient.Get(ctx).Delete(...) ... }) return f } We need to build some support code, like implement namespace scoped clients and inject them into context: like the tbd servingclient package. the caveat is that each test should be independent of the previous. Because if you depend on ordering, it will fail, or it will be filtered out if only Should is run... etc the caveat is that each test should be independent of the previous. Because if you depend on ordering, it will fail, or it will be filtered out if only Should is run... etc In the prior example I have control over the test sequence since they naturally tie into golang's test lifecycle. This is the point I'm trying to make - this separation makes this more cumbersome ie. another example asserting which filesystem properties in our runtime contracts. func TestRuntimePaths(gotest *testing.T) { t := globals.NewT(gotest) var s v1.Service = someTestService() t.ServingClient.Create(s) // once ready fetch it's runtime environment env := http.Get(s.Status.URL + "/runtime") for _, path := range runtimev1.MustFilesystemPaths { t.Must(path, checkPath(env, path)) } for _, path := range runtimev1.MayFilesystemPaths { t.May(path, checkPath(env, path)) } } bump I am trading cumbersome with composability and decoupled tests. I would write the TestRuntimePaths like this: func TestRuntimePaths(t *testing.T) { ctx, env := global.Environment() // Use the existing helpers to make a KSVC. env.Prerequisite(ctx, t, features.ServiceIsCreatedAndReady(s)) // once KSVC is ready fetch it's runtime environment env := http.Get(s.Status.URL + "/runtime") // RuntimePaths is the feature we are trying to test. env.Test(ctx, runtime.RuntimePaths(env)) env.Finish() } // ...in a runtime features package... func RuntimePaths(env http.Response) *feature.Feature { f := new(feature.Feature) s := f.Stable("RuntimePaths") for _, path := range runtimev1.MustFilesystemPaths { s.Must("have "+path , checkPath(env, path)) } for _, path := range runtimev1.MayFilesystemPaths { s.May("have"+path, checkPath(env, path)) } return f } This might look like more code, but what we can do is break down the thing we are trying to assert is a feature (the payload of the runtime contract) from the how we get that response. So with some edits to the test entry point, we can test the runtime or whatever to another way of getting that data, say, a kn environment result or something like that, or a CloudRun curl on another kind of endpoint. The example above leads me to more questions/thoughts: 1) Use of Feature in lieu of a list of steps in Prerequisites env.Prerequisite(ctx, t, features.ServiceIsCreatedAndReady(s)) Prerequisite accepting a Feature seems weird since there could be levelled requirements and/or feature state assertions that I wouldn't expect. 2) Generating state from prerequisite steps & using them in assertions var s v1.Service = someTestService() env.Prerequisite(ctx, t, features.ServiceIsCreatedAndReady(s)) rt := http.Get(s.Status.URL + "/runtime") someTestService() in my mind just returns a v1.Service go struct. If that's the case is features.ServiceIsCreatedAndReady(s) meant to mutate s with the intent of setting a valid s.Status.URL? env.Prerequisite and env.Test both take unique contexts so it seems like there's no hand off of state between the two functions. Meaning Prerequisite doesn't allow for the context to be mutated. This leads to two ways to consume dependencies for an assertion - closures (in your example above) and the context function argument that's originally passed to Test (which I've seen include clients/informers etc.). I'm not advocating we let env.Prerequisite modify the context. In order for that to work the context keys need to be known by the prereq & assertion steps - which I'd argue is coupling them together. Maybe a workaround would be to dynamically create assertions with a context keys as input but :shrug: this become fairly complex. 3) Coalescing of Features Steps I was sold on the fact that we could describe features in a fluent style. I like this because having a central place to see all the aspects of a feature set is great. This is a pattern that originated from networking conformance. ie. thus for a runtime conformance I'd like to write func RuntimeV1Conformance() feature.Feature { f := new(feature.Feature) f.Stable("filesystem", TestRuntimePaths) f.Stable("http", TestRuntimeHTTP) f.Stable("http-upgrade", TestRuntimeHTTPUpgrade) // etc.. } In the example the test setup/runner TestRuntimePaths has specific logic that's unique only to RuntimePaths. Given that I'm not sure we can still describe a holistic RuntimeV1Conformance feature if I needed to create unique services for different aspects of the runtime conformance. It seems like I have two options: Figure out how to use something like features.ServiceIsCreatedAndReady(s) N times in my test runner con: changing a test requires changes in two places setting up N services (on a single context?) to be consumed by N different assertions requires coordination Create the services in each assertion func(ctx.Context, t*testing.T) con: things that are written as steps can't be used here - ie. features.ServiceIsCreatedAndReady so there's going to be code duplication 4) t.Parallel() Serving and Networking use t.Parallel() quite extensively and they're invoked at the feature state boundary. ie. Kingress/basic is parallel wrt. KIngress/websocket. I'm not sure how to do this with feature.Feature or why by default step assertions are all parallelized (https://github.com/knative-sandbox/reconciler-test/pull/52) cc @slinkydeveloper I'm beginning to wonder if feature.Feature should be called something else - feature.Set because the stable/alpha/beta calls are scoped to a feature. I'm not sure how to do this with feature.Feature or why by default step assertions are all parallelized (#52) cc @slinkydeveloper @dprotaso since you want to parallelize all tests, why not? I'm not sure I get your question @dprotaso since you want to parallelize all tests, why not? I'm not sure I get your question Shouldn't test authors control the flow? What if assertions have side-effects on the remote object being tested? This seems relevant to #3 - To guarantee test isolation I would have to change my feature definition either by: Taking what would be in feature.Setup and moving it into every Step - ie. creating N Knative Services instead of one. Creating duplicate feature.Feature definitions with different Steps on each Shouldn't test authors control the flow? What if assertions have side-effects on the remote object being tested? An assertion by definition shouldn't have a side effect right? Are you sure that particular assertion shouldn't live in the setup phase? An assertion by definition shouldn't have a side effect right? Yup - I agree. Are you sure that particular assertion shouldn't live in the setup phase? Ideally it should - but then I'm forced to split my feature.Feature definition. With the side effects being: the same issue I mentioned above where some arguments are passed via closures and some via context env.Test() doesn't run in parallel to other env.Test() invocations - splitting features would actually slow things down Ideally it should - but then I'm forced to split my feature.Feature definition Can you provide a use case for that? It might help me reason on the issue I mentioned it all here see point 2): https://github.com/knative-sandbox/reconciler-test/issues/37#issuecomment-733445763 Another hypothetical kingress := new(feature.Feature) // kingress.Stable("websocket").Must("receive traffic", AssertWebSocketTraffic()) kingiress.Stable("http2").Must("receive traffic", AssertHTTP2Traffic()) If each Stable feature requires a different KIngress configuration where should they be setup? Inside Assert*Traffic calls? con: not really using feature.Setup anymore so you can't re-use steps Setup on the feature? con: all the assertions are block on all the KIngresses to be ready, how do you transfer the right ingress endpoint to the assertion? Split the feature into two? con: more boilerplate, tests run linearly unless you add more Test funcs I mentioned it all here see point 2) This is kinda similar to the discussion happening here: https://github.com/knative-sandbox/reconciler-test/issues/51. My understanding is that to generate state you should use Setup. If you need to generate state, then assert, then generate state and then assert again you need to develop 2 features. I see this even more clearly in your sample with kingress, where yes there is more boilerplate, but it clearly shows how websocket and http2 testing should be 2 different features. Maybe what might be useful is something like "feature template": in your case the setup and teardown is the same except (i guess) one step to trigger the state change to stimulate the kingress. But, assuming you create feature websocket and http2, you could mostly share the same setup and teardown. So what we could do is to create a sort of "template" that you can use to define common setup and teardown for more features: the websocket and http2 features will eventually share the same kingress data plane feature template. But still, "at runtime", those features then are executed separately, so the setup and teardown steps are in fact repeated. The name feature.Feature could be wrong, I have been more focused on how these things are composable and the signatures of the steps and feature providers. Use of Feature in lieu of a list of steps in Prerequisites Ville wanted a setup step that was external to the Test or Feature to get the env to a state that is required by orthogonal to the test you would like to preform. In Eventing this would be installing a class of Broker, in Serving this would be configuring the ingress type. To test out the usage I thought it would be handy to use feature.Feature (bikeshed the name) as a shortcut to make a env.Prerequisite in the same signature as env.Test, it is for the author, reporter and debugger's convence to make it clear what is being tested or asserted inside the Prerequisites phase is independent but required for the Test phase. Generating state from prerequisite steps & using them in assertions I think var s v1.Service = someTestService() should have been var s *v1.Service = someTestService() and it will work out. This leads to two ways to consume dependencies for an assertion - closures (in your example above) and the context function argument that's originally passed to Test (which I've seen include clients/informers etc.). The framework should provide the hooks but make no opinion on which way is best. For some cases having some magic thing in context might be the best, like a set of namespaced clients setup in some Prerequisite phase serving likes to use. I'm not sure we can still describe a holistic RuntimeV1Conformance feature. I don't think you can or want to. I would see this written as a series of Test calls on an environment with a list of features that are required to pass conformance: // RuntimeV1Conformance func TestRuntimeV1Conformance(t *testing.T) { ctx, env := global.Environment() env.Test(ctx, t, conformance.Feature1()) env.Test(ctx, t, conformance.Feature2()) env.Test(ctx, t, conformance.Feature3()) env.Test(ctx, t, conformance.Feature4()) <... etc> env.Finish() } The results of this will be collected into a report that is more easily consumable and understood because the Test phase focused on an aspect of conformance. (that is my hope) RuntimeV1Conformance feature if I needed to create unique services for different aspects of the runtime conformance I would see this as something like: // RuntimeV1Conformance func TestRuntimeV1Conformance(t *testing.T) { ctx, env := global.Environment() s := "ksvc-name" env.Precondition(ctx, t, conformance.GivenSerivce(s)) env.Test(ctx, t, conformance.Feature1(s)) env.Test(ctx, t, conformance.Feature2(s)) env.Test(ctx, t, conformance.Feature3(s)) env.Test(ctx, t, conformance.Feature4(s)) <... etc> env.Finish() } t.Parallel() Big plus one, we need to support this, the runner code is total PoC. I am trying to focus on the following things: Features and Steps are venderable cross projects. Several Features can be run on a single env. StepFn has no external dependencies, even to the framework unless you are opting into some base feature. I have ran into some struggles so far with the PoC. One being the timing and isolation of the Steps makes it hard to pass results, so the step and feature isolation cause you to think about how to compose the test so it has no or few dependencies. This results in each step having a bit more code than you might expect, but that also results in the steps being composable in new ways than you originally wrote them in, as an example: // TestBrokerAsMiddleware func TestBrokerAsMiddleware(t *testing.T) { t.Parallel() ctx, env := global.Environment( knative.WithKnativeNamespace(system.Namespace()), knative.WithLoggingConfig, knative.WithTracingConfig, k8s.WithEventListener, ) // Install and wait for a Ready Broker. env.Prerequisite(ctx, t, features.BrokerGoesReady("default", "MTChannelBroker")) // Test that a Broker can act as middleware. env.Test(ctx, t, features.BrokerAsMiddleware("default")) env.Finish() } Here, I need a Broker to be ready, but the focus of the test is not a ready Broker. I wanted to write a feature that assumes for a ready Broker of a given name, I can pass events through it. This means that I can vendor the features and reuse this in other downstream repos can leverage this same test: // <... In eventing-rabbitmq ...> // TestBrokerAsMiddleware func TestBrokerAsMiddleware(t *testing.T) { t.Parallel() ctx, env := global.Environment( knative.WithKnativeNamespace(system.Namespace()), knative.WithLoggingConfig, knative.WithTracingConfig, k8s.WithEventListener, ) // Create a RabbitmqCluster in the env, the CO that creates the underlying RabbitMQ Broker (not knative). env.Prerequisite(ctx, t, rabbitfeatures.RabbitMQBrokerIsCreated()) // Install and wait for a Ready Broker. env.Prerequisite(ctx, t, features.BrokerGoesReady("default", "RabbitMQBroker")) // Test that a Broker can act as middleware. env.Test(ctx, t, features.BrokerAsMiddleware("default")) env.Finish() } So the downstream repo has opt'ed to include this test, but they only have to add the test entry point, not the features. The reasoning here is likely the downstream needs to do some additional setup like above. then generate state and then assert again you need to develop 2 features. Then I'm confused by the feature state methods (Alpha, Beta, Stable etc.) on the Feature struct. To my third point above and following observation I think it's important to be able to go to a single place to see a group of related features together and their level of maturity. ie. like the KIngress example A template approach would work - but I was pointing out in 2) that things neither clearly defined nor consistent I would need to know more about KIngress to be able to answer in full, but I have been assuming a Feature tests a contained set of functionality, and variations would be passed down to it. I think you can compose the test you are wanting to, and you will have to do a bit more work when trying to special case a particular feature of an implementation, if that is not generally testable by all implementations. 1) Use of Feature in lieu of a list of steps in Prerequisites it is for the author, reporter and debugger's convence to make it clear what is being tested or asserted inside the Prerequisites phase is independent but required for the Test phase. Whatever goes here shouldn't have state (alpha, beta, stable) or levels (must, should, may) decorations since those could be skipped via environment flags etc. ie. What if I was only testing alpha features of a broker - the broker ready feature passed to a prereq wouldn't run https://github.com/knative/eventing/blob/fe1b34c4c084eaca1f964e09bd791a428c7bb8cf/test/rekt/features/broker_feature.go#L42-L44 2) Generating state from prerequisite steps & using them in assertions I think var s v1.Service = someTestService() should have been var s *v1.Service = someTestService() and it will work out. The pointer doesn't change my interpretation - s still needs to be mutated or refetched to get additional info for the subsequent steps. 3) Coalescing of Features Steps I'm not sure we can still describe a holistic RuntimeV1Conformance feature. I don't think you can or want to. I would see this written as a series of Test calls on an environment with a list of features that are required to pass conformance: We do this in networking currently and it good for discoverability. A side effect is diffs become very clear https://github.com/knative/networking/pull/277/files. Another example RuntimeV1Conformance feature if I needed to create unique services for different aspects of the runtime conformance I would see this as something like: // RuntimeV1Conformance func TestRuntimeV1Conformance(t *testing.T) { ctx, env := global.Environment() s := "ksvc-name" env.Precondition(ctx, t, conformance.GivenSerivce(s)) env.Test(ctx, t, conformance.Feature1(s)) env.Test(ctx, t, conformance.Feature2(s)) env.Test(ctx, t, conformance.Feature3(s)) env.Test(ctx, t, conformance.Feature4(s)) <... etc> env.Finish() } I see only one service in this example. Also we want to group relevant conformance features to be run together to make it easier for downstream folks to run the correct set of tests. Right now in networking this done using an exported RunConformance function. The results of this will be collected into a report that is more easily consumable and understood because the Test phase focused on an aspect of conformance. (that is my hope) Can you elaborate? I would need to know more about KIngress to be able to answer in full, but I have been assuming a Feature tests a contained set of functionality, and variations would be passed down to it. I think you can compose the test you are wanting to, and you will have to do a bit more work when trying to special case a particular feature of an implementation, if that is not generally testable by all implementations. I would need to know more about KIngress to be able to answer in full, but I have been assuming a Feature tests a contained set of functionality, and variations would be passed down to it. It's conformance so there's are no variations on the features being tested. The only thing that differs is the ingress installation and maybe configuring some global test properties (ie. does DNS work, which ingress class etc.) I have made several issues to continue the discussion in the forks we have made: for 1): Framework should report Prerequisite features with Asserts #64 for 2): Framework needs to help test authors understand how to pass state between steps #66 for 3): Framework needs a way to group Features into Sets #65 The results of this will be collected into a report that is more easily consumable and understood because the Test phase focused on an aspect of conformance. (that is my hope) Can you elaborate? I think we need to define this a bit more. The thinking is the runner can get metadata out of the feature and generate data around the run of the test and conformance is somehow different, but I think we are waiting for the conformance group to ask for data. Tagging @nak3 @ZhiminXiang @tcnghia - can someone with networking start taking a look at the current framework and the open issues Scott created. It may be worth doing a small POC with a single KIngress feature.
gharchive/issue
2020-11-16T20:50:50
2025-04-01T06:39:18.057264
{ "authors": [ "dprotaso", "n3wscott", "slinkydeveloper" ], "repo": "knative-sandbox/reconciler-test", "url": "https://github.com/knative-sandbox/reconciler-test/issues/37", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
613164128
Port Gitlab source to adapter/v2 Problem Gitlab source still uses knative.dev/eventing/pkg/adapter module, which uses the cloudevents/sdk-go v1. We should port it to knative.dev/eventing/pkg/adapter/v2, which uses the new sdk-go Persona: Event developer/Event producer Exit Criteria Gitlab source should not use anymore knative.dev/eventing/pkg/adapter Time Estimate (optional): 1 /assign
gharchive/issue
2020-05-06T09:06:12
2025-04-01T06:39:18.083161
{ "authors": [ "slinkydeveloper" ], "repo": "knative/eventing-contrib", "url": "https://github.com/knative/eventing-contrib/issues/1200", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
557040976
Removal of sources-controller after 0.12.0 Based on the change here: https://github.com/knative/eventing/pull/2448 It is time for operator to shine, in terms of upgrade without manual interruption. I can implement this /assign aliok @aliok PLZ be advised that I am thinking of extending the scope of our tests-on-latest-eventing t verify the upgrade: https://github.com/knative/eventing-operator/pull/93. Hope we do not overlap our work. Some notes: Reconciliation will be updating existing resources and create new resources It won't be able to delete the old resources (that is removed in eventing manifest) So, I wrote a small script where I feed the 0.12 manifest (gsutil cp gs://knative-releases/eventing/previous/v0.12.0/eventing.yaml ./) and latest nightly manifest (gsutil cp gs://knative-nightly/eventing/latest/eventing.yaml ./) to make sure we will cover all resources that were deleted. This is the output: Unable to find 1.2.0 resource in nightly: apiVersion:v1 kind:ServiceAccount name:eventing-source-controller Unable to find 1.2.0 resource in nightly: apiVersion:rbac.authorization.k8s.io/v1 kind:ClusterRole name:knative-eventing-source-controller Unable to find 1.2.0 resource in nightly: apiVersion:rbac.authorization.k8s.io/v1 kind:ClusterRoleBinding name:eventing-source-controller Unable to find 1.2.0 resource in nightly: apiVersion:rbac.authorization.k8s.io/v1 kind:ClusterRoleBinding name:eventing-source-controller-resolver Unable to find 1.2.0 resource in nightly: apiVersion:apps/v1 kind:Deployment name:sources-controller These are the things we need to delete in the operator More changes needed after https://github.com/knative/eventing/pull/2519
gharchive/issue
2020-01-29T18:33:03
2025-04-01T06:39:18.087881
{ "authors": [ "aliok", "houshengbo" ], "repo": "knative/eventing-operator", "url": "https://github.com/knative/eventing-operator/issues/90", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2184242838
[main] Upgrade to latest dependencies GKE fixes in hack -dprotaso /cc knative/serving-writers knative/eventing-writers /assign knative/serving-writers knative/eventing-writers Produced by: knative-extensions/knobots/actions/update-deps /retest /lgtm /approve
gharchive/pull-request
2024-03-13T15:04:54
2025-04-01T06:39:18.125054
{ "authors": [ "dprotaso", "knative-automation" ], "repo": "knative/pkg", "url": "https://github.com/knative/pkg/pull/2987", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
492410858
Renaming Importer to Source Renaming Importer metrics to Source /assign @yanweiguo /lgtm /approve /assign @srinivashegde86 for test/OWNERS approval /lgtm /approve
gharchive/pull-request
2019-09-11T18:58:17
2025-04-01T06:39:18.126933
{ "authors": [ "nachocano", "vaikas-google", "yanweiguo" ], "repo": "knative/pkg", "url": "https://github.com/knative/pkg/pull/662", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
178655041
Grab bag of minor improvements This is a grab bag of minor fixes that things slightly easier on me: expand range of transformers allowed (I'm trying to get to GHC 8 via stack/lts-7) unit tests are now much easier to add add trivial stack.yml add test cases which are expected and known to fail. A sort of TODO list (I plan on working on these, and it's a nice place to add future ones) bump version number and copyright year clean up warnings and linter warnings in tests bug fix+tests: heredoc logic is wrong I can split these into multiple PRs or cut out commits that you don't like: let me know what you prefer! Added one more bug fix: the logic in heredoc was wrong. Also included tests. The logic was reversed - it's the quoted heredoc that does no expansion. Here's the spec from man bash: The format of here-documents is: <<[-]word here-document delimiter No parameter and variable expansion, command substitution, arith- metic expansion, or pathname expansion is performed on word. If any characters in word are quoted, the delimiter is the result of quote removal on word, and the lines in the here-document are not expanded. If word is unquoted, all lines of the here-document are subjected to parameter expansion, command substitution, and arith- metic expansion, the character sequence \<newline> is ignored, and \ must be used to quote the characters \, $, and `. Thanks a lot for the PR! :D I think it's pretty clear that I haven't touched this library since Stack was released, but I'll try to build it again and review the PR this weekend (or Monday). Thanks! I took a look at fixing the arithmetic and it's way beyond my understanding of parsec. If you're interested in fixing that, that would be very awesome 😁. If I had to guess, it seems to look for 4 sets of parens but unsure. I'll take a look. It might be related to #15 My guess though is that the parser at https://github.com/knrafto/language-bash/blob/master/src/Language/Bash/Parse/Word.hs#L157 is the culprit Fixed it! I can't push to your branch, but here's a patch: From 2b34962d7ed4bb1c5e874b164da6d8cfb7e98d8f Mon Sep 17 00:00:00 2001 From: Kyle Raftogianis <knrafto@gmail.com> Date: Sun, 25 Sep 2016 20:21:29 -0700 Subject: [PATCH] Fix parsing of arithmetic expressions (fixes #15) --- src/Language/Bash/Parse/Word.hs | 6 +++--- tests/Tests.hs | 16 +++++++++++----- 2 files changed, 14 insertions(+), 8 deletions(-) diff --git a/src/Language/Bash/Parse/Word.hs b/src/Language/Bash/Parse/Word.hs index 111e606..9e407b0 100644 --- a/src/Language/Bash/Parse/Word.hs +++ b/src/Language/Bash/Parse/Word.hs @@ -154,10 +154,10 @@ backquote = Backquote <$> matchedPair '`' '`' False escape -- | Parse an arithmetic expression. arith :: Stream s m Char => ParsecT s u m String -arith = B.toString <$> parens <?> "arithmetic expression" +arith = B.toString <$> arithPart <?> "arithmetic expression" where - parens = B.many inner - inner = B.matchedPair '(' ')' parens + arithPart = B.many inner + inner = B.noneOf "()" <|> B.char '(' <+> arithPart <+> B.char ')' -- | Parse a parenthesized substitution. subst :: Stream s m Char => ParsecT s u m String diff --git a/tests/Tests.hs b/tests/Tests.hs index d5fcab9..4a2b01c 100644 --- a/tests/Tests.hs +++ b/tests/Tests.hs @@ -91,15 +91,21 @@ unittests = testGroup "Unit tests" heredocDelim = "EOF", heredocDelimQuoted = True, hereDocument = expandString "asd\\`\n"}]) - + , tp "echo $((2 + 2))" + (Command + (SimpleCommand [] [expandString "echo", [ArithSubst "2 + 2"]]) + []) + , tp "((2 + 2))" + (Command (Arith "2 + 2") []) + , tp "echo $(((2 + 2)))" + (Command + (SimpleCommand [] [expandString "echo", [ArithSubst "(2 + 2)"]]) + []) ] failingtests :: TestTree failingtests = testGroup "Failing tests" (map expectFail - [ - tp "echo $((2+2))" - (Command (Arith "2 + 2") []) - ]) + []) tests :: TestTree tests = testGroup "Tests" [properties, unittests, failingtests] -- 2.7.4 (Apple Git-66) The rest of the PR looks good. I'll merge and add the arith fix, and then release on Hackage. Thanks for your help! Thanks for the review and the fix!!
gharchive/pull-request
2016-09-22T16:16:59
2025-04-01T06:39:18.192257
{ "authors": [ "knrafto", "pbiggar" ], "repo": "knrafto/language-bash", "url": "https://github.com/knrafto/language-bash/pull/16", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1495172613
Async improvements Check: https://developers.home-assistant.io/docs/asyncio_working_with_async/ E.g. need to add some await when pulling states for price and energy. no need
gharchive/issue
2022-12-13T21:16:12
2025-04-01T06:39:18.194275
{ "authors": [ "knudsvik" ], "repo": "knudsvik/EnergyScore", "url": "https://github.com/knudsvik/EnergyScore/issues/36", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
279612631
How are you so awesome? This is a bug that needs immediate solving. Are you simply a mountain of raw programming power, or did you do some automated translation and then fill in the gaps? I mean even if the latter, still a mountain of raw programming power. Sorry to open a bug ticket for such a stupid reason but seriously this is a great find. Hopefully now I can kiss IKVM goodbye. You're absolutely right. This library is a work of art, and it is way too obscure and hard to find. Wow!! I am very happy to see the compliments. I worked thousands of hours on this project, something close to 6 months of work, just for fun and curiosity, when I started, I had no knowledge about NLP, I just wanted to learn. All code that exists was written by hand, without tools, I was simply solving one problem after another, to the point of being 100% compatible with the original version in Java. I'm just passionate about programming!
gharchive/issue
2017-12-06T03:20:28
2025-04-01T06:39:18.196478
{ "authors": [ "TechnikEmpire", "TheQuack45", "knuppe" ], "repo": "knuppe/SharpNL", "url": "https://github.com/knuppe/SharpNL/issues/39", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1727828890
Fix kind image loading for MacOS See: https://github.com/kubernetes-sigs/kind/pull/2957 Fixes test following #1054 @cpanato @imjasonh Will this MR be merged ? I am running into the same issue when trying to use Ko on a M series mac
gharchive/pull-request
2023-05-26T15:24:04
2025-04-01T06:39:18.197922
{ "authors": [ "aidy", "psankar" ], "repo": "ko-build/ko", "url": "https://github.com/ko-build/ko/pull/1057", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
127137729
Improving the Koa ecosystem by moving all Koa modules to the koajs organization! With GitHub's new organization administration, I want to try moving all koa middleware/plugins to the org. The goals are: To have shared maintenance of all code. Have all middleware/plugins be in one location. Create teams so everyone knows who to ask for help. Allow members to create repos within the organization. I'm not sure how this will work. I remember last time, you had to be an owner of the organization for you to transfer a repository to an organization. The steps will be something like: Let us know what repository you'd like to transfer We'll create a team for that repository if there isn't one yet We'll add that repository to that team We'll add you to that team If anything, just transfer the repository to me and I'll add it to the organization. Comment here so I remember to add it to the organization! Let me know if anyone wants to transfer any repositories. First one i'd like to transfer is koa-convert :) NOTE: IF YOU'RE IN THIS ORGANIZATION, PLEASE SETUP 2-FACTOR AUTHENTICATION! Agree! I will transfer all koa relate middlewares/plugins to koajs org. I want to transfer coderhaoxin/koa-redis-cache to this org. Maybe cache team is suitable. @coderhaoxin are you able to transfer? if not, what's the error message? I'd like that, too. We can manage teams with an organization and this makes things easier for people to find good quality (and actively maintained) middlewares. I'd transfer https://github.com/m4nuC/async-busboy if that makes sense. @m4nuC It seems that async-busboy is not a koa middleware >_< I don't think it makes sense to transfer async-busboy since its application isn't tied to Koa. @haoxin yes just a module (It's basically co-body for multipart with koa2 support). It could be made a middleware if that makes more sense. On 18 Jan 2016 13:11, haoxin notifications@github.com wrote: @m4nuChttps://github.com/m4nuC It seems that async-busboy is not a koa middleware >_< — Reply to this email directly or view it on GitHubhttps://github.com/koajs/koa/issues/638#issuecomment-172427379. @jonathanong Do we really want all middleware that works in the org though? Doesn't having it in the org signify that the org is going to maintain it? Just because it works doesn't mean it's maintained. some that would be nice in the org... koa-graphql @chentsulin koa-bearer-token @chentsulin koa-jwt @stiang koa-joi-schema @simplyianm We could vote on whether a module is deemed worthy of being in the org or not not sure if github sends notifications for @mentions on edits/updates so here are some more :P koa-resourcer @aheckmann koa-router @alexmingoia We could vote on whether a module is deemed worthy of being in the org or not fair enough, this can get messy in threads. maybe we could get a survey platform? not sure what's good out there... if anyone knows, throw around some suggestions :) @jonathanong When I try to transfer koa-redis-cache to koajs, cache isn't in the teams checklist . ^ Maybe it's a secret team for some odd reason... :P Great idea, I’d be happy to transfer koa-jwt. @jonathanong, can you please confirm that you want koa-jwt in the org? @coderhaoxin but you're in the team and see it, right? @stiang i'll let others decide that. i don't understand JWT :) +1 from me :+1: @jonathanong Got it, but I’m not really sure who is authorized to say whether I should transfer koa-jwt :) @tejasmanohar, you have already requested it, and I see that you are part of the koajs org. Are you in a position to OK the transfer? Nah, let others vote on it and an Owner (who can actually create teams + transfer) decide :) :+1: for the idea :+1: for a voting system. This will also provide insight of what is actually being used (i.e. I didn't even know about koa-jwt!) what about this: if more than one person wants to maintain it (ex. if someone wants to help @stiang maintain that repository), then it's in. also, if someone could figure out how to easily transfer repos to the org, that woudl be great. otherwise, you can transfer it to me (or another owner) and we'll transfer it here if more than one person wants to maintain it (ex. if someone wants to help @stiang maintain that repository), then it's in. @jonathanong :+1: I agree. @stiang I'm all in to help maintain koa-jwt so you can transfer. My projects depend on it anyways :P sweet. @stiang @tejasmanohar added you both to a new @koajs/jwt team. @stiang if you have trouble transferring it over to the org, transfer it to me and i'll transfer it to this org. @jonathanong Great! I’ve transferred it now, looks like it’s already available at https://github.com/koajs/koa-jwt I suppose it should be renamed to just "jwt" to adhere to the naming scheme used by koajs. Also, although I was presented with a list of koajs teams to give permission to, the koa-jwt group was not among them, so I currently don’t have write access to the repo. Could you please add the koa-jwt team manually? @tejasmanohar Very happy that you are willing to help! I’ve fallen behind on PRs and issues lately, so the project could really benefit from some fresh attention. interesting. when you transferred it over, it was added to no teams, so i had to add it to a team manually. Yeah, I could have selected another team, but since the relevant one wasn’t displayed I didn’t add it to any. I tried reloading the page and retrying the transfer again, but I still got just a subset of the koajs teams to select from. Not sure what’s going on there. I renamed the repo to "jwt". If I am a member of a team, am I free to add new teams as well? For example, we have a few koa projects and middleware over at gh://pebble but I wouldn't want to move anything if I can't bring several developers along with it. Does it make sense to (1) create a Pebble team and add people and repos or does it make more sense to (2) stay distributed just leave repos where they are? I'm leaning towards option 2. I don't mind the idea of just everyone having access, if people do weird shit it'll just self-correct haha. That said I don't see a huge problem with using the wiki to list projects, that's still a more condensed view than browsing koajs/* @tj @jonathanong can you guys add me to @koajs and so I can have collab on https://github.com/koajs/ratelimit? I rewrote it the other day and I want to push it up and release a new major version for koa@next :+1: (it even has support for whitelist/blacklist, and all tests pass with flying sparkles :sparkles:) @jonathanong How does moving popular Koa middleware into this org impact the general guidelines we already follow with repos in this org? For example, is it ok to transfer a middleware that's using Babel? @niftylettuce I don't think we should merge that PR yet if koa-convert does the trick. That's the general procedure that all the rest of the middleware in this org has been following. i don't care about code style as long as you have a linter to enforce one. middleware should be transpiled before publishing anyways, at least for node v4+. i don't see why you need to write middleware that needs to be transpiled though. A transpiler should not be required when using a middleware; it should only be application-specific. That being, I personally don't care if one is used, as long as npm install does not pull Babel or anything like that as dependency. @jonathanong ah ok. fair enough. @yanickrochon yep, I didn't mean including the full babel-runtime in the package's distribution or anything like that but precompiling instead. that said, I thought we decided not to use Babel in middleware in this org so I thought I'd mention that this may be an issue when/if we transfer external middleware here :) @tejasmanohar all good :smile: @jonathanong we can contribute koa-pagination and koa-requestid. Additionally, we are preparing the release of an error mapper which allows registering different mappers to process each error class differently. @thomseddon might be open to transfer koa-oauth-server, which we currently help maintain too. I see many people jumping in here, advocating for their projects... and while it's legit, may I suggest to all, no offense intended, that the most downloaded (i.e. popular) get's pulled into the koajs organisation first? I too have koa-* middleware and one of them has more downloads than some proposed here. Yet, I consider that some middlewares should be promoted first. Like koa-gzip, koa-cors, koa-proxy, koa-timeout, koa-passport, etc. m2c I’d be happy to transfe: koa-errorhandler @jonathanong koa-ip @jonathanong koa-mongo @jonathanong koa-scheme @jonathanong koa-router-validator @jonathanong koa-router-cache @jonathanong co-cache @jonathanong Tip: you can use npm-user-downloads check your packages downloads ranking. @yanickrochon I think we should take this opportunity not to find which packages have more downloads than the others but to build consensus around them and improve the whole koa ecosystem. I'd be happy to work together on any of the ones I mentioned in case multiple ones from the community are available for the same purpose. Sometimes the multitude of packages on npm comes from the horrible search alone :) @ruimarinho oh, but I totally agree! My point was that I don't want everybody promoting their own koa-* repositories just for the sake that it has "koa-" as prefix! It's more of a prevention notice, so we don't need to trigger any rejection complex or whatever (LOL) That being said, usually, the most used (i.e. downloaded) are the ones that should be maintained first (logically). There are cases, however, where you are right and we should take this opportunity to correct, or influence the use of a package over another. For example, there are a few caching middlewares for koa already, some of them perform almost the same thing. All and all, my point is that community approved packages should be promoted first, so we don't end up with duplicated packages offering the same features yet again. @yanickrochon agreed! Main areas where I believe there is quite some overlap are caching, routing and error handling. i've setup some teams here: https://github.com/orgs/koajs/teams if you want to help with some teams, let me know if you think the teams could use better organization (new team, add a module, etc) let me know i'm sure some modules, even in the org, are not properly organized. looking at the teams might help you guys think about other modules to add to the org! also i noticed a lot of PRs for Koa v2 support. if you see one and want to help, let me know. i don't plan on touching those PRs myself for a while. @jonathanong we (Pebble) have a few repos we'd like to transfer. You want me to transfer to you first or just directly to koajs? https://github.com/pebble/koa-resourcer https://github.com/pebble/koa-joi-router https://github.com/pebble/koa-resourcer-docs https://github.com/pebble/koa-bunyan-logger @aheckmann you're an owner you should be able to transfer yourself (to koajs) cool, will do. thanks I can transfer https://github.com/tunnckoCore/koa-ip-filter if you want. I have few more, but they are very outdated and I'm trying to update them soon. Now i'm working on total refactor of koa-better-body, using koa-body-parsers under the hood. https://github.com/koa-modules/ ? @fundon It's better to give a list you want to transfer, not a org link :smile: @coderhaoxin Ok :smile: Great idea, I’d be happy to transfer koa-generator. @jonathanong @coderhaoxin @fundon @coderhaoxin's an owner so he can transfer those modules over @i5ting :+1: for the generator because we don't have one. anyone else willing to help maintain it? @i5ting Could you translate Chinese into English in the koa-generator/Readme? (Or I can do that several days later :smile:) @coderhaoxin 哈哈,一共也没几个中文,我抽空就整理了,目前测试还差一些,在补充呢 https://github.com/17koa/koa-generator @fundon Some modules are duplicate! Such as: koajs/locales vs koa-modules/i18n for i18n koajs/static vs koa-modules/serve-static koajs/override-method vs koa-modules/methodoverride That will be confused for the users, IMO :) @coderhaoxin Should we add more detail for them? Suggestions? I want transfer my middleware webpack-koa2-middleware to koa community. I'm going to transfer: https://github.com/tunnckoCore/koa-better-body https://github.com/tunnckoCore/koa-better-serve https://github.com/tunnckoCore/koa-better-ratelimit (v3 in progress, sry there's no readme currently) https://github.com/tunnckoCore/koa-better-router https://github.com/tunnckocore/koa-rest-router https://github.com/tunnckoCore/koa-ip-filter As soon as possible. :) It's only a testing framework but I don't know if you guys are interested in having: https://github.com/chrsalx/koa-test. I would gladly transfer.
gharchive/issue
2016-01-18T01:08:31
2025-04-01T06:39:18.249795
{ "authors": [ "aheckmann", "chrsalx", "coderhaoxin", "fengmk2", "fundon", "i5ting", "iyuq", "jonathanong", "m4nuC", "niftylettuce", "nswbmw", "ruimarinho", "stiang", "stojanovic", "tejasmanohar", "tj", "tunnckoCore", "yanickrochon" ], "repo": "koajs/koa", "url": "https://github.com/koajs/koa/issues/638", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
496712437
Incorrect Efficientnets' FLOPs in ImageNet Classification Leaderboard Hi, I think in ImageNet Classification Leaderboard, Efficientnet-B7' FLOPS should be 37G (37B in the paper) instead of 37000G. Same for other versions. +1 It should be 0.37 because the column is in GFLOPS
gharchive/issue
2019-09-21T23:21:35
2025-04-01T06:39:18.261794
{ "authors": [ "M4gicT0", "g1910", "thanhmvu" ], "repo": "kobiso/Computer-Vision-Leaderboard", "url": "https://github.com/kobiso/Computer-Vision-Leaderboard/issues/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1556590447
kodadot.xyz not loading in nova wallet What happened? When trying to open kodadot.xyz in Nova Wallet DApp browser it is not loading. Error from the dev console is attached. Please reproduce in steps Open Nova Wallet app Go to Browser tab Put "kodadot" into the search field Tap on found KodaDot option Browser is opened but nothing loads Expected Behavior DApp browser for KodaDot is loaded and wallet account is requested What browsers are you seeing the problem on? Mobile iOS Safari (WebKit) At which address did you encounter bug? kodadot.xyz Are you logged in? No Which wallet you are using? Nova Wallet At which chain did you encounter bug? Basilisk, MoonSama, RMRK Screenshots https://user-images.githubusercontent.com/570634/214567060-eef8921d-4aa2-4b9e-81c9-15589d97de1d.MP4 Relevant log output No response Payment link for reward No response Code of Conduct [X] I agree to follow this project's Code of Conduct Yeah not working on iOS probably. I tried it too on my iPhone. Hey @ERussel is this still thing? We did few updates recently. I don't have iOS tho. Hey @yangwao still can reproduce on iOS. I think it is related to some part of the code that treats Safari with polkadot js extension at the same time in the wrong way. hey @ERussel if you can check now, it should be available on https://beta.kodadot.xyz
gharchive/issue
2023-01-25T12:47:30
2025-04-01T06:39:18.283394
{ "authors": [ "ERussel", "JustLuuuu", "yangwao" ], "repo": "kodadot/nft-gallery", "url": "https://github.com/kodadot/nft-gallery/issues/4827", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1136914729
#2344 Rework navbar Thank you for your contribution to the KodaDot NFT gallery. 👇 _ Let's make a quick check before the contribution. PR type [ ] Bugfix [ ] Feature [x] Refactoring What's new? [x] PR closes #2344 [ ] Before submitting Pull Request, please make sure: [x] My contribution builds clean without any errors or warnings [x] I've merged recent default branch -- main and I've no conflicts [x] I've tried to respect high code quality standards [x] I've didn't break any original functionality [x] I've posted a screenshot of demonstrated change in this PR Optional [ ] I've tested it at </rmrk/collection/26902bc2f7c20c546a-1FVG7> [ ] I've tested PR on mobile and everything seems works [ ] I found edge cases [ ] I've written some unit tests 🧪 Had issue bounty label? [x] Fill up your KSM address: Payout Community participation [x] Are you at KodaDot Discord? Screenshot [x] My fix has changed something on UI; a screenshot is best to understand changes for others. @roiLeo doneee Doesn't take me anywhere :| https://user-images.githubusercontent.com/5887929/153859039-058ef7cc-8117-4fb0-a390-1173b6dc49ac.mov Doesn't take me anywhere :| Screen.Recording.2022-02-14.at.12.49.00.mov found the issue, fixing it @yangwao check now love you! 😘 pay 200 usd 😍 Perfect, I’ve sent the payout 💵 $200 @ 161.6 USD/KSM ~ 1.238 $KSM 🧗 EzGc4s9PgCPx1YnF3fqzhLzVHpHMTL4LWPScwpDrR8JKgSU 🔗 0x0f683cefd64ee30ab53b073df12d9707b4c7ba07fb225984257d00d9c993d093 🪅 Let’s grab another issue and get rewarded! 🪄 github.com/kodadot/nft-gallery/issues
gharchive/pull-request
2022-02-14T07:58:44
2025-04-01T06:39:18.293969
{ "authors": [ "prachi00", "yangwao" ], "repo": "kodadot/nft-gallery", "url": "https://github.com/kodadot/nft-gallery/pull/2348", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1879187004
docs: updated style_guide Thank you for your contribution to the KodaDot - One Stop Shop for Polkadot NFTs. 👇 __ Let's make a quick check before the contribution. PR Type [ ] Bugfix [ ] Feature [ ] Refactoring [x] Documentation related to #7106 Context [ ] Closes #<issue_number> [ ] Requires deployment <snek/rubick/worker> Before submitting pull request, please make sure: [x] My contribution builds clean without any errors or warnings [x] I've merged recent default branch -- main and I've no conflicts [x] I've tried to respect high code quality standards [x] I've didn't break any original functionality Optional [ ] I've tested it at </ksm/collection> [ ] I've tested PR on mobile [ ] I've written unit tests 🧪 [ ] I've found edge cases Did your issue had any of the "$" label on it? [x] Fill up your DOT address: Payout Community participation [ ] Are you at KodaDot Ecosystem Telegram? Screenshot 📸 [ ] My fix has changed something on UI; a screenshot is best to understand changes for others. Copilot Summary 🤖 Generated by Copilot at ec90a29 Updated STYLE_GUIDE.md to fix errors and enhance readability. 🤖 Generated by Copilot at ec90a29 Style guide refined sentence, example, format Cut like autumn leaves pay 10 usd 😍 Perfect, I’ve sent the payout 💵 $10 @ 4.25 USD/DOT ~ 2.353 $DOT 🧗 13rFRPVKjJzQXVC8ZqHZv5YMmwmk4MU7z4HeYk218hEMpQXH 🔗 0x27e11c7b0e13c2027faa9fc9b688a601194ed26fc93bd47c40e66cc790148651 🪅 Let’s grab another issue and get rewarded! 🪄 github.com/kodadot/nft-gallery/issues
gharchive/pull-request
2023-09-03T18:14:46
2025-04-01T06:39:18.304780
{ "authors": [ "stephenjason89", "yangwao" ], "repo": "kodadot/nft-gallery", "url": "https://github.com/kodadot/nft-gallery/pull/7118", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2442112043
Correct the link for Getting Started with SwiftUI Course in README.md Correct the link for Getting Started with SwiftUI Course in README.md Dear @jellodiil , I have a minor update link for "Getting Started with SwiftUI Course" in README.md. Could you please help me to review it? Thanks, Lam
gharchive/pull-request
2024-08-01T10:38:55
2025-04-01T06:39:18.307175
{ "authors": [ "lamtrinhdev" ], "repo": "kodecocodes/m3-suii-materials", "url": "https://github.com/kodecocodes/m3-suii-materials/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
172791582
Fix typo in readme Hope this saves someone else some time. Thanks @mikesplain! Sorry for slow reaction, I must have missed a notification on this 😇 😇 😇 Np!
gharchive/pull-request
2016-08-23T19:57:10
2025-04-01T06:39:18.322829
{ "authors": [ "mikesplain", "rjeczalik" ], "repo": "koding/vagrantutil", "url": "https://github.com/koding/vagrantutil/pull/9", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
231881879
[plugin/tiling] ToogleWindow function don't check for floating window Edit: Second time I open an issue without typing a word XD But basically the ToggleWindow function don't check if the window passed is floating, this can lead to crashes when calling one of the function on one of these. Obviously one shouldn't call these function floating windows but it can happen by pressing the wrong binding etc.. and we should probably not crash. Specifically in many places it check for (!Node && !Node->Parent) leading to a crash Uh, do you have a line of code or steps to reproduce a crash? I don't see a particular problem in the ToggleWindow function. Probably just something I'm overlooking. chunkc window --toggle float chunkc window --toggle split on the same window produces one reliably. Well that is a major fuckup, wonder how late it was when I wrote that line of code. It's supposed to be if(!Node || !Node->Parent) so we can short-circuit. I have tested this and it should now be fixed. Marking this as a bug even tho it is fixed, for future references.
gharchive/issue
2017-05-28T17:20:46
2025-04-01T06:39:18.325662
{ "authors": [ "choco", "koekeishiya" ], "repo": "koekeishiya/chunkwm", "url": "https://github.com/koekeishiya/chunkwm/issues/64", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
127661544
Tools like PopClip do not work any more Since #85 and the removal of menu-fix I cannot use PopClip any more. I suspect other tools relaying on an interactive overlay may be broken as well. kwm switches to the PopClip application and thereby losing the context to act on. I really hate the layer system implemented in OSX.. Will see if something can be done about this in the future. Thanks, because until there is a fix I need to unload kwm. I have a fix working atm, that still allows for menubar/dock/context menus to work, but I'm not sure if it may break focusing of other window types. Basically it has to do with: enum _CGCommonWindowLevelKey { kCGBaseWindowLevelKey = 0, kCGMinimumWindowLevelKey = 1, kCGDesktopWindowLevelKey = 2, kCGBackstopMenuLevelKey = 3, kCGNormalWindowLevelKey = 4, kCGFloatingWindowLevelKey = 5, kCGTornOffMenuWindowLevelKey = 6, kCGDockWindowLevelKey = 7, kCGMainMenuWindowLevelKey = 8, kCGStatusWindowLevelKey = 9, kCGModalPanelWindowLevelKey = 10, kCGPopUpMenuWindowLevelKey = 11, kCGDraggingWindowLevelKey = 12, kCGScreenSaverWindowLevelKey = 13, kCGMaximumWindowLevelKey = 14, kCGOverlayWindowLevelKey = 15, kCGHelpWindowLevelKey = 16, kCGUtilityWindowLevelKey = 17, kCGDesktopIconWindowLevelKey = 18, kCGNumberOfWindowLevelKeys = 19 Cool, thank you very much for the quick turnaround! It appears to work for me.
gharchive/issue
2016-01-20T11:21:56
2025-04-01T06:39:18.328782
{ "authors": [ "herrbischoff", "koekeishiya" ], "repo": "koekeishiya/kwm", "url": "https://github.com/koekeishiya/kwm/issues/93", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
430085394
Added table styles Table styles added. Closes #5! I made some minute changes. Thanks so much! Ready to ship? Changes look great! :+1: :shipit:
gharchive/pull-request
2019-04-06T23:02:55
2025-04-01T06:39:18.337515
{ "authors": [ "gavinhenderson", "kognise" ], "repo": "kognise/water.css", "url": "https://github.com/kognise/water.css/pull/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1666925921
v3.0.1 testnet params and checkpoints New testnet parameters
gharchive/pull-request
2023-04-13T18:24:33
2025-04-01T06:39:18.338295
{ "authors": [ "koh-gt" ], "repo": "koh-gt/ferrite-core", "url": "https://github.com/koh-gt/ferrite-core/pull/48", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
514943686
"pam_authenticate failed" what drives this message? 2019-10-30 15:51:03.239 ERROR 20613 --- [nio-8443-exec-4] o.o.a.s.CustomAuthenticationProvider : PAM authentication failed: pam_authenticate failed : Authentication failure -- org.jvnet.libpam.PAMException: pam_authenticate failed : Authentication failure at org.jvnet.libpam.PAM.check(PAM.java:106) at org.jvnet.libpam.PAM.authenticate(PAM.java:124) Getting this in my program. Tryied on three different linux OS same result. Can't seem to figure out what this message is telling me. Did I miss a step somewhere? pamtester -v login **** authenticate This works... but fails via the java program running as the same user. Tried on Fedora 27, Fedora 29, even a Oracle Linux 7 I had kicking around. ok finally found the problem. Never fails I search at a problem but the minute I have to document it into an issue like this I find it. So my problem was this: Collection<? extends GrantedAuthority> authorities = Collections.singleton(new SimpleGrantedAuthority("ROLE_USER")); return new UsernamePasswordAuthenticationToken(authentication.getPrincipal(), authentication.getCredentials(), authorities); vs authentication.getCredentials()); Without the ROLES at the end it authenticates once fine (which I missed in the logs) then blanks out the password. Then tries a second attempt with a null password. Which is what I saw above. When it was a wrong password the thing would fail on first try. So resolution... pay attention to the logs more closely. Posting here in case anyone else falls into the same trap but the issue can be closed.
gharchive/issue
2019-10-30T19:57:00
2025-04-01T06:39:18.343043
{ "authors": [ "MikeyCarter" ], "repo": "kohsuke/libpam4j", "url": "https://github.com/kohsuke/libpam4j/issues/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1511722354
v1.8.8 弹幕菜单显示不全 这个。 隔壁前端🐏了没来,来了我问问他要怎么解决。 解决了!用新版本把高度拉长点。
gharchive/issue
2022-12-27T11:41:30
2025-04-01T06:39:18.389997
{ "authors": [ "LoveLoliii", "MistEO" ], "repo": "kokolokksk/catcat-dm-react", "url": "https://github.com/kokolokksk/catcat-dm-react/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2616376908
1.21.1 Support! https://github.com/PrismarineJS/mineflayer/commit/3f1f0a3fef9aef3113d2de62d70e6e42410b0b44 With the new release 4.23.0, mineflayer dropped 1.21.1 support In testing i found out that 1.21 doesn't work????? But 1.21.1 does sorry I implemented it myself #255
gharchive/pull-request
2024-10-27T07:02:22
2025-04-01T06:39:18.391566
{ "authors": [ "1223nij", "MaxRobinsonTheGreat" ], "repo": "kolbytn/mindcraft", "url": "https://github.com/kolbytn/mindcraft/pull/249", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1258848293
Improve attribute parsing from monsters.txt I rationalized how we parse attributes from monsters.txt. Rather than having 36 different attributes and calling a MonsterData constructor with 28 parameters, parse the parameters into 1 data structure and call a constructor with 4 parameters. The constructor still needs to extract all the attributes out and set the expected fields, but adding a new attribute now no longer requires adding a parameter to the constructor: add parsing in MonsterData, add a field to the MonsterData class and set it in the constructor, and, as before, use that field as needed. Perhaps a field in the Monster Proxy Record. And so on. I normalized the order of the fields in monsters.txt: Atk, Def, HP, Init, ... or something. I added Article, Group, Elemental Resistance as attribute, complete with Monster Proxy Record field. I normalize string you get from the Monster.attributes proxy record field I added "test fix-monster-attributes" to read monsters.txt line by line and replace the attributes (field 4) with the normalized version. I ran it on monsters.txt and found 998 (!) monsters that needed to be updated. This is included. I made Phys: and Elem: allow Expressions, rather than just ints I wrote a test suite for the string -> attribute map and attribute map -> string methods. I wrote a test showing that the MonsterData constructor parses an attribute string. Coming in a follow-up PR: Improve MonsterManuelManager to extract "Article" and include it in the "attributes" string it generates for new monsters. Provide MonsterManuelManager with a utility to visit every page and extract the Articles for every monster, saving them in a map from monster ID to Attribute map Given that, it will be trivial to make "test fix-monster-articles" to add that field to every monster in Manuel. Ready for review. I am deferring importing "article" from MonsterManuelManager until the next PR. Checking monsters.txt for lines of differing length (then removing those where Poison was added) gives: 263 full-length mirror 1553 bigmirror.gif NOCOPY LUCKY Scale: 0 Cap: ? Init: 50 P: construct 263 full-length mirror 1553 bigmirror.gif NOCOPY LUCKY Scale: 0 Cap: 10000 Init: 50 P: construct 292 Gnollish Crossdresser 19 dk_cross.gif Atk: 11 Def: 9 HP: 5 Init: 60 E: sleaze EA: sleaze Meat: 22 P: humanoid gnoll teeth (5) frilly skirt (25) maiden wig (25) frilly shirt (c5) Gnollish Crossdress (c0) 292 Gnollish Crossdresser 19 dk_cross.gif Atk: 11 Def: 9 HP: 5 Init: 60 Meat: 22 P: humanoid E: sleaze gnoll teeth (5) frilly skirt (25) maiden wig (25) frilly shirt (c5) Gnollish Crossdress (c0) 378 Knott Slanding 617 slanding.gif NOCOPY ULTRARARE NOMANUEL Atk: 150 Def: 135 HP: 150 Init: 10000 E: cold Meat: 970-1342 P: beast Dallas Dynasty Falcon Crest shield (100) 378 Knott Slanding 617 slanding.gif NOCOPY NOMANUEL ULTRARARE Atk: 150 Def: 135 HP: 150 Init: 10000 Meat: 1156 P: beast E: cold Dallas Dynasty Falcon Crest shield (100) 388 lihc 32 lich.gif Atk: 20 Def: 18 HP: 11 Init: 60 ED: spooky EA: spooky EA: sleaze Meat: 36 P: undead lihc eye (30) 388 lihc 32 lich.gif Atk: 20 Def: 18 HP: 11 Init: 60 Meat: 36 P: undead E: sleaze lihc eye (30) 570 scary pirate 307 zompirate.gif Atk: 150 Def: 135 HP: 200 Init: 10000 E: spooky Meat: 45 P: pirate Wiki: "scary pirate (cursed)" cursed breeches (n0) cursed cutlass (n0) cursed eyepatch (n0) folder (skull and crossbones) (c0) 570 scary pirate 307 zompirate.gif Atk: 150 Def: 135 HP: 200 Init: 10000 Meat: 45 P: pirate E: spooky Wiki: "scary pirate (cursed)" cursed breeches (n0) cursed cutlass (n0) cursed eyepatch (n0) folder (skull and crossbones) (c0) 638 Stephen Spookyraven 1565 steven.gif NOCOPY Scale: 5 Cap: 300 Init: 100 Meat: 500 P: undead Stephen's secret formula (100) Stephen's lab coat (100) 638 Stephen Spookyraven 1565 steven.gif NOCOPY Scale: 5 Cap: 300 Init: 100 Meat: 500 P: undead Stephen's secret formula (100) Stephen's lab coat (100) 681 The Thing in the Basement 2164 drippything1.gif NOMANUEL Atk: [300] Def: [300] HP: [320] Init: -10000 P: horror DRIPPY The Eye of the Thing in the Basement (n100) The Fingernail of the Thing in the Basement (n100) 681 The Thing in the Basement 2164 drippything1.gif NOMANUEL Atk: [300] Def: [300] HP: [320] Init: -10000 P: horror DRIPPY The Eye of the Thing in the Basement (n100) The Fingernail of the Thing in the Basement (n100) 1240 furious cow 1926 awolcow1.gif WANDERER Scale: 0 Cap: ? Floor: ? Init: -10000 P: demon 1240 furious cow 1926 awolcow1.gif WANDERER Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: demon 1241 furious giant cow 1929 awolcow2.gif WANDERER Scale: 20 Cap: ? Floor: ? Init: -10000 P: demon 1241 furious giant cow 1929 awolcow2.gif WANDERER Scale: 20 Cap: 10000 Floor: 10 Init: -10000 P: demon 1242 ungulith 1932 awolcow3.gif WANDERER Scale: 40 Cap: ? Floor: ? Init: -10000 P: demon 1242 ungulith 1932 awolcow3.gif WANDERER Scale: 40 Cap: 10000 Floor: 10 Init: -10000 P: demon 1244 emaciated rodeo clown 1927 awolclown1.gif WANDERER Scale: 0 Cap: ? Floor: ? Init: -10000 P: horror 1244 emaciated rodeo clown 1927 awolclown1.gif WANDERER Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: horror 1245 menacing rodeo clown 1930 awolclown2.gif WANDERER Scale: 20 Cap: ? Floor: ? Init: -10000 P: horror 1245 menacing rodeo clown 1930 awolclown2.gif WANDERER Scale: 20 Cap: 10000 Floor: 10 Init: -10000 P: horror 1246 grizzled rodeo clown 1933 awolclown3.gif WANDERER Scale: 40 Cap: ? Floor: ? Init: -10000 P: horror 1246 grizzled rodeo clown 1933 awolclown3.gif WANDERER Scale: 40 Cap: 10000 Floor: 10 Init: -10000 P: horror 1248 aggressive grass snake 1928 awolsnake1.gif WANDERER Scale: 0 Cap: ? Floor: ? Init: -10000 P: beast SNAKE 1248 aggressive grass snake 1928 awolsnake1.gif WANDERER Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: beast SNAKE 1249 prince snake 1931 awolsnake2.gif WANDERER Scale: 20 Cap: ? Floor: ? Init: -10000 E: sleaze P: beast SNAKE snake oil (100) snake oil (100) 1249 prince snake 1931 awolsnake2.gif WANDERER Scale: 20 Cap: 10000 Floor: 10 Init: -10000 P: beast SNAKE E: sleaze snake oil (100) snake oil (100) 1250 king snake 1934 awolsnake3.gif WANDERER Scale: 40 Cap: ? Floor: ? Init: -10000 P: beast SNAKE snake oil (100) snake oil (100) snake oil (100) 1250 king snake 1934 awolsnake3.gif WANDERER Scale: 40 Cap: 10000 Floor: 10 Init: -10000 P: beast SNAKE snake oil (100) snake oil (100) snake oil (100) 1258 Villainous Minion 2024 bond_minion1.gif,bond_minion2.gif,bond_minion3.gif,bond_minion4.gif Scale: -2 Cap: ? Floor: ? Init: 50 P: dude 1258 Villainous Minion 2024 bond_minion1.gif,bond_minion2.gif,bond_minion3.gif,bond_minion4.gif Scale: -2 Cap: 10000 Floor: 10 Init: 50 P: dude 1259 Villainous Henchperson 2025 bond_sidekick1.gif,bond_sidekick2.gif,bond_sidekick3.gif,bond_sidekick4.gif,bond_sidekick5.gif,bond_sidekick6.gif,bond_sidekick7.gif,bond_sidekick8.gif,bond_sidekick9.gif NOCOPY Scale: 10 Cap: ? Floor: ? Init: -10000 P: dude 1259 Villainous Henchperson 2025 bond_sidekick1.gif,bond_sidekick2.gif,bond_sidekick3.gif,bond_sidekick4.gif,bond_sidekick5.gif,bond_sidekick6.gif,bond_sidekick7.gif,bond_sidekick8.gif,bond_sidekick9.gif NOCOPY Scale: 10 Cap: 10000 Floor: 10 Init: -10000 P: dude 1260 Villainous Villain 2026 bond_villain1.gif,bond_villain2.gif,bond_villain3.gif,bond_villain4.gif,bond_villain5.gif,bond_villain6.gif,bond_villain7.gif NOCOPY Scale: 11 Cap: ? Floor: ? Init: -10000 P: dude 1260 Villainous Villain 2026 bond_villain1.gif,bond_villain2.gif,bond_villain3.gif,bond_villain4.gif,bond_villain5.gif,bond_villain6.gif,bond_villain7.gif NOCOPY Scale: 11 Cap: 10000 Floor: 10 Init: -10000 P: dude 1288 invader bullet 2137 invaderbullet.gif WANDERER Scale: 20 Cap: ? Floor: ? Init: 10000 P: construct 1288 invader bullet 2137 invaderbullet.gif WANDERER Scale: 20 Cap: 10000 Floor: 10 Init: 10000 P: construct 1515 Black Crayon Beast 1187 cray_beast.gif FREE Scale: 0 Cap: ? Init: -10000 P: beast crayon shavings (100) 1515 Black Crayon Beast 1187 cray_beast.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: beast crayon shavings (100) 1516 Black Crayon Beetle 1191 cray_bug.gif FREE Scale: 0 Cap: ? Init: -10000 P: bug crayon shavings (100) 1516 Black Crayon Beetle 1191 cray_bug.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: bug crayon shavings (100) 1517 Black Crayon Constellation 1206 cray_const.gif FREE Scale: 0 Cap: ? Init: -10000 P: constellation crayon shavings (100) 1517 Black Crayon Constellation 1206 cray_const.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: constellation crayon shavings (100) 1518 Black Crayon Crimbo Elf 1201 cray_elf.gif FREE Scale: 0 Cap: ? Init: -10000 P: elf crayon shavings (100) 1518 Black Crayon Crimbo Elf 1201 cray_elf.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: elf crayon shavings (100) 1519 Black Crayon Demon 1194 cray_demon.gif FREE Scale: 0 Cap: ? Init: -10000 P: demon crayon shavings (100) 1519 Black Crayon Demon 1194 cray_demon.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: demon crayon shavings (100) 1520 Black Crayon Elemental 1205 cray_elemental.gif FREE Scale: 0 Cap: ? Init: -10000 P: elemental crayon shavings (100) 1520 Black Crayon Elemental 1205 cray_elemental.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: elemental crayon shavings (100) 1521 Black Crayon Fish 1196 cray_fish.gif FREE Scale: 0 Cap: ? Init: -10000 P: fish crayon shavings (100) 1521 Black Crayon Fish 1196 cray_fish.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: fish crayon shavings (100) 1522 Black Crayon Flower 1199 cray_plant.gif FREE Scale: 0 Cap: ? Init: -10000 P: plant crayon shavings (100) 1522 Black Crayon Flower 1199 cray_plant.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: plant crayon shavings (100) 1523 Black Crayon Frat Orc 1193 cray_orc.gif FREE Scale: 0 Cap: ? Init: -10000 P: orc crayon shavings (100) 1523 Black Crayon Frat Orc 1193 cray_orc.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: orc crayon shavings (100) 1524 Black Crayon Goblin 1197 cray_goblin.gif FREE Scale: 0 Cap: ? Init: -10000 P: goblin crayon shavings (100) 1524 Black Crayon Goblin 1197 cray_goblin.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: goblin crayon shavings (100) 1525 Black Crayon Golem 1188 cray_construct.gif FREE Scale: 0 Cap: ? Init: -10000 P: construct crayon shavings (100) 1525 Black Crayon Golem 1188 cray_construct.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: construct crayon shavings (100) 1526 Black Crayon Hippy 1192 cray_hippy.gif FREE Scale: 0 Cap: ? Init: -10000 P: hippy crayon shavings (100) 1526 Black Crayon Hippy 1192 cray_hippy.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: hippy crayon shavings (100) 1527 Black Crayon Hobo 1207 cray_hobo.gif FREE Scale: 0 Cap: ? Init: -10000 P: hobo crayon shavings (100) 1527 Black Crayon Hobo 1207 cray_hobo.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: hobo crayon shavings (100) 1528 Black Crayon Man 1186 cray_dude.gif FREE Scale: 0 Cap: ? Init: -10000 P: dude crayon shavings (100) 1528 Black Crayon Man 1186 cray_dude.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: dude crayon shavings (100) 1529 Black Crayon Manloid 1190 cray_humanoid.gif FREE Scale: 0 Cap: ? Init: -10000 P: humanoid crayon shavings (100) 1529 Black Crayon Manloid 1190 cray_humanoid.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: humanoid crayon shavings (100) 1530 Black Crayon Mer-kin 1202 cray_merkin.gif FREE Scale: 0 Cap: ? Init: -10000 P: mer-kin crayon shavings (100) 1530 Black Crayon Mer-kin 1202 cray_merkin.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: mer-kin crayon shavings (100) 1531 Black Crayon Penguin 1204 cray_penguin.gif FREE Scale: 0 Cap: ? Init: -10000 P: penguin crayon shavings (100) 1531 Black Crayon Penguin 1204 cray_penguin.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: penguin crayon shavings (100) 1532 Black Crayon Pirate 1198 cray_pirate.gif FREE Scale: 0 Cap: ? Init: -10000 P: pirate crayon shavings (100) 1532 Black Crayon Pirate 1198 cray_pirate.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: pirate crayon shavings (100) 1533 Black Crayon Shambling Monstrosity 1195 cray_horror.gif FREE Scale: 0 Cap: ? Init: -10000 P: horror crayon shavings (100) 1533 Black Crayon Shambling Monstrosity 1195 cray_horror.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: horror crayon shavings (100) 1534 Black Crayon Slime 1203 cray_slime.gif FREE Scale: 0 Cap: ? Init: -10000 P: slime crayon shavings (100) 1534 Black Crayon Slime 1203 cray_slime.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: slime crayon shavings (100) 1535 Black Crayon Spiraling Shape 1200 cray_weird.gif FREE Scale: 0 Cap: ? Init: -10000 P: weird crayon shavings (100) 1535 Black Crayon Spiraling Shape 1200 cray_weird.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: weird crayon shavings (100) 1536 Black Crayon Undead Thing 1189 cray_undead.gif FREE Scale: 0 Cap: ? Init: -10000 P: undead crayon shavings (100) 1536 Black Crayon Undead Thing 1189 cray_undead.gif FREE Scale: 0 Cap: 10000 Init: -10000 P: undead crayon shavings (100) 1622 gummi plesiosaur 1491 plesio.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: beast small gummi fin (0) gummi sword (0) 1622 gummi plesiosaur 1491 plesio.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: beast small gummi fin (0) gummi sword (0) 1623 lemonhead minnow 1492 lemonfish.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: fish lemonhead caviar (0) lemony scales (0) lemony scales (0) 1623 lemonhead minnow 1492 lemonfish.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: fish lemonhead caviar (0) lemony scales (0) lemony scales (0) 1624 school of gummi piranhas 1490 gummifish.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: fish gummi fang (0) chocolate cow bone (0) 1624 school of gummi piranhas 1490 gummifish.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: fish gummi fang (0) chocolate cow bone (0) 1625 chocolate hare 1494 chocohare.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: beast chocolate rabbit's foot (0) candy carrot (0) 1625 chocolate hare 1494 chocohare.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: beast chocolate rabbit's foot (0) candy carrot (0) 1626 chocolate-cherry prairie dog 1495 ccprairie.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: beast candy mountain oyster (0) chocolate-stained collar (0) 1626 chocolate-cherry prairie dog 1495 ccprairie.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: beast candy mountain oyster (0) chocolate-stained collar (0) 1627 Rock Pop weasel 1493 popweasel.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: beast mulberry (0) weasel stomping pants (0) 1627 Rock Pop weasel 1493 popweasel.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: beast mulberry (0) weasel stomping pants (0) 1628 candied pecan tree 1496 pecantree.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: plant candied pecan (0) candy stick (0) 1628 candied pecan tree 1496 pecantree.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: plant candied pecan (0) candy stick (0) 1629 licorice snake 1497 licosnake.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: beast SNAKE anise-flavored venom (0) licorice whip (0) 1629 licorice snake 1497 licosnake.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: beast SNAKE anise-flavored venom (0) licorice whip (0) 1630 tricksy pixie 1498 trixiepixie.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: humanoid sour powder (0) pixie axie (0) 1630 tricksy pixie 1498 trixiepixie.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: humanoid sour powder (0) pixie axie (0) 1631 fire truck 1499 firetruck.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: construct fireman's lunch (0) fire hose (0) 1631 fire truck 1499 firetruck.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: construct fireman's lunch (0) fire hose (0) 1632 ice cream truck 1500 icecreamtruck.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: construct ice cream sandwich (0) plain paper hat (0) 1632 ice cream truck 1500 icecreamtruck.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: construct ice cream sandwich (0) plain paper hat (0) 1633 monster hearse 1502 monsterhearse.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: construct nachos of the night (0) skull gearshift knob (0) 1633 monster hearse 1502 monsterhearse.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: construct nachos of the night (0) skull gearshift knob (0) 1634 sewer tanker 1501 sewertruck.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: construct plumber's lunch (0) &quot;honey&quot; dipper (0) 1634 sewer tanker 1501 sewertruck.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: construct plumber's lunch (0) &quot;honey&quot; dipper (0) 1635 sketchy van 1503 sketchyvan.gif NOCOPY Scale: ? Cap: ? Floor: ? Init: -10000 P: construct can of Adultwitch&trade; (0) Sketcherz&trade; (0) 1635 sketchy van 1503 sketchyvan.gif NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: construct can of Adultwitch&trade; (0) Sketcherz&trade; (0) 1913 Ancient Skeleton with Skin still on it 1962 skinskeleton.gif Scale: 3 Cap: ? Floor: 20 Init: -10000 P: undead compounded experience (c100) time residue (c0) 1913 Ancient Skeleton with Skin still on it 1962 skinskeleton.gif Scale: 3 Cap: 10000 Floor: 20 Init: -10000 P: undead compounded experience (c100) time residue (c0) 1914 Apathetic Tyrannosaurus 1959 apatheticdino.gif Scale: 3 Cap: ? Floor: 20 Init: -10000 P: beast compounded experience (c100) time residue (c0) 1914 Apathetic Tyrannosaurus 1959 apatheticdino.gif Scale: 3 Cap: 10000 Floor: 20 Init: -10000 P: beast compounded experience (c100) time residue (c0) 1915 Assembly Elemental 1960 assemblyelem.gif Scale: 5 Cap: ? Floor: 20 Init: -10000 P: elemental compounded experience (c100) time residue (c0) 1915 Assembly Elemental 1960 assemblyelem.gif Scale: 5 Cap: 10000 Floor: 20 Init: -10000 P: elemental compounded experience (c100) time residue (c0) 1916 Cro-Magnon Gnoll 1961 cromaggnoll.gif Scale: 7 Cap: ? Floor: 20 Init: -10000 P: humanoid compounded experience (c100) time residue (c0) 1916 Cro-Magnon Gnoll 1961 cromaggnoll.gif Scale: 7 Cap: 10000 Floor: 20 Init: -10000 P: humanoid compounded experience (c100) time residue (c0) 1917 Krakrox the Barbarian 1964 krakrox.gif Scale: 25 Cap: ? Floor: 40 Init: -10000 P: dude compounded experience (c100) time residue (c100) time residue (c100) time residue (c100) 1917 Krakrox the Barbarian 1964 krakrox.gif Scale: 25 Cap: 10000 Floor: 40 Init: -10000 P: dude compounded experience (c100) time residue (c100) time residue (c100) time residue (c100) 1919 Wooly Duck 1963 woolyduck.gif Scale: 5 Cap: ? Floor: 20 Init: -10000 P: beast compounded experience (c100) time residue (c0) 1919 Wooly Duck 1963 woolyduck.gif Scale: 5 Cap: 10000 Floor: 20 Init: -10000 P: beast compounded experience (c100) time residue (c0) 1939 GNG-3-R 1992 gng3r.gif NOCOPY BOSS Scale: ? Cap: ? Floor: ? Init: 300 P: construct industrial frosting (n100) 1939 GNG-3-R 1992 gng3r.gif BOSS NOCOPY Scale: 0 Cap: 10000 Floor: 10 Init: 300 P: construct industrial frosting (n100) 1942 LOV Enforcer 2009 lovenforcer.gif NOCOPY Scale: 5 Cap: ? Floor: ? Init: -10000 P: dude Phys: 75 LOV Elixir #3 (c0) 1942 LOV Enforcer 2009 lovenforcer.gif NOCOPY Scale: 5 Cap: 10000 Floor: 10 Init: -10000 P: dude Phys: 75 LOV Elixir #3 (c0) 1943 LOV Engineer 2010 lovengineer.gif NOCOPY Scale: 5 Cap: ? Floor: ? Init: -10000 P: dude Phys: 25 LOV Elixir #6 (c0) 1943 LOV Engineer 2010 lovengineer.gif NOCOPY Scale: 5 Cap: 10000 Floor: 10 Init: -10000 P: dude Phys: 25 LOV Elixir #6 (c0) 1944 LOV Equivocator 2011 lovequivocator.gif NOCOPY Scale: 5 Cap: ? Floor: ? Init: 500 P: dude Phys: 50 LOV Elixir #9 (p0) 1944 LOV Equivocator 2011 lovequivocator.gif NOCOPY Scale: 5 Cap: 10000 Floor: 10 Init: 500 P: dude Phys: 50 LOV Elixir #9 (p0) 1947 hostile plant 2013 sgplanta1.gif,sgplanta2.gif,sgplanta3.gif,sgplanta4.gif,sgplanta5.gif,sgplanta6.gif,sgplanta7.gif,sgplanta8.gif,sgplanta9.gif,sgplanta10.gif,sgplanta11.gif,sgplanta12.gif,sgplanta13.gif,sgplanta14.gif,sgplanta15.gif,sgplanta16.gif,sgplanta17.gif,sgplanta18.gif,sgplanta19.gif,sgplanta20.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] P: plant edible alien plant bit alien plant fibers alien plant goo 1947 hostile plant 2013 sgplanta1.gif,sgplanta2.gif,sgplanta3.gif,sgplanta4.gif,sgplanta5.gif,sgplanta6.gif,sgplanta7.gif,sgplanta8.gif,sgplanta9.gif,sgplanta10.gif,sgplanta11.gif,sgplanta12.gif,sgplanta13.gif,sgplanta14.gif,sgplanta15.gif,sgplanta16.gif,sgplanta17.gif,sgplanta18.gif,sgplanta19.gif,sgplanta20.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: [60+10*pref(_spacegatePlanetIndex)] P: plant edible alien plant bit alien plant fibers alien plant goo 1948 large hostile plant 2014 sgplantb1.gif,sgplantb2.gif,sgplantb3.gif,sgplantb4.gif,sgplantb5.gif,sgplantb6.gif,sgplantb7.gif,sgplantb8.gif,sgplantb9.gif,sgplantb10.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] Init: -10000 P: plant edible alien plant bit edible alien plant bit alien plant fibers alien plant fibers alien plant goo 1948 large hostile plant 2014 sgplantb1.gif,sgplantb2.gif,sgplantb3.gif,sgplantb4.gif,sgplantb5.gif,sgplantb6.gif,sgplantb7.gif,sgplantb8.gif,sgplantb9.gif,sgplantb10.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: -10000 P: plant edible alien plant bit edible alien plant bit alien plant fibers alien plant fibers alien plant goo 1949 exotic hostile plant 2015 sgplantc1.gif,sgplantc2.gif,sgplantc3.gif Scale: [10+25*pref(_spacegatePlanetIndex)] NOCOPY Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] P: plant edible alien plant bit edible alien plant bit edible alien plant bit alien plant fibers alien plant fibers alien plant fibers alien plant goo alien plant pod 1949 exotic hostile plant 2015 sgplantc1.gif,sgplantc2.gif,sgplantc3.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: [60+10*pref(_spacegatePlanetIndex)] P: plant edible alien plant bit edible alien plant bit edible alien plant bit alien plant fibers alien plant fibers alien plant fibers alien plant goo alien plant pod 1950 small hostile animal 2016 sganimala1.gif,sganimala2.gif,sganimala3.gif,sganimala4.gif,sganimala5.gif,sganimala6.gif,sganimala7.gif,sganimala8.gif,sganimala9.gif,sganimala10.gif,sganimala11.gif,sganimala12.gif,sganimala13.gif,sganimala14.gif,sganimala15.gif,sganimala16.gif,sganimala17.gif,sganimala18.gif,sganimala19.gif,sganimala20.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] P: beast alien meat alien toenails alien animal goo 1950 small hostile animal 2016 sganimala1.gif,sganimala2.gif,sganimala3.gif,sganimala4.gif,sganimala5.gif,sganimala6.gif,sganimala7.gif,sganimala8.gif,sganimala9.gif,sganimala10.gif,sganimala11.gif,sganimala12.gif,sganimala13.gif,sganimala14.gif,sganimala15.gif,sganimala16.gif,sganimala17.gif,sganimala18.gif,sganimala19.gif,sganimala20.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: [60+10*pref(_spacegatePlanetIndex)] P: beast alien meat alien toenails alien animal goo 1951 large hostile animal 2017 sganimalb1.gif,sganimalb2.gif,sganimalb3.gif,sganimalb4.gif,sganimalb5.gif,sganimalb6.gif,sganimalb7.gif,sganimalb8.gif,sganimalb9.gif,sganimalb10.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] P: beast alien meat alien meat alien toenails alien toenails alien animal goo 1951 large hostile animal 2017 sganimalb1.gif,sganimalb2.gif,sganimalb3.gif,sganimalb4.gif,sganimalb5.gif,sganimalb6.gif,sganimalb7.gif,sganimalb8.gif,sganimalb9.gif,sganimalb10.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: [60+10*pref(_spacegatePlanetIndex)] P: beast alien meat alien meat alien toenails alien toenails alien animal goo 1952 exotic hostile animal 2018 sganimalc1.gif,sganimalc2.gif,sganimalc3.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] P: beast alien meat alien meat alien meat alien toenails alien toenails alien toenails alien animal goo alien animal milk 1952 exotic hostile animal 2018 sganimalc1.gif,sganimalc2.gif,sganimalc3.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: [60+10*pref(_spacegatePlanetIndex)] P: beast alien meat alien meat alien meat alien toenails alien toenails alien toenails alien animal goo alien animal milk 1953 Spant drone 2019 sgspantdrone.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] P: bug spant chitin spant chitin spant tendon spant tendon 1953 Spant drone 2019 sgspantdrone.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: [60+10*pref(_spacegatePlanetIndex)] P: bug spant chitin spant chitin spant tendon spant tendon 1954 Spant soldier 2020 sgspantwarrior.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] P: bug spant chitin spant chitin spant tendon spant tendon spant spear 1954 Spant soldier 2020 sgspantwarrior.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: [60+10*pref(_spacegatePlanetIndex)] P: bug spant chitin spant chitin spant tendon spant tendon spant spear 1955 Murderbot drone 2021 sgmbdrone.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] P: construct murderbot component casing murderbot monofilament murderbot monofilament murderbot power cell murderbot power cell murderbot memory chip 1955 Murderbot drone 2021 sgmbdrone.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: [60+10*pref(_spacegatePlanetIndex)] P: construct murderbot component casing murderbot monofilament murderbot monofilament murderbot power cell murderbot power cell murderbot memory chip 1956 Murderbot soldier 2022 sgmb.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] P: construct murderbot component casing murderbot component casing murderbot monofilament murderbot monofilament murderbot power cell murderbot power cell murderbot memory chip murderbot memory chip murderbot memory chip murderbot plasma rifle 1956 Murderbot soldier 2022 sgmb.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: [60+10*pref(_spacegatePlanetIndex)] P: construct murderbot component casing murderbot component casing murderbot monofilament murderbot monofilament murderbot power cell murderbot power cell murderbot memory chip murderbot memory chip murderbot memory chip murderbot plasma rifle 1957 hostile intelligent alien 2023 sgalienb1.gif,sgalienb2.gif,sgalienb3.gif,sgalienb4.gif,sgalienb5.gif,sgalienb6.gif,sgalienb7.gif,sgalienb8.gif,sgalienb9.gif,sgalienb10.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: ? Init: [60+10*pref(_spacegatePlanetIndex)] P: humanoid 1957 hostile intelligent alien 2023 sgalienb1.gif,sgalienb2.gif,sgalienb3.gif,sgalienb4.gif,sgalienb5.gif,sgalienb6.gif,sgalienb7.gif,sgalienb8.gif,sgalienb9.gif,sgalienb10.gif NOCOPY Scale: [10+25*pref(_spacegatePlanetIndex)] Cap: 10000 Floor: 10 Init: [60+10*pref(_spacegatePlanetIndex)] P: humanoid 1998 God Lobster 2088 godlob.gif,godlob_scepter.gif,godlob_ring.gif,godlob_rod.gif,godlob_cape.gif,godlob_crown.gif NOCOPY FREE Scale: [10*equipped(God Lobster's Scepter)+20*equipped(God Lobster's Ring)+30*equipped(God Lobster's Rod)+40*equipped(God Lobster's Robe)+50*equipped(God Lobster's Crown)] Cap: 1500 Floor: ? Init: 1000 P: horror 1998 God Lobster 2088 godlob.gif,godlob_scepter.gif,godlob_ring.gif,godlob_rod.gif,godlob_cape.gif,godlob_crown.gif NOCOPY FREE Scale: [10*equipped(God Lobster's Scepter)+20*equipped(God Lobster's Ring)+30*equipped(God Lobster's Rod)+40*equipped(God Lobster's Robe)+50*equipped(God Lobster's Crown)] Cap: 1500 Floor: 10 Init: 1000 P: horror 2015 sausage goblin 2104 sausagegoblin.gif FREE Scale: [1+2*pref(_sausageFights)] Cap: ? Floor: ? Init: -10000 Meat: 69 E: sleaze P: goblin magical sausage casing (n100) red-hot sausage fork (0) bag of sausage links (n0) 2015 sausage goblin 2104 sausagegoblin.gif FREE Scale: [1+2*pref(_sausageFights)] Cap: 10000 Floor: 10 Init: -10000 Meat: 69 P: goblin E: sleaze magical sausage casing (n100) red-hot sausage fork (0) bag of sausage links (n0) 2051 holographic army 1629 holoarmy.gif NOCOPY Scale: 30 Floor: ? Init: -10000 P: weird holo-bomber (0) holo-platoon (0) holo-tank (0) 2051 holographic army 1629 holoarmy.gif NOCOPY Scale: 30 Floor: 10 Init: -10000 P: weird holo-bomber (0) holo-platoon (0) holo-tank (0) 2053 Xiblaxian political prisoner 1630 polprisoner.gif NOCOPY Scale: 40 Floor: ? Init: -10000 P: weird residual zeal (n100) little red .epub file (c0) 2053 Xiblaxian political prisoner 1630 polprisoner.gif NOCOPY Scale: 40 Floor: 10 Init: -10000 P: weird residual zeal (n100) little red .epub file (c0) 2109 Drunken Tropical Vacationer 1772 kok_drunk.gif Scale: 0 Cap: ? Init: -10000 P: dude Meat: 200 Kokomo Resort Brand Suntan Oil (0) Kokomo Resort Chip (0) Kokomo Resort Chip (0) Kokomo Resort Chip (0) captured boozles (c0) 2109 Drunken Tropical Vacationer 1772 kok_drunk.gif Scale: 0 Cap: 10000 Init: -10000 Meat: 200 P: dude Kokomo Resort Brand Suntan Oil (0) Kokomo Resort Chip (0) Kokomo Resort Chip (0) Kokomo Resort Chip (0) captured boozles (c0) 2110 Lovestruck Tropical Honeymooners 1771 kok_lovers.gif Scale: -5 Cap: ? Init: -10000 P: dude Meat: 250 Afternoon Delight (0) Kokomo Resort Chip (0) Kokomo Resort Chip (0) 2110 Lovestruck Tropical Honeymooners 1771 kok_lovers.gif Scale: -5 Cap: 10000 Init: -10000 Meat: 250 P: dude Afternoon Delight (0) Kokomo Resort Chip (0) Kokomo Resort Chip (0) 2111 Resort Waiter 1773 kok_waiter.gif Scale: 5 Cap: ? Init: -10000 P: dude Meat: 100 Kokomo Resort Order Pad (0) Kokomo Resort Chip (0) Kokomo Resort Chip (0) drinks tray (c0) 2111 Resort Waiter 1773 kok_waiter.gif Scale: 5 Cap: 10000 Init: -10000 Meat: 100 P: dude Kokomo Resort Order Pad (0) Kokomo Resort Chip (0) Kokomo Resort Chip (0) drinks tray (c0) 2112 Brick Mulligan, the Bartender 1774 kok_tender.gif NOCOPY Scale: 5 Cap: ? Floor: ? Init: -10000 P: dude lemon (0) lime (0) orange (0) bottle of gin (0) bottle of vodka (0) bottle of rum (0) coconut shell (0) little paper umbrella (0) magical ice cubes (0) eyebrow lifter (c0) 2112 Brick Mulligan, the Bartender 1774 kok_tender.gif NOCOPY Scale: 5 Cap: 10000 Floor: 10 Init: -10000 P: dude lemon (0) lime (0) orange (0) bottle of gin (0) bottle of vodka (0) bottle of rum (0) coconut shell (0) little paper umbrella (0) magical ice cubes (0) eyebrow lifter (c0) 2334 mutant circuit-soldering elf 751 antlerelf.gif Scale: 8 Floor: ? Init: 150 P: elf elven moonshine (0) miniature antlers (0) 2334 mutant circuit-soldering elf 751 antlerelf.gif Scale: 8 Floor: 10 Init: 150 P: elf elven moonshine (0) miniature antlers (0) 2335 mutant cookie-baking elf 749 elfblob.gif Scale: 6 Floor: ? Init: 75 P: elf festive holiday hat (0) 2335 mutant cookie-baking elf 749 elfblob.gif Scale: 6 Floor: 10 Init: 75 P: elf festive holiday hat (0) 2336 mutant doll-dressing elf 748 elflimbs.gif Scale: 6 Floor: ? Init: 75 P: elf cheap elven gloves (0) 2336 mutant doll-dressing elf 748 elflimbs.gif Scale: 6 Floor: 10 Init: 75 P: elf cheap elven gloves (0) 2337 mutant gift-wrapping elf 747 elfclaw.gif Scale: 6 Floor: ? Init: 75 P: elf elven socks (0) 2337 mutant gift-wrapping elf 747 elfclaw.gif Scale: 6 Floor: 10 Init: 75 P: elf elven socks (0) 2338 mutant whistle-carving elf 750 elfhulk.gif Scale: 7 Floor: ? Init: 100 P: elf elven <i>limbos</i> gingerbread (0) elven whittling knife (0) 2338 mutant whistle-carving elf 750 elfhulk.gif Scale: 7 Floor: 10 Init: 100 P: elf elven <i>limbos</i> gingerbread (0) elven whittling knife (0) 2348 deadwood tree 864 shiv_dead.gif Scale: 5 Cap: 200 Floor: ? Init: -10000 P: plant petrified wood (5) spooky bark (10) spooky sap (5) 2348 deadwood tree 864 shiv_dead.gif Scale: 5 Cap: 200 Floor: 10 Init: -10000 P: plant petrified wood (5) spooky bark (10) spooky sap (5) 2349 fur tree 862 shiv_fur.gif Scale: 5 Cap: 200 Floor: ? Init: -10000 P: plant petrified wood (5) spooky bark (10) spooky sap (5) wolfman mask (20) 2349 fur tree 862 shiv_fur.gif Scale: 5 Cap: 200 Floor: 10 Init: -10000 P: plant petrified wood (5) spooky bark (10) spooky sap (5) wolfman mask (20) 2350 hangman's tree 865 shiv_hangman.gif Scale: 5 Cap: 200 Floor: ? Init: -10000 P: plant petrified wood (5) spooky bark (10) spooky sap (5) 2350 hangman's tree 865 shiv_hangman.gif Scale: 5 Cap: 200 Floor: 10 Init: -10000 P: plant petrified wood (5) spooky bark (10) spooky sap (5) 2351 pumpkin tree 863 shiv_pumpkin.gif Scale: 5 Cap: 200 Floor: ? Init: -10000 P: plant petrified wood (5) pumpkinhead mask (20) spooky bark (10) spooky sap (5) 2351 pumpkin tree 863 shiv_pumpkin.gif Scale: 5 Cap: 200 Floor: 10 Init: -10000 P: plant petrified wood (5) pumpkinhead mask (20) spooky bark (10) spooky sap (5) 2352 toilet-papered tree 861 shiv_tp.gif Scale: 5 Cap: 200 Floor: ? Init: -10000 P: plant mummy costume (20) petrified wood (5) spooky bark (10) spooky sap (5) 2352 toilet-papered tree 861 shiv_tp.gif Scale: 5 Cap: 200 Floor: 10 Init: -10000 P: plant mummy costume (20) petrified wood (5) spooky bark (10) spooky sap (5) 2358 bolt-cuttin' elf 900 elf_boltcutters.gif Scale: ? Cap: ? Floor: ? Init: -10000 P: elf elf resistance button (100) handful of headless bolts (c100) pair of bolt cutters (0) 2358 bolt-cuttin' elf 900 elf_boltcutters.gif Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: elf elf resistance button (100) handful of headless bolts (c100) pair of bolt cutters (0) 2359 monkey wrenchin' elf 894 elf_wrench.gif Scale: ? Cap: ? Floor: ? Init: -10000 P: elf elf resistance button (100) throwing wrench (0) wrench handle (c100) 2359 monkey wrenchin' elf 894 elf_wrench.gif Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: elf elf resistance button (100) throwing wrench (0) wrench handle (c100) 2360 propaganda-spewin' elf 901 elf_propaganda.gif Scale: ? Cap: ? Floor: ? Init: -10000 P: elf bottle of agitprop ink (c100) elf resistance button (100) poison pen (0) 2360 propaganda-spewin' elf 901 elf_propaganda.gif Scale: 0 Cap: 10000 Floor: 10 Init: -10000 P: elf bottle of agitprop ink (c100) elf resistance button (100) poison pen (0) 2423 Hammered Yam Golem -57 drunkyam.gif NOCOPY WANDERER NOMANUEL Scale: -3 MLMult: 0 marshmallow flamb&eacute; (c100) 2423 Hammered Yam Golem -57 drunkyam.gif NOCOPY NOMANUEL WANDERER Scale: -3 MLMult: 0 marshmallow flamb&eacute; (c100) 2497 depressing French accordionist 1452 wanderacc1.gif WANDERER Scale: -3 Floor: ? Init: -10000 P: dude Bal-musette accordion (a0) 2497 depressing French accordionist 1452 wanderacc1.gif WANDERER Scale: -3 Floor: 10 Init: -10000 P: dude Bal-musette accordion (a0) 2498 lively Cajun accordionist 1453 wanderacc2.gif WANDERER Scale: -3 Floor: ? Init: -10000 P: dude Cajun accordion (a0) 2498 lively Cajun accordionist 1453 wanderacc2.gif WANDERER Scale: -3 Floor: 10 Init: -10000 P: dude Cajun accordion (a0) 2499 quirky indie-rock accordionist 1454 wanderacc3.gif WANDERER Scale: -3 Floor: ? Init: -10000 P: dude quirky accordion (a0) 2499 quirky indie-rock accordionist 1454 wanderacc3.gif WANDERER Scale: -3 Floor: 10 Init: -10000 P: dude quirky accordion (a0) 2518 Warbear Officer 1467 warbear21.gif,warbear22.gif,warbear23.gif Scale: ? Cap: ? Floor: ? Init: 50 P: beast warbear whosit (n100) warbear whosit (n100) warbear requisition box (f0) warbear badge (f0) 2518 Warbear Officer 1467 warbear21.gif,warbear22.gif,warbear23.gif Scale: 0 Cap: 10000 Floor: 10 Init: 50 P: beast warbear whosit (n100) warbear whosit (n100) warbear requisition box (f0) warbear badge (f0) 2519 High-Ranking Warbear Officer 1468 warbear31.gif,warbear32.gif,warbear33.gif Scale: ? Cap: ? Floor: ? Init: 100 P: beast warbear whosit (n100) warbear whosit (n100) warbear whosit (n100) warbear officer requisition box (f0) 2519 High-Ranking Warbear Officer 1468 warbear31.gif,warbear32.gif,warbear33.gif Scale: 0 Cap: 10000 Floor: 10 Init: 100 P: beast warbear whosit (n100) warbear whosit (n100) warbear whosit (n100) warbear officer requisition box (f0) 2552 imagining how your life would be better if you'd made different decisions 1995 badvibe3.gif Scale: +3 Cap: 10000 Floor: 20 Init: -10000 P: weird negative lump (0) chakra sludge (0) 2552 imagining how your life would be better if you'd made different decisions 1995 badvibe3.gif Scale: 3 Cap: 10000 Floor: 20 Init: -10000 P: weird negative lump (0) chakra sludge (0) I note that: Scale: ? Cap: ? Floor: ? is replaced with Scale: 0 Cap: 10000 Floor: 10 Not sure if deliberate or correct. Gnollish Crossdresser had both an E: sleaze and an EA: sleaze. Probably an error. Some spaces were collapsed -- good. I note that: Scale: ? Cap: ? Floor: ? is replaced with Scale: 0 Cap: 10000 Floor: 10 Not sure if deliberate or correct. It is deliberate for the MonsterData to have an int for those -- and it intentionally uses the default if the value is "?" - unknown. Yet another place where I should defer evaluation (and defaulting) of "?" to an int in the constructor. Gnollish Crossdresser had both an E: sleaze and an EA: sleaze. Probably an error. Yes. Its element is sleaze and it has a sleaze attack. It should just be E: DataFileMechanicsTest just checks the number of delimited fields in a line, for various data files. The previous iteration of monsters.txt was too complex for that. Once this work is done we might consider whether the new version is simpler and thus a candidate for that test. I'll monitor but this will help remind me or start a sidebar discussion about whether DataFileMechanicsTest has any utility as a test. I stuck in some logging: > test fix-monster-attributes Monster: "Gnollish Crossdresser" has both EA: sleaze and sleaze Monster: "Knott Slanding" has meat range 970-1342 Monster: "lihc" has both EA: spooky and sleaze 919 monster attributes changed Looks like you found all the "weird" monsters. :) I fixed Scale, Cap, and Floor to allow either an int, an expression, or ? - which means unknown so use default. Those are all calculated at monster expression time. But the attribute has the original form. I added a test for those three factors with "?". I also reran the "fix" script on the original monsters.txt and included the fixed file. Take another look, please? DataFileMechanicsTest just checks the number of delimited fields in a line, for various data files. The previous iteration of monsters.txt was too complex for that. Once this work is done we might consider whether the new version is simpler and thus a candidate for that test. I'll monitor but this will help remind me or start a sidebar discussion about whether DataFileMechanicsTest has any utility as a test. The new version is not "simpler" in that the number of delimited fields did not change; there are 4 required fields and then an arbitrary number of tab-separated item drops. Now, if DataFileMechanicsTest wants to test the CONTENTS of the various fields, it could certainly validate whether the "attributes" field of this file are correct. Of course, running KoLmafia will do that for you. :) I moved the duplicate attribute check up in front of the switch statement which handles them. Monster: "lihc" has multiple values for the "EA:" attribute. Monster: "large hostile plant" has multiple values for the "Init:" attribute. This means that I can't be specific about the two values (since we haven't parsed them yet), but it would catch the error with the two Init: values. If we add support for multiple EA:, this will have to change - just as the code for parsing EA: will have to change. :) I fixed the large hostile plant to have the formula. midgleyc requested changes on behalf of https://github.com/orgs/kolmafia/teams/staff 4 hours ago large hostile plant had Init present twice (one formula, one number), and has been collapsed to one (number). This was probably a mistake in the initial data? Other than the plant, this all seems sensible. I fixed that, but I don't seem to have the option to "resolve" the request?
gharchive/pull-request
2022-06-02T22:16:52
2025-04-01T06:39:18.417957
{ "authors": [ "Veracity0", "jaadams5", "midgleyc" ], "repo": "kolmafia/kolmafia", "url": "https://github.com/kolmafia/kolmafia/pull/793", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2116490100
could you help me fix them? @kolos26 The places I marked were connected at the top of the plane (yellow line), but there were still errors after I changed them, and I couldn't do it precisely because it was difficult to change it on the map. Thank you! Hainan Airlines and Hong Kong Airlines belong to the same group, so their fuselages are the same I will try to fix them :) sorry for not reacting for your messages, i had tones of homeworks @kolos26 maybe i have a idea to fix. it if i can not fix,i will tell you.and under my promotion, about a thousand people play your livery plugin in chinese bilibili! @kolos26 maybe i have a idea to fix. it if i can not fix,i will tell you.and under my promotion, about a thousand people play your livery plugin in chinese bilibili! Wow thats cool, thanks a lot! One of the hardest things is to monitor non contributor users. Thank You For Supporting Us! Btw this reminded me how huge China is... @kolos26 maybe i have a idea to fix. it if i can not fix,i will tell you.and under my promotion, about a thousand people play your livery plugin in chinese bilibili! Wow thats cool, thanks a lot! One of the hardest things is to monitor non contributor users. Thank You For Supporting Us! Btw this reminded me how huge China is... please you fix them my idea is wrong:(
gharchive/issue
2024-02-03T13:50:32
2025-04-01T06:39:18.427543
{ "authors": [ "1414841886", "kolos26" ], "repo": "kolos26/GEOFS-LiverySelector", "url": "https://github.com/kolos26/GEOFS-LiverySelector/issues/512", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1694902567
SendTransaction request Description Closes #18 Type of change [ ] New feature (non-breaking change which adds functionality) Checklist [ ] My code follows the style guidelines of this project [ ] I have performed a self-review of my code [ ] I have commented on my code, particularly in hard-to-understand areas [ ] I have made corresponding changes to the documentation [ ] My changes generate no new warnings [ ] I have added tests that prove my fix is effective or that my feature works [ ] New and existing unit tests pass locally with my changes Pull Request Test Coverage Report for Build 6a728f28715d8899f38398a23266491a59653a37-PR-34 2 of 2 (100.0%) changed or added relevant lines in 2 files are covered. No unchanged relevant lines lost coverage. Overall coverage remained the same at 100.0% Totals Change from base Build ad932605898b88c4966fadf8167963d66e3ca2c5: 0% Covered Lines: 133 Relevant Lines: 133 💛 - Coveralls
gharchive/pull-request
2023-05-03T21:58:28
2025-04-01T06:39:18.436621
{ "authors": [ "Odraxs", "coveralls" ], "repo": "kommitters/soroban.ex", "url": "https://github.com/kommitters/soroban.ex/pull/34", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }