id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
866070660
Connection keeps on pending after several plays Having multiple wavesurfer instances, after played several songs connection eventually get stuck on "pending" status. Here is a pen with the test: https://codepen.io/diegoigle/pen/OJWrZvO I'm using Chrome lastest build. Can someone help? Make sure all the instances are paused or stopped before on the play event. add for (var i = 0; i < waveforms.length; i++) { waveforms[i].pause(); } to the playerPlay var playerPlay = function(buttonIdAttr){ currentPlaying['status'] = 'PLAYING'; currentPlaying['id'] = buttonIdAttr; for (var i = 0; i < waveforms.length; i++) { waveforms[i].pause(); } waveforms[currentPlaying['id']].play(); } Thank for your help. Unfortunately it keeps happening, though now it takes a little more retries to get it pending. I updated the pen with your code: https://codepen.io/diegoigle/pen/OJWrZvO https://stackoverflow.com/questions/23679968/chrome-hangs-after-certain-amount-of-data-transfered-waiting-for-available-soc
gharchive/issue
2021-04-23T12:35:53
2025-04-01T04:34:44.848237
{ "authors": [ "diegoiglesias", "entonbiba", "leeqkrtkxp" ], "repo": "katspaugh/wavesurfer.js", "url": "https://github.com/katspaugh/wavesurfer.js/issues/2249", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1438342703
ReferenceError: self is not defined Hi, After running a global npm update I’m not able to use the package anymore with the following error: self is not defined I’m wrapping the package in a Vue component for a nuxt app. I tried to reverse back to the old version that worked previously (5.1.0) but it also gave an error (window is not defined). I guess this is due to the update of other modules/webpack. The current installed version is 6.4.0. I saw that other people experienced the same problem: https://github.com/katspaugh/wavesurfer.js/issues/2350 https://github.com/katspaugh/wavesurfer.js/issues/751 https://github.com/ShiiRochi/wavesurfer-react/issues/51 I tried to dynamically import wavesurfer.js in mounted and also to load it as a plugin only in client mode but still no success. It produce a new error: defu__WEBPACK_IMPORTED_MODULE_3__ is not a function Vscode give a tooltip hint on the import statement: Could not find a declaration file for module ‘wavesurfer.js’. ‘/Users/xx/xx/node_modules/wavesurfer.js/dist/wavesurfer.js’ implicitly has an ‘any’ type. I want to be clear that before the npm update everything worked perfectly. Thanks for your help, I was able to make it work with dynamic import once I deleted the node_modules folder and re-run npm install.
gharchive/issue
2022-11-07T13:09:30
2025-04-01T04:34:44.852811
{ "authors": [ "dany68" ], "repo": "katspaugh/wavesurfer.js", "url": "https://github.com/katspaugh/wavesurfer.js/issues/2615", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
188187782
How is the position of the cursor determined Can someone please tell me what variables controls the position of the cursor on the canvas? I have added the regions plugin to my project, and when I have added multiple regions, it becomes difficult to control the position of the cursor by clicking different areas on the canvas. When one region is currently playing (looping) and I click outside of the region, I expect the cursor to move to that region, and start playing at this point. Also I am having trouble switching between regions, playing one after the other, by clicking within the region I want to play next. I would very much like to delve into the code as such to fix these issues, but I need some help figuring out how the cursor is being controlled, and whether the cursor position is tied to the position where the player will continue playing etc. It seems that there are more than one variable controlling the cursor, the start play position and being within a region or not - and all these comes into effect resulting in a user experience that leaves much to be desired. I would of course appreciate any help on the matter of figuring out how it all ties together. BR, Harriet The inconsistency you describe sounds like a bug in the way region looping is implemented. The cursor shouldn't be controlled by anything except for the audioprogress event. Has the problem been solved?
gharchive/issue
2016-11-09T08:10:16
2025-04-01T04:34:44.856064
{ "authors": [ "harrietsCreativeDabblets", "katspaugh", "wangzhaoming132" ], "repo": "katspaugh/wavesurfer.js", "url": "https://github.com/katspaugh/wavesurfer.js/issues/858", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
210470423
module is not defined, cordova.plugins deprecated meanwhile cordova.plugins is deprecated. How do i get a reference to the backgroundMode object? I tried cordova.require('de.appplant.cordova.plugin.background-mode') but with no success # solved my own problem: var backgroundMode = cordova.require("cordova-plugin-background-mode.BackgroundMode"); did it for me
gharchive/issue
2017-02-27T12:29:25
2025-04-01T04:34:44.858949
{ "authors": [ "der-jochen" ], "repo": "katzer/cordova-plugin-background-mode", "url": "https://github.com/katzer/cordova-plugin-background-mode/issues/254", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1325654678
chore: import repository https://github.com/kavishkhanna/standard-usage-schemas.git this commit will trigger a pipeline to generate the CI/CD configuration which will create a second commit on this Pull Request before it auto merges [APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: To complete the pull request process, please assign You can assign the PR to them by writing /assign in a comment when ready. The full list of commands accepted by this bot can be found here. Needs approval from an approver in each of these files: OWNERS Approvers can indicate their approval by writing /approve in a comment Approvers can cancel approval by writing /approve cancel in a comment
gharchive/pull-request
2022-08-02T10:38:43
2025-04-01T04:34:44.889072
{ "authors": [ "kavishkhanna" ], "repo": "kavishkhanna/jx3-eks-asm", "url": "https://github.com/kavishkhanna/jx3-eks-asm/pull/2", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
951223435
Redesign bottom navigation + header to follow Website 2.0 design UI Design New bottom navigation Navigation menu (shown when clicking Menu) Implementation model Navigation menu disclosure is controlled by Headless UI (bundlephobia) ❓ Disclosure or Popover? Tasks [x] Redesign bottom navigation [ ] Add overflow menu to navigation I think Popover makes more sense in this case. Is this bottom navigation applied in desktop view too? I know that the priority is mobile first, but I find this bottom navigation a bit inconvenient in desktop view So maybe we can make it to just show in mobile and for desktop, showing header menu just like in https://kawalcovid19.id/ @ekamuktia The design on the Figma file only has a mobile version, so yes, this bottom nav will be present on both mobile and desktop. Although for the desktop, I think we can move the menu button to the global header instead, since we also have this design on the Figma file: Will think more about this tomorrow.
gharchive/issue
2021-07-23T03:46:38
2025-04-01T04:34:44.895849
{ "authors": [ "ekamuktia", "resir014", "zainfathoni" ], "repo": "kawalcovid19/wargabantuwarga.com", "url": "https://github.com/kawalcovid19/wargabantuwarga.com/issues/346", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
569181341
Transportation section Transportation finishing touches depends on the venue and hotels we select. In the meantime, can we complete the small section below on the site: "NFTA's Metro Rail provides a direct line from UB's South Campus to downtown Canalside. From metro stops, it's just a quick walk to many of Buffalo's entertainment and dining options. " Potential images:  https://commons.wikimedia.org/wiki/File:Main_Street_in_Buffalo_NY_with_light_rail_and_Shea's_Performing_Arts_Center_(2015).jpg https://commons.wikimedia.org/wiki/File:Theater_Station_at_Sheas.jpg https://commons.wikimedia.org/wiki/File:20150827_61_NFTA_Light_Rail_at_Fountain_Plaza_(21990211710).jpg https://commons.wikimedia.org/wiki/File:Buffalo_light_rail_train_laying_over_at_Erie_Canal_Harbor_station_(2015).jpg I'm sorry @cmklotzb I am not sure I understand what you mean on "the small section below on the site" Important to remember "completeness" will be done by a different group of folks, so no one reading the proposal brochure expects the decision on where the venue will be to be included. Compare for example the proposal for Pittsburgh and the Pittsburgh actual conference Never mind @cmklotzb I see what you mean now.
gharchive/issue
2020-02-21T20:59:25
2025-04-01T04:34:44.903089
{ "authors": [ "cmklotzb", "kayiwa" ], "repo": "kayiwa/c4lbuffaloproposal", "url": "https://github.com/kayiwa/c4lbuffaloproposal/issues/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
299578709
Revert stream to regular map construction Getting weird errors: java.lang.IllegalStateException: Duplicate key kbasesearchengine.search.ObjectData@2ab1c576 Note that ObjectData is the value Urrg this went to master instead of develop
gharchive/pull-request
2018-02-23T02:28:32
2025-04-01T04:34:44.926721
{ "authors": [ "MrCreosote" ], "repo": "kbase/KBaseSearchEngine", "url": "https://github.com/kbase/KBaseSearchEngine/pull/200", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1865982324
Update fastapi & pydantic Pydantic v2 has a number of breaking changes. Also see https://github.com/kbase/collections/issues/428 I didn't think of that, but looking at the code changes, they shouldn't break anything since microtrait doesn't actually use the models any more, just the field names. Therefore the imports just need to work. Would it be ok if I asked you to bump pydantic next time you work on the tools? I've literally never run them so testing them is a bit of a pain for me currently (which at some point I need to fix)
gharchive/pull-request
2023-08-24T22:36:41
2025-04-01T04:34:44.928316
{ "authors": [ "MrCreosote" ], "repo": "kbase/collections", "url": "https://github.com/kbase/collections/pull/429", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1284321839
Implement and evaluate PSRAM QSPI controller First, just implement something that can write and read the PSRAM. Write test code that fills whole PSRAM with pseudo-random numbers, then verify. Measure performance, latency. Investigate how to tie it in to the n64 bus PIO. Latency test: Set up a test harness that, when a trigger gpio goes high, loads 2 bytes from the PSRAM, outputs the result on gpio0->gpio15. Measure latency with a logic analyzer. Not relevant anymore
gharchive/issue
2022-06-24T23:44:15
2025-04-01T04:34:44.932407
{ "authors": [ "kbeckmann" ], "repo": "kbeckmann/PicoCart64", "url": "https://github.com/kbeckmann/PicoCart64/issues/3", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
107315535
Cow<'static, str> for strings; no excessive lifetimes Increasing performance and easing away the lifetimes. https://gist.github.com/llogiq/4557032bdf18a2cff136 Cow works by offering both the &'static str and the owned String. Maybe even use the lifetime of the App or Arg instead of 'static. In which would maintain one lifetime at each struct. I'm game for it, let's try and it see how it works out! Worst case scenario, we have a concept to shelf for a future iteration :+1: @kbknapp I did try it, and reverted it to &'a See my rewrite. :D Ah, my mistake! :)
gharchive/issue
2015-09-19T06:21:41
2025-04-01T04:34:44.945192
{ "authors": [ "james-darkfox", "kbknapp" ], "repo": "kbknapp/clap-rs", "url": "https://github.com/kbknapp/clap-rs/issues/258", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
316671256
Removed dynamic evaluate functions closes #166 I think I found a solution - https://github.com/GoogleChrome/puppeteer/blob/v1.0.0/lib/ExecutionContext.js#L42 const result = await handle.jsonValue().catch(error => undefined); we should change the line https://github.com/kblok/puppeteer-sharp/blob/master/lib/PuppeteerSharp/ExecutionContext.cs#L33 var result = await handle.JsonValue<T>(); and make it catch the exception and return null Make sense, we should also add the EvaluateBigObject test here. PuppeteerSharp.Tests.Puppeteer.PuppeteerLaunchTests.UserDataDirOption [FAIL] System.Net.WebSockets.WebSocketException : The 'System.Net.WebSockets.InternalClientWebSocket' instance cannot be used for communication because it has been transitioned into the 'Aborted' state. Conflicted @Meir017
gharchive/pull-request
2018-04-23T05:59:50
2025-04-01T04:34:44.948116
{ "authors": [ "Meir017", "kblok" ], "repo": "kblok/puppeteer-sharp", "url": "https://github.com/kblok/puppeteer-sharp/pull/167", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1578444404
Are there any real examples of using rich text and promise toasts? I'm looking through the docs and I can't figure out how to use rich text in my toasts. Are there any real examples? If you scroll down to the bottom of the docs, there should be a button called "Rich content" that shows you a demo of rich text as well as how to implement it. You need to create a Svelte component (say RichText.svelte) which contains your rich text (i.e. any HTML markup). Then just import it wherever you're creating your toasts and call toast with toast(RichText). Let me know if that works for you or if you have a different use case that's not addressed by the docs!
gharchive/issue
2023-02-09T19:10:13
2025-04-01T04:34:44.964490
{ "authors": [ "jakubdonovan", "kbrgl" ], "repo": "kbrgl/svelte-french-toast", "url": "https://github.com/kbrgl/svelte-french-toast/issues/23", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
485301190
Expand all favorites groups button Is your feature request related to a problem? Please describe. Sometimes, my favorites disappear or become collapsed, leaving me to expand them all out again. Describe the solution you'd like Provide an Expand All button like the existing Collapse All button, but that will recursively expand all favorites groups instead. I would love that also! On macOS this seems to work already, just hold option (aka Alt) while clicking on the expand arrow. The normal file explorer in VSCode works the same way. On macOS this seems to work already, just hold option (aka Alt) while clicking on the expand arrow. The normal file explorer in VSCode works the same way. I think it works only for subfolders/subgroups in a group? I cannot expand all folders/groups with one click if they aren't nested, or am I missing something? Ah, now I understood. Yes, what you describe is currently not possible. (To not clutter the UI maybe the "collapse all" button could become a toggle for both expanding/collapsing all.)
gharchive/issue
2019-08-26T15:37:02
2025-04-01T04:34:44.981492
{ "authors": [ "arfordweb", "fharper", "halo" ], "repo": "kdcro101/vscode-favorites", "url": "https://github.com/kdcro101/vscode-favorites/issues/43", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1810138530
Fix foreign grants and functions ownership chyběly schéma granty na cizí projekty @pivnicek neměl by si chvilku? abych to mohl mergnout
gharchive/pull-request
2023-07-18T15:13:07
2025-04-01T04:34:45.026147
{ "authors": [ "ondrajodas" ], "repo": "keboola/project-migration-tool", "url": "https://github.com/keboola/project-migration-tool/pull/16", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1852664408
SOX-353 vytvareni tokenu Original PR: https://github.com/keboola/storage-api-php-client/pull/1163 Reverts keboola/storage-api-php-client#1167 KBC: https://github.com/keboola/connection/pull/4574 Guest token vyřešen v connection Nejde to approvnou zajcovi, ale říkal, že je to ok :)
gharchive/pull-request
2023-08-16T07:24:01
2025-04-01T04:34:45.032535
{ "authors": [ "tomasfejfar", "zajca" ], "repo": "keboola/storage-api-php-client", "url": "https://github.com/keboola/storage-api-php-client/pull/1168", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1809282103
chore: Remove all non-editable keys from resources Provide a description of what has been changed The field creationTimestamp is auto-generated by kubernetes. This field is ignored during creation/update because it's non-editable by clients. Checklist [x] I have verified that my change is according to the deprecations & breaking changes policy [x] Commits are signed with Developer Certificate of Origin (DCO - learn more) [x] README is updated with new configuration values (if applicable) learn more [x] A PR is opened to update KEDA core (repo) (if applicable, ie. when deployment manifests are modified) Thanks for the fix ❤️
gharchive/pull-request
2023-07-18T07:07:47
2025-04-01T04:34:45.036250
{ "authors": [ "JorTurFer", "alexmanno" ], "repo": "kedacore/charts", "url": "https://github.com/kedacore/charts/pull/498", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
443118618
Débug outside cluster Someone could write a walkthrough on how to debug Keda outside the cluster using vscode. aminebizid- I have hard a hard time doing this due to the metrics adapter. I could not get a version of it running stand alone outside the cluster. Without this, KEDA is debuggable as a stand alone app. So a workaround is to comment the adapter setup portion in main.go and start it up. Could you show us an example of vscode launch.json example of launch.json for debugging Keda controller { "version": "0.2.0", "configurations": [ { "name": "Launch", "type": "go", "request": "launch", "mode": "auto", "program": "${workspaceFolder}/cmd/manager/main.go", "env": {"OPERATOR_NAME": "keda-operator", "WATCH_NAMESPACE": ""}, "args": ["--kubeconfig=/path/to/config/file"] } ] }
gharchive/issue
2019-05-12T15:48:43
2025-04-01T04:34:45.038250
{ "authors": [ "Aarthisk", "aminebizid", "nisan270390" ], "repo": "kedacore/keda", "url": "https://github.com/kedacore/keda/issues/190", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
689180356
Make controllers/ golint compatible util/status.go : Comments added for exported functions: "SetStatusConditions" and "UpdateScaledObjectStatus". util/string_lists.go : Comments added for exported functions: "Contains" and "Remove". scaledjob_controller : Comments added for exported methods: "SetupWithManager" and "Reconcile". An else block readjusted to make it golint compliant. scaledobject_controller.go : Comments added to exported methods: "SetupWithManager" and "Reconcile". Two else blocks readjusted to make it golint compliant. Checklist [x] Commits are signed with Developer Certificate of Origin (DCO) [ ] Tests have been added [ ] A PR is opened to update the documentation on https://github.com/kedacore/keda-docs [ ] Changelog has been updated Fixes #1055 Hi folks, This is my first time contributing to an open source project. I have done things to the best of my knowledge. Please let me know if something is missing or was done incorrectly. Any and all constructive criticism is welcome.
gharchive/pull-request
2020-08-31T12:47:21
2025-04-01T04:34:45.041979
{ "authors": [ "Mystigan" ], "repo": "kedacore/keda", "url": "https://github.com/kedacore/keda/pull/1065", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1299783356
Migrate Azure Log Analytics test to Go. Signed-off-by: Vighnesh Shenoy vshenoy@microsoft.com Provide a description of what has been changed Checklist [x] Commits are signed with Developer Certificate of Origin (DCO - learn more) [x] Tests have been added [x] A PR is opened to update our Helm chart (repo) (if applicable, ie. when deployment manifests are modified) [x] A PR is opened to update the documentation on (repo) (if applicable) [x] Changelog has been updated and is aligned with our changelog requirements Fixes #3258 Relates to #2737 /run-e2e log_analytics /run-e2e log_analytics /run-e2e log_analytics
gharchive/pull-request
2022-07-09T23:02:54
2025-04-01T04:34:45.046889
{ "authors": [ "v-shenoy" ], "repo": "kedacore/keda", "url": "https://github.com/kedacore/keda/pull/3345", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2195492510
Add prototype Event Management calls and Event Proxy impl files Describe the contribution This PR add prototype implementations for the cFS Event Proxy design. This design intends to abstract the cFS EVS API calls through the Event Management core in BPLib. A related PR exists for BPLib, here: https://github.com/keegan-moore/bplib/pull/1 . This design change causes the following in the BP app: During app initialization BP now calls BPL_EVM_Initialize instead of CFE_EVS_Register BPL_EVM_Initialize takes a struct of function pointers to the cFS Event Proxy implementation When the BP app wants to generate an event, the app should now use BPL_EVM_SendEvent Testing performed Steps taken to test the contribution: Build steps make native.distclean 2.make ENABLE_TESTS=false native.install Execution steps '...' cd build-native-9.4.0/exe/cpu1 ./core-cpu1 Expected behavior changes EVS's CFE_EVS_Register should no longer generate the following warning in the ES System Log, since the event filter array is now appropriately sized: 1980-012-14:03:20.55879 CFE_EVS_Register: Filter limit truncated to 8 A clear and concise description of how this contribution will change behavior and level of impact. API Changes: During app initialization BP now calls BPL_EVM_Initialize instead of CFE_EVS_Register When the BP app wants to generate an event, the app should now use BPL_EVM_SendEvent Behavior Change: EVS's CFE_EVS_Register should no longer generate a warning in the ES System Log during BP app init No other changes to behavior System(s) tested on Hardware: PC / x86_64 OS: Ubuntu 22.04 Versions: cFE 6.7 (equuleus) release candidate Additional context N/A Third party code N/A Contributor Info - All information REQUIRED for consideration of pull request Keegan Moore, NASA/GSFC Code 582 (Flight Software Systems) Before this change set is complete, we'd need to scrape through BP and look for all calls to CFE_EVS_SendEvent, replacing them with calls to BPL_EVM_SendEvent .
gharchive/pull-request
2024-03-19T16:59:43
2025-04-01T04:34:45.057385
{ "authors": [ "keegan-moore" ], "repo": "keegan-moore/bp", "url": "https://github.com/keegan-moore/bp/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2464536145
[bug] [bitcoin-core test] 'NOP OVER 1' Issue The following bitcoin-core test is failing to give the correct result : NOP OVER 1 This code should be failing due to a INVALID_STACK_OPERATION. Replicate To replicate the issue you can use the latest main branch with the following command: scarb cairo-run '[[],5132112,3,[],87231948791857,6]' Please can I be assigned to this issue @ShantelPeters For sure, thank you! Let me know if you have any issues. @b-j-roberts Thank you sir … I’ll get to work immediately.
gharchive/issue
2024-08-14T00:07:59
2025-04-01T04:34:45.073047
{ "authors": [ "ShantelPeters", "b-j-roberts" ], "repo": "keep-starknet-strange/shinigami", "url": "https://github.com/keep-starknet-strange/shinigami/issues/135", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
70737607
Decipher/Cipher links not correct when URL not traditional We're running redmine at https://redmine.company.com/redmine, not just at https://redmine.company.com. When we try to encode/decode, the URL gets rewritten incorrectly. For example: https://redmine.company.com/redmine/projects/testproject/wiki gets rewritten to https://redmine.company.com/projects/testproject/wiki. Dear @zlacelle First of all, thank you for your feedback. And yes, we kind of have that URL creation hard-coded for a Redmine deployment in / (not suitable for other deployments). So, as soon as we have the opportunity/time, we will fix that and let you know. OR, if you wish to, you can provide a fix by submitting a pull request, which we will happily review and integrate with the base code. Great, thanks for replying. I just changed the hardcoded value, but I'll see if I can learn a bit of Ruby and come up with a fix.
gharchive/issue
2015-04-24T16:48:54
2025-04-01T04:34:45.091131
{ "authors": [ "hsilva-keep", "zlacelle" ], "repo": "keeps/redmine_wikicipher", "url": "https://github.com/keeps/redmine_wikicipher/issues/23", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2455772600
Can't fill out the Sentry dialog when in a dialog Tapping in a text field doesn't show the keyboard - it comes up briefly and then goes away when something steals focus Ref: cfe751488d0969e4d43d68cb02baf0c4bacb28e8 Well it seems to work OK on the About Page now. Could reproduce it locally in the iOS Simulator in XCode, but when I tested it was working using this dummy button on the AboutPage. Tried removing 100dvh and .sentry-error-embed-wrapper locally, but still couldn't get it to work properly like it did for real. If it occurs again, next steps might be debugging on my phone for real by connecting it to my laptop. Ref: https://docs.sentry.io/platforms/javascript/guides/react/user-feedback/#crash-report-modal Also ref: 26092add5089698a0c4954f6822b447d98570f0f (Just to be safe)
gharchive/issue
2024-08-08T13:11:17
2025-04-01T04:34:45.129258
{ "authors": [ "keithamoss" ], "repo": "keithamoss/demsausage", "url": "https://github.com/keithamoss/demsausage/issues/3091", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
665935783
一句傻逼送给你,不成敬意 you seems like a fool, without respect~ and then... Your mother exploded on the spot 老哥 冷静 calm down ... it is only a joke 大哥冷静 大哥,稳住 Please don't spit fragrance in the community 大哥冷静冷静 认真你就输了
gharchive/issue
2020-07-27T02:50:45
2025-04-01T04:34:45.169954
{ "authors": [ "12yy296", "Jiang-xs", "Naozumi520", "Suckson", "kav012375", "mind-c", "yuluowuying" ], "repo": "kelseyhightower/nocode", "url": "https://github.com/kelseyhightower/nocode/issues/4108", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
256520882
locking dependencies of editable vcs requirements When we pip install then pip freeze an editable vcs requirement, the frozen requirements include all dependencies. With pipenv the dependencies of editable vcs requirements do not appear in Pipfile.lock. Example: $ pipenv install -e git+https://github.com/sbidoul/subdirtest.git#egg=subdirtest\&subdirectory=setup ... $ cat Pipfile.lock ... "default": { "subdirtest": { "editable": true, "git": "https://github.com/sbidoul/subdirtest.git", "subdirectory": "setup", "ref": "7b18afe980a42bfe8d1a96b5ed52e05f58d9ef6c" } }, ... $ pipenv run pip freeze click==6.7 pkg-resources==0.0.0 -e git+https://github.com/sbidoul/subdirtest.git@7b18afe980a42bfe8d1a96b5ed52e05f58d9ef6c#egg=subdirtest&subdirectory=setup Since the installed package depends on click I expect to see it in the locked dependencies. pip-tools lacks this capability (https://github.com/kennethreitz/pipenv/issues/508#issuecomment-328345816) This is definately out of scope for now — until pip-tools grows this ability — I'd raise the issue over there. is there an issue in pip-tools for this that I can follow? if not, can you make one @sbidoul? Problem is that I don't know pip-tools so I can't create a meaningful issue in that project. I'd need to do some research on the pip-tools api and how it is used by pipenv. Issue title: "Resolve dependencies for editable requirements, required by pipenv" contents: the contents of this issue. pip-tools is poorly maintained, so i doubt it'll happen. I'm committer on it though, so if you write a patch to get this to work, I'll merge it! @kennethreitz is it supposed to be implemented in 7.3.2? yep! Hm, looks like its not working for me with 7.3.5 (the command in the description above). oh i dont' know if it works for subdirectories, i know it works for VCSs in general now though. I just tried pipenv install -e git+https://github.com/kennethreitz/requests#egg=requests and I don't get the dependencies in the lockfile. ah, it works with the other form. v7.6.3 released, which has this confirmed working! Almost there: $ pipenv install -e git+https://github.com/kennethreitz/requests#egg=requests Installing -e git+https://github.com/kennethreitz/requests#egg=requests… Obtaining requests from git+https://github.com/kennethreitz/requests#egg=requests Updating /home/sbi-local/.local/share/virtualenvs/pipenvtest-egFzzt6y/src/requests clone Requirement already satisfied: chardet<3.1.0,>=3.0.2 in /home/sbi-local/.local/share/virtualenvs/pipenvtest-egFzzt6y/lib/python2.7/site-packages (from requests) Requirement already satisfied: idna<2.7,>=2.5 in /home/sbi-local/.local/share/virtualenvs/pipenvtest-egFzzt6y/lib/python2.7/site-packages (from requests) Requirement already satisfied: urllib3<1.23,>=1.21.1 in /home/sbi-local/.local/share/virtualenvs/pipenvtest-egFzzt6y/lib/python2.7/site-packages (from requests) Requirement already satisfied: certifi>=2017.4.17 in /home/sbi-local/.local/share/virtualenvs/pipenvtest-egFzzt6y/lib/python2.7/site-packages (from requests) Installing collected packages: requests Running setup.py develop for requests Successfully installed requests Adding -e git+https://github.com/kennethreitz/requests#egg=requests to Pipfile's [packages]… Locking [dev-packages] dependencies… Locking [packages] dependencies… Traceback (most recent call last): File "/home/sbi-local/.local/bin/pipenv", line 11, in <module> sys.exit(cli()) File "/home/sbi-local/.local/venvs/pipenv/lib/python2.7/site-packages/pipenv/vendor/click/core.py", line 722, in __call__ return self.main(*args, **kwargs) File "/home/sbi-local/.local/venvs/pipenv/lib/python2.7/site-packages/pipenv/vendor/click/core.py", line 697, in main rv = self.invoke(ctx) File "/home/sbi-local/.local/venvs/pipenv/lib/python2.7/site-packages/pipenv/vendor/click/core.py", line 1066, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/home/sbi-local/.local/venvs/pipenv/lib/python2.7/site-packages/pipenv/vendor/click/core.py", line 895, in invoke return ctx.invoke(self.callback, **ctx.params) File "/home/sbi-local/.local/venvs/pipenv/lib/python2.7/site-packages/pipenv/vendor/click/core.py", line 535, in invoke return callback(*args, **kwargs) File "/home/sbi-local/.local/venvs/pipenv/local/lib/python2.7/site-packages/pipenv/cli.py", line 1500, in install do_lock(system=system) File "/home/sbi-local/.local/venvs/pipenv/local/lib/python2.7/site-packages/pipenv/cli.py", line 942, in do_lock project=project File "/home/sbi-local/.local/venvs/pipenv/local/lib/python2.7/site-packages/pipenv/utils.py", line 133, in resolve_deps extra_constraints = best_matches_from(dep[len('-e '):], which=which, which_pip=which_pip, project=project) File "/home/sbi-local/.local/venvs/pipenv/local/lib/python2.7/site-packages/pipenv/utils.py", line 106, in best_matches_from setup_py_path = os.path.abspath(os.sep.join([result, 'setup.py'])) TypeError: sequence item 0: expected string, NoneType found Ugh, in Python 3 this doesn't result in a type error.
gharchive/issue
2017-09-10T16:15:53
2025-04-01T04:34:45.183418
{ "authors": [ "AlJohri", "kennethreitz", "sbidoul" ], "repo": "kennethreitz/pipenv", "url": "https://github.com/kennethreitz/pipenv/issues/512", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
128419712
Inconsistent behaviour with https between urllib3 and requestsion. In trying to make a https connection as so: (trustmile-api-p2710)jBeast:trustmile james$ python -c "import requests; r = requests.get('https://devapi.trustmile.com/static/index.html', verify=True)" Traceback (most recent call last): File "<string>", line 1, in <module> File "/Users/james/.virtualenvs/trustmile-api-p2710/lib/python2.7/site-packages/requests/api.py", line 67, in get return request('get', url, params=params, **kwargs) File "/Users/james/.virtualenvs/trustmile-api-p2710/lib/python2.7/site-packages/requests/api.py", line 53, in request return session.request(method=method, url=url, **kwargs) File "/Users/james/.virtualenvs/trustmile-api-p2710/lib/python2.7/site-packages/requests/sessions.py", line 468, in request resp = self.send(prep, **send_kwargs) File "/Users/james/.virtualenvs/trustmile-api-p2710/lib/python2.7/site-packages/requests/sessions.py", line 576, in send r = adapter.send(request, **kwargs) File "/Users/james/.virtualenvs/trustmile-api-p2710/lib/python2.7/site-packages/requests/adapters.py", line 447, in send raise SSLError(e, request=request) requests.exceptions.SSLError: ("bad handshake: Error([('SSL routines', 'ssl3_get_server_certificate', 'certificate verify failed')],)",) I have latest necessary packages installed: (trustmile-api-p2710)jBeast:trustmile james$ pip freeze | egrep '(requests|urllib3|certifi)' certifi==2015.11.20.1 requests==2.9.1 urllib3==1.14 However, urllib works in the same environment: import urllib3 import certifi http = urllib3.PoolManager( cert_reqs='CERT_REQUIRED', # Force certificate check. ca_certs=certifi.where(), # Path to the Certifi bundle. ) # You're ready to make verified HTTPS requests. try: r = http.request('GET', 'https://devapi.trustmile.com/static/index.html') print r.status except urllib3.exceptions.SSLError as e: print e Output is status 200. Am I missing something? I tried to set the relevant environment variables. The cert is a legit one and it's root cert is in the cacerts.pem in certifi package. Thanks again for an awesome package I use all the time Kenneth et al! Out of interest, what version of OpenSSL are you using? Run python -c 'import ssl; print ssl.OPENSSL_VERSION' OpenSSL 0.9.8zg 14 July 2015 Interesting. Does this problem occur if you don't pass verify=True? Yep, same issue Hmm. What other packages do you have installed? Ah, that must be it, created a clean env with just certifi, requests, and urllib3 in it. Will investigate further. I have a lot installed. ok, fixed in my env. Not sure which was the culprit but this worked pip uninstall backports.ssl-match-hostname pip uninstall pyOpenSSL Yeah, so PyOpenSSL is the likely culprit. Requests, unlike urllib3, will automatically try to use PyOpenSSL if it's present. Can you reinstall it, and then run your urllib3 script with these two lines added to the top: import urllib3.contrib.pyopenssl urllib3.contrib.pyopenssl.inject_into_urllib3() Your script would look like this: import urllib3 import urllib3.contrib.pyopenssl import certifi urllib3.contrib.pyopenssl.inject_into_urllib3() http = urllib3.PoolManager( cert_reqs='CERT_REQUIRED', # Force certificate check. ca_certs=certifi.where(), # Path to the Certifi bundle. ) # You're ready to make verified HTTPS requests. try: r = http.request('GET', 'https://devapi.trustmile.com/static/index.html') print r.status except urllib3.exceptions.SSLError as e: print e
gharchive/issue
2016-01-24T19:41:01
2025-04-01T04:34:45.189284
{ "authors": [ "Lukasa", "jorourke" ], "repo": "kennethreitz/requests", "url": "https://github.com/kennethreitz/requests/issues/2978", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
278078073
Pretrained Model on Sports1M Dear @kenshohara, Thank you very much for your fantastic repository. Do you have any pretrained model on Sports1M? No, I have not tried training on Sports1M because Sports1M is too big to utilize easily. The size of Kinetics is not quite big (2TB storage is sufficient), and include sufficient training data to train 3D CNNs. If you do not have specific reason for using Sports1M, I think Kinetics pretrained models are useful. Thank for your time and response. I agree with you there. Pretrained 3D CNN models on Kinetics are welcome.
gharchive/issue
2017-11-30T10:56:35
2025-04-01T04:34:45.206975
{ "authors": [ "ahkarami", "kenshohara" ], "repo": "kenshohara/3D-ResNets-PyTorch", "url": "https://github.com/kenshohara/3D-ResNets-PyTorch/issues/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
419720179
NODE_ENV undefined with typescript and nodemon cross-env version: 5.2.0 node version: 10.12.0 npm (or yarn) version: 6.4.1 Relevant code or config "dev:ts": "cross-env NODE_ENV=production tsc-watch --onSuccess \"node ./node_modules/nodemon/bin/nodemon.js dist/main.js\"" "dev:ts": "tsc-watch --onSuccess \"cross-env NODE_ENV=production node ./node_modules/nodemon/bin/nodemon.js dist/main.js\"" Tried both commands above but console.log(process.env.NODE_ENV) prints undefined. Any help is appreciated. Can't reproduce. The environment looks fine. Sorry for not responding to this earlier. It's been long enough that I'm going to close this issue. If you're still experiencing problems, please open a new issue. @minotaurrr 17 thumbs up i'm encountering same issue, did you found a fix?
gharchive/issue
2019-03-11T22:47:17
2025-04-01T04:34:45.210133
{ "authors": [ "cristian-eriomenco", "geastwood", "kentcdodds", "minotaurrr" ], "repo": "kentcdodds/cross-env", "url": "https://github.com/kentcdodds/cross-env/issues/198", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
344045829
JSS style blocks overwhelm prettyDOM / debug output Describe the feature you'd like: The automatic debug output / prettyDOM is hard to pick through with JSS, since a rather large style block is added to the DOM with JSS classes. There should be a way to avoid JSS output (the material-ui library in our case dumps out 10,000 lines of CSS, which is thought-provoking in itself, but here we are). Suggested implementation: I was thinking of an option to prune style blocks from the output; @kentcdodds suggested that defaulting the debug output to the container instead of the documentElement might be a better idea. Thinking about it more, what if we use document.body? That would avoid this issue without making it harder/confusing for portals. That would work well for my purposes; the JSS style blocks are in the document head. I know browsers are lenient with Hmmm... I actually think this change will need to be in react-testing-library because the way it works right now is it'll log out everything in the container that's the basis for the queries and in react-testing-library that's set to the documentElement. I think the best solution would probably be to just change that to document.body. I don't think anyone puts portals anywhere other than the document.body so we should be safe to make that change here. Would you like to do that? Thanks @kentcdodds! Submitted a [react-testing-lib PR|https://github.com/kentcdodds/react-testing-library/pull/145].
gharchive/issue
2018-07-24T13:49:30
2025-04-01T04:34:45.214800
{ "authors": [ "kentcdodds", "timbonicus" ], "repo": "kentcdodds/dom-testing-library", "url": "https://github.com/kentcdodds/dom-testing-library/issues/81", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
388472336
feat(queryByCurrentValue) What: It adds queryByCurrentValue, getByCurrentValue, queryAllByCurrentValue and getAllByCurrentValue. Why: getByValue cannot get elements with value property (it only checks DOM value attributes) (Discussed at #166) How: Added a set of (get|query)(All)?ByCurrentValue methods. Checklist: [ ] Documentation [x] Tests [ ] Ready to be merged [ ] Added myself to contributors table Feel free to give any feedbacks. Closes #158 Thanks for the review. I made changes regarding your three comments. Looks good! Just need docs and a test to execute line 358 (the error state for queryAllBy...) to keep coverage at 100%. Also it might be nice to add an integration test for fireEvent.change() with getCurrentValue. @alexkrolick I've added a test to check if it throws. @kentcdodds I've renamed and updates the README.md. I didn't know how to edit it, so it's kind of a draft. Where do you want to put the section in the document? And how do you want to explain it to users? Take a look at the doc and let me know. Besides, I'm not a native English speaker, so pretty much worried about it 😅 @kentcdodds done! I've used container instead of document.body since I found out other parts of the doc used container. Looks like we're missing some coverage. You can open the coverage report in the browser in coverage/lcov-report to see what's missing. @kentcdodds coverage 100% done Updated the doc. Thanks for pointing it out @alexkrolick :tada: This PR is included in version 3.14.0 :tada: The release is available on: npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket: Awesome! Thank you :) This merge made my day!!!
gharchive/pull-request
2018-12-07T01:49:45
2025-04-01T04:34:45.223763
{ "authors": [ "alexkrolick", "eunjae-lee", "kentcdodds" ], "repo": "kentcdodds/dom-testing-library", "url": "https://github.com/kentcdodds/dom-testing-library/pull/169", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
383496938
Change behaviour regarding controlled mode Possible changes regarding - https://github.com/kentcdodds/react-toggled/issues/26 Is there a blocker for this PR? Sorry, this slipped away from me somehow. Also I never even used this project (it was an experiment/demo) so I'm totally the wrong person to maintain it. Anyone want to take it over? At the company I'm working for we use this library and make use of the patterns shown here. I'd be interested in helping to maintain this in case others are in the same boat. Totally, it's all yours @therealparmesh. To transition this over to you, I'm going to remove my npm and github token from travis which will disable the auto-releases. You can re-enable that yourself if you like. Please let me know your npm username and I'll add you as a collaborator there. This may also be helpful to you: https://kcd.im/write-oss Let me know if you have any other questions. @kentcdodds I will do my best to find other quality maintainers and to keep this to the standard that you're setting. My npm username is therealparmesh. Awesome. I've added you to npm and github. I'm going to stop watching the repo now. Good luck! @Andarist Do you still think this PR is worth moving forward on?
gharchive/pull-request
2018-11-22T11:36:52
2025-04-01T04:34:45.251372
{ "authors": [ "Andarist", "kentcdodds", "therealparmesh" ], "repo": "kentcdodds/react-toggled", "url": "https://github.com/kentcdodds/react-toggled/pull/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
118034540
Slide height percentage of parent e.g. 100% Hi, Sorry couldn't find any other issues relating to this. Great plugin BTW. What is the correct way to have 100% height on slides? The parent container has a height that is a relative and the slide is relative to the container e.g. 100% Slick adds many internal elements between the parent and the slides so the slide is no longer relative to the parent. Without hacking a height onto the slide or having knowledge of the internal workings of Slick and the additional containers it produces, what is the best way around this issue? I was having the same issue with slides that had no content, only background images that covered the area - this was to be used as a non-interactive enhancement piece in the background of another slide, but I couldn't get it working with 100% height so had to use another technique. Thanks, Jarrod Closing due to not following CONTRIBUTING.md works great for me with this additional css: .slick-list, .slick-track { height: 100%; } I tried the above solution and didn't work, but this does work! .slick-track { display: flex; } .slick-track .slick-slide { display: flex; height: auto; } I tried the above solution and didn't work, but this does work! .slick-track { display: flex; } .slick-track .slick-slide { display: flex; height: auto; } Wow so glad that I found this! Thanks! .slick-track { display: flex; } .slick-track .slick-slide { display: flex; height: auto; } thanks If anyone is experiencing weird slick height (very tall), the following should work (the other solutions weren't what I was looking for): .slick-slide{ height: auto; } None of the above solutions work for me.
gharchive/issue
2015-11-20T12:45:05
2025-04-01T04:34:45.256638
{ "authors": [ "KhaledMohamedP", "McGern", "MsMatki", "frankstepanski", "kenwheeler", "lunow", "rudestewing", "vasilzhigilei" ], "repo": "kenwheeler/slick", "url": "https://github.com/kenwheeler/slick/issues/1905", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
259191204
slidesToShow, slidesToScroll set to 0 crashes browsers When slidesToShow and / or slidesToScroll is set to 0 or "" the browser crashes. The provided jsfiddle only sets slidesToShow to 0 when breakpoint is below 500. So be sure to resize the window to the size. I did this on purpose so you don't get a crashed browser without seeing the jsfiddle ;) ==================================================================== http://jsfiddle.net/h367Lerw/2/ ==================================================================== Steps to reproduce the problem Open jsfiddle Resize browser window to match the breakpoint of 500 Experience the crash ==================================================================== What is the expected behaviour? The slider should either throw a error, or ignore the settings. ==================================================================== What is observed behaviour? The browser crashes ==================================================================== More Details Which browsers/versions does it happen on? Chrome 60 Chrome 61 Safari 11.0 Firefox 55 Firefox 56 Which jQuery/Slick version are you using? jQuery (edge) Slick (master) Did this work before? Don't know Related Task: #3016 I'm getting the same issue!
gharchive/issue
2017-09-20T14:47:37
2025-04-01T04:34:45.263869
{ "authors": [ "elz1", "simoozzay" ], "repo": "kenwheeler/slick", "url": "https://github.com/kenwheeler/slick/issues/3133", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
400976580
Saving/Loading/resuming agent state and weights Hello, I noticed that loading saved weights results in test of untrained agent (using the breakoutagent example). I read that saving just the weights is not enough since the agent/model is not optimized. Is there a solution? Is using the keras model.save/load_model functions solve this? Ref: https://keras.io/getting-started/faq/#how-can-i-use-pre-trained-models-in-keras Hey jheffez, ¿How did you solve this issue? I have trained an agent in a custom environment and after this process, I would want to use the trained and saved weights to restore the agent (I think that load_weights should be the way) and be able to deploy it into production; passing him the observations and getting the actions. ¿Is this doable with keras-rl actually? I recall that this could be done with keras as your post (using the predict methods available and feeding the x parameter). However, I don't see this possibility in the keras-rl docs. Any help would be appreciated. Thanks in advance, Would like help with the same, receiving an action prediction for a single observation instance in production setting. Can this be done with Keras-RL?
gharchive/issue
2019-01-19T08:57:36
2025-04-01T04:34:45.281521
{ "authors": [ "Eriz11", "jarlva", "krumiaa" ], "repo": "keras-rl/keras-rl", "url": "https://github.com/keras-rl/keras-rl/issues/288", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
205530388
Keras CTC Model Save Discrepency I am trying to implement a simple BLSTM-CTC model using Keras (backend Tensorflow). I am testing this with a small dataset of online handwriting samples (316 Train data with 10 distinct characters and 4 words). each of these samples have 401 timesteps and at each timestep I have 16 features. So the input is a Numpy array of dimension [316,401,16]. My network is implemented successfully as suggested by this example. My code is as follows def ctc_lambda_func(self,args): y_pred, labels, input_length, label_length = args # the 2 is critical here since the first couple outputs of the RNN # tend to be garbage: y_pred = y_pred[:, 2:, :] return K.ctc_batch_cost(labels, y_pred, input_length, label_length) self.inputlayer=Input(name='input', shape=[timesteps,features], dtype='float32') self.labels = Input(name='the_labels', shape=[maxstringlen], dtype='float32') self.input_length = Input(name='input_length', shape=[1], dtype='int64') self.label_length = Input(name='label_length', shape=[1], dtype='int64') self.lstm_1 = LSTM(rnn_size, return_sequences=True, init='he_normal', name='LSTM1')(self.inputlayer) self.lstm_1b = LSTM(rnn_size, return_sequences=True, go_backwards=True, init='he_normal', name='LSTM1_b')(self.inputlayer) self.gru1_merged = merge([self.lstm_1, self.lstm_1b], mode='sum') self.out=TimeDistributed(Dense(nbclasses,name="dense2",activation="softmax"))(self.gru1_merged) self.loss_out = Lambda(self.ctc_lambda_func, output_shape=(1,), name='ctc')([self.out, self.labels, self.input_length, self.label_length]) self.optimizer = RMSprop(lr=0.001, rho=0.9, epsilon=1e-08, decay=0.0) self.final=Model(input=[self.inputlayer,self.labels,self.input_length,self.label_length],output=self.loss_out) self.final.compile(loss={'ctc': lambda y_true, out: out}, optimizer=self.optimizer) This network is compiled successfully. Now I am running it and saving with for e in range(nbepochs): self.final.fit(x,y,batch_size=64,nb_epoch=1,verbose=1) self.final.save_weights("weights.h5") #Loading with the following lines after creating the network again self.final.load_weights("weights.h5") When I am running the training everything is working fine. The CTC error is reducing as expected. But when I am trying to load the model from a previous state then it is not restoring from the last saved state. Say I have executed 10 epochs Epoch 1/1 316/316 [==============================] - 2s - loss: 11.5716 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 11.2032 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 11.1163 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 10.9920 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 10.9337 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 10.8708 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 10.7654 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 10.7335 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 10.6119 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 10.4657 Model saved But when I am loading it again Epoch 1/1 316/316 [==============================] - 2s - loss: 12.5233 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 11.8969 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 11.7075 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 11.5625 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 11.4145 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 11.2987 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 11.1833 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 11.0691 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 10.9600 Model saved Epoch 1/1 316/316 [==============================] - 1s - loss: 10.8676 Model saved Clearly there is something wrong as the loaded model is not starting from an error around 10.4657. I also tried to save the whole model with save() and load_model() which gave me an error "KeyError: CTC Lambda Func not found". I am totally in dark. Is the Lmbda layer creating any problem? Please help if possible. Thank you for your time. Is it possible that you have to set the start epoch to prevent it from starting from scratch? @Cerno-b what do you mean ? I do not understand I was thinking it could be related to this question: https://stackoverflow.com/questions/52476191/what-does-initial-epoch-in-keras-mean/52478034
gharchive/issue
2017-02-06T09:39:53
2025-04-01T04:34:45.294614
{ "authors": [ "Cerno-b", "ibrahimabarry01", "xisnu" ], "repo": "keras-team/keras", "url": "https://github.com/keras-team/keras/issues/5286", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
607143602
Fix incorrect loss_weights error message, clarify docs Summary loss_weights argument to model.fit only accepts python lists, so no numpy arrays. clarify docs for that. Also the error message was incorrect, asking for a "list of dicts" instead of a "list or dict" as the docs ask for. PR Overview No special requirements, minor change to docs, error message n This PR requires new unit tests [y/n] (make sure tests are included) ? This PR requires to update the documentation [y/n] (make sure the docs are up-to-date) n This PR is backwards compatible [y/n] n This PR changes the current API [y/n] (all API changes need to be approved by fchollet) Edit: Force-pushed to reflow text to fit in 85 cols "list OR dict" is very different from "list OF dictS" It's not obvious at all given that most methods take numpy "lists" (arrays). Superseded by https://github.com/tensorflow/tensorflow/pull/41859
gharchive/pull-request
2020-04-26T23:39:30
2025-04-01T04:34:45.298798
{ "authors": [ "MingweiSamuel" ], "repo": "keras-team/keras", "url": "https://github.com/keras-team/keras/pull/14008", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1945836278
Add named_product test utility to generate test cases with names. This is the be used with absl.parameterized.named_parameters and instead of absl.parameterized.product. It creates testcases that have intuitive names. Tests in numpy_test.py and dtypes_test.py that were using product now use named_product. Also standardized the way we create KerasTensors in numpy_test.py for readability. Thanks for the PR! Thanks for the review, François! Two comments about this PR: The named_product thing is purely cosmetic. When a test fail, you do see the parameters used. However, when you have a high number of combination it's nice to immediately see the failing combinations without having to drill into each test. (For instance, you can see that all "int" types are failing). I put this PR up separately but my goal is to use it with upcoming sparse ops tests. Codecov Report All modified lines are covered by tests :white_check_mark: Comparison is base (36322e7) 57.96% compared to head (c852228) 57.99%. Additional details and impacted files @@ Coverage Diff @@ ## master #18626 +/- ## ========================================== + Coverage 57.96% 57.99% +0.02% ========================================== Files 334 334 Lines 32504 32524 +20 Branches 6348 6353 +5 ========================================== + Hits 18841 18861 +20 Misses 12231 12231 Partials 1432 1432 Flag Coverage Δ keras 57.99% <100.00%> (+0.02%) :arrow_up: keras-numpy 57.99% <100.00%> (+0.02%) :arrow_up: Flags with carried forward coverage won't be shown. Click here to find out more. Files Coverage Δ keras/testing/test_utils.py 100.00% <100.00%> (ø) :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
gharchive/pull-request
2023-10-16T18:23:48
2025-04-01T04:34:45.309718
{ "authors": [ "codecov-commenter", "hertschuh" ], "repo": "keras-team/keras", "url": "https://github.com/keras-team/keras/pull/18626", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1657915366
Removing footnote stripping/tagging footnote origin points What is your OS and architecture? Windows is not supported and Mac OS arm64 is not yet supported. For non-supported OS, you can use Docker (https://grobid.readthedocs.io/en/latest/Grobid-docker/) Ubuntu 22.04.2 What is your Java version (java --version)? Expected versions are currently Java 8, 9 and 10. JDK 11.0.18 Hey! First off, I'd like to thank you for creating this tool. After testing a number of tools to markup and extract text from pdfs, I can quite confidently say that yours is the best for both usability and accuracy - it's helped me with my current project immensely, so thanks again. For context, my current project relates to Australian legal cases and academic texts in the legal field - I have been working on code which extracts the text of these documents and converts them to an audio file using ESPNet. GROBID has been doing the heavy lifting for my text extraction, and on the whole works beautifully, but it is somewhat overzealous on footnote classification for legal documents. Occasionally, it will strip paragraphs of the body of the case out as footnotes, moving them to the footer. I was wondering if there is a way to stop GROBID from stripping footnotes altogether - I have other ways to process them if necessary. I have tried implementing custom TEI schemas, but can only succeed in moving the location of the footnotes, not keeping them in the body/their original position. Alternatively, if there is a way to tag the original position of the footnotes, I could reinsert them by post-processing the .xml. Thank you in advance for any help you may be able to provide, and thanks again for creating this tool. Hi @TheDeponent ! Thank you very much for the kind words on grobid and happy that it can be useful for your project. The legal field is still challenging for Grobid, not really represented in the training data for the moment. But even in general, there's this issue of paragraph incorrectly labeled as footnote, which is an error very visible and not really possible to recover with the final XML only. Unfortunately at this time, the addition of coordinates does not include the element note, I need to add it - I note it for the next release. If you're courageous, one "hacking" solution is to retrain a segmentation model without the "foot note" structure, I think it should work: modify the segmentation training data XML parser to consider <note place="footnote"> tag as <body>: Change this: https://github.com/kermitt2/grobid/blob/master/grobid-trainer/src/main/java/org/grobid/trainer/sax/TEISegmentationSaxParser.java#L157 the line 157 into these 3 lines: currentTag = "<body>"; upperTag = currentTag; upperQname = "body"; retrain the segmentation model (this command will also recompile for the above change): > ./gradlew train_segmentation After some hours (it depends on the number of CPU of your machine - usually half day), you get a new segmentation model and there should not be any foot note section anymore, they will be included as paragraph in the normal body text. @kermitt2 Thanks so much for your response. Even with the lack of legal documents in the training data, the error rate is still quite low, on average in my testing it misidentifies 5 paragraphs in a legal document containing 200 numbered paragraphs. I will attempt retraining the segmentation model over the next few days, and will
gharchive/issue
2023-04-06T19:10:13
2025-04-01T04:34:45.322435
{ "authors": [ "TheDeponent", "kermitt2" ], "repo": "kermitt2/grobid", "url": "https://github.com/kermitt2/grobid/issues/996", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
233018412
Using a method instead of a template This is really just for discussion. In the CMS for https://www.raise.me I actually don't use templates at all with the Shortcode gem. I set the template for every shortcode to '<%= @content %> and generate a presenter that returns the content I want. I actually have a presenter factory class that take in a name and a block and uses that block to generate the content. Here is that class: https://gist.github.com/toddnestor/d6c94655e98ce5fbe0e65843262b0b58 I want to discuss if there is a good way to incorporate something like this as a feature in the shortcode gem. I'd guess there are other people who need to be able to basically tie a method to a shortcode in this way like I do, but maybe not. I have no problem if this doesn't become a feature, just want to open up a discussion about it. So any thoughts about something like this that allows just tying a block of code to a shortcode instead of using a template? Obviously my presenter factory class was a workaround, it seems to me that within this plugin there could probably be a more direct way to tie a shortcode to a method if we wanted to incorporate that feature. Perhaps the shortcode arrays could be refactored to be hashes with the names as the keys and the value being a hash of options, one of those options could be to use a template or a method, or something along those lines. I'll have to think about it some more.
gharchive/issue
2017-06-01T21:38:55
2025-04-01T04:34:45.341956
{ "authors": [ "toddnestor" ], "repo": "kernow/shortcode", "url": "https://github.com/kernow/shortcode/issues/50", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2370779404
🛑 FX blog (kty) is down In f14038a, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in 027da02 after 6 minutes.
gharchive/issue
2024-06-24T17:46:12
2025-04-01T04:34:45.498021
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/10038", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2400258179
🛑 FX blog (kty) is down In 5d0298f, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in f56a6af after 6 minutes.
gharchive/issue
2024-07-10T09:42:09
2025-04-01T04:34:45.500586
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/14338", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2417367110
🛑 FX blog (kty) is down In 59f3dd7, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in b471263 after 5 minutes.
gharchive/issue
2024-07-18T20:46:42
2025-04-01T04:34:45.502894
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/16531", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2439864220
🛑 FX corretor (kty) is down In 7a638b7, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in 32cff21 after 6 minutes.
gharchive/issue
2024-07-31T11:54:45
2025-04-01T04:34:45.505228
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/19519", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2445891033
🛑 FX blog (kty) is down In b6a7bf1, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in c51a9ab after 6 minutes.
gharchive/issue
2024-08-02T23:51:44
2025-04-01T04:34:45.507536
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/20075", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2456099944
🛑 FX corretor (kty) is down In eddf782, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in d0d1d4f after 6 minutes.
gharchive/issue
2024-08-08T15:31:38
2025-04-01T04:34:45.509855
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/21309", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2479622500
🛑 FX blog (kty) is down In 48cbd6b, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in 887a5c3 after 17 minutes.
gharchive/issue
2024-08-22T02:13:56
2025-04-01T04:34:45.512336
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/24234", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2530429310
🛑 FX corretor (kty) is down In 7e5f418, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in 6406271 after 8 minutes.
gharchive/issue
2024-09-17T08:34:59
2025-04-01T04:34:45.514658
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/29552", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1659798684
🛑 FX net (kty) is down In 174deb1, FX net (kty) (https://foxter.net.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX net (kty) is back up in 05a8cf9.
gharchive/issue
2023-04-09T08:38:08
2025-04-01T04:34:45.516919
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/3355", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2589609327
🛑 FX blog (kty) is down In 5c10683, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in e96950f after 13 minutes.
gharchive/issue
2024-10-15T18:54:34
2025-04-01T04:34:45.519223
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/34536", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2592687315
🛑 FX blog (kty) is down In 3c3630f, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in 9f8b2b6 after 21 minutes.
gharchive/issue
2024-10-16T18:16:55
2025-04-01T04:34:45.521673
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/34689", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2609102013
🛑 FX blog (kty) is down In b970882, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in b86432b after 6 minutes.
gharchive/issue
2024-10-23T15:54:58
2025-04-01T04:34:45.524183
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/35937", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2673931542
🛑 FX blog (kty) is down In 82039ac, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down: HTTP code: 0 Response time: 0 ms Resolved: FX blog (kty) is back up in 9154fba after 6 minutes.
gharchive/issue
2024-11-19T23:50:44
2025-04-01T04:34:45.526456
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/41002", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2719086746
🛑 FX corretor (kty) is down In e562732, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in e9c0378 after 36 minutes.
gharchive/issue
2024-12-05T01:04:25
2025-04-01T04:34:45.528775
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/43780", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1670020653
🛑 FX net (kty) is down In d7277fa, FX net (kty) (https://foxter.net.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX net (kty) is back up in e06b15f.
gharchive/issue
2023-04-16T15:38:33
2025-04-01T04:34:45.530999
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/4387", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2728935951
🛑 FX corretor (kty) is down In bbecc9c, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX corretor (kty) is back up in 2ffee96 after 6 minutes.
gharchive/issue
2024-12-10T03:59:39
2025-04-01T04:34:45.533320
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/44682", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1672016911
🛑 FX net (kty) is down In 3c9ae63, FX net (kty) (https://foxter.net.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX net (kty) is back up in 5104eec.
gharchive/issue
2023-04-17T21:48:53
2025-04-01T04:34:45.535710
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/4556", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1676671964
🛑 FX net (kty) is down In 05806ca, FX net (kty) (https://foxter.net.br) was down: HTTP code: 0 Response time: 0 ms Resolved: FX net (kty) is back up in 702d3b8.
gharchive/issue
2023-04-20T12:57:25
2025-04-01T04:34:45.537956
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/4906", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1950272891
🛑 FX kty hom (kty) is down In af8179e, FX kty hom (kty) (https://foxter-hom.konecty.com) was down: HTTP code: 503 Response time: 307 ms Resolved: FX kty hom (kty) is back up in 99a6253 after 6 minutes.
gharchive/issue
2023-10-18T17:20:34
2025-04-01T04:34:45.540248
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/6437", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2225440028
🛑 FX portal api (kty) is down In 80f3477, FX portal api (kty) (https://api.foxterciaimobiliaria.com.br) was down: HTTP code: 404 Response time: 222 ms Resolved: FX portal api (kty) is back up in 2d87875 after 1 hour, 11 minutes.
gharchive/issue
2024-04-04T12:59:45
2025-04-01T04:34:45.542572
{ "authors": [ "keviocastro" ], "repo": "keviocastro/upptime", "url": "https://github.com/keviocastro/upptime/issues/8017", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1093825933
Add xwx.moe Important: Please read all instructions carefully. Select the appropriate category for what this PR is about This PR is: [x] Adding a new domain [ ] Updating existing domain size [ ] Changing domain name [ ] Removing existing domain from list [ ] Website code changes (512kb.club site) [ ] Other not listed Do not tick a checkbox if you haven’t performed its action. Honesty is indispensable for a smooth review process. [x] I used the uncompressed size of the site [x] I have included a link to the GTMetrix report [x] The domain is in the correct alphabetical order [x] This site is not a ultra lightweight site [x] The following information is filled identical to the data file - domain: xwx.moe url: https://xwx.moe/ size: 324 last_checked: 2022-01-04 GTMetrix Report (English homepage): https://gtmetrix.com/reports/xwx.moe/8ZF0k2IG/ GTMetrix Report (Esperanto homepage): https://gtmetrix.com/reports/xwx.moe/eKSG8HZY/ GTMetrix Report (Language selection): https://gtmetrix.com/reports/xwx.moe/FXNrknyQ/ I used the English homepage's (https://xwx.moe/en) size for the entry, since it's the real content, and is larger than the "language select" page (https://xwx.moe/). @kevquirk Not sure how to do this one.
gharchive/pull-request
2022-01-04T23:10:21
2025-04-01T04:34:45.549349
{ "authors": [ "JadedCtrl", "kmutahar" ], "repo": "kevquirk/512kb.club", "url": "https://github.com/kevquirk/512kb.club/pull/620", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
328235114
Can't access my public files on Linux or keybase.pub I can't access my public files on Linux or keybase.pub, which has brought down my site smiller.keybase.pub. I can access files on my Mac, but on Linux I see: ➜ ~ keybase fs ls /keybase/public/smiller171 ▶ ERROR Error checking merkle tree: tried 100 roots, but seqno not found Just submitted logs which also had errors. ▶ INFO ignoring UI logs: context deadline exceeded Hint: You are currently not seeing messages from other users and the system. Users in the 'systemd-journal' group can see all messages. Pass -q to turn off this notice. No journal files were opened due to insufficient permissions. ▶ ERROR Error reading logs: Journalctl exited with an error: exit status 1 ------------ Success! Your log ID is: 08d5535fd3033f797d59461c Here's a URL to submit new bug reports containing this ID: https://github.com/keybase/client/issues/new?body=[write%20something%20useful%20and%20descriptive%20here]%0A%0Amy%20log%20id:%2008d5535fd3033f797d59461c Thanks! ------------ I'm submitting from a Chromebook using Crostini, but since I can't access from keybase.pub either the issue isn't Chromebook specific. (I didn't log in on the Chomebook until after the issue appeared) Thanks @smiller171. I see the bug, it'll take us a little while to fix it. We'll follow up with you. Sorry for the inconvenience! @smiller171 we are already working on a fix. cc @maxtaco. BTW, we have two fixes coming out for this problem, the first fix might solve it, which should be out sooner. Thank you. keybase/kbfs#1597 should fix this once it's merged. (Though it won't fix the underlying issue, which will still happen if the last write to a folder happens to have been from a revoked device that got unlucky w.r.t. the global merkle tree. @maxtaco is going to work on that fix soon.) @strib sounds like currently the answer to avoid this is to write to the directory from a different device before revoking a device? If I make a new write to the folder from my Mac will that restore access? @strib sounds like currently the answer to avoid this is to write to the directory from a different device before revoking a device? If I make a new write to the folder from my Mac will that restore access? The current problem is because one of the devices you revoked in February got unlucky, and the revoke operation itself got stuck in a long line of server operations. (We expect there to be only one or two server operations in between when you click "revoke" and the server confirms it, but in this case there were more than 300.) However, this was not the device that last wrote to your public folder, that was the other device you revoked in February. Usually Keybase clients only need to run the (currently buggy) verification process for the device that actually did the last write, but due to a different bug, it was doing them on all revoked devices, and hitting a problem with the key that didn't do the write. So in short, no, writing from your Mac won't help current Linux clients, because they will still be verifying all your keys unnecessarily. keybase/kbfs#1597 stops them from verifying all keys, which will fix the issue as soon as I get it reviewed by someone and merged. If the folder happened to have been written last by the problematic key, then yes writing to it from a different device before revoking would fix it. But you shouldn't have to change your behavior to adapt to our bugs, so I would just go about business as usual, and we'll fix the bugs ASAP. Is there any reason that keybase.pub should have only just now run into this issue when the revocations were a while ago? Was one of these bugs only recently introduced? Also, my current behavior is to revoke devices after I reformat/wipe them, Is it better to just leave them as they are since the private key data can't be retrieved anyway? Is there any reason that keybase.pub should have only just now run into this issue when the revocations were a while ago? Was one of these bugs only recently introduced? Yes, the bugs were introduced by us this week. We are tightening checks on data written by revoked keys, and there are a few corner cases, like this one, that we messed up. Also, my current behavior is to revoke devices after I reformat/wipe them, Is it better to just leave them as they are since the private key data can't be retrieved anyway? I think it's always better to revoke devices that won't get used again. There's a new linux build out with a fix, and https://smiller171.keybase.pub/ works again. Please close this out if everything looks ok on your end. Thanks, and sorry for the inconvenience. Thanks for getting a fix out so fast! I'll take a look shortly
gharchive/issue
2018-05-31T17:47:01
2025-04-01T04:34:45.569026
{ "authors": [ "maxtaco", "smiller171", "strib" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/12134", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
131738838
I/O error on one particular user's public listing When i do an: /keybase/public $ ls USERNAME for the USERNAME mentioned in the logs, I get an I/O error. my log id: cf323fd57709d213cce9541c The user's key was reset, so doing: keybase track USERNAME was able to fix it. Great, glad it works now!
gharchive/issue
2016-02-05T19:50:13
2025-04-01T04:34:45.571279
{ "authors": [ "strib", "xxv" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/1912", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
132851324
Dragging a file into /keybase/private/ on OS X Finder gives an "Newer version of file already exists in this location" error. Dragging a file into /keybase/private/ or /keybase/public/ with Finder gives the error "A newer version of this file already exists". Pressing "Stop" or "Replace" dismisses the dialog, but no file remains. Screencast: https://keybase.pub/rodbegbie/keybase_bugreport_screens/copy_to_private.gif my log id: 89d93dbb2c0d802d9e725b1c Aside, after being thrown by this a few times, I realized the problem was that /keybase/private/rodbegbie/ wasn't mounted and that's where I was supposed to be dragging files -- but the error remains that the dialog is confusing and unhelpful. Thanks for the report! Is this on OSX Yosemite or El Cap? I can't seem to reproduce it on Yosemite. From the log, it doesn't look like KBFS is returning that error code. I think what's happening is this: Entries under public/ or private/ correspond to (combinations of) normalized Keybase usernames. Normalized Keybase username are all lower case. To let people use whatever case they want, we fake out symlinks when someone gives us a non-normalized name, that points to the normalized version. So in this case, we are pretending there is a walter.JPG symlink, pointing to walter.jpg. When the OS tries to write to walter.jpg (according to your log), we reject it with an I/O error. The Finder must be choosing to ignore that error, and instead tries to overwrite the fake symlink we made for walter.JPG, which looks like it already exists. I've seen El Cap do weird stuff like this before. I agree this is confusing. I'll try to figure out a way around it. But I'm curious what happens if you try to drag in a file which has an all-lower-case name. I'm guessing you'll get an IO error. ^ (FYI, an IO error will show up as a mysterious error code 36.) You are correct! The Finder can’t complete the operation because some data in “walter.jpg” can’t be read or written. (Error code -36) Ok cool, thanks. Maybe we can head off invalid file names earlier, and avoid showing symlinks for them. I'll file an issue for it on our internal tracker. Also note that in upcoming versions of KBFS, you'll get nicer error pop-ups with better error messages than the OS lets us have, when you hit situations like this.
gharchive/issue
2016-02-10T23:46:57
2025-04-01T04:34:45.576564
{ "authors": [ "rodbegbie", "strib" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/1980", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
638931561
Error Code 202 / BAD SESSION? Cannot login on iOS or keybase.io Attempts to login to Keybase via iOS app leads to the following error: Also, I cannot login via keybase.io; that error shows as: Edit: also posted to keybase-issues I it is the same issue with timeout to SSL handshake I it is the same issue with timeout to SSL handshake Do you have a sense of the cause of the timeout? I haven't seen this problem reported by anyone else but I still cannot access my account. @njm2112 I have the timeout exp for current ios apps. Here is related ticket and comment that client suppose to have a new version to fix the problem : https://github.com/keybase/keybase-issues/issues/3917#issuecomment-646724838
gharchive/issue
2020-06-15T15:11:47
2025-04-01T04:34:45.580793
{ "authors": [ "miry", "njm2112" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/24224", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1507447631
KBFS doesn't work on macOS X 11 Big Sur (still works with <=6.0.1) (Reporting this separately as to point out that latest release for newer macOS broke the functionality for older OS versions) Encountering already known issue with Keybase startup where KBFS fails to install. Used to work on Intel 16inch 2019 MacBookPro (Big Sur 11.6), but latest official release will error out (which is 6.0.3-20221212203703+608e46df72 at the time of reporting this). Some good news: was able to restore full functionality by downgrading to 6.0.1-20220606162955+0ee8d3021e (a version that I could still acquire from the official site via https://prerelease.keybase.io/darwin-test/index.html). Relates to: https://github.com/keybase/client/issues/25226 (where macOS X 13 Ventura users claim that latest release fixed their issue) What was tried to get 6.0.3 working (without any luck): The machine was restarted (as prompted) kbfs was uninstalled, Keybase restarted and kbfs reinstalled My log id: 576775d4181cb5637d7b851c This is essentially what I've tried to document on https://github.com/keybase/client/issues/24366#issuecomment-1361858895. Briefly — because the Keybase app for macOS includes its own bundled version of macfuse, requiring a kernel extensionhat is highly specific to different macOS versions, it's not reasonable to expect that subsequent versions of the Keybase app will alternatively break support of either older or newer versions of macOS — depending on how exactly it was compiled. FYI, the plain, unpatched version of macfuse supports all macOS versions from 10.9 to 13. The problem is not really with macfuse; it's with the way specific versions of macfuse are bundled with the Keybase app installer. These require manual intervention by a maintainer with the correct permissions (namely, access to the Keybase signing key for distributing macOS apps...) and are thus 'unfixable' by the public-at-large. Also note that the major difference between 6.0.1 and 6.0.3 is the bumped version of macfuse, in order to support not only macOS Ventura, but better support for Apple Silicon chips. There doesn't seem to be any fundamental, game-breaking changes otherwise, as far as I can see (and understand) the code. I actually messed up the report and it seems that the breaking issue was introduced when going from 6.0.0 to 6.0.1 :| Update the title and the description to reflect that as well. The takeaway from all this is probably that when a breaking change for older OS versions is introduced, the "Your Keybase app is out of dated, please update now" message should not be presented to the users of said systems. Another thing this would be to list older releases on the site (and maybe segregate them based on OS versions). Good News YES ! @allanpaiste just like you said downgrading to to 6.0.1-20220606201228+4a1945f0a3 got me working on macOS 11.74 kbfs file integration yay!
gharchive/issue
2022-12-22T08:20:53
2025-04-01T04:34:45.589256
{ "authors": [ "GwynethLlewelyn", "allanpaiste", "paulhugel" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/25379", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2181375265
arm64 linux build? hey, why you aren't shipping, .deb, .rpm, and other packages for arm64? with asahi linux getting traction, more and more people will use that. I am trying to build it at my machine, I was able to build it (cli and a desktop app in electron that gave a lot of alarms at selinux), will try to discover all the steps and share here just in case someone want to do it.
gharchive/issue
2024-03-12T11:24:01
2025-04-01T04:34:45.590656
{ "authors": [ "patricuspl", "pulpo" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/26328", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
209296007
Write lock conflict Log id: d58f77ad249bcad098ab411c Details on what happens when attempting to delete a file while transferring a folder on OS X. Operation eventually finished without additional user interaction. Thanks for the report. Hmm, that's an odd error message, it doesn't look like it's coming from KBFS. What programs were you using for transfer, and what showed you that error message? I do see some timeout errors in your keybase log, but those are expected in the current release under some situations (we should be fixing them soon). I don't know how they'd end up showing you a lock error though. This was a drag-and-drop via the Finder for the transfer, and a Cmd-Del in the Finder folder for the delete. The message I believe was via the normal thing-used-to-be-known-as-a-growl notification. Attempts to replicate a couple times so far have failed (Log ID 5f306ec74e18857120bacb1c): Drag and drop large archive Delete file from active transfer folder Works shrug On 2/21/17 4:52 PM, Jeremy Stribling wrote: Thanks for the report. Hmm, that's an odd error message, it doesn't look like it's coming from KBFS. What programs were you using for transfer, and what showed you that error message? I do see some timeout errors in your keybase log, but those are expected in the current release under some situations (we should be fixing them soon). I don't know how they'd end up showing you a lock error though. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/keybase/client/issues/5969#issuecomment-281508811, or mute the thread https://github.com/notifications/unsubscribe-auth/AB4qxiMZT9y-0p0oOrLqRbVuGt8v27-jks5re2qngaJpZM4MH9pv. Closing due to non-reproducibility.
gharchive/issue
2017-02-21T22:41:57
2025-04-01T04:34:45.596382
{ "authors": [ "itys", "strib" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/5969", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
246588862
How to remove the K disk KBFS (K:) 10.0 GB You can uninstall Dokan in the control panel.
gharchive/issue
2017-07-30T14:58:13
2025-04-01T04:34:45.597306
{ "authors": [ "liyuhang1997", "zanderz" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/7946", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
269938095
KBFS Keeps Requesting I Rekey I have a number of devices connected to my account and I am asked on a regular basis to rekey using my paperkey on each device. I have done this but I keep getting asked to rekey. Not sure why the rekey doesn't appear to be working. Any assistance would be much appreciated! Thanks. my log id: e1f5f13a80be26bd9ccf671c Hi @jeffbristow. It looks like the folder in question can only be rekeyed by your devices Home Server and Macbook Air. Can you log into Keybase from one of those devices? We're asking for your paper key out of desperation I guess, but we should be more intelligent about that. Hmm, ok then I'm confused about what's going on, it seems our server-side information might have gotten a little out-of-sync somehow. Can you try writing a small file to the folder that's badged as needing a rekey? Oh, actually I think I see now. The confusion is over a shared folder with a user who reset their account before we handled resets automatically. I'll mark that folder as ignorable and you shouldn't see the badge anymore. I was able to make a text file named sample.txt in folder "private/user1,myuser" on Home Server. I also was able to access the file from a computer I added this morning called DellUltrabook and there were no issues, other than keybase shows that this folder needs to be rekeyed. (I can share the actual folder name through email). Will this also end the prompts to rekey? I hope so. It seems like your device did do a fair amount of rekeying today on a bunch of folders, so hopefully things are stable besides this one folder. If this doesn't fix it, we'll take another look at the prompt stuff. (This is a server-side fix, and will need to wait for the next server deploy, not sure when that will be.) Thanks, sounds good. Same issue here; log ID 5d0132248073eb85c001b51c $ keybase rekey status TLFs need rekeying: private/<snip-1>,johanstokking private/<snip-2>,johanstokking private/johanstokking Devices that can rekey: desktop Mac mobile iPhone 7 I'm on Mac but no luck getting rid of the rekey message. Does it take time or do you need to do anything? I'm on Linux, Ubuntu 16.04, Keybase version 1.0.36-20171130235940+7d909be3c. I launched the Desktop application and saw the red dot, indicating something needed my attention. I went to the private folder, and tried to rekey it with my paper secret. In my case, the paper secret did nothing, nor was there any validation saying what happened. It just left me with the need to rekey it again. I repeated this a few times, restarted the application, etc, and still did not work. I revoked my paper secret and tried to rekey with a new one and it still didn't work. Eventually I logged in to my account in a browser to make sure my passphrase to login still worked. It did... I restarted the desktop application once again, and the message to need to rekey the private folder went away, but now there is still a red dot, indicating something needs my attention. It feels like there is a synchronization issue, and it's a little bit frustrating that the application didn't really indicate anything, eg, no real validation was working. I deduced all my answers from just fiddling... Obviously this is not good. Two things; I think we have a bug that slows down rekeying (cc: @songgao) and also we have a bug with badging (cc: @mmaxim) probably related to chat. We can look into the former if key keybase log send, and the latter, we're already looking into. @nicorellius I can confirm it took about 50 minutes to rekey your private folder after you added the lion_2017-12-03 device. It seems like there was an issue on the robot2 device you used to provision it. Can you please keybase log send from robot2 so we track it down? Also, a keybase log send from lion_2017-12-03 would be useful as well. Thanks! @strib Thanks for the feedback. Yes, that sounds about right. I just sent logs from robot2 and lion_2017-12-03. Let me know if you need anything else. Thanks. Thanks, looking at the logs now. robot2 logs: 0982f731822d3f6e928f0c1c lion_2017-12-03 logs: 5689aa35103f4f0fc666961c It seems like before the provision, robot2 lost connectivity to our servers (maybe due to some transient DNS thing), and then got into a situation where it was backing off reconnect attempts: 2017-12-02T23:12:23.787200-08:00 ▶ [DEBU kbfs keybase_service_base.go:352] 12a MDServerRemote: CheckReachability(): failed to connect, reconnecting: dial tcp: lookup mdserver-1.kbfs.keybaseapi.com on 127.0.1.1:53: server misbehaving ... 2017-12-02T23:12:23.787568-08:00 ▶ [DEBU kbfs connection.go:619] 12e (CONN MDServerRemote fb905c5e) starting random backoff: 57m46.90297598s Then the computer went to sleep for many hours, and when it came up I think it was still in this backoff period (which may be triggered based on the waking time, rather than the wall clock time?): 2017-12-03T07:23:15.016555-08:00 ▶ [DEBU kbfs notify_keyfamily.go:34] 148 Key family for user 27392ecdbf4bf98b78ef248b9bdc5c19 changed 2017-12-03T07:23:19.104299-08:00 ▶ [INFO kbfs asm_amd64.s:2337] 149 CheckForRekeys: checking for rekeys @songgao: perhaps a CheckForRekeys call should use rpc.FireNow()? How is it possible the backoff is ever 57 minutes? Why not cap it at something way lower? @mmaxim: is it to avoid a thundering herd when mdservers restart. But any active mdserver request is supposed to break through the backoff, we're just not doing it for rekey requests right now. But 57 minutes just seems way too high to solve that particular problem. It seems like something lower could still solve that problem, and lower the bar for missing this break through the backoff deal. @songgao: perhaps a CheckForRekeys call should use rpc.FireNow()? Good point; I'll put up a PR.
gharchive/issue
2017-10-31T12:20:45
2025-04-01T04:34:45.609241
{ "authors": [ "jeffbristow", "johanstokking", "maxtaco", "mmaxim", "nicorellius", "songgao", "strib" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/issues/9312", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
465944182
Airdrop profile star popup @keybase/picnicsquad CC @keybase/react-hackers @keybase/design Here's the overlay for the airdrop star on the profile page: Kb.FloatingMenu didn't have support for setting a textColor or backgroundColor, so those new props are added. Dividers are missing on the mobile action sheet. Otherwise LGTM! Design: @keybase/picnicsquad @keybase/react-hackers bumpity bump
gharchive/pull-request
2019-07-09T19:18:47
2025-04-01T04:34:45.612427
{ "authors": [ "cecileboucheron", "cjb" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/pull/18312", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
164889219
gui quit on service quit fixes flow issues and listens to the new rpc @keybase/react-hackers :shipit: So this might not do anything for macOS cause we terminate the app process from ctl stop (see https://github.com/keybase/client/blob/master/go/client/cmd_ctl_stop_osx.go#L77). But this is good for linux/windows, so we might want to add a comment in here to that effect.
gharchive/pull-request
2016-07-11T17:19:12
2025-04-01T04:34:45.614349
{ "authors": [ "MarcoPolo", "chrisnojima", "gabriel" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/pull/3410", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
186703106
Zanderz/core 4038 I tested it, it works. If you invoke the installer on the command line and add nodriver=1, it will skip the drivers. Our updater may still be irritating for driver-switching people but that will be another ticket. @maxtaco @oconnor663 LGTM
gharchive/pull-request
2016-11-02T03:21:27
2025-04-01T04:34:45.615565
{ "authors": [ "oconnor663", "zanderz" ], "repo": "keybase/client", "url": "https://github.com/keybase/client/pull/4770", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
341815432
Use correct field separator in keywords.txt The Arduino IDE requires the use of a single true tab separator between the keyword name and identifier. When spaces are used rather than a true tab the keyword is not highlighted. Reference: https://github.com/arduino/Arduino/wiki/Arduino-IDE-1.5:-Library-specification#keywords Thanks!
gharchive/pull-request
2018-07-17T08:25:24
2025-04-01T04:34:45.629292
{ "authors": [ "obra", "per1234" ], "repo": "keyboardio/KeyboardioScanner", "url": "https://github.com/keyboardio/KeyboardioScanner/pull/15", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
182440594
Strange behavior on Windows 8 Here is my definition: module EXP-SYNTAX syntax Ids ::= List{Id, ";"} endmodule module EXP imports EXP-SYNTAX endmodule Compiling works, but the following test fails. Here is the content I'm trying to parse: a, b Here is my output: >kast text.exp [Error] Internal: Uncaught exception thrown of type NoClassDefFoundError. Please rerun your program with the --debug flag to generate a stack trace, and file a bug report at https://github.com/kframework/k/issues Here is my output with --debug: >kast text.exp --debug java.lang.NoClassDefFoundError: scala/reflect/runtime/package$ at org.kframework.meta.Reflection$.<init>(Reflection.scala:7) at org.kframework.meta.Reflection$.<clinit>(Reflection.scala) at org.kframework.meta.Down.apply(Down.scala:29) at org.kframework.meta.Down.apply(Down.scala:11) at scala.Option.map(Option.scala:146) at org.kframework.attributes.Att.get(Att.scala:24) at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte r.processProductions(KSyntax2GrammarStatesFilter.java:142) at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte r.lambda$getGrammar$92(KSyntax2GrammarStatesFilter.java:51) at java.util.HashMap.forEach(Unknown Source) at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte r.getGrammar(KSyntax2GrammarStatesFilter.java:51) at org.kframework.parser.concrete2kore.ParseInModule.getGrammar(ParseInM odule.java:93) at org.kframework.parser.concrete2kore.ParseInModule.parseStringTerm(Par seInModule.java:127) at org.kframework.parser.concrete2kore.ParseInModule.parseString(ParseIn Module.java:101) at org.kframework.parser.concrete2kore.ParseInModule.parseString(ParseIn Module.java:87) at org.kframework.kompile.CompiledDefinition.lambda$getParser$a68fb235$1 (CompiledDefinition.java:161) at org.kframework.kast.KastFrontEnd.run(KastFrontEnd.java:105) at org.kframework.main.FrontEnd.main(FrontEnd.java:52) at org.kframework.main.Main.runApplication(Main.java:110) at org.kframework.main.Main.runApplication(Main.java:100) at org.kframework.main.Main.main(Main.java:52) Caused by: java.lang.ClassNotFoundException: scala.reflect.runtime.package$ at java.net.URLClassLoader.findClass(Unknown Source) at java.lang.ClassLoader.loadClass(Unknown Source) at sun.misc.Launcher$AppClassLoader.loadClass(Unknown Source) at java.lang.ClassLoader.loadClass(Unknown Source) ... 20 more java.lang.NoClassDefFoundError: scala/reflect/runtime/package$ at org.kframework.meta.Reflection$.<init>(Reflection.scala:7) at org.kframework.meta.Reflection$.<clinit>(Reflection.scala) at org.kframework.meta.Down.apply(Down.scala:29) at org.kframework.meta.Down.apply(Down.scala:11) at scala.Option.map(Option.scala:146) at org.kframework.attributes.Att.get(Att.scala:24) at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte r.processProductions(KSyntax2GrammarStatesFilter.java:142) at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte r.lambda$getGrammar$92(KSyntax2GrammarStatesFilter.java:51) at java.util.HashMap.forEach(Unknown Source) at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte r.getGrammar(KSyntax2GrammarStatesFilter.java:51) at org.kframework.parser.concrete2kore.ParseInModule.getGrammar(ParseInM odule.java:93) at org.kframework.parser.concrete2kore.ParseInModule.parseStringTerm(Par seInModule.java:127) at org.kframework.parser.concrete2kore.ParseInModule.parseString(ParseIn Module.java:101) at org.kframework.parser.concrete2kore.ParseInModule.parseString(ParseIn Module.java:87) at org.kframework.kompile.CompiledDefinition.lambda$getParser$a68fb235$1 (CompiledDefinition.java:161) at org.kframework.kast.KastFrontEnd.run(KastFrontEnd.java:105) at org.kframework.main.FrontEnd.main(FrontEnd.java:52) at org.kframework.main.Main.runApplication(Main.java:110) at org.kframework.main.Main.runApplication(Main.java:100) at org.kframework.main.Main.main(Main.java:52) Caused by: java.lang.ClassNotFoundException: scala.reflect.runtime.package$ at java.net.URLClassLoader.findClass(Unknown Source) at java.lang.ClassLoader.loadClass(Unknown Source) at sun.misc.Launcher$AppClassLoader.loadClass(Unknown Source) at java.lang.ClassLoader.loadClass(Unknown Source) ... 20 more [Error] Internal: Uncaught exception thrown of type NoClassDefFoundError. Please rerun your program with the --debug flag to generate a stack trace, and file a bug report at https://github.com/kframework/k/issues OS: Windows 8.1 Pro, x64 K Framework version: 4.0.0 Java version: >java -version java version "1.8.0_101" Java(TM) SE Runtime Environment (build 1.8.0_101-b13) Java HotSpot(TM) 64-Bit Server VM (build 25.101-b13, mixed mode) Kompiler version: >kompile --version K framework version 4.0.0 Git revision: d310c7a Git branch: v4.0.0 Build date: Thu Jul 28 05:10:26 EEST 2016 Hi, I just tried this on master, revision: 1377911, and it works correctly. No ugly stack trace. I think it got fixed at one point. On another note: you have to provide the start symbol for the programs, in the configuration: configuration <k> $PGM:Ids </k> If nothing is declared, all sorts are taken into consideration, except for list sorts, since a single element in the program is always ambiguous (single element, or the list with one element). I tried this on win 10, x64. Let me know if the problem persists on your computer though. @Iaroslav-Mazur, do you still have this problem with the latest master? @radumereuta, shall we close this? I know for sure that Iaroslav just switched to Linux and he doesn't have this problem anymore. If it's fixed on a newer version of K then you can probably close it. I will take that as a yes.
gharchive/issue
2016-10-12T05:45:46
2025-04-01T04:34:45.724840
{ "authors": [ "Iaroslav-Mazur", "andreiarusoaie", "cos", "radumereuta" ], "repo": "kframework/k", "url": "https://github.com/kframework/k/issues/2218", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
517722683
Incorrect set variable de-duplication During substitution normalization, we de-duplicate element variable substitutions in this way: x = t₁ ∧ ... ∧ x = tₙ x = (t₁ ∧ ... ∧ tₙ) Set variables are de-duplicated in the same way, which is not correct. Set variables should be de-duplicated this way: @x = t₁ ∧ ... ∧ @x = tₙ @x = t₁ ∧ t₁ = t₂ ∧ ... ∧ t₁ = tₙ These transformations are equivalent for element variables. See also: https://runtimeverification.slack.com/archives/CC360GUTG/p1572914597064200 During the iteration planning meeting, we discussed this issue, but could not come to a resolution. We don't actually need to implement set variable deduplication right now because it is not required by any semantics. We should implement an error message in case it is required. In Kore.Step.Simplification.SubstitutionSimplifier.deduplicateSubstitution, the second argument is the set of substitutions. The keys of that Map are the substitution variable, and the NonEmpty list of values are all the assignments for that variable. We should add an assertion that if any key is a SetVar, then its list of assignments isSingleton (isSingleton = isJust . getSingleton). Fixed in #1224.
gharchive/issue
2019-11-05T11:43:37
2025-04-01T04:34:45.729118
{ "authors": [ "ttuegel" ], "repo": "kframework/kore", "url": "https://github.com/kframework/kore/issues/1207", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
422309234
Show Axiom Location Reviewer checklist [ ] Test coverage: stack test --coverage [ ] Public API documentation: stack haddock [ ] Style conformance: stylish-haskell @ttuegel I either addressed or replied to every comment. Let me know what you think. @vladciobanu The only thing I'm left wondering about is the cases where the attribute parser doesn't fail on malformed input, such as: if the Location attribute or Source attribute has no arguments, or if the Location(_,_,_,_) string is missing either pair. I have seen examples where either Source or Location are empty, and I don't think we want to FAIL on these. Maybe we want to produce a message with Pretty saying "missing X", just so it's obvious? Either way, I don't think I want to spend the time investigating this issue in the front-end quite yet. Although the type doesn't quite say the same story, its parser only allows "all-or-nothing": https://github.com/kframework/kore/blob/62b1fb0a87459b9b1d50e14b155b707fd6a66ef8/kore/src/Kore/Attribute/Location.hs#L86 I guess the correct representation for Location would be Maybe (LineColumn, LineColumn), but again, I'm not sure it's really a relevant distinction. Whether the attribute is present and empty, or missing is really not that important.
gharchive/pull-request
2019-03-18T16:17:52
2025-04-01T04:34:45.733188
{ "authors": [ "ttuegel", "vladciobanu" ], "repo": "kframework/kore", "url": "https://github.com/kframework/kore/pull/514", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
114953269
客户端已经停止使用,请前往douban.fm收听,感谢理解 hi,最近打开客户端,界面显示标题上的信息。封面是douban.fm的二维码。请问是douban封了这个客户端还是其它原因? 用了这么久,现在停了,我只能给作者加个🌟了。。。 这个客户端确实很好。有点舍不得。研究了一下,发现其实是app_name=radio_desktop_win的接口停止支持了。有个小技巧,可以把接口里的radio_desktop_win替换成radio。比如,可以用fiddler的AutoResponder功能来实现。 配置fiddler规则如下: regex:http://www.douban.com/j/app/radio/people(.+)(radio_desktop_win)(.+) http://douban.fm/j/mine/playlist$1radio$3 但是,貌似douban.fm官网上不会有歌曲的历史记录。 我修复了不能播放的问题,怎么打包为exe的安装文件? @LiteCoder 参考appveyor.yml里的after_build一节 @kfstorm thx 明天看看.
gharchive/issue
2015-11-04T02:10:11
2025-04-01T04:34:45.739392
{ "authors": [ "LiteCoder", "haozhu233", "iahu", "kfstorm" ], "repo": "kfstorm/DoubanFM", "url": "https://github.com/kfstorm/DoubanFM/issues/6", "license": "MS-PL", "license_type": "permissive", "license_source": "github-api" }
2116782174
🛑 NC is down In 47c3d05, NC ($NC) was down: HTTP code: 0 Response time: 0 ms Resolved: NC is back up in 5b9a259 after 18 hours, 6 minutes.
gharchive/issue
2024-02-04T01:11:00
2025-04-01T04:34:45.742743
{ "authors": [ "khadanja" ], "repo": "khadanja/upptime", "url": "https://github.com/khadanja/upptime/issues/444", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2506159513
TrollStore iOS17 Support? hi, thanks for this great apps! i want to ask sumn. did you plan any support for trollstore on iOS17? after the release of trollrestore, i think this will be a thing. I think Trollstore Lite is very similar to Trollstore in terms or url scheme and other aspects, and you can just install and use LiveContainer in Trollstore Lite like how you do in Trollstore Im not a dev or anything, but why would you use livecontainer if you can use trollstore ? Trollstore has unlimited signing already, plus jit Im not a dev or anything, but why would you use livecontainer if you can use trollstore ? Trollstore has unlimited signing already, plus jit Probably for multiple versions of app Im not a dev or anything, but why would you use livecontainer if you can use trollstore ? Trollstore has unlimited signing already, plus jit yes, i want to install and use multiple version of same app in in a phone, live container seems great to do it.. I just recalled that TrollStore can enable JIT for LiveContainer, so you can just sideload LiveContainer with TrollStore without JITLess mode. It will automatically jump to TrollStore for JIT and launch the app you want.
gharchive/issue
2024-09-04T19:53:03
2025-04-01T04:34:45.786645
{ "authors": [ "csdamien", "dejunedd", "hugeBlack", "veygax" ], "repo": "khanhduytran0/LiveContainer", "url": "https://github.com/khanhduytran0/LiveContainer/issues/151", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2429145606
🛑 Fvideo is down In f03c3e8, Fvideo (https://www.fvideo.ro) was down: HTTP code: 0 Response time: 0 ms Resolved: Fvideo is back up in dd913b7 after 10 minutes.
gharchive/issue
2024-07-25T06:38:34
2025-04-01T04:34:45.789387
{ "authors": [ "kharloss" ], "repo": "kharloss/monitorizare_websites", "url": "https://github.com/kharloss/monitorizare_websites/issues/229", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1521126738
onScan is called again when ReaderWidget is shown after stopImageStream() is used Below is an example app to demonstrate the issue. The camera preview is shown. If a barcode / QR code is scanned, the onScan callback is called. In the callback: stopImageStream() is called. The scanned text is assigned to _text. The camera preview is removed and the result view is shown. If the "Back to camera" button is pressed, the camera preview is shown. onScan is called immediately and the result view is shown. (This is wrong!) If the "Back to camera" button is pressed, the camera preview is shown. onScan is not called this time. import 'package:flutter/material.dart'; import 'package:camera/camera.dart'; import 'package:flutter_zxing/flutter_zxing.dart'; void main() { runApp(const App()); } class App extends StatefulWidget { const App({super.key}); @override State<App> createState() => _AppState(); } class _AppState extends State<App> { String? _text; CameraController? _controller; @override Widget build(BuildContext context) { final text = _text; final scanned = text != null && text.isNotEmpty; return MaterialApp( home: Scaffold( body: scanned ? Center( child: Column( mainAxisAlignment: MainAxisAlignment.center, children: [ Text(text), ElevatedButton( onPressed: () => setState(() => _text = null), child: const Text('Back to camera'), ), ], ), ) : ReaderWidget( onControllerCreated: (controller) { _controller = controller; }, onScan: (code) async { await _controller?.stopImageStream(); // This causes the issue. setState(() => _text = code.text); }, ), ), ); } } I used stopImageStream() because I had experienced duplicate calls to the scan callback in another plugin. I'm not sure if it can happen with flutter_zxing too, but I wanted to make sure onScan() was called only once. I don't think it is a critical issue, but it took me hours to figure out that stopImageStream() was causing it, and I still have no idea why it is wrong to use it there. I feel that something is wrong in the package and it is better that the issue is fixed so that other developers won't experience it. Here is a log that was printed in the console when a QR code was scanned. I/flutter ( 4417): onScan I/Camera ( 4417): startPreview I/Camera ( 4417): CameraCaptureSession onConfigured I/Camera ( 4417): Updating builder settings D/Camera ( 4417): Updating builder with feature: ExposureLockFeature D/Camera ( 4417): Updating builder with feature: ExposurePointFeature D/Camera ( 4417): Updating builder with feature: ZoomLevelFeature D/Camera ( 4417): Updating builder with feature: AutoFocusFeature D/Camera ( 4417): Updating builder with feature: NoiseReductionFeature I/Camera ( 4417): updateNoiseReduction | currentSetting: fast D/Camera ( 4417): Updating builder with feature: FocusPointFeature D/Camera ( 4417): Updating builder with feature: ResolutionFeature D/Camera ( 4417): Updating builder with feature: SensorOrientationFeature D/Camera ( 4417): Updating builder with feature: FlashFeature D/Camera ( 4417): Updating builder with feature: ExposureOffsetFeature D/Camera ( 4417): Updating builder with feature: FpsRangeFeature I/Camera ( 4417): refreshPreviewCaptureSession I/Camera ( 4417): CameraCaptureSession onClosed I/Camera ( 4417): dispose I/Camera ( 4417): close I/Camera ( 4417): open | onClosed The first line is a print call I added before stopImageStream() in the onScan callback. It seems strange that "startPreview" appears when the image stream is stopped. Thank you for bringing these issues to my attention. I really appreciate the time you took to report this and for helping me improve. I'll be sure to take a look and get back to you as soon as possible. Best regards, Khoren Hi there, I just wanted to let you know that I have updated the example in the repository. Could you please take a look and see if everything is working as expected? Thank you! @khoren93 I'm sorry if my explanation was unclear. I actually don't know if there was anything wrong in the example of the package itself. The issue is about the strange behaviour that the scan result remains and triggers the onScan callback immediately next time if stopImageStream() is used when scanning is successful. I didn't mean the example in the repository should be improved, but it is great that the example is better now. Hi @kaboc In my case, I have not experienced duplicate calls with the scan callback when using flutter_zxing, so I would not recommend using stopImageStream().
gharchive/issue
2023-01-05T17:42:16
2025-04-01T04:34:45.818341
{ "authors": [ "kaboc", "khoren93" ], "repo": "khoren93/flutter_zxing", "url": "https://github.com/khoren93/flutter_zxing/issues/71", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2493230641
[no-jira]:Translation strings ready 📲 What All languages ready :warning: Please install the to ensure uploads and comments are reliably processed by Codecov. Codecov Report Attention: Patch coverage is 0% with 4 lines in your changes missing coverage. Please review. Project coverage is 67.66%. Comparing base (bd46335) to head (f9461cf). Files with missing lines Patch % Lines ...i/activities/compose/projectpage/CheckoutScreen.kt 0.00% 2 Missing :warning: .../activities/compose/projectpage/AddOnsContainer.kt 0.00% 1 Missing :warning: ...kstarter/ui/views/compose/checkout/BonusSupport.kt 0.00% 1 Missing :warning: :exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality. Additional details and impacted files @@ Coverage Diff @@ ## master #2117 +/- ## ========================================= Coverage 67.66% 67.66% - Complexity 2189 2190 +1 ========================================= Files 367 367 Lines 22563 22563 Branches 3236 3236 ========================================= + Hits 15267 15268 +1 Misses 5561 5561 + Partials 1735 1734 -1 :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here.
gharchive/pull-request
2024-08-29T00:42:59
2025-04-01T04:34:45.856097
{ "authors": [ "Arkariang", "codecov-commenter" ], "repo": "kickstarter/android-oss", "url": "https://github.com/kickstarter/android-oss/pull/2117", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
506720598
Update to latest swiper 5.1.0 version Is it possible to update the package to the latest version of swiper. @kidjp85 🙏 please 5.1.0 now since it fixes errors with ios 13 As by documentation you install the latest version of swiper yourself npm install --save react-id-swiper@latest swiper@latest. If there are any errors with the latest swiper version maybe you should post the error here. Hi @groomain , sorry for the late response. I will try my best to update new version within this week. Newest version was released!
gharchive/issue
2019-10-14T15:19:59
2025-04-01T04:34:45.858928
{ "authors": [ "fxOne", "groomain", "kidjp85", "lhtdesignde" ], "repo": "kidjp85/react-id-swiper", "url": "https://github.com/kidjp85/react-id-swiper/issues/349", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1361693020
Fix compilation Should solve https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/blue/organizations/jenkins/KIE%2Fkogito%2Fmain%2Fnightly%2Fkogito-operator-deploy/detail/kogito-operator-deploy/111/pipeline/ Many thanks for submitting your Pull Request :heart:! Please make sure your PR meets the following requirements: [ ] You have read the contributors' guide [ ] Pull Request title is properly formatted: [KOGITO-XYZ] Subject [ ] Pull Request contains a link to the JIRA issue [ ] Pull Request contains a description of the issue [ ] Pull Request does not include fixes for issues other than the main ticket [ ] Your feature/bug fix has a unit test that verifies it [ ] You've ran make before-pr and everything is working accordingly [ ] You've tested the new feature/bug fix in an actual OpenShift cluster [ ] You've added a RELEASE_NOTES.md entry regarding this change How to retest this PR or trigger a specific build: Run operator BDD testing Please add comment: /jenkins test Run RHPAM operator BDD testing Please add comment: /jenkins RHPAM test /jenkins test PR job #634 was: FAILURE Possible explanation: Pipeline failure or project build failure You can find test error logs here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/634//artifact/test/logs/error/ Please look here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/634/display/redirect See console log: Console Logs DEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/factory.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/data_index.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/explainability.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/completion/completion.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/shared/resource_checks.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/task_console.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/supportingservices.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/kogitobuildtype_converter.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/runtime_flag.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/project/use_project.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/test/common.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/config_converter.goDEBU[2022-09-05T10:33:02Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/project/factory.goDEBU[2022-09-05T10:33:02Z] found 0 boxes../../../../go/pkg/mod/github.com/googleapis/gnostic@v0.5.1/jsonschema/models.go:19:8: missing go.sum entry for module providing package gopkg.in/yaml.v3 (imported by github.com/googleapis/gnostic/openapiv2); to add: go get github.com/googleapis/gnostic/openapiv2@v0.5.1make: *** [Makefile:173: build-cli] Error 1[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Push Operator Image(s) to Openshift Registry)Stage "Push Operator Image(s) to Openshift Registry" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Run BDD tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] stage[Pipeline] { (Running smoke tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/634/api/json[Pipeline] readJSON[Pipeline] sh+ tail -n 50+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/634/consoleText /jenkins test PR job #635 was: FAILURE Possible explanation: Pipeline failure or project build failure You can find test error logs here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/635//artifact/test/logs/error/ Please look here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/635/display/redirect See console log: Console Logs DEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/project/new_project.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/artifact_converter.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/runtimetype_converter.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/deploy/factory.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/probe_flag.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/util/common_util.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/kogitoprobe_converter.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/install.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/supportingservices.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/gitsource_converter.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/runtime_flag.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/kogito_infra.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/explainability.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/config_flag.goDEBU[2022-09-05T13:14:18Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/image_flag.goDEBU[2022-09-05T13:14:18Z] found 0 boxes../../../../go/pkg/mod/github.com/googleapis/gnostic@v0.5.1/jsonschema/models.go:19:8: missing go.sum entry for module providing package gopkg.in/yaml.v3 (imported by github.com/googleapis/gnostic/openapiv2); to add: go get github.com/googleapis/gnostic/openapiv2@v0.5.1make: *** [Makefile:173: build-cli] Error 1[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Push Operator Image(s) to Openshift Registry)Stage "Push Operator Image(s) to Openshift Registry" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Run BDD tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] stage[Pipeline] { (Running smoke tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/635/api/json[Pipeline] readJSON[Pipeline] sh I'll try it later locally. @vaibhavjainwiz might be able to help as well. /jenkins test PR job #639 was: FAILURE Possible explanation: Pipeline failure or project build failure You can find test error logs here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/639//artifact/test/logs/error/ Please look here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/639/display/redirect See console log: Console Logs DEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/remove/factory.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/service/buildresource_service.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/service/runtime_service.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/shared/install_services.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/test/common.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/completion/factory.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/infraresource_flag.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/install.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/core/resource_manager.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/context/command_context.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/monitoring_converter.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/deploy/factory.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/kogito_infra.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/data_index.goDEBU[2022-09-06T13:03:27Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/project.goDEBU[2022-09-06T13:03:27Z] found 0 boxes../../../../go/pkg/mod/github.com/googleapis/gnostic@v0.5.1/jsonschema/models.go:19:8: missing go.sum entry for module providing package gopkg.in/yaml.v3 (imported by github.com/googleapis/gnostic/openapiv2); to add: go get github.com/googleapis/gnostic/openapiv2@v0.5.1make: *** [Makefile:173: build-cli] Error 1[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Push Operator Image(s) to Openshift Registry)Stage "Push Operator Image(s) to Openshift Registry" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Run BDD tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] stage[Pipeline] { (Running smoke tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/639/api/json[Pipeline] readJSON[Pipeline] sh /jenkins test PR job #640 was: FAILURE Possible explanation: Pipeline failure or project build failure You can find test error logs here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/640//artifact/test/logs/error/ Please look here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/640/display/redirect See console log: Console Logs DEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/context/logger.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/project/use_project.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/remove/supportingservices.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/service/runtime_service.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/main.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/trusty_ui.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/remove/factory.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/context/command_context.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/deploy/deploy_service.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/build_flag.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/infraresource_flag.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/monitoring_flag.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/gitsource_converter.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/kogito_infra.goDEBU[2022-09-06T13:37:42Z] *parser.Parser#Run parsing=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/kubeconfig.goDEBU[2022-09-06T13:37:42Z] found 0 boxes../../../../go/pkg/mod/github.com/googleapis/gnostic@v0.5.1/jsonschema/models.go:19:8: missing go.sum entry for module providing package gopkg.in/yaml.v3 (imported by github.com/googleapis/gnostic/openapiv2); to add: go get github.com/googleapis/gnostic/openapiv2@v0.5.1make: *** [Makefile:173: build-cli] Error 1[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Push Operator Image(s) to Openshift Registry)Stage "Push Operator Image(s) to Openshift Registry" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Run BDD tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] stage[Pipeline] { (Running smoke tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/640/api/json[Pipeline] readJSON[Pipeline] sh /jenkins test PR job #642 was: ABORTED Possible explanation: Most probably a timeout, please review You can find test error logs here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/642//artifact/test/logs/error/ Please look here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/642/display/redirect See console log: Console Logs go: downloading k8s.io/klog/v2 v2.9.0go: downloading golang.org/x/tools v0.1.2go: downloading golang.org/x/sys v0.0.0-20210616094352-59db8d763f22----> Generating client code [versioned]----> Generating client code [listers]----> Generating client code [informers]Aborted by IAM#tradisso@redhat.comSending interrupt signal to processsignal: terminatedFailed to generate informersmake: *** [Makefile:122: generate] Error 1script returned exit code 2[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Build Kogito CLI)Stage "Build Kogito CLI" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Push Operator Image(s) to Openshift Registry)Stage "Push Operator Image(s) to Openshift Registry" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Run BDD tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] stage[Pipeline] { (Running smoke tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/642/api/json[Pipeline] readJSON[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/642/consoleText+ tail -n 50 /jenkins test
gharchive/pull-request
2022-09-05T09:58:07
2025-04-01T04:34:45.916614
{ "authors": [ "kie-ci", "radtriste", "spolti" ], "repo": "kiegroup/kogito-operator", "url": "https://github.com/kiegroup/kogito-operator/pull/1272", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
172498864
option to create without destination manifest / just detination manifest For DCDM's to save time can we have an option to option to create ~without destination manifest ~ just detination manifest? Hey, thanks for raising this as well - we need to do this. I didn't think of adding that option, I think that's actually the best/easiest way to do it for now. We can probably either alter move.py, or create a verify.py script to do the destionation manifest check if needs be. Also, doing a manual check with md5deep is very simple if it comes to it md5deep -ler drivename > manifest.md5 - then we can compare afterwards. also lol I only fixed this after @AnjaMahler is not even with the IFI anymore! :(
gharchive/issue
2016-08-22T16:39:59
2025-04-01T04:34:45.926471
{ "authors": [ "AnjaMahler", "kieranjol" ], "repo": "kieranjol/IFIscripts", "url": "https://github.com/kieranjol/IFIscripts/issues/70", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
177841373
moveit.py - add progress indicator As we've moved from md5deep, we need a progress indicator. This or something similar might work @ecodonohoe ?? Yep, I'll start looking into the progress package and any other options I come across. Awesome! On 19 Sep 2016 6:35 p.m., "ecodonohoe" notifications@github.com wrote: Yep, I'll start looking into the progress package and any other options I come across. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/kieranjol/IFIscripts/issues/91#issuecomment-248063251, or mute the thread https://github.com/notifications/unsubscribe-auth/ABEyvpjUmBd6PK540K0bV0tuQ6rZhlftks5qrsfmgaJpZM4KAufY . Thanks @ecodonohoe for adding this - closing.
gharchive/issue
2016-09-19T17:07:33
2025-04-01T04:34:45.930698
{ "authors": [ "ecodonohoe", "kieranjol" ], "repo": "kieranjol/IFIscripts", "url": "https://github.com/kieranjol/IFIscripts/issues/91", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
910626601
AnkiConnect failed to fetch I added https://killergerbah.github.io to ankiconnect but it doesn't work. Do you know how to fix it? @yona19 Some questions: What browser are you using? Do you have any ad blockers enabled? Can you paste the logs that appear when trying the AnkiConnect URL in asbplayer? Go to Chrome menu -> Other tools -> Developer console to see logs. I changed the browser to Chrome and it worked. Thank you I really appreciate it Glad it works now. Curious what browser you were using? I know with Brave you have to add the asbplayer site to a whitelist. Sorry for the late response. I was using brave.
gharchive/issue
2021-06-03T15:47:37
2025-04-01T04:34:45.964557
{ "authors": [ "killergerbah", "yona19" ], "repo": "killergerbah/asbplayer", "url": "https://github.com/killergerbah/asbplayer/issues/52", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2751247372
TagBot trigger issue This issue is used to trigger TagBot; feel free to unsubscribe. If you haven't already, you should update your TagBot.yml to include issue comment triggers. Please see this post on Discourse for instructions and more details. If you'd like for me to do this for you, comment TagBot fix on this issue. I'll open a PR within a few hours, please be patient! Triggering TagBot for merged registry pull request: https://github.com/JuliaRegistries/General/pull/121509
gharchive/issue
2024-12-19T19:25:48
2025-04-01T04:34:45.979639
{ "authors": [ "JuliaTagBot" ], "repo": "kingaa/PartiallyObservedMarkovProcesses.jl", "url": "https://github.com/kingaa/PartiallyObservedMarkovProcesses.jl/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
373757752
whatanime.ga has moved to trace.moe Hello, I'm (was) the owner of whatanime.ga A few days ago, the .ga domain was taken down by freenom T_T, so the site has moved to https://trace.moe if you are still actively supporting this project, please update the URL, thank you m(_ _;)m Thanks for let me know! 😃 I will update the URL.
gharchive/issue
2018-10-25T03:22:41
2025-04-01T04:34:45.981679
{ "authors": [ "kingdido999", "soruly" ], "repo": "kingdido999/atogatari", "url": "https://github.com/kingdido999/atogatari/issues/20", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
163566269
Write Dockerfile that installs node & slackit slackit already has a Dockerfile but it looks like it has a few issues. It might be easier to have a quick Dockerfile that installs node and then installs slackit, before running it. Once this is up, we can continue with https://github.com/kingsreach/kingsreach-online/issues/23 Can't believe I accidentally called it slackit everywhere. Such a mungbeen 😧. Done.
gharchive/issue
2016-07-03T15:14:17
2025-04-01T04:34:45.989486
{ "authors": [ "adamkdean" ], "repo": "kingsreach/kingsreach-slackit", "url": "https://github.com/kingsreach/kingsreach-slackit/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1627010371
Trouble with API calls - most certainly user error. Hey mate, First of all, thanks so much for this template and for all the work you've put in. I've got it mostly up and running on my machine. I'd say you can certainly preemptively mark this issue as 'user error' haha. Basically, I'm just having trouble with my API calls. I apologize for reaching out, but I've been trying for a good few days and just cannot work it out! Are you able to take a look at my screenshots below and let me know if I'm missing something obvious? My env My index.js My getData.js My console response: and finally, my terminal response from my npm run dev (this all runs fine until the request happens). Also, goes without saying, but if you can't be bothered looking at all this, just tell me to piss off and mark this as resolved haha! You've done enough just making the template. Thank you for the kind words! I just looked into this; It's partly my fault too that I ignored how the REST client actually works. I've pushed a new commit that fixes how REST client is created and it works now. The issue on your end was you weren't passing the path correctly. Here's the fix for that: userRoutes.get("/api/rest-test", async (req, res) => { const {client} = await clientProvider.restClient({ isOnline: false, req, res }); const data = await client.get({ path: "blogs" }); console.log(data.body.blogs) res.status(200).send({ blogs: data.body.blogs }); }); Oh wow! Thanks so much for the quick response. I'm somewhat glad that is wasn't 100% user error haha. I'll give this a go after work! Thanks again for everything!
gharchive/issue
2023-03-16T09:07:01
2025-04-01T04:34:45.994752
{ "authors": [ "kinngh", "ohermans1" ], "repo": "kinngh/shopify-node-express-mongodb-app", "url": "https://github.com/kinngh/shopify-node-express-mongodb-app/issues/66", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
520073628
packet: Use Route53 FQDN attribute instead of name The name attribute only refers to the name used to create the DNS record. The values accepted are either the subdomain to be created on the dns_zone or the FQDN. For example, if you do something like: resource "aws_route53_record" "example" { zone_id = "xxx" # let's say it is the zone id for awesome.lokomotive-k8s.com name = "deleteme" type = "A" ttl = 300 # private IPv4 address for etcd records = ["10.0.0.1"] } output "name" { value = "${aws_route53_record.example.name}" } output "fqdn" { value = "${aws_route53_record.example.fqdn}" } You will see: fqdn = deleteme.awesome.lokomotive-k8s.com name = deleteme If you use the FQDN for the name attribute in the example resource, however, the name and fqdn will match. This is what we are currently doing, for example here: https://github.com/kinvolk/lokomotive-kubernetes/blob/da1cdcfaec23acd727193f53b45ef40753960534/packet/flatcar-linux/kubernetes/controllers.tf#L8 However, the correct and reliable way to get the FQDN of a domain is to get the fqdn attribute, and not the name. This patch just simply changes to that, so we use what we really want and not depend on how the record was created. Although in this case it is a no-op. @invidian the name is whatever you used for the name, if you included the rest of the FQDN then it has it, if no, it doesn't. For example: resource "aws_route53_record" "example" { zone_id = "xxx" # let's say it is the zone id for awesome.lokomotive-k8s.com name = "deleteme" type = "A" ttl = 300 # private IPv4 address for etcd records = ["10.0.0.1"] } output "name" { value = "${aws_route53_record.example.name}" } output "fqdn" { value = "${aws_route53_record.example.fqdn}" } You will see: fqdn = deleteme.awesome.lokomotive-k8s.com name = deleteme Therefore, it will not be the FQDN. We happen to use the FQDN for the name (see https://github.com/kinvolk/lokomotive-kubernetes/blob/da1cdcfaec23acd727193f53b45ef40753960534/packet/flatcar-linux/kubernetes/controllers.tf#L8), therefore is not an issue in our case, but it is still the proper fix. I'll update the commit message to be more clear about this @invidian updated the PR comment. PTAL :) It seems that AWS provider has a mess regarding this thing. What is the mess, exactly? The docs seem quite clear that the fqdn attr is the way to know the fqdn of the record, IMHO. There is a lot of bug requests about the behavior of those 2 properties as far as I saw. I went trough them and trough documentation and honestly I'm still not sure if I should put the FQDN in name field or not without testing :smile: haha, was not aware of that. It makes even more sense to use the guaranteed and documented way to get the fqdn, then :-)
gharchive/pull-request
2019-11-08T15:22:23
2025-04-01T04:34:46.001272
{ "authors": [ "invidian", "rata" ], "repo": "kinvolk/lokomotive-kubernetes", "url": "https://github.com/kinvolk/lokomotive-kubernetes/pull/91", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
794370760
docs/quickstarts/aks: rephrase Kubernetes version note AKS currently provides 1.18 (and other platforms 1.19). It's easy that this note gets out of date so let's rephrase it so it's more generic. fixed a typo in the commit message fixed a typo in the commit message
gharchive/pull-request
2021-01-26T16:23:06
2025-04-01T04:34:46.002701
{ "authors": [ "iaguis" ], "repo": "kinvolk/lokomotive", "url": "https://github.com/kinvolk/lokomotive/pull/1344", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2461994848
Q: Custom App + Access Token Hi, how can I use this gem if I only have the x-access token from a custom install like described here: https://shopify.dev/docs/apps/build/authentication-authorization/access-tokens/generate-app-access-tokens-admin The shopify_app expects an whole oAuth that I cant seem to get working in case of a single user custom app. You'll need to create custom app for the store and then grab access token and API credentials from it. These credentials will be sufficient to set api_key and secret in ShopifyApp initializer. And then you can add the following method to Shop model: class Shop < ActiveRecord::Base include ShopifyApp::ShopSessionStorageWithScopes def self.system new( shopify_domain: "MYSHOPIFY_DOMAIN", shopify_token: "API_ACCESS_TOKEN_FOR_CUSTOM_APP" ) end end Using this method, you should be able to make API calls like this: Shop.system.with_shopify_session do GetOrder.call(id: order.shopify_gid) end Thank you for answering my Question. I made a worarkound by using a custom class: class MyShop def self.with_session(&block) ShopifyAPI::Auth::Session.temp(shop: "xxx.myshopify.com", access_token: "xxxxx-token", &block) end end So I can use it that way: MyShop.with_session do GetOrder.call(id: order.shopify_gid) end Is this any better or worse? Since I didnt even really install the shoppify_app gem after I couldnt get it to work in first case, but just pulled it in with your gem. Any comment is highly appreciated as I'd like to enhance the documentation as very many questions in the official forums seem to come from Auth with token isntead of o-Auth as you need token auth for every custom app that is not to be sold/ deployed. Documented both approaches. Thanks for sharing! There seems to be no way to authenticate custom apps via shopify_app gem. I end up spinning my own authentication with devise. However, shopify_app gem still might be helpful for handling webhooks, for example.
gharchive/issue
2024-08-12T22:26:27
2025-04-01T04:34:46.033090
{ "authors": [ "kbachl", "kirillplatonov" ], "repo": "kirillplatonov/shopify_graphql", "url": "https://github.com/kirillplatonov/shopify_graphql/issues/45", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
715332473
Python packages requiring rebuilds Send a PR bumping the second field in the version file. [x] asciinema @mmatongo [x] py-sip @sdsddsd1 [x] pyqt5 @sdsddsd1 [x] pyqtwebengine @sdsddsd1 [x] qemu @jedavies-dev [x] qutebrowser @sdsddsd1 [x] youtube-dl @Vouivre These packages require rebuild for Python 3.9. Done Thanks everyone.
gharchive/issue
2020-10-06T04:40:02
2025-04-01T04:34:46.041112
{ "authors": [ "dylanaraps", "mmatongo" ], "repo": "kisslinux/community", "url": "https://github.com/kisslinux/community/issues/1355", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
468803789
just --build=x86_64-pc-linux-musl It can handle --host and --target automatically Tested and works on my machine. Merged.
gharchive/pull-request
2019-07-16T18:35:59
2025-04-01T04:34:46.043275
{ "authors": [ "konimex", "protonesso" ], "repo": "kissx/packages", "url": "https://github.com/kissx/packages/pull/29", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
100990618
Problem with new users Hi, So I got taskboard installed fine. Everything seems to work but when I created a user things went wrong. When the new user logged in he doesn't see any taskboard. The user can't logout etc. This is the console output: GET http://----------.----.net/api/users/current 500 (Internal Server Error)(anonymous function) @ libs.min.js:507 q @ libs.min.js:503 g @ libs.min.js:500 (anonymous function) @ libs.min.js:532 q.$eval @ libs.min.js:546 q.$digest @ libs.min.js:543 q.$apply @ libs.min.js:547 g @ libs.min.js:501 u @ libs.min.js:505 w.onload @ libs.min.js:506 I'm running this on Apache, PHP/5.5.9, pdo_sqlite is loaded , AllowOverride is on etc. I have read and done #3 and #7 The same problem is in your own demo version on your site http://taskboard.matthewross.me/demo/#/boards When I create there a new user I got the same problem. Thanks. Christian I just duplicated this, and it seems to be an issue when the user is created without a default board. I'll get it fixed soon!
gharchive/issue
2015-08-14T11:06:30
2025-04-01T04:34:46.047385
{ "authors": [ "CIJeu", "kiswa" ], "repo": "kiswa/TaskBoard", "url": "https://github.com/kiswa/TaskBoard/issues/210", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1806865604
how to hot reload sql functions when develop, i add a new function in the sql file -- :name get-user-by-id :? :1 -- :doc returns a user object by id, or nil if not present SELECT * FROM user_profile WHERE id = :id; then, how can i hot reload the sql file, and exec the function? everytime when i add a new sql function , i should re-run (def query-fn (:db.sql/query-fn state/system)) It would be better to use a (defn query-fn [] (:db.sql/query-fn state/system)), this way it'll grab the fresh state of the system when it's called. Yes, that solves this problem,thanks
gharchive/issue
2023-07-17T03:06:07
2025-04-01T04:34:46.049230
{ "authors": [ "vinurs", "yogthos" ], "repo": "kit-clj/kit", "url": "https://github.com/kit-clj/kit/issues/106", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
999442684
Revert "Add the enter key support in grid block to create the default text bl…" Reverts kitconcept/volto-blocks-grid#25 @iFlameing I ran the acceptance tests twice and they still fail. Is it possible that we are missing another commit here to amend the Cypress tests?
gharchive/pull-request
2021-09-17T14:51:51
2025-04-01T04:34:46.055811
{ "authors": [ "tisto" ], "repo": "kitconcept/volto-blocks-grid", "url": "https://github.com/kitconcept/volto-blocks-grid/pull/26", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }