id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
866070660
|
Connection keeps on pending after several plays
Having multiple wavesurfer instances, after played several songs connection eventually get stuck on "pending" status.
Here is a pen with the test:
https://codepen.io/diegoigle/pen/OJWrZvO
I'm using Chrome lastest build.
Can someone help?
Make sure all the instances are paused or stopped before on the play event.
add
for (var i = 0; i < waveforms.length; i++) {
waveforms[i].pause();
}
to the playerPlay
var playerPlay = function(buttonIdAttr){
currentPlaying['status'] = 'PLAYING';
currentPlaying['id'] = buttonIdAttr;
for (var i = 0; i < waveforms.length; i++) {
waveforms[i].pause();
}
waveforms[currentPlaying['id']].play();
}
Thank for your help. Unfortunately it keeps happening, though now it takes a little more retries to get it pending.
I updated the pen with your code:
https://codepen.io/diegoigle/pen/OJWrZvO
https://stackoverflow.com/questions/23679968/chrome-hangs-after-certain-amount-of-data-transfered-waiting-for-available-soc
|
gharchive/issue
| 2021-04-23T12:35:53 |
2025-04-01T04:34:44.848237
|
{
"authors": [
"diegoiglesias",
"entonbiba",
"leeqkrtkxp"
],
"repo": "katspaugh/wavesurfer.js",
"url": "https://github.com/katspaugh/wavesurfer.js/issues/2249",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1438342703
|
ReferenceError: self is not defined
Hi,
After running a global npm update I’m not able to use the package anymore with the following error: self is not defined
I’m wrapping the package in a Vue component for a nuxt app.
I tried to reverse back to the old version that worked previously (5.1.0) but it also gave an error (window is not defined). I guess this is due to the update of other modules/webpack.
The current installed version is 6.4.0.
I saw that other people experienced the same problem:
https://github.com/katspaugh/wavesurfer.js/issues/2350
https://github.com/katspaugh/wavesurfer.js/issues/751
https://github.com/ShiiRochi/wavesurfer-react/issues/51
I tried to dynamically import wavesurfer.js in mounted and also to load it as a plugin only in client mode but still no success.
It produce a new error:
defu__WEBPACK_IMPORTED_MODULE_3__ is not a function
Vscode give a tooltip hint on the import statement:
Could not find a declaration file for module ‘wavesurfer.js’. ‘/Users/xx/xx/node_modules/wavesurfer.js/dist/wavesurfer.js’ implicitly has an ‘any’ type.
I want to be clear that before the npm update everything worked perfectly.
Thanks for your help,
I was able to make it work with dynamic import once I deleted the node_modules folder and re-run npm install.
|
gharchive/issue
| 2022-11-07T13:09:30 |
2025-04-01T04:34:44.852811
|
{
"authors": [
"dany68"
],
"repo": "katspaugh/wavesurfer.js",
"url": "https://github.com/katspaugh/wavesurfer.js/issues/2615",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
188187782
|
How is the position of the cursor determined
Can someone please tell me what variables controls the position of the cursor on the canvas?
I have added the regions plugin to my project, and when I have added multiple regions, it becomes difficult to control the position of the cursor by clicking different areas on the canvas.
When one region is currently playing (looping) and I click outside of the region, I expect the cursor to move to that region, and start playing at this point.
Also I am having trouble switching between regions, playing one after the other, by clicking within the region I want to play next.
I would very much like to delve into the code as such to fix these issues, but I need some help figuring out how the cursor is being controlled, and whether the cursor position is tied to the position where the player will continue playing etc.
It seems that there are more than one variable controlling the cursor, the start play position and being within a region or not - and all these comes into effect resulting in a user experience that leaves much to be desired.
I would of course appreciate any help on the matter of figuring out how it all ties together.
BR, Harriet
The inconsistency you describe sounds like a bug in the way region looping is implemented. The cursor shouldn't be controlled by anything except for the audioprogress event.
Has the problem been solved?
|
gharchive/issue
| 2016-11-09T08:10:16 |
2025-04-01T04:34:44.856064
|
{
"authors": [
"harrietsCreativeDabblets",
"katspaugh",
"wangzhaoming132"
],
"repo": "katspaugh/wavesurfer.js",
"url": "https://github.com/katspaugh/wavesurfer.js/issues/858",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
210470423
|
module is not defined, cordova.plugins deprecated
meanwhile cordova.plugins is deprecated.
How do i get a reference to the backgroundMode object?
I tried cordova.require('de.appplant.cordova.plugin.background-mode') but with no success #
solved my own problem:
var backgroundMode = cordova.require("cordova-plugin-background-mode.BackgroundMode"); did it for me
|
gharchive/issue
| 2017-02-27T12:29:25 |
2025-04-01T04:34:44.858949
|
{
"authors": [
"der-jochen"
],
"repo": "katzer/cordova-plugin-background-mode",
"url": "https://github.com/katzer/cordova-plugin-background-mode/issues/254",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1325654678
|
chore: import repository https://github.com/kavishkhanna/standard-usage-schemas.git
this commit will trigger a pipeline to generate the CI/CD configuration which will create a second commit on this Pull Request before it auto merges
[APPROVALNOTIFIER] This PR is NOT APPROVED
This pull-request has been approved by:
To complete the pull request process, please assign
You can assign the PR to them by writing /assign in a comment when ready.
The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
OWNERS
Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment
|
gharchive/pull-request
| 2022-08-02T10:38:43 |
2025-04-01T04:34:44.889072
|
{
"authors": [
"kavishkhanna"
],
"repo": "kavishkhanna/jx3-eks-asm",
"url": "https://github.com/kavishkhanna/jx3-eks-asm/pull/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
951223435
|
Redesign bottom navigation + header to follow Website 2.0 design
UI Design
New bottom navigation
Navigation menu (shown when clicking Menu)
Implementation model
Navigation menu disclosure is controlled by Headless UI (bundlephobia)
❓ Disclosure or Popover?
Tasks
[x] Redesign bottom navigation
[ ] Add overflow menu to navigation
I think Popover makes more sense in this case.
Is this bottom navigation applied in desktop view too?
I know that the priority is mobile first, but I find this bottom navigation a bit inconvenient in desktop view
So maybe we can make it to just show in mobile and for desktop, showing header menu just like in https://kawalcovid19.id/
@ekamuktia The design on the Figma file only has a mobile version, so yes, this bottom nav will be present on both mobile and desktop.
Although for the desktop, I think we can move the menu button to the global header instead, since we also have this design on the Figma file:
Will think more about this tomorrow.
|
gharchive/issue
| 2021-07-23T03:46:38 |
2025-04-01T04:34:44.895849
|
{
"authors": [
"ekamuktia",
"resir014",
"zainfathoni"
],
"repo": "kawalcovid19/wargabantuwarga.com",
"url": "https://github.com/kawalcovid19/wargabantuwarga.com/issues/346",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
569181341
|
Transportation section
Transportation finishing touches depends on the venue and hotels we select.
In the meantime, can we complete the small section below on the site:
"NFTA's Metro Rail provides a direct line from UB's South Campus to downtown Canalside. From metro stops, it's just a quick walk to many of Buffalo's entertainment and dining options. "
Potential images:
https://commons.wikimedia.org/wiki/File:Main_Street_in_Buffalo_NY_with_light_rail_and_Shea's_Performing_Arts_Center_(2015).jpg
https://commons.wikimedia.org/wiki/File:Theater_Station_at_Sheas.jpg
https://commons.wikimedia.org/wiki/File:20150827_61_NFTA_Light_Rail_at_Fountain_Plaza_(21990211710).jpg
https://commons.wikimedia.org/wiki/File:Buffalo_light_rail_train_laying_over_at_Erie_Canal_Harbor_station_(2015).jpg
I'm sorry @cmklotzb I am not sure I understand what you mean on "the small section below on the site" Important to remember "completeness" will be done by a different group of folks, so no one reading the proposal brochure expects the decision on where the venue will be to be included. Compare for example the proposal for Pittsburgh and the Pittsburgh actual conference
Never mind @cmklotzb I see what you mean now.
|
gharchive/issue
| 2020-02-21T20:59:25 |
2025-04-01T04:34:44.903089
|
{
"authors": [
"cmklotzb",
"kayiwa"
],
"repo": "kayiwa/c4lbuffaloproposal",
"url": "https://github.com/kayiwa/c4lbuffaloproposal/issues/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
299578709
|
Revert stream to regular map construction
Getting weird errors:
java.lang.IllegalStateException: Duplicate key kbasesearchengine.search.ObjectData@2ab1c576
Note that ObjectData is the value
Urrg this went to master instead of develop
|
gharchive/pull-request
| 2018-02-23T02:28:32 |
2025-04-01T04:34:44.926721
|
{
"authors": [
"MrCreosote"
],
"repo": "kbase/KBaseSearchEngine",
"url": "https://github.com/kbase/KBaseSearchEngine/pull/200",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1865982324
|
Update fastapi & pydantic
Pydantic v2 has a number of breaking changes.
Also see https://github.com/kbase/collections/issues/428
I didn't think of that, but looking at the code changes, they shouldn't break anything since microtrait doesn't actually use the models any more, just the field names. Therefore the imports just need to work.
Would it be ok if I asked you to bump pydantic next time you work on the tools? I've literally never run them so testing them is a bit of a pain for me currently (which at some point I need to fix)
|
gharchive/pull-request
| 2023-08-24T22:36:41 |
2025-04-01T04:34:44.928316
|
{
"authors": [
"MrCreosote"
],
"repo": "kbase/collections",
"url": "https://github.com/kbase/collections/pull/429",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1284321839
|
Implement and evaluate PSRAM QSPI controller
First, just implement something that can write and read the PSRAM.
Write test code that fills whole PSRAM with pseudo-random numbers, then verify.
Measure performance, latency.
Investigate how to tie it in to the n64 bus PIO.
Latency test:
Set up a test harness that, when a trigger gpio goes high, loads 2 bytes from the PSRAM, outputs the result on gpio0->gpio15. Measure latency with a logic analyzer.
Not relevant anymore
|
gharchive/issue
| 2022-06-24T23:44:15 |
2025-04-01T04:34:44.932407
|
{
"authors": [
"kbeckmann"
],
"repo": "kbeckmann/PicoCart64",
"url": "https://github.com/kbeckmann/PicoCart64/issues/3",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
107315535
|
Cow<'static, str> for strings; no excessive lifetimes
Increasing performance and easing away the lifetimes.
https://gist.github.com/llogiq/4557032bdf18a2cff136
Cow works by offering both the &'static str and the owned String.
Maybe even use the lifetime of the App or Arg instead of 'static. In which would maintain one lifetime at each struct.
I'm game for it, let's try and it see how it works out! Worst case scenario, we have a concept to shelf for a future iteration :+1:
@kbknapp I did try it, and reverted it to &'a See my rewrite. :D
Ah, my mistake! :)
|
gharchive/issue
| 2015-09-19T06:21:41 |
2025-04-01T04:34:44.945192
|
{
"authors": [
"james-darkfox",
"kbknapp"
],
"repo": "kbknapp/clap-rs",
"url": "https://github.com/kbknapp/clap-rs/issues/258",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
316671256
|
Removed dynamic evaluate functions
closes #166
I think I found a solution - https://github.com/GoogleChrome/puppeteer/blob/v1.0.0/lib/ExecutionContext.js#L42
const result = await handle.jsonValue().catch(error => undefined);
we should change the line
https://github.com/kblok/puppeteer-sharp/blob/master/lib/PuppeteerSharp/ExecutionContext.cs#L33
var result = await handle.JsonValue<T>();
and make it catch the exception and return null
Make sense, we should also add the EvaluateBigObject test here.
PuppeteerSharp.Tests.Puppeteer.PuppeteerLaunchTests.UserDataDirOption [FAIL]
System.Net.WebSockets.WebSocketException : The 'System.Net.WebSockets.InternalClientWebSocket' instance cannot be used for communication because it has been transitioned into the 'Aborted' state.
Conflicted @Meir017
|
gharchive/pull-request
| 2018-04-23T05:59:50 |
2025-04-01T04:34:44.948116
|
{
"authors": [
"Meir017",
"kblok"
],
"repo": "kblok/puppeteer-sharp",
"url": "https://github.com/kblok/puppeteer-sharp/pull/167",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1578444404
|
Are there any real examples of using rich text and promise toasts?
I'm looking through the docs and I can't figure out how to use rich text in my toasts. Are there any real examples?
If you scroll down to the bottom of the docs, there should be a button called "Rich content" that shows you a demo of rich text as well as how to implement it. You need to create a Svelte component (say RichText.svelte) which contains your rich text (i.e. any HTML markup). Then just import it wherever you're creating your toasts and call toast with toast(RichText).
Let me know if that works for you or if you have a different use case that's not addressed by the docs!
|
gharchive/issue
| 2023-02-09T19:10:13 |
2025-04-01T04:34:44.964490
|
{
"authors": [
"jakubdonovan",
"kbrgl"
],
"repo": "kbrgl/svelte-french-toast",
"url": "https://github.com/kbrgl/svelte-french-toast/issues/23",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
485301190
|
Expand all favorites groups button
Is your feature request related to a problem? Please describe.
Sometimes, my favorites disappear or become collapsed, leaving me to expand them all out again.
Describe the solution you'd like
Provide an Expand All button like the existing Collapse All button, but that will recursively expand all favorites groups instead.
I would love that also!
On macOS this seems to work already, just hold option (aka Alt) while clicking on the expand arrow. The normal file explorer in VSCode works the same way.
On macOS this seems to work already, just hold option (aka Alt) while clicking on the expand arrow. The normal file explorer in VSCode works the same way.
I think it works only for subfolders/subgroups in a group? I cannot expand all folders/groups with one click if they aren't nested, or am I missing something?
Ah, now I understood. Yes, what you describe is currently not possible.
(To not clutter the UI maybe the "collapse all" button could become a toggle for both expanding/collapsing all.)
|
gharchive/issue
| 2019-08-26T15:37:02 |
2025-04-01T04:34:44.981492
|
{
"authors": [
"arfordweb",
"fharper",
"halo"
],
"repo": "kdcro101/vscode-favorites",
"url": "https://github.com/kdcro101/vscode-favorites/issues/43",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1810138530
|
Fix foreign grants and functions ownership
chyběly schéma granty na cizí projekty
@pivnicek neměl by si chvilku? abych to mohl mergnout
|
gharchive/pull-request
| 2023-07-18T15:13:07 |
2025-04-01T04:34:45.026147
|
{
"authors": [
"ondrajodas"
],
"repo": "keboola/project-migration-tool",
"url": "https://github.com/keboola/project-migration-tool/pull/16",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1852664408
|
SOX-353 vytvareni tokenu
Original PR: https://github.com/keboola/storage-api-php-client/pull/1163
Reverts keboola/storage-api-php-client#1167
KBC: https://github.com/keboola/connection/pull/4574
Guest token vyřešen v connection
Nejde to approvnou zajcovi, ale říkal, že je to ok :)
|
gharchive/pull-request
| 2023-08-16T07:24:01 |
2025-04-01T04:34:45.032535
|
{
"authors": [
"tomasfejfar",
"zajca"
],
"repo": "keboola/storage-api-php-client",
"url": "https://github.com/keboola/storage-api-php-client/pull/1168",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1809282103
|
chore: Remove all non-editable keys from resources
Provide a description of what has been changed
The field creationTimestamp is auto-generated by kubernetes. This field is ignored during creation/update because it's non-editable by clients.
Checklist
[x] I have verified that my change is according to the deprecations & breaking changes policy
[x] Commits are signed with Developer Certificate of Origin (DCO - learn more)
[x] README is updated with new configuration values (if applicable) learn more
[x] A PR is opened to update KEDA core (repo) (if applicable, ie. when deployment manifests are modified)
Thanks for the fix ❤️
|
gharchive/pull-request
| 2023-07-18T07:07:47 |
2025-04-01T04:34:45.036250
|
{
"authors": [
"JorTurFer",
"alexmanno"
],
"repo": "kedacore/charts",
"url": "https://github.com/kedacore/charts/pull/498",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
443118618
|
Débug outside cluster
Someone could write a walkthrough on how to debug Keda outside the cluster using vscode.
aminebizid- I have hard a hard time doing this due to the metrics adapter. I could not get a version of it running stand alone outside the cluster. Without this, KEDA is debuggable as a stand alone app. So a workaround is to comment the adapter setup portion in main.go and start it up.
Could you show us an example of vscode launch.json
example of launch.json for debugging Keda controller
{
"version": "0.2.0",
"configurations": [
{
"name": "Launch",
"type": "go",
"request": "launch",
"mode": "auto",
"program": "${workspaceFolder}/cmd/manager/main.go",
"env": {"OPERATOR_NAME": "keda-operator", "WATCH_NAMESPACE": ""},
"args": ["--kubeconfig=/path/to/config/file"]
}
]
}
|
gharchive/issue
| 2019-05-12T15:48:43 |
2025-04-01T04:34:45.038250
|
{
"authors": [
"Aarthisk",
"aminebizid",
"nisan270390"
],
"repo": "kedacore/keda",
"url": "https://github.com/kedacore/keda/issues/190",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
689180356
|
Make controllers/ golint compatible
util/status.go : Comments added for exported functions: "SetStatusConditions" and "UpdateScaledObjectStatus".
util/string_lists.go : Comments added for exported functions: "Contains" and "Remove".
scaledjob_controller : Comments added for exported methods: "SetupWithManager" and "Reconcile". An else block readjusted to make it golint compliant.
scaledobject_controller.go : Comments added to exported methods: "SetupWithManager" and "Reconcile". Two else blocks readjusted to make it golint compliant.
Checklist
[x] Commits are signed with Developer Certificate of Origin (DCO)
[ ] Tests have been added
[ ] A PR is opened to update the documentation on https://github.com/kedacore/keda-docs
[ ] Changelog has been updated
Fixes #1055
Hi folks,
This is my first time contributing to an open source project. I have done things to the best of my knowledge. Please let me know if something is missing or was done incorrectly.
Any and all constructive criticism is welcome.
|
gharchive/pull-request
| 2020-08-31T12:47:21 |
2025-04-01T04:34:45.041979
|
{
"authors": [
"Mystigan"
],
"repo": "kedacore/keda",
"url": "https://github.com/kedacore/keda/pull/1065",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1299783356
|
Migrate Azure Log Analytics test to Go.
Signed-off-by: Vighnesh Shenoy vshenoy@microsoft.com
Provide a description of what has been changed
Checklist
[x] Commits are signed with Developer Certificate of Origin (DCO - learn more)
[x] Tests have been added
[x] A PR is opened to update our Helm chart (repo) (if applicable, ie. when deployment manifests are modified)
[x] A PR is opened to update the documentation on (repo) (if applicable)
[x] Changelog has been updated and is aligned with our changelog requirements
Fixes #3258
Relates to #2737
/run-e2e log_analytics
/run-e2e log_analytics
/run-e2e log_analytics
|
gharchive/pull-request
| 2022-07-09T23:02:54 |
2025-04-01T04:34:45.046889
|
{
"authors": [
"v-shenoy"
],
"repo": "kedacore/keda",
"url": "https://github.com/kedacore/keda/pull/3345",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2195492510
|
Add prototype Event Management calls and Event Proxy impl files
Describe the contribution
This PR add prototype implementations for the cFS Event Proxy design.
This design intends to abstract the cFS EVS API calls through the Event Management core in BPLib.
A related PR exists for BPLib, here: https://github.com/keegan-moore/bplib/pull/1 .
This design change causes the following in the BP app:
During app initialization BP now calls BPL_EVM_Initialize instead of CFE_EVS_Register
BPL_EVM_Initialize takes a struct of function pointers to the cFS Event Proxy implementation
When the BP app wants to generate an event, the app should now use BPL_EVM_SendEvent
Testing performed
Steps taken to test the contribution:
Build steps
make native.distclean
2.make ENABLE_TESTS=false native.install
Execution steps '...'
cd build-native-9.4.0/exe/cpu1
./core-cpu1
Expected behavior changes
EVS's CFE_EVS_Register should no longer generate the following warning in the ES System Log, since the event filter array is now appropriately sized:
1980-012-14:03:20.55879 CFE_EVS_Register: Filter limit truncated to 8
A clear and concise description of how this contribution will change behavior and level of impact.
API Changes:
During app initialization BP now calls BPL_EVM_Initialize instead of CFE_EVS_Register
When the BP app wants to generate an event, the app should now use BPL_EVM_SendEvent
Behavior Change:
EVS's CFE_EVS_Register should no longer generate a warning in the ES System Log during BP app init
No other changes to behavior
System(s) tested on
Hardware: PC / x86_64
OS: Ubuntu 22.04
Versions: cFE 6.7 (equuleus) release candidate
Additional context
N/A
Third party code
N/A
Contributor Info - All information REQUIRED for consideration of pull request
Keegan Moore, NASA/GSFC Code 582 (Flight Software Systems)
Before this change set is complete, we'd need to scrape through BP and look for all calls to CFE_EVS_SendEvent, replacing them with calls to BPL_EVM_SendEvent .
|
gharchive/pull-request
| 2024-03-19T16:59:43 |
2025-04-01T04:34:45.057385
|
{
"authors": [
"keegan-moore"
],
"repo": "keegan-moore/bp",
"url": "https://github.com/keegan-moore/bp/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2464536145
|
[bug] [bitcoin-core test] 'NOP OVER 1'
Issue
The following bitcoin-core test is failing to give the correct result :
NOP OVER 1
This code should be failing due to a INVALID_STACK_OPERATION.
Replicate
To replicate the issue you can use the latest main branch with the following command:
scarb cairo-run '[[],5132112,3,[],87231948791857,6]'
Please can I be assigned to this issue
@ShantelPeters For sure, thank you! Let me know if you have any issues.
@b-j-roberts Thank you sir …
I’ll get to work immediately.
|
gharchive/issue
| 2024-08-14T00:07:59 |
2025-04-01T04:34:45.073047
|
{
"authors": [
"ShantelPeters",
"b-j-roberts"
],
"repo": "keep-starknet-strange/shinigami",
"url": "https://github.com/keep-starknet-strange/shinigami/issues/135",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
70737607
|
Decipher/Cipher links not correct when URL not traditional
We're running redmine at https://redmine.company.com/redmine, not just at https://redmine.company.com. When we try to encode/decode, the URL gets rewritten incorrectly. For example: https://redmine.company.com/redmine/projects/testproject/wiki gets rewritten to https://redmine.company.com/projects/testproject/wiki.
Dear @zlacelle
First of all, thank you for your feedback. And yes, we kind of have that URL creation hard-coded for a Redmine deployment in / (not suitable for other deployments).
So, as soon as we have the opportunity/time, we will fix that and let you know. OR, if you wish to, you can provide a fix by submitting a pull request, which we will happily review and integrate with the base code.
Great, thanks for replying. I just changed the hardcoded value, but I'll see if I can learn a bit of Ruby and come up with a fix.
|
gharchive/issue
| 2015-04-24T16:48:54 |
2025-04-01T04:34:45.091131
|
{
"authors": [
"hsilva-keep",
"zlacelle"
],
"repo": "keeps/redmine_wikicipher",
"url": "https://github.com/keeps/redmine_wikicipher/issues/23",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2455772600
|
Can't fill out the Sentry dialog when in a dialog
Tapping in a text field doesn't show the keyboard - it comes up briefly and then goes away when something steals focus
Ref: cfe751488d0969e4d43d68cb02baf0c4bacb28e8
Well it seems to work OK on the About Page now.
Could reproduce it locally in the iOS Simulator in XCode, but when I tested it was working using this dummy button on the AboutPage.
Tried removing 100dvh and .sentry-error-embed-wrapper locally, but still couldn't get it to work properly like it did for real.
If it occurs again, next steps might be debugging on my phone for real by connecting it to my laptop.
Ref: https://docs.sentry.io/platforms/javascript/guides/react/user-feedback/#crash-report-modal
Also ref: 26092add5089698a0c4954f6822b447d98570f0f (Just to be safe)
|
gharchive/issue
| 2024-08-08T13:11:17 |
2025-04-01T04:34:45.129258
|
{
"authors": [
"keithamoss"
],
"repo": "keithamoss/demsausage",
"url": "https://github.com/keithamoss/demsausage/issues/3091",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
665935783
|
一句傻逼送给你,不成敬意
you seems like a fool, without respect~
and then...
Your mother exploded on the spot
老哥 冷静
calm down ... it is only a joke
大哥冷静
大哥,稳住
Please don't spit fragrance in the community
大哥冷静冷静
认真你就输了
|
gharchive/issue
| 2020-07-27T02:50:45 |
2025-04-01T04:34:45.169954
|
{
"authors": [
"12yy296",
"Jiang-xs",
"Naozumi520",
"Suckson",
"kav012375",
"mind-c",
"yuluowuying"
],
"repo": "kelseyhightower/nocode",
"url": "https://github.com/kelseyhightower/nocode/issues/4108",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
256520882
|
locking dependencies of editable vcs requirements
When we pip install then pip freeze an editable vcs requirement, the frozen requirements include all dependencies.
With pipenv the dependencies of editable vcs requirements do not appear in Pipfile.lock.
Example:
$ pipenv install -e git+https://github.com/sbidoul/subdirtest.git#egg=subdirtest\&subdirectory=setup
...
$ cat Pipfile.lock
...
"default": {
"subdirtest": {
"editable": true,
"git": "https://github.com/sbidoul/subdirtest.git",
"subdirectory": "setup",
"ref": "7b18afe980a42bfe8d1a96b5ed52e05f58d9ef6c"
}
},
...
$ pipenv run pip freeze
click==6.7
pkg-resources==0.0.0
-e git+https://github.com/sbidoul/subdirtest.git@7b18afe980a42bfe8d1a96b5ed52e05f58d9ef6c#egg=subdirtest&subdirectory=setup
Since the installed package depends on click I expect to see it in the locked dependencies.
pip-tools lacks this capability (https://github.com/kennethreitz/pipenv/issues/508#issuecomment-328345816)
This is definately out of scope for now — until pip-tools grows this ability — I'd raise the issue over there.
is there an issue in pip-tools for this that I can follow? if not, can you make one @sbidoul?
Problem is that I don't know pip-tools so I can't create a meaningful issue in that project. I'd need to do some research on the pip-tools api and how it is used by pipenv.
Issue title: "Resolve dependencies for editable requirements, required by pipenv"
contents: the contents of this issue.
pip-tools is poorly maintained, so i doubt it'll happen.
I'm committer on it though, so if you write a patch to get this to work, I'll merge it!
@kennethreitz is it supposed to be implemented in 7.3.2?
yep!
Hm, looks like its not working for me with 7.3.5 (the command in the description above).
oh i dont' know if it works for subdirectories, i know it works for VCSs in general now though.
I just tried pipenv install -e git+https://github.com/kennethreitz/requests#egg=requests and I don't get the dependencies in the lockfile.
ah, it works with the other form.
v7.6.3 released, which has this confirmed working!
Almost there:
$ pipenv install -e git+https://github.com/kennethreitz/requests#egg=requests Installing -e git+https://github.com/kennethreitz/requests#egg=requests…
Obtaining requests from git+https://github.com/kennethreitz/requests#egg=requests
Updating /home/sbi-local/.local/share/virtualenvs/pipenvtest-egFzzt6y/src/requests clone
Requirement already satisfied: chardet<3.1.0,>=3.0.2 in /home/sbi-local/.local/share/virtualenvs/pipenvtest-egFzzt6y/lib/python2.7/site-packages (from requests)
Requirement already satisfied: idna<2.7,>=2.5 in /home/sbi-local/.local/share/virtualenvs/pipenvtest-egFzzt6y/lib/python2.7/site-packages (from requests)
Requirement already satisfied: urllib3<1.23,>=1.21.1 in /home/sbi-local/.local/share/virtualenvs/pipenvtest-egFzzt6y/lib/python2.7/site-packages (from requests)
Requirement already satisfied: certifi>=2017.4.17 in /home/sbi-local/.local/share/virtualenvs/pipenvtest-egFzzt6y/lib/python2.7/site-packages (from requests)
Installing collected packages: requests
Running setup.py develop for requests
Successfully installed requests
Adding -e git+https://github.com/kennethreitz/requests#egg=requests to Pipfile's [packages]…
Locking [dev-packages] dependencies…
Locking [packages] dependencies…
Traceback (most recent call last):
File "/home/sbi-local/.local/bin/pipenv", line 11, in <module>
sys.exit(cli())
File "/home/sbi-local/.local/venvs/pipenv/lib/python2.7/site-packages/pipenv/vendor/click/core.py", line 722, in __call__
return self.main(*args, **kwargs)
File "/home/sbi-local/.local/venvs/pipenv/lib/python2.7/site-packages/pipenv/vendor/click/core.py", line 697, in main
rv = self.invoke(ctx)
File "/home/sbi-local/.local/venvs/pipenv/lib/python2.7/site-packages/pipenv/vendor/click/core.py", line 1066, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/home/sbi-local/.local/venvs/pipenv/lib/python2.7/site-packages/pipenv/vendor/click/core.py", line 895, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/home/sbi-local/.local/venvs/pipenv/lib/python2.7/site-packages/pipenv/vendor/click/core.py", line 535, in invoke
return callback(*args, **kwargs)
File "/home/sbi-local/.local/venvs/pipenv/local/lib/python2.7/site-packages/pipenv/cli.py", line 1500, in install
do_lock(system=system)
File "/home/sbi-local/.local/venvs/pipenv/local/lib/python2.7/site-packages/pipenv/cli.py", line 942, in do_lock
project=project
File "/home/sbi-local/.local/venvs/pipenv/local/lib/python2.7/site-packages/pipenv/utils.py", line 133, in resolve_deps
extra_constraints = best_matches_from(dep[len('-e '):], which=which, which_pip=which_pip, project=project)
File "/home/sbi-local/.local/venvs/pipenv/local/lib/python2.7/site-packages/pipenv/utils.py", line 106, in best_matches_from
setup_py_path = os.path.abspath(os.sep.join([result, 'setup.py']))
TypeError: sequence item 0: expected string, NoneType found
Ugh, in Python 3 this doesn't result in a type error.
|
gharchive/issue
| 2017-09-10T16:15:53 |
2025-04-01T04:34:45.183418
|
{
"authors": [
"AlJohri",
"kennethreitz",
"sbidoul"
],
"repo": "kennethreitz/pipenv",
"url": "https://github.com/kennethreitz/pipenv/issues/512",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
128419712
|
Inconsistent behaviour with https between urllib3 and requestsion.
In trying to make a https connection as so:
(trustmile-api-p2710)jBeast:trustmile james$ python -c "import requests; r = requests.get('https://devapi.trustmile.com/static/index.html', verify=True)"
Traceback (most recent call last):
File "<string>", line 1, in <module>
File "/Users/james/.virtualenvs/trustmile-api-p2710/lib/python2.7/site-packages/requests/api.py", line 67, in get
return request('get', url, params=params, **kwargs)
File "/Users/james/.virtualenvs/trustmile-api-p2710/lib/python2.7/site-packages/requests/api.py", line 53, in request
return session.request(method=method, url=url, **kwargs)
File "/Users/james/.virtualenvs/trustmile-api-p2710/lib/python2.7/site-packages/requests/sessions.py", line 468, in request
resp = self.send(prep, **send_kwargs)
File "/Users/james/.virtualenvs/trustmile-api-p2710/lib/python2.7/site-packages/requests/sessions.py", line 576, in send
r = adapter.send(request, **kwargs)
File "/Users/james/.virtualenvs/trustmile-api-p2710/lib/python2.7/site-packages/requests/adapters.py", line 447, in send
raise SSLError(e, request=request)
requests.exceptions.SSLError: ("bad handshake: Error([('SSL routines', 'ssl3_get_server_certificate', 'certificate verify failed')],)",)
I have latest necessary packages installed:
(trustmile-api-p2710)jBeast:trustmile james$ pip freeze | egrep '(requests|urllib3|certifi)'
certifi==2015.11.20.1
requests==2.9.1
urllib3==1.14
However, urllib works in the same environment:
import urllib3
import certifi
http = urllib3.PoolManager(
cert_reqs='CERT_REQUIRED', # Force certificate check.
ca_certs=certifi.where(), # Path to the Certifi bundle.
)
# You're ready to make verified HTTPS requests.
try:
r = http.request('GET', 'https://devapi.trustmile.com/static/index.html')
print r.status
except urllib3.exceptions.SSLError as e:
print e
Output is status 200.
Am I missing something? I tried to set the relevant environment variables. The cert is a legit one and it's root cert is in the cacerts.pem in certifi package.
Thanks again for an awesome package I use all the time Kenneth et al!
Out of interest, what version of OpenSSL are you using? Run python -c 'import ssl; print ssl.OPENSSL_VERSION'
OpenSSL 0.9.8zg 14 July 2015
Interesting. Does this problem occur if you don't pass verify=True?
Yep, same issue
Hmm. What other packages do you have installed?
Ah, that must be it, created a clean env with just certifi, requests, and urllib3 in it. Will investigate further. I have a lot installed.
ok, fixed in my env. Not sure which was the culprit but this worked
pip uninstall backports.ssl-match-hostname
pip uninstall pyOpenSSL
Yeah, so PyOpenSSL is the likely culprit. Requests, unlike urllib3, will automatically try to use PyOpenSSL if it's present. Can you reinstall it, and then run your urllib3 script with these two lines added to the top:
import urllib3.contrib.pyopenssl
urllib3.contrib.pyopenssl.inject_into_urllib3()
Your script would look like this:
import urllib3
import urllib3.contrib.pyopenssl
import certifi
urllib3.contrib.pyopenssl.inject_into_urllib3()
http = urllib3.PoolManager(
cert_reqs='CERT_REQUIRED', # Force certificate check.
ca_certs=certifi.where(), # Path to the Certifi bundle.
)
# You're ready to make verified HTTPS requests.
try:
r = http.request('GET', 'https://devapi.trustmile.com/static/index.html')
print r.status
except urllib3.exceptions.SSLError as e:
print e
|
gharchive/issue
| 2016-01-24T19:41:01 |
2025-04-01T04:34:45.189284
|
{
"authors": [
"Lukasa",
"jorourke"
],
"repo": "kennethreitz/requests",
"url": "https://github.com/kennethreitz/requests/issues/2978",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
278078073
|
Pretrained Model on Sports1M
Dear @kenshohara,
Thank you very much for your fantastic repository. Do you have any pretrained model on Sports1M?
No, I have not tried training on Sports1M because Sports1M is too big to utilize easily.
The size of Kinetics is not quite big (2TB storage is sufficient),
and include sufficient training data to train 3D CNNs.
If you do not have specific reason for using Sports1M,
I think Kinetics pretrained models are useful.
Thank for your time and response. I agree with you there. Pretrained 3D CNN models on Kinetics are welcome.
|
gharchive/issue
| 2017-11-30T10:56:35 |
2025-04-01T04:34:45.206975
|
{
"authors": [
"ahkarami",
"kenshohara"
],
"repo": "kenshohara/3D-ResNets-PyTorch",
"url": "https://github.com/kenshohara/3D-ResNets-PyTorch/issues/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
419720179
|
NODE_ENV undefined with typescript and nodemon
cross-env version: 5.2.0
node version: 10.12.0
npm (or yarn) version: 6.4.1
Relevant code or config
"dev:ts": "cross-env NODE_ENV=production tsc-watch --onSuccess \"node ./node_modules/nodemon/bin/nodemon.js dist/main.js\""
"dev:ts": "tsc-watch --onSuccess \"cross-env NODE_ENV=production node ./node_modules/nodemon/bin/nodemon.js dist/main.js\""
Tried both commands above but console.log(process.env.NODE_ENV) prints undefined.
Any help is appreciated.
Can't reproduce. The environment looks fine.
Sorry for not responding to this earlier. It's been long enough that I'm going to close this issue. If you're still experiencing problems, please open a new issue.
@minotaurrr 17 thumbs up i'm encountering same issue, did you found a fix?
|
gharchive/issue
| 2019-03-11T22:47:17 |
2025-04-01T04:34:45.210133
|
{
"authors": [
"cristian-eriomenco",
"geastwood",
"kentcdodds",
"minotaurrr"
],
"repo": "kentcdodds/cross-env",
"url": "https://github.com/kentcdodds/cross-env/issues/198",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
344045829
|
JSS style blocks overwhelm prettyDOM / debug output
Describe the feature you'd like:
The automatic debug output / prettyDOM is hard to pick through with JSS, since a rather large style block is added to the DOM with JSS classes. There should be a way to avoid JSS output (the material-ui library in our case dumps out 10,000 lines of CSS, which is thought-provoking in itself, but here we are).
Suggested implementation:
I was thinking of an option to prune style blocks from the output; @kentcdodds suggested that defaulting the debug output to the container instead of the documentElement might be a better idea.
Thinking about it more, what if we use document.body? That would avoid this issue without making it harder/confusing for portals.
That would work well for my purposes; the JSS style blocks are in the document head. I know browsers are lenient with
Hmmm... I actually think this change will need to be in react-testing-library because the way it works right now is it'll log out everything in the container that's the basis for the queries and in react-testing-library that's set to the documentElement. I think the best solution would probably be to just change that to document.body. I don't think anyone puts portals anywhere other than the document.body so we should be safe to make that change here.
Would you like to do that?
Thanks @kentcdodds! Submitted a [react-testing-lib PR|https://github.com/kentcdodds/react-testing-library/pull/145].
|
gharchive/issue
| 2018-07-24T13:49:30 |
2025-04-01T04:34:45.214800
|
{
"authors": [
"kentcdodds",
"timbonicus"
],
"repo": "kentcdodds/dom-testing-library",
"url": "https://github.com/kentcdodds/dom-testing-library/issues/81",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
388472336
|
feat(queryByCurrentValue)
What: It adds queryByCurrentValue, getByCurrentValue, queryAllByCurrentValue and getAllByCurrentValue.
Why: getByValue cannot get elements with value property (it only checks DOM value attributes)
(Discussed at #166)
How: Added a set of (get|query)(All)?ByCurrentValue methods.
Checklist:
[ ] Documentation
[x] Tests
[ ] Ready to be merged
[ ] Added myself to contributors table
Feel free to give any feedbacks.
Closes #158
Thanks for the review. I made changes regarding your three comments.
Looks good! Just need docs and a test to execute line 358 (the error state for queryAllBy...) to keep coverage at 100%.
Also it might be nice to add an integration test for fireEvent.change() with getCurrentValue.
@alexkrolick I've added a test to check if it throws.
@kentcdodds I've renamed and updates the README.md. I didn't know how to edit it, so it's kind of a draft. Where do you want to put the section in the document? And how do you want to explain it to users? Take a look at the doc and let me know. Besides, I'm not a native English speaker, so pretty much worried about it 😅
@kentcdodds done! I've used container instead of document.body since I found out other parts of the doc used container.
Looks like we're missing some coverage. You can open the coverage report in the browser in coverage/lcov-report to see what's missing.
@kentcdodds coverage 100% done
Updated the doc. Thanks for pointing it out @alexkrolick
:tada: This PR is included in version 3.14.0 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
Awesome! Thank you :) This merge made my day!!!
|
gharchive/pull-request
| 2018-12-07T01:49:45 |
2025-04-01T04:34:45.223763
|
{
"authors": [
"alexkrolick",
"eunjae-lee",
"kentcdodds"
],
"repo": "kentcdodds/dom-testing-library",
"url": "https://github.com/kentcdodds/dom-testing-library/pull/169",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
383496938
|
Change behaviour regarding controlled mode
Possible changes regarding - https://github.com/kentcdodds/react-toggled/issues/26
Is there a blocker for this PR?
Sorry, this slipped away from me somehow. Also I never even used this project (it was an experiment/demo) so I'm totally the wrong person to maintain it. Anyone want to take it over?
At the company I'm working for we use this library and make use of the patterns shown here. I'd be interested in helping to maintain this in case others are in the same boat.
Totally, it's all yours @therealparmesh.
To transition this over to you, I'm going to remove my npm and github token from travis which will disable the auto-releases. You can re-enable that yourself if you like.
Please let me know your npm username and I'll add you as a collaborator there.
This may also be helpful to you: https://kcd.im/write-oss
Let me know if you have any other questions.
@kentcdodds I will do my best to find other quality maintainers and to keep this to the standard that you're setting.
My npm username is therealparmesh.
Awesome. I've added you to npm and github.
I'm going to stop watching the repo now. Good luck!
@Andarist Do you still think this PR is worth moving forward on?
|
gharchive/pull-request
| 2018-11-22T11:36:52 |
2025-04-01T04:34:45.251372
|
{
"authors": [
"Andarist",
"kentcdodds",
"therealparmesh"
],
"repo": "kentcdodds/react-toggled",
"url": "https://github.com/kentcdodds/react-toggled/pull/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
118034540
|
Slide height percentage of parent e.g. 100%
Hi,
Sorry couldn't find any other issues relating to this. Great plugin BTW.
What is the correct way to have 100% height on slides? The parent container has a height that is a relative and the slide is relative to the container e.g. 100%
Slick adds many internal elements between the parent and the slides so the slide is no longer relative to the parent. Without hacking a height onto the slide or having knowledge of the internal workings of Slick and the additional containers it produces, what is the best way around this issue?
I was having the same issue with slides that had no content, only background images that covered the area - this was to be used as a non-interactive enhancement piece in the background of another slide, but I couldn't get it working with 100% height so had to use another technique.
Thanks,
Jarrod
Closing due to not following CONTRIBUTING.md
works great for me with this additional css:
.slick-list,
.slick-track {
height: 100%;
}
I tried the above solution and didn't work, but this does work!
.slick-track {
display: flex;
}
.slick-track .slick-slide {
display: flex;
height: auto;
}
I tried the above solution and didn't work, but this does work!
.slick-track {
display: flex;
}
.slick-track .slick-slide {
display: flex;
height: auto;
}
Wow so glad that I found this!
Thanks!
.slick-track {
display: flex;
}
.slick-track .slick-slide {
display: flex;
height: auto;
}
thanks
If anyone is experiencing weird slick height (very tall), the following should work (the other solutions weren't what I was looking for):
.slick-slide{
height: auto;
}
None of the above solutions work for me.
|
gharchive/issue
| 2015-11-20T12:45:05 |
2025-04-01T04:34:45.256638
|
{
"authors": [
"KhaledMohamedP",
"McGern",
"MsMatki",
"frankstepanski",
"kenwheeler",
"lunow",
"rudestewing",
"vasilzhigilei"
],
"repo": "kenwheeler/slick",
"url": "https://github.com/kenwheeler/slick/issues/1905",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
259191204
|
slidesToShow, slidesToScroll set to 0 crashes browsers
When slidesToShow and / or slidesToScroll is set to 0 or "" the browser crashes.
The provided jsfiddle only sets slidesToShow to 0 when breakpoint is below 500. So be sure to resize the window to the size. I did this on purpose so you don't get a crashed browser without seeing the jsfiddle ;)
====================================================================
http://jsfiddle.net/h367Lerw/2/
====================================================================
Steps to reproduce the problem
Open jsfiddle
Resize browser window to match the breakpoint of 500
Experience the crash
====================================================================
What is the expected behaviour?
The slider should either throw a error, or ignore the settings.
====================================================================
What is observed behaviour?
The browser crashes
====================================================================
More Details
Which browsers/versions does it happen on?
Chrome 60
Chrome 61
Safari 11.0
Firefox 55
Firefox 56
Which jQuery/Slick version are you using?
jQuery (edge)
Slick (master)
Did this work before?
Don't know
Related Task: #3016
I'm getting the same issue!
|
gharchive/issue
| 2017-09-20T14:47:37 |
2025-04-01T04:34:45.263869
|
{
"authors": [
"elz1",
"simoozzay"
],
"repo": "kenwheeler/slick",
"url": "https://github.com/kenwheeler/slick/issues/3133",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
400976580
|
Saving/Loading/resuming agent state and weights
Hello, I noticed that loading saved weights results in test of untrained agent (using the breakoutagent example). I read that saving just the weights is not enough since the agent/model is not optimized. Is there a solution? Is using the keras model.save/load_model functions solve this?
Ref: https://keras.io/getting-started/faq/#how-can-i-use-pre-trained-models-in-keras
Hey jheffez,
¿How did you solve this issue? I have trained an agent in a custom environment and after this process, I would want to use the trained and saved weights to restore the agent (I think that load_weights should be the way) and be able to deploy it into production; passing him the observations and getting the actions.
¿Is this doable with keras-rl actually? I recall that this could be done with keras as your post (using the predict methods available and feeding the x parameter). However, I don't see this possibility in the keras-rl docs.
Any help would be appreciated. Thanks in advance,
Would like help with the same, receiving an action prediction for a single observation instance in production setting. Can this be done with Keras-RL?
|
gharchive/issue
| 2019-01-19T08:57:36 |
2025-04-01T04:34:45.281521
|
{
"authors": [
"Eriz11",
"jarlva",
"krumiaa"
],
"repo": "keras-rl/keras-rl",
"url": "https://github.com/keras-rl/keras-rl/issues/288",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
205530388
|
Keras CTC Model Save Discrepency
I am trying to implement a simple BLSTM-CTC model using Keras (backend Tensorflow). I am testing this with a small dataset of online handwriting samples (316 Train data with 10 distinct characters and 4 words). each of these samples have 401 timesteps and at each timestep I have 16 features. So the input is a Numpy array of dimension [316,401,16]. My network is implemented successfully as suggested by this example. My code is as follows
def ctc_lambda_func(self,args):
y_pred, labels, input_length, label_length = args
# the 2 is critical here since the first couple outputs of the RNN
# tend to be garbage:
y_pred = y_pred[:, 2:, :]
return K.ctc_batch_cost(labels, y_pred, input_length, label_length)
self.inputlayer=Input(name='input', shape=[timesteps,features], dtype='float32')
self.labels = Input(name='the_labels', shape=[maxstringlen], dtype='float32')
self.input_length = Input(name='input_length', shape=[1], dtype='int64')
self.label_length = Input(name='label_length', shape=[1], dtype='int64')
self.lstm_1 = LSTM(rnn_size, return_sequences=True, init='he_normal', name='LSTM1')(self.inputlayer)
self.lstm_1b = LSTM(rnn_size, return_sequences=True, go_backwards=True, init='he_normal', name='LSTM1_b')(self.inputlayer)
self.gru1_merged = merge([self.lstm_1, self.lstm_1b], mode='sum')
self.out=TimeDistributed(Dense(nbclasses,name="dense2",activation="softmax"))(self.gru1_merged)
self.loss_out = Lambda(self.ctc_lambda_func, output_shape=(1,), name='ctc')([self.out, self.labels, self.input_length, self.label_length])
self.optimizer = RMSprop(lr=0.001, rho=0.9, epsilon=1e-08, decay=0.0)
self.final=Model(input=[self.inputlayer,self.labels,self.input_length,self.label_length],output=self.loss_out)
self.final.compile(loss={'ctc': lambda y_true, out: out}, optimizer=self.optimizer)
This network is compiled successfully. Now I am running it and saving with
for e in range(nbepochs):
self.final.fit(x,y,batch_size=64,nb_epoch=1,verbose=1)
self.final.save_weights("weights.h5")
#Loading with the following lines after creating the network again
self.final.load_weights("weights.h5")
When I am running the training everything is working fine. The CTC error is reducing as expected. But when I am trying to load the model from a previous state then it is not restoring from the last saved state.
Say I have executed 10 epochs
Epoch 1/1
316/316 [==============================] - 2s - loss: 11.5716
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 11.2032
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 11.1163
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 10.9920
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 10.9337
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 10.8708
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 10.7654
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 10.7335
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 10.6119
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 10.4657
Model saved
But when I am loading it again
Epoch 1/1
316/316 [==============================] - 2s - loss: 12.5233
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 11.8969
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 11.7075
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 11.5625
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 11.4145
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 11.2987
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 11.1833
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 11.0691
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 10.9600
Model saved
Epoch 1/1
316/316 [==============================] - 1s - loss: 10.8676
Model saved
Clearly there is something wrong as the loaded model is not starting from an error around 10.4657. I also tried to save the whole model with save() and load_model() which gave me an error "KeyError: CTC Lambda Func not found".
I am totally in dark. Is the Lmbda layer creating any problem? Please help if possible. Thank you for your time.
Is it possible that you have to set the start epoch to prevent it from starting from scratch?
@Cerno-b what do you mean ? I do not understand
I was thinking it could be related to this question: https://stackoverflow.com/questions/52476191/what-does-initial-epoch-in-keras-mean/52478034
|
gharchive/issue
| 2017-02-06T09:39:53 |
2025-04-01T04:34:45.294614
|
{
"authors": [
"Cerno-b",
"ibrahimabarry01",
"xisnu"
],
"repo": "keras-team/keras",
"url": "https://github.com/keras-team/keras/issues/5286",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
607143602
|
Fix incorrect loss_weights error message, clarify docs
Summary
loss_weights argument to model.fit only accepts python lists, so no numpy arrays. clarify docs for that. Also the error message was incorrect, asking for a "list of dicts" instead of a "list or dict" as the docs ask for.
PR Overview
No special requirements, minor change to docs, error message
n This PR requires new unit tests [y/n] (make sure tests are included)
? This PR requires to update the documentation [y/n] (make sure the docs are up-to-date)
n This PR is backwards compatible [y/n]
n This PR changes the current API [y/n] (all API changes need to be approved by fchollet)
Edit:
Force-pushed to reflow text to fit in 85 cols
"list OR dict" is very different from "list OF dictS"
It's not obvious at all given that most methods take numpy "lists" (arrays).
Superseded by https://github.com/tensorflow/tensorflow/pull/41859
|
gharchive/pull-request
| 2020-04-26T23:39:30 |
2025-04-01T04:34:45.298798
|
{
"authors": [
"MingweiSamuel"
],
"repo": "keras-team/keras",
"url": "https://github.com/keras-team/keras/pull/14008",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1945836278
|
Add named_product test utility to generate test cases with names.
This is the be used with absl.parameterized.named_parameters and instead of absl.parameterized.product. It creates testcases that have intuitive names.
Tests in numpy_test.py and dtypes_test.py that were using product now use named_product.
Also standardized the way we create KerasTensors in numpy_test.py for readability.
Thanks for the PR!
Thanks for the review, François!
Two comments about this PR:
The named_product thing is purely cosmetic. When a test fail, you do see the parameters used. However, when you have a high number of combination it's nice to immediately see the failing combinations without having to drill into each test. (For instance, you can see that all "int" types are failing).
I put this PR up separately but my goal is to use it with upcoming sparse ops tests.
Codecov Report
All modified lines are covered by tests :white_check_mark:
Comparison is base (36322e7) 57.96% compared to head (c852228) 57.99%.
Additional details and impacted files
@@ Coverage Diff @@
## master #18626 +/- ##
==========================================
+ Coverage 57.96% 57.99% +0.02%
==========================================
Files 334 334
Lines 32504 32524 +20
Branches 6348 6353 +5
==========================================
+ Hits 18841 18861 +20
Misses 12231 12231
Partials 1432 1432
Flag
Coverage Δ
keras
57.99% <100.00%> (+0.02%)
:arrow_up:
keras-numpy
57.99% <100.00%> (+0.02%)
:arrow_up:
Flags with carried forward coverage won't be shown. Click here to find out more.
Files
Coverage Δ
keras/testing/test_utils.py
100.00% <100.00%> (ø)
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
|
gharchive/pull-request
| 2023-10-16T18:23:48 |
2025-04-01T04:34:45.309718
|
{
"authors": [
"codecov-commenter",
"hertschuh"
],
"repo": "keras-team/keras",
"url": "https://github.com/keras-team/keras/pull/18626",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1657915366
|
Removing footnote stripping/tagging footnote origin points
What is your OS and architecture? Windows is not supported and Mac OS arm64 is not yet supported. For non-supported OS, you can use Docker (https://grobid.readthedocs.io/en/latest/Grobid-docker/)
Ubuntu 22.04.2
What is your Java version (java --version)? Expected versions are currently Java 8, 9 and 10.
JDK 11.0.18
Hey!
First off, I'd like to thank you for creating this tool. After testing a number of tools to markup and extract text from pdfs, I can quite confidently say that yours is the best for both usability and accuracy - it's helped me with my current project immensely, so thanks again.
For context, my current project relates to Australian legal cases and academic texts in the legal field - I have been working on code which extracts the text of these documents and converts them to an audio file using ESPNet. GROBID has been doing the heavy lifting for my text extraction, and on the whole works beautifully, but it is somewhat overzealous on footnote classification for legal documents. Occasionally, it will strip paragraphs of the body of the case out as footnotes, moving them to the footer.
I was wondering if there is a way to stop GROBID from stripping footnotes altogether - I have other ways to process them if necessary. I have tried implementing custom TEI schemas, but can only succeed in moving the location of the footnotes, not keeping them in the body/their original position.
Alternatively, if there is a way to tag the original position of the footnotes, I could reinsert them by post-processing the .xml.
Thank you in advance for any help you may be able to provide, and thanks again for creating this tool.
Hi @TheDeponent !
Thank you very much for the kind words on grobid and happy that it can be useful for your project.
The legal field is still challenging for Grobid, not really represented in the training data for the moment. But even in general, there's this issue of paragraph incorrectly labeled as footnote, which is an error very visible and not really possible to recover with the final XML only.
Unfortunately at this time, the addition of coordinates does not include the element note, I need to add it - I note it for the next release.
If you're courageous, one "hacking" solution is to retrain a segmentation model without the "foot note" structure, I think it should work:
modify the segmentation training data XML parser to consider <note place="footnote"> tag as <body>:
Change this:
https://github.com/kermitt2/grobid/blob/master/grobid-trainer/src/main/java/org/grobid/trainer/sax/TEISegmentationSaxParser.java#L157
the line 157 into these 3 lines:
currentTag = "<body>";
upperTag = currentTag;
upperQname = "body";
retrain the segmentation model (this command will also recompile for the above change):
> ./gradlew train_segmentation
After some hours (it depends on the number of CPU of your machine - usually half day), you get a new segmentation model and there should not be any foot note section anymore, they will be included as paragraph in the normal body text.
@kermitt2 Thanks so much for your response. Even with the lack of legal documents in the training data, the error rate is still quite low, on average in my testing it misidentifies 5 paragraphs in a legal document containing 200 numbered paragraphs.
I will attempt retraining the segmentation model over the next few days, and will
|
gharchive/issue
| 2023-04-06T19:10:13 |
2025-04-01T04:34:45.322435
|
{
"authors": [
"TheDeponent",
"kermitt2"
],
"repo": "kermitt2/grobid",
"url": "https://github.com/kermitt2/grobid/issues/996",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
233018412
|
Using a method instead of a template
This is really just for discussion. In the CMS for https://www.raise.me I actually don't use templates at all with the Shortcode gem. I set the template for every shortcode to '<%= @content %> and generate a presenter that returns the content I want.
I actually have a presenter factory class that take in a name and a block and uses that block to generate the content.
Here is that class: https://gist.github.com/toddnestor/d6c94655e98ce5fbe0e65843262b0b58
I want to discuss if there is a good way to incorporate something like this as a feature in the shortcode gem.
I'd guess there are other people who need to be able to basically tie a method to a shortcode in this way like I do, but maybe not. I have no problem if this doesn't become a feature, just want to open up a discussion about it.
So any thoughts about something like this that allows just tying a block of code to a shortcode instead of using a template?
Obviously my presenter factory class was a workaround, it seems to me that within this plugin there could probably be a more direct way to tie a shortcode to a method if we wanted to incorporate that feature.
Perhaps the shortcode arrays could be refactored to be hashes with the names as the keys and the value being a hash of options, one of those options could be to use a template or a method, or something along those lines.
I'll have to think about it some more.
|
gharchive/issue
| 2017-06-01T21:38:55 |
2025-04-01T04:34:45.341956
|
{
"authors": [
"toddnestor"
],
"repo": "kernow/shortcode",
"url": "https://github.com/kernow/shortcode/issues/50",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2370779404
|
🛑 FX blog (kty) is down
In f14038a, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in 027da02 after 6 minutes.
|
gharchive/issue
| 2024-06-24T17:46:12 |
2025-04-01T04:34:45.498021
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/10038",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2400258179
|
🛑 FX blog (kty) is down
In 5d0298f, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in f56a6af after 6 minutes.
|
gharchive/issue
| 2024-07-10T09:42:09 |
2025-04-01T04:34:45.500586
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/14338",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2417367110
|
🛑 FX blog (kty) is down
In 59f3dd7, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in b471263 after 5 minutes.
|
gharchive/issue
| 2024-07-18T20:46:42 |
2025-04-01T04:34:45.502894
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/16531",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2439864220
|
🛑 FX corretor (kty) is down
In 7a638b7, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX corretor (kty) is back up in 32cff21 after 6 minutes.
|
gharchive/issue
| 2024-07-31T11:54:45 |
2025-04-01T04:34:45.505228
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/19519",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2445891033
|
🛑 FX blog (kty) is down
In b6a7bf1, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in c51a9ab after 6 minutes.
|
gharchive/issue
| 2024-08-02T23:51:44 |
2025-04-01T04:34:45.507536
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/20075",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2456099944
|
🛑 FX corretor (kty) is down
In eddf782, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX corretor (kty) is back up in d0d1d4f after 6 minutes.
|
gharchive/issue
| 2024-08-08T15:31:38 |
2025-04-01T04:34:45.509855
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/21309",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2479622500
|
🛑 FX blog (kty) is down
In 48cbd6b, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in 887a5c3 after 17 minutes.
|
gharchive/issue
| 2024-08-22T02:13:56 |
2025-04-01T04:34:45.512336
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/24234",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2530429310
|
🛑 FX corretor (kty) is down
In 7e5f418, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX corretor (kty) is back up in 6406271 after 8 minutes.
|
gharchive/issue
| 2024-09-17T08:34:59 |
2025-04-01T04:34:45.514658
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/29552",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1659798684
|
🛑 FX net (kty) is down
In 174deb1, FX net (kty) (https://foxter.net.br) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX net (kty) is back up in 05a8cf9.
|
gharchive/issue
| 2023-04-09T08:38:08 |
2025-04-01T04:34:45.516919
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/3355",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2589609327
|
🛑 FX blog (kty) is down
In 5c10683, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in e96950f after 13 minutes.
|
gharchive/issue
| 2024-10-15T18:54:34 |
2025-04-01T04:34:45.519223
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/34536",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2592687315
|
🛑 FX blog (kty) is down
In 3c3630f, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in 9f8b2b6 after 21 minutes.
|
gharchive/issue
| 2024-10-16T18:16:55 |
2025-04-01T04:34:45.521673
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/34689",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2609102013
|
🛑 FX blog (kty) is down
In b970882, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in b86432b after 6 minutes.
|
gharchive/issue
| 2024-10-23T15:54:58 |
2025-04-01T04:34:45.524183
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/35937",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2673931542
|
🛑 FX blog (kty) is down
In 82039ac, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in 9154fba after 6 minutes.
|
gharchive/issue
| 2024-11-19T23:50:44 |
2025-04-01T04:34:45.526456
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/41002",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2719086746
|
🛑 FX corretor (kty) is down
In e562732, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX corretor (kty) is back up in e9c0378 after 36 minutes.
|
gharchive/issue
| 2024-12-05T01:04:25 |
2025-04-01T04:34:45.528775
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/43780",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1670020653
|
🛑 FX net (kty) is down
In d7277fa, FX net (kty) (https://foxter.net.br) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX net (kty) is back up in e06b15f.
|
gharchive/issue
| 2023-04-16T15:38:33 |
2025-04-01T04:34:45.530999
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/4387",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2728935951
|
🛑 FX corretor (kty) is down
In bbecc9c, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX corretor (kty) is back up in 2ffee96 after 6 minutes.
|
gharchive/issue
| 2024-12-10T03:59:39 |
2025-04-01T04:34:45.533320
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/44682",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1672016911
|
🛑 FX net (kty) is down
In 3c9ae63, FX net (kty) (https://foxter.net.br) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX net (kty) is back up in 5104eec.
|
gharchive/issue
| 2023-04-17T21:48:53 |
2025-04-01T04:34:45.535710
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/4556",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1676671964
|
🛑 FX net (kty) is down
In 05806ca, FX net (kty) (https://foxter.net.br) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX net (kty) is back up in 702d3b8.
|
gharchive/issue
| 2023-04-20T12:57:25 |
2025-04-01T04:34:45.537956
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/4906",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1950272891
|
🛑 FX kty hom (kty) is down
In af8179e, FX kty hom (kty) (https://foxter-hom.konecty.com) was down:
HTTP code: 503
Response time: 307 ms
Resolved: FX kty hom (kty) is back up in 99a6253 after 6 minutes.
|
gharchive/issue
| 2023-10-18T17:20:34 |
2025-04-01T04:34:45.540248
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/6437",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2225440028
|
🛑 FX portal api (kty) is down
In 80f3477, FX portal api (kty) (https://api.foxterciaimobiliaria.com.br) was down:
HTTP code: 404
Response time: 222 ms
Resolved: FX portal api (kty) is back up in 2d87875 after 1 hour, 11 minutes.
|
gharchive/issue
| 2024-04-04T12:59:45 |
2025-04-01T04:34:45.542572
|
{
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/8017",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1093825933
|
Add xwx.moe
Important: Please read all instructions carefully.
Select the appropriate category for what this PR is about
This PR is:
[x] Adding a new domain
[ ] Updating existing domain size
[ ] Changing domain name
[ ] Removing existing domain from list
[ ] Website code changes (512kb.club site)
[ ] Other not listed
Do not tick a checkbox if you haven’t performed its action. Honesty is indispensable for a smooth review process.
[x] I used the uncompressed size of the site
[x] I have included a link to the GTMetrix report
[x] The domain is in the correct alphabetical order
[x] This site is not a ultra lightweight site
[x] The following information is filled identical to the data file
- domain: xwx.moe
url: https://xwx.moe/
size: 324
last_checked: 2022-01-04
GTMetrix Report (English homepage): https://gtmetrix.com/reports/xwx.moe/8ZF0k2IG/
GTMetrix Report (Esperanto homepage): https://gtmetrix.com/reports/xwx.moe/eKSG8HZY/
GTMetrix Report (Language selection): https://gtmetrix.com/reports/xwx.moe/FXNrknyQ/
I used the English homepage's (https://xwx.moe/en) size for the entry, since it's the real content, and is larger than the "language select" page (https://xwx.moe/).
@kevquirk Not sure how to do this one.
|
gharchive/pull-request
| 2022-01-04T23:10:21 |
2025-04-01T04:34:45.549349
|
{
"authors": [
"JadedCtrl",
"kmutahar"
],
"repo": "kevquirk/512kb.club",
"url": "https://github.com/kevquirk/512kb.club/pull/620",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
328235114
|
Can't access my public files on Linux or keybase.pub
I can't access my public files on Linux or keybase.pub, which has brought down my site smiller.keybase.pub.
I can access files on my Mac, but on Linux I see:
➜ ~ keybase fs ls /keybase/public/smiller171
▶ ERROR Error checking merkle tree: tried 100 roots, but seqno not found
Just submitted logs which also had errors.
▶ INFO ignoring UI logs: context deadline exceeded
Hint: You are currently not seeing messages from other users and the system.
Users in the 'systemd-journal' group can see all messages. Pass -q to
turn off this notice.
No journal files were opened due to insufficient permissions.
▶ ERROR Error reading logs: Journalctl exited with an error: exit status 1
------------
Success! Your log ID is:
08d5535fd3033f797d59461c
Here's a URL to submit new bug reports containing this ID:
https://github.com/keybase/client/issues/new?body=[write%20something%20useful%20and%20descriptive%20here]%0A%0Amy%20log%20id:%2008d5535fd3033f797d59461c
Thanks!
------------
I'm submitting from a Chromebook using Crostini, but since I can't access from keybase.pub either the issue isn't Chromebook specific. (I didn't log in on the Chomebook until after the issue appeared)
Thanks @smiller171. I see the bug, it'll take us a little while to fix it. We'll follow up with you. Sorry for the inconvenience!
@smiller171 we are already working on a fix.
cc @maxtaco.
BTW, we have two fixes coming out for this problem, the first fix might solve it, which should be out sooner. Thank you.
keybase/kbfs#1597 should fix this once it's merged. (Though it won't fix the underlying issue, which will still happen if the last write to a folder happens to have been from a revoked device that got unlucky w.r.t. the global merkle tree. @maxtaco is going to work on that fix soon.)
@strib sounds like currently the answer to avoid this is to write to the directory from a different device before revoking a device? If I make a new write to the folder from my Mac will that restore access?
@strib sounds like currently the answer to avoid this is to write to the directory from a different device before revoking a device? If I make a new write to the folder from my Mac will that restore access?
The current problem is because one of the devices you revoked in February got unlucky, and the revoke operation itself got stuck in a long line of server operations. (We expect there to be only one or two server operations in between when you click "revoke" and the server confirms it, but in this case there were more than 300.) However, this was not the device that last wrote to your public folder, that was the other device you revoked in February. Usually Keybase clients only need to run the (currently buggy) verification process for the device that actually did the last write, but due to a different bug, it was doing them on all revoked devices, and hitting a problem with the key that didn't do the write.
So in short, no, writing from your Mac won't help current Linux clients, because they will still be verifying all your keys unnecessarily. keybase/kbfs#1597 stops them from verifying all keys, which will fix the issue as soon as I get it reviewed by someone and merged.
If the folder happened to have been written last by the problematic key, then yes writing to it from a different device before revoking would fix it. But you shouldn't have to change your behavior to adapt to our bugs, so I would just go about business as usual, and we'll fix the bugs ASAP.
Is there any reason that keybase.pub should have only just now run into this issue when the revocations were a while ago? Was one of these bugs only recently introduced?
Also, my current behavior is to revoke devices after I reformat/wipe them, Is it better to just leave them as they are since the private key data can't be retrieved anyway?
Is there any reason that keybase.pub should have only just now run into this issue when the revocations were a while ago? Was one of these bugs only recently introduced?
Yes, the bugs were introduced by us this week. We are tightening checks on data written by revoked keys, and there are a few corner cases, like this one, that we messed up.
Also, my current behavior is to revoke devices after I reformat/wipe them, Is it better to just leave them as they are since the private key data can't be retrieved anyway?
I think it's always better to revoke devices that won't get used again.
There's a new linux build out with a fix, and https://smiller171.keybase.pub/ works again. Please close this out if everything looks ok on your end. Thanks, and sorry for the inconvenience.
Thanks for getting a fix out so fast! I'll take a look shortly
|
gharchive/issue
| 2018-05-31T17:47:01 |
2025-04-01T04:34:45.569026
|
{
"authors": [
"maxtaco",
"smiller171",
"strib"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/12134",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
131738838
|
I/O error on one particular user's public listing
When i do an:
/keybase/public $ ls USERNAME
for the USERNAME mentioned in the logs, I get an I/O error.
my log id: cf323fd57709d213cce9541c
The user's key was reset, so doing:
keybase track USERNAME
was able to fix it.
Great, glad it works now!
|
gharchive/issue
| 2016-02-05T19:50:13 |
2025-04-01T04:34:45.571279
|
{
"authors": [
"strib",
"xxv"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/1912",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
132851324
|
Dragging a file into /keybase/private/ on OS X Finder gives an "Newer version of file already exists in this location" error.
Dragging a file into /keybase/private/ or /keybase/public/ with Finder gives the error "A newer version of this file already exists". Pressing "Stop" or "Replace" dismisses the dialog, but no file remains.
Screencast: https://keybase.pub/rodbegbie/keybase_bugreport_screens/copy_to_private.gif
my log id: 89d93dbb2c0d802d9e725b1c
Aside, after being thrown by this a few times, I realized the problem was that /keybase/private/rodbegbie/ wasn't mounted and that's where I was supposed to be dragging files -- but the error remains that the dialog is confusing and unhelpful.
Thanks for the report! Is this on OSX Yosemite or El Cap? I can't seem to reproduce it on Yosemite.
From the log, it doesn't look like KBFS is returning that error code. I think what's happening is this:
Entries under public/ or private/ correspond to (combinations of) normalized Keybase usernames.
Normalized Keybase username are all lower case.
To let people use whatever case they want, we fake out symlinks when someone gives us a non-normalized name, that points to the normalized version.
So in this case, we are pretending there is a walter.JPG symlink, pointing to walter.jpg. When the OS tries to write to walter.jpg (according to your log), we reject it with an I/O error. The Finder must be choosing to ignore that error, and instead tries to overwrite the fake symlink we made for walter.JPG, which looks like it already exists. I've seen El Cap do weird stuff like this before.
I agree this is confusing. I'll try to figure out a way around it. But I'm curious what happens if you try to drag in a file which has an all-lower-case name. I'm guessing you'll get an IO error.
^ (FYI, an IO error will show up as a mysterious error code 36.)
You are correct!
The Finder can’t complete the operation because some data in “walter.jpg” can’t be read or written.
(Error code -36)
Ok cool, thanks. Maybe we can head off invalid file names earlier, and avoid showing symlinks for them. I'll file an issue for it on our internal tracker.
Also note that in upcoming versions of KBFS, you'll get nicer error pop-ups with better error messages than the OS lets us have, when you hit situations like this.
|
gharchive/issue
| 2016-02-10T23:46:57 |
2025-04-01T04:34:45.576564
|
{
"authors": [
"rodbegbie",
"strib"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/1980",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
638931561
|
Error Code 202 / BAD SESSION? Cannot login on iOS or keybase.io
Attempts to login to Keybase via iOS app leads to the following error:
Also, I cannot login via keybase.io; that error shows as:
Edit: also posted to keybase-issues
I it is the same issue with timeout to SSL handshake
I it is the same issue with timeout to SSL handshake
Do you have a sense of the cause of the timeout? I haven't seen this problem reported by anyone else but I still cannot access my account.
@njm2112 I have the timeout exp for current ios apps.
Here is related ticket and comment that client suppose to have a new version to fix the problem : https://github.com/keybase/keybase-issues/issues/3917#issuecomment-646724838
|
gharchive/issue
| 2020-06-15T15:11:47 |
2025-04-01T04:34:45.580793
|
{
"authors": [
"miry",
"njm2112"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/24224",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1507447631
|
KBFS doesn't work on macOS X 11 Big Sur (still works with <=6.0.1)
(Reporting this separately as to point out that latest release for newer macOS broke the functionality for older OS versions)
Encountering already known issue with Keybase startup where KBFS fails to install.
Used to work on Intel 16inch 2019 MacBookPro (Big Sur 11.6), but latest official release will error out (which is 6.0.3-20221212203703+608e46df72 at the time of reporting this).
Some good news: was able to restore full functionality by downgrading to 6.0.1-20220606162955+0ee8d3021e (a version that I could still acquire from the official site via https://prerelease.keybase.io/darwin-test/index.html).
Relates to: https://github.com/keybase/client/issues/25226 (where macOS X 13 Ventura users claim that latest release fixed their issue)
What was tried to get 6.0.3 working (without any luck):
The machine was restarted (as prompted)
kbfs was uninstalled, Keybase restarted and kbfs reinstalled
My log id: 576775d4181cb5637d7b851c
This is essentially what I've tried to document on https://github.com/keybase/client/issues/24366#issuecomment-1361858895.
Briefly — because the Keybase app for macOS includes its own bundled version of macfuse, requiring a kernel extensionhat is highly specific to different macOS versions, it's not reasonable to expect that subsequent versions of the Keybase app will alternatively break support of either older or newer versions of macOS — depending on how exactly it was compiled.
FYI, the plain, unpatched version of macfuse supports all macOS versions from 10.9 to 13. The problem is not really with macfuse; it's with the way specific versions of macfuse are bundled with the Keybase app installer. These require manual intervention by a maintainer with the correct permissions (namely, access to the Keybase signing key for distributing macOS apps...) and are thus 'unfixable' by the public-at-large.
Also note that the major difference between 6.0.1 and 6.0.3 is the bumped version of macfuse, in order to support not only macOS Ventura, but better support for Apple Silicon chips. There doesn't seem to be any fundamental, game-breaking changes otherwise, as far as I can see (and understand) the code.
I actually messed up the report and it seems that the breaking issue was introduced when going from 6.0.0 to 6.0.1 :| Update the title and the description to reflect that as well.
The takeaway from all this is probably that when a breaking change for older OS versions is introduced, the "Your Keybase app is out of dated, please update now" message should not be presented to the users of said systems.
Another thing this would be to list older releases on the site (and maybe segregate them based on OS versions).
Good News YES ! @allanpaiste just like you said downgrading to to 6.0.1-20220606201228+4a1945f0a3 got me working on macOS 11.74 kbfs file integration yay!
|
gharchive/issue
| 2022-12-22T08:20:53 |
2025-04-01T04:34:45.589256
|
{
"authors": [
"GwynethLlewelyn",
"allanpaiste",
"paulhugel"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/25379",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2181375265
|
arm64 linux build?
hey, why you aren't shipping, .deb, .rpm, and other packages for arm64?
with asahi linux getting traction, more and more people will use that.
I am trying to build it at my machine, I was able to build it (cli and a desktop app in electron that gave a lot of alarms at selinux), will try to discover all the steps and share here just in case someone want to do it.
|
gharchive/issue
| 2024-03-12T11:24:01 |
2025-04-01T04:34:45.590656
|
{
"authors": [
"patricuspl",
"pulpo"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/26328",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
209296007
|
Write lock conflict
Log id: d58f77ad249bcad098ab411c
Details on what happens when attempting to delete a file while transferring a folder on OS X. Operation eventually finished without additional user interaction.
Thanks for the report. Hmm, that's an odd error message, it doesn't look like it's coming from KBFS. What programs were you using for transfer, and what showed you that error message?
I do see some timeout errors in your keybase log, but those are expected in the current release under some situations (we should be fixing them soon). I don't know how they'd end up showing you a lock error though.
This was a drag-and-drop via the Finder for the transfer, and a Cmd-Del
in the Finder folder for the delete. The message I believe was via the
normal thing-used-to-be-known-as-a-growl notification.
Attempts to replicate a couple times so far have failed (Log ID
5f306ec74e18857120bacb1c):
Drag and drop large archive
Delete file from active transfer folder
Works shrug
On 2/21/17 4:52 PM, Jeremy Stribling wrote:
Thanks for the report. Hmm, that's an odd error message, it doesn't
look like it's coming from KBFS. What programs were you using for
transfer, and what showed you that error message?
I do see some timeout errors in your keybase log, but those are
expected in the current release under some situations (we should be
fixing them soon). I don't know how they'd end up showing you a lock
error though.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/keybase/client/issues/5969#issuecomment-281508811,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AB4qxiMZT9y-0p0oOrLqRbVuGt8v27-jks5re2qngaJpZM4MH9pv.
Closing due to non-reproducibility.
|
gharchive/issue
| 2017-02-21T22:41:57 |
2025-04-01T04:34:45.596382
|
{
"authors": [
"itys",
"strib"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/5969",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
246588862
|
How to remove the K disk
KBFS (K:) 10.0 GB
You can uninstall Dokan in the control panel.
|
gharchive/issue
| 2017-07-30T14:58:13 |
2025-04-01T04:34:45.597306
|
{
"authors": [
"liyuhang1997",
"zanderz"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/7946",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
269938095
|
KBFS Keeps Requesting I Rekey
I have a number of devices connected to my account and I am asked on a regular basis to rekey using my paperkey on each device. I have done this but I keep getting asked to rekey. Not sure why the rekey doesn't appear to be working. Any assistance would be much appreciated! Thanks.
my log id: e1f5f13a80be26bd9ccf671c
Hi @jeffbristow. It looks like the folder in question can only be rekeyed by your devices Home Server and Macbook Air. Can you log into Keybase from one of those devices?
We're asking for your paper key out of desperation I guess, but we should be more intelligent about that.
Hmm, ok then I'm confused about what's going on, it seems our server-side information might have gotten a little out-of-sync somehow. Can you try writing a small file to the folder that's badged as needing a rekey?
Oh, actually I think I see now. The confusion is over a shared folder with a user who reset their account before we handled resets automatically. I'll mark that folder as ignorable and you shouldn't see the badge anymore.
I was able to make a text file named sample.txt in folder "private/user1,myuser" on Home Server.
I also was able to access the file from a computer I added this morning called DellUltrabook and there were no issues, other than keybase shows that this folder needs to be rekeyed.
(I can share the actual folder name through email).
Will this also end the prompts to rekey?
I hope so. It seems like your device did do a fair amount of rekeying today on a bunch of folders, so hopefully things are stable besides this one folder. If this doesn't fix it, we'll take another look at the prompt stuff. (This is a server-side fix, and will need to wait for the next server deploy, not sure when that will be.)
Thanks, sounds good.
Same issue here; log ID 5d0132248073eb85c001b51c
$ keybase rekey status
TLFs need rekeying:
private/<snip-1>,johanstokking
private/<snip-2>,johanstokking
private/johanstokking
Devices that can rekey:
desktop Mac
mobile iPhone 7
I'm on Mac but no luck getting rid of the rekey message.
Does it take time or do you need to do anything?
I'm on Linux, Ubuntu 16.04, Keybase version 1.0.36-20171130235940+7d909be3c. I launched the Desktop application and saw the red dot, indicating something needed my attention. I went to the private folder, and tried to rekey it with my paper secret. In my case, the paper secret did nothing, nor was there any validation saying what happened. It just left me with the need to rekey it again.
I repeated this a few times, restarted the application, etc, and still did not work. I revoked my paper secret and tried to rekey with a new one and it still didn't work. Eventually I logged in to my account in a browser to make sure my passphrase to login still worked. It did...
I restarted the desktop application once again, and the message to need to rekey the private folder went away, but now there is still a red dot, indicating something needs my attention. It feels like there is a synchronization issue, and it's a little bit frustrating that the application didn't really indicate anything, eg, no real validation was working. I deduced all my answers from just fiddling... Obviously this is not good.
Two things; I think we have a bug that slows down rekeying (cc: @songgao) and also we have a bug with badging (cc: @mmaxim) probably related to chat. We can look into the former if key keybase log send, and the latter, we're already looking into.
@nicorellius I can confirm it took about 50 minutes to rekey your private folder after you added the lion_2017-12-03 device. It seems like there was an issue on the robot2 device you used to provision it. Can you please keybase log send from robot2 so we track it down? Also, a keybase log send from lion_2017-12-03 would be useful as well. Thanks!
@strib Thanks for the feedback. Yes, that sounds about right.
I just sent logs from robot2 and lion_2017-12-03. Let me know if you need anything else. Thanks.
Thanks, looking at the logs now.
robot2 logs: 0982f731822d3f6e928f0c1c
lion_2017-12-03 logs: 5689aa35103f4f0fc666961c
It seems like before the provision, robot2 lost connectivity to our servers (maybe due to some transient DNS thing), and then got into a situation where it was backing off reconnect attempts:
2017-12-02T23:12:23.787200-08:00 ▶ [DEBU kbfs keybase_service_base.go:352] 12a MDServerRemote: CheckReachability(): failed to connect, reconnecting: dial tcp: lookup mdserver-1.kbfs.keybaseapi.com on 127.0.1.1:53: server misbehaving
...
2017-12-02T23:12:23.787568-08:00 ▶ [DEBU kbfs connection.go:619] 12e (CONN MDServerRemote fb905c5e) starting random backoff: 57m46.90297598s
Then the computer went to sleep for many hours, and when it came up I think it was still in this backoff period (which may be triggered based on the waking time, rather than the wall clock time?):
2017-12-03T07:23:15.016555-08:00 ▶ [DEBU kbfs notify_keyfamily.go:34] 148 Key family for user 27392ecdbf4bf98b78ef248b9bdc5c19 changed
2017-12-03T07:23:19.104299-08:00 ▶ [INFO kbfs asm_amd64.s:2337] 149 CheckForRekeys: checking for rekeys
@songgao: perhaps a CheckForRekeys call should use rpc.FireNow()?
How is it possible the backoff is ever 57 minutes? Why not cap it at something way lower?
@mmaxim: is it to avoid a thundering herd when mdservers restart. But any active mdserver request is supposed to break through the backoff, we're just not doing it for rekey requests right now.
But 57 minutes just seems way too high to solve that particular problem. It seems like something lower could still solve that problem, and lower the bar for missing this break through the backoff deal.
@songgao: perhaps a CheckForRekeys call should use rpc.FireNow()?
Good point; I'll put up a PR.
|
gharchive/issue
| 2017-10-31T12:20:45 |
2025-04-01T04:34:45.609241
|
{
"authors": [
"jeffbristow",
"johanstokking",
"maxtaco",
"mmaxim",
"nicorellius",
"songgao",
"strib"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/issues/9312",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
465944182
|
Airdrop profile star popup
@keybase/picnicsquad CC @keybase/react-hackers @keybase/design
Here's the overlay for the airdrop star on the profile page:
Kb.FloatingMenu didn't have support for setting a textColor or backgroundColor, so those new props are added.
Dividers are missing on the mobile action sheet. Otherwise LGTM!
Design:
@keybase/picnicsquad @keybase/react-hackers bumpity bump
|
gharchive/pull-request
| 2019-07-09T19:18:47 |
2025-04-01T04:34:45.612427
|
{
"authors": [
"cecileboucheron",
"cjb"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/pull/18312",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
164889219
|
gui quit on service quit
fixes flow issues and listens to the new rpc
@keybase/react-hackers
:shipit:
So this might not do anything for macOS cause we terminate the app process from ctl stop (see https://github.com/keybase/client/blob/master/go/client/cmd_ctl_stop_osx.go#L77).
But this is good for linux/windows, so we might want to add a comment in here to that effect.
|
gharchive/pull-request
| 2016-07-11T17:19:12 |
2025-04-01T04:34:45.614349
|
{
"authors": [
"MarcoPolo",
"chrisnojima",
"gabriel"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/pull/3410",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
186703106
|
Zanderz/core 4038
I tested it, it works. If you invoke the installer on the command line and add nodriver=1, it will skip the drivers. Our updater may still be irritating for driver-switching people but that will be another ticket.
@maxtaco @oconnor663
LGTM
|
gharchive/pull-request
| 2016-11-02T03:21:27 |
2025-04-01T04:34:45.615565
|
{
"authors": [
"oconnor663",
"zanderz"
],
"repo": "keybase/client",
"url": "https://github.com/keybase/client/pull/4770",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
341815432
|
Use correct field separator in keywords.txt
The Arduino IDE requires the use of a single true tab separator between the keyword name and identifier. When spaces are used rather than a true tab the keyword is not highlighted.
Reference:
https://github.com/arduino/Arduino/wiki/Arduino-IDE-1.5:-Library-specification#keywords
Thanks!
|
gharchive/pull-request
| 2018-07-17T08:25:24 |
2025-04-01T04:34:45.629292
|
{
"authors": [
"obra",
"per1234"
],
"repo": "keyboardio/KeyboardioScanner",
"url": "https://github.com/keyboardio/KeyboardioScanner/pull/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
182440594
|
Strange behavior on Windows 8
Here is my definition:
module EXP-SYNTAX
syntax Ids ::= List{Id, ";"}
endmodule
module EXP
imports EXP-SYNTAX
endmodule
Compiling works, but the following test fails. Here is the content I'm trying to parse:
a, b
Here is my output:
>kast text.exp
[Error] Internal: Uncaught exception thrown of type NoClassDefFoundError.
Please rerun your program with the --debug flag to generate a stack trace, and
file a bug report at https://github.com/kframework/k/issues
Here is my output with --debug:
>kast text.exp --debug
java.lang.NoClassDefFoundError: scala/reflect/runtime/package$
at org.kframework.meta.Reflection$.<init>(Reflection.scala:7)
at org.kframework.meta.Reflection$.<clinit>(Reflection.scala)
at org.kframework.meta.Down.apply(Down.scala:29)
at org.kframework.meta.Down.apply(Down.scala:11)
at scala.Option.map(Option.scala:146)
at org.kframework.attributes.Att.get(Att.scala:24)
at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte
r.processProductions(KSyntax2GrammarStatesFilter.java:142)
at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte
r.lambda$getGrammar$92(KSyntax2GrammarStatesFilter.java:51)
at java.util.HashMap.forEach(Unknown Source)
at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte
r.getGrammar(KSyntax2GrammarStatesFilter.java:51)
at org.kframework.parser.concrete2kore.ParseInModule.getGrammar(ParseInM
odule.java:93)
at org.kframework.parser.concrete2kore.ParseInModule.parseStringTerm(Par
seInModule.java:127)
at org.kframework.parser.concrete2kore.ParseInModule.parseString(ParseIn
Module.java:101)
at org.kframework.parser.concrete2kore.ParseInModule.parseString(ParseIn
Module.java:87)
at org.kframework.kompile.CompiledDefinition.lambda$getParser$a68fb235$1
(CompiledDefinition.java:161)
at org.kframework.kast.KastFrontEnd.run(KastFrontEnd.java:105)
at org.kframework.main.FrontEnd.main(FrontEnd.java:52)
at org.kframework.main.Main.runApplication(Main.java:110)
at org.kframework.main.Main.runApplication(Main.java:100)
at org.kframework.main.Main.main(Main.java:52)
Caused by: java.lang.ClassNotFoundException: scala.reflect.runtime.package$
at java.net.URLClassLoader.findClass(Unknown Source)
at java.lang.ClassLoader.loadClass(Unknown Source)
at sun.misc.Launcher$AppClassLoader.loadClass(Unknown Source)
at java.lang.ClassLoader.loadClass(Unknown Source)
... 20 more
java.lang.NoClassDefFoundError: scala/reflect/runtime/package$
at org.kframework.meta.Reflection$.<init>(Reflection.scala:7)
at org.kframework.meta.Reflection$.<clinit>(Reflection.scala)
at org.kframework.meta.Down.apply(Down.scala:29)
at org.kframework.meta.Down.apply(Down.scala:11)
at scala.Option.map(Option.scala:146)
at org.kframework.attributes.Att.get(Att.scala:24)
at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte
r.processProductions(KSyntax2GrammarStatesFilter.java:142)
at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte
r.lambda$getGrammar$92(KSyntax2GrammarStatesFilter.java:51)
at java.util.HashMap.forEach(Unknown Source)
at org.kframework.parser.concrete2kore.kernel.KSyntax2GrammarStatesFilte
r.getGrammar(KSyntax2GrammarStatesFilter.java:51)
at org.kframework.parser.concrete2kore.ParseInModule.getGrammar(ParseInM
odule.java:93)
at org.kframework.parser.concrete2kore.ParseInModule.parseStringTerm(Par
seInModule.java:127)
at org.kframework.parser.concrete2kore.ParseInModule.parseString(ParseIn
Module.java:101)
at org.kframework.parser.concrete2kore.ParseInModule.parseString(ParseIn
Module.java:87)
at org.kframework.kompile.CompiledDefinition.lambda$getParser$a68fb235$1
(CompiledDefinition.java:161)
at org.kframework.kast.KastFrontEnd.run(KastFrontEnd.java:105)
at org.kframework.main.FrontEnd.main(FrontEnd.java:52)
at org.kframework.main.Main.runApplication(Main.java:110)
at org.kframework.main.Main.runApplication(Main.java:100)
at org.kframework.main.Main.main(Main.java:52)
Caused by: java.lang.ClassNotFoundException: scala.reflect.runtime.package$
at java.net.URLClassLoader.findClass(Unknown Source)
at java.lang.ClassLoader.loadClass(Unknown Source)
at sun.misc.Launcher$AppClassLoader.loadClass(Unknown Source)
at java.lang.ClassLoader.loadClass(Unknown Source)
... 20 more
[Error] Internal: Uncaught exception thrown of type NoClassDefFoundError.
Please rerun your program with the --debug flag to generate a stack trace, and
file a bug report at https://github.com/kframework/k/issues
OS: Windows 8.1 Pro, x64
K Framework version: 4.0.0
Java version:
>java -version
java version "1.8.0_101"
Java(TM) SE Runtime Environment (build 1.8.0_101-b13)
Java HotSpot(TM) 64-Bit Server VM (build 25.101-b13, mixed mode)
Kompiler version:
>kompile --version
K framework version 4.0.0
Git revision: d310c7a
Git branch: v4.0.0
Build date: Thu Jul 28 05:10:26 EEST 2016
Hi, I just tried this on master, revision: 1377911, and it works correctly. No ugly stack trace. I think it got fixed at one point.
On another note: you have to provide the start symbol for the programs, in the configuration:
configuration <k> $PGM:Ids </k>
If nothing is declared, all sorts are taken into consideration, except for list sorts, since a single element in the program is always ambiguous (single element, or the list with one element).
I tried this on win 10, x64. Let me know if the problem persists on your computer though.
@Iaroslav-Mazur, do you still have this problem with the latest master?
@radumereuta, shall we close this?
I know for sure that Iaroslav just switched to Linux and he doesn't have this problem anymore. If it's fixed on a newer version of K then you can probably close it.
I will take that as a yes.
|
gharchive/issue
| 2016-10-12T05:45:46 |
2025-04-01T04:34:45.724840
|
{
"authors": [
"Iaroslav-Mazur",
"andreiarusoaie",
"cos",
"radumereuta"
],
"repo": "kframework/k",
"url": "https://github.com/kframework/k/issues/2218",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
517722683
|
Incorrect set variable de-duplication
During substitution normalization, we de-duplicate element variable substitutions in this way:
x = t₁ ∧ ... ∧ x = tₙ
x = (t₁ ∧ ... ∧ tₙ)
Set variables are de-duplicated in the same way, which is not correct. Set variables should be de-duplicated this way:
@x = t₁ ∧ ... ∧ @x = tₙ
@x = t₁ ∧ t₁ = t₂ ∧ ... ∧ t₁ = tₙ
These transformations are equivalent for element variables.
See also: https://runtimeverification.slack.com/archives/CC360GUTG/p1572914597064200
During the iteration planning meeting, we discussed this issue, but could not come to a resolution. We don't actually need to implement set variable deduplication right now because it is not required by any semantics. We should implement an error message in case it is required.
In Kore.Step.Simplification.SubstitutionSimplifier.deduplicateSubstitution, the second argument is the set of substitutions. The keys of that Map are the substitution variable, and the NonEmpty list of values are all the assignments for that variable. We should add an assertion that if any key is a SetVar, then its list of assignments isSingleton (isSingleton = isJust . getSingleton).
Fixed in #1224.
|
gharchive/issue
| 2019-11-05T11:43:37 |
2025-04-01T04:34:45.729118
|
{
"authors": [
"ttuegel"
],
"repo": "kframework/kore",
"url": "https://github.com/kframework/kore/issues/1207",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
422309234
|
Show Axiom Location
Reviewer checklist
[ ] Test coverage: stack test --coverage
[ ] Public API documentation: stack haddock
[ ] Style conformance: stylish-haskell
@ttuegel I either addressed or replied to every comment. Let me know what you think.
@vladciobanu The only thing I'm left wondering about is the cases where the attribute parser doesn't fail on malformed input, such as: if the Location attribute or Source attribute has no arguments, or if the Location(_,_,_,_) string is missing either pair.
I have seen examples where either Source or Location are empty, and I don't think we want to FAIL on these. Maybe we want to produce a message with Pretty saying "missing X", just so it's obvious? Either way, I don't think I want to spend the time investigating this issue in the front-end quite yet.
Although the type doesn't quite say the same story, its parser only allows "all-or-nothing": https://github.com/kframework/kore/blob/62b1fb0a87459b9b1d50e14b155b707fd6a66ef8/kore/src/Kore/Attribute/Location.hs#L86
I guess the correct representation for Location would be Maybe (LineColumn, LineColumn), but again, I'm not sure it's really a relevant distinction. Whether the attribute is present and empty, or missing is really not that important.
|
gharchive/pull-request
| 2019-03-18T16:17:52 |
2025-04-01T04:34:45.733188
|
{
"authors": [
"ttuegel",
"vladciobanu"
],
"repo": "kframework/kore",
"url": "https://github.com/kframework/kore/pull/514",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
114953269
|
客户端已经停止使用,请前往douban.fm收听,感谢理解
hi,最近打开客户端,界面显示标题上的信息。封面是douban.fm的二维码。请问是douban封了这个客户端还是其它原因?
用了这么久,现在停了,我只能给作者加个🌟了。。。
这个客户端确实很好。有点舍不得。研究了一下,发现其实是app_name=radio_desktop_win的接口停止支持了。有个小技巧,可以把接口里的radio_desktop_win替换成radio。比如,可以用fiddler的AutoResponder功能来实现。
配置fiddler规则如下:
regex:http://www.douban.com/j/app/radio/people(.+)(radio_desktop_win)(.+)
http://douban.fm/j/mine/playlist$1radio$3
但是,貌似douban.fm官网上不会有歌曲的历史记录。
我修复了不能播放的问题,怎么打包为exe的安装文件?
@LiteCoder 参考appveyor.yml里的after_build一节
@kfstorm thx 明天看看.
|
gharchive/issue
| 2015-11-04T02:10:11 |
2025-04-01T04:34:45.739392
|
{
"authors": [
"LiteCoder",
"haozhu233",
"iahu",
"kfstorm"
],
"repo": "kfstorm/DoubanFM",
"url": "https://github.com/kfstorm/DoubanFM/issues/6",
"license": "MS-PL",
"license_type": "permissive",
"license_source": "github-api"
}
|
2116782174
|
🛑 NC is down
In 47c3d05, NC ($NC) was down:
HTTP code: 0
Response time: 0 ms
Resolved: NC is back up in 5b9a259 after 18 hours, 6 minutes.
|
gharchive/issue
| 2024-02-04T01:11:00 |
2025-04-01T04:34:45.742743
|
{
"authors": [
"khadanja"
],
"repo": "khadanja/upptime",
"url": "https://github.com/khadanja/upptime/issues/444",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2506159513
|
TrollStore iOS17 Support?
hi, thanks for this great apps!
i want to ask sumn. did you plan any support for trollstore on iOS17? after the release of trollrestore, i think this will be a thing.
I think Trollstore Lite is very similar to Trollstore in terms or url scheme and other aspects, and you can just install and use LiveContainer in Trollstore Lite like how you do in Trollstore
Im not a dev or anything, but why would you use livecontainer if you can use trollstore ? Trollstore has unlimited signing already, plus jit
Im not a dev or anything, but why would you use livecontainer if you can use trollstore ? Trollstore has unlimited signing already, plus jit
Probably for multiple versions of app
Im not a dev or anything, but why would you use livecontainer if you can use trollstore ? Trollstore has unlimited signing already, plus jit
yes, i want to install and use multiple version of same app in in a phone, live container seems great to do it..
I just recalled that TrollStore can enable JIT for LiveContainer, so you can just sideload LiveContainer with TrollStore without JITLess mode. It will automatically jump to TrollStore for JIT and launch the app you want.
|
gharchive/issue
| 2024-09-04T19:53:03 |
2025-04-01T04:34:45.786645
|
{
"authors": [
"csdamien",
"dejunedd",
"hugeBlack",
"veygax"
],
"repo": "khanhduytran0/LiveContainer",
"url": "https://github.com/khanhduytran0/LiveContainer/issues/151",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2429145606
|
🛑 Fvideo is down
In f03c3e8, Fvideo (https://www.fvideo.ro) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Fvideo is back up in dd913b7 after 10 minutes.
|
gharchive/issue
| 2024-07-25T06:38:34 |
2025-04-01T04:34:45.789387
|
{
"authors": [
"kharloss"
],
"repo": "kharloss/monitorizare_websites",
"url": "https://github.com/kharloss/monitorizare_websites/issues/229",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1521126738
|
onScan is called again when ReaderWidget is shown after stopImageStream() is used
Below is an example app to demonstrate the issue.
The camera preview is shown.
If a barcode / QR code is scanned, the onScan callback is called. In the callback:
stopImageStream() is called.
The scanned text is assigned to _text.
The camera preview is removed and the result view is shown.
If the "Back to camera" button is pressed, the camera preview is shown.
onScan is called immediately and the result view is shown. (This is wrong!)
If the "Back to camera" button is pressed, the camera preview is shown.
onScan is not called this time.
import 'package:flutter/material.dart';
import 'package:camera/camera.dart';
import 'package:flutter_zxing/flutter_zxing.dart';
void main() {
runApp(const App());
}
class App extends StatefulWidget {
const App({super.key});
@override
State<App> createState() => _AppState();
}
class _AppState extends State<App> {
String? _text;
CameraController? _controller;
@override
Widget build(BuildContext context) {
final text = _text;
final scanned = text != null && text.isNotEmpty;
return MaterialApp(
home: Scaffold(
body: scanned
? Center(
child: Column(
mainAxisAlignment: MainAxisAlignment.center,
children: [
Text(text),
ElevatedButton(
onPressed: () => setState(() => _text = null),
child: const Text('Back to camera'),
),
],
),
)
: ReaderWidget(
onControllerCreated: (controller) {
_controller = controller;
},
onScan: (code) async {
await _controller?.stopImageStream(); // This causes the issue.
setState(() => _text = code.text);
},
),
),
);
}
}
I used stopImageStream() because I had experienced duplicate calls to the scan callback in another plugin. I'm not sure if it can happen with flutter_zxing too, but I wanted to make sure onScan() was called only once.
I don't think it is a critical issue, but it took me hours to figure out that stopImageStream() was causing it, and I still have no idea why it is wrong to use it there. I feel that something is wrong in the package and it is better that the issue is fixed so that other developers won't experience it.
Here is a log that was printed in the console when a QR code was scanned.
I/flutter ( 4417): onScan
I/Camera ( 4417): startPreview
I/Camera ( 4417): CameraCaptureSession onConfigured
I/Camera ( 4417): Updating builder settings
D/Camera ( 4417): Updating builder with feature: ExposureLockFeature
D/Camera ( 4417): Updating builder with feature: ExposurePointFeature
D/Camera ( 4417): Updating builder with feature: ZoomLevelFeature
D/Camera ( 4417): Updating builder with feature: AutoFocusFeature
D/Camera ( 4417): Updating builder with feature: NoiseReductionFeature
I/Camera ( 4417): updateNoiseReduction | currentSetting: fast
D/Camera ( 4417): Updating builder with feature: FocusPointFeature
D/Camera ( 4417): Updating builder with feature: ResolutionFeature
D/Camera ( 4417): Updating builder with feature: SensorOrientationFeature
D/Camera ( 4417): Updating builder with feature: FlashFeature
D/Camera ( 4417): Updating builder with feature: ExposureOffsetFeature
D/Camera ( 4417): Updating builder with feature: FpsRangeFeature
I/Camera ( 4417): refreshPreviewCaptureSession
I/Camera ( 4417): CameraCaptureSession onClosed
I/Camera ( 4417): dispose
I/Camera ( 4417): close
I/Camera ( 4417): open | onClosed
The first line is a print call I added before stopImageStream() in the onScan callback. It seems strange that "startPreview" appears when the image stream is stopped.
Thank you for bringing these issues to my attention. I really appreciate the time you took to report this and for helping me improve. I'll be sure to take a look and get back to you as soon as possible.
Best regards,
Khoren
Hi there,
I just wanted to let you know that I have updated the example in the repository. Could you please take a look and see if everything is working as expected?
Thank you!
@khoren93
I'm sorry if my explanation was unclear. I actually don't know if there was anything wrong in the example of the package itself. The issue is about the strange behaviour that the scan result remains and triggers the onScan callback immediately next time if stopImageStream() is used when scanning is successful. I didn't mean the example in the repository should be improved, but it is great that the example is better now.
Hi @kaboc
In my case, I have not experienced duplicate calls with the scan callback when using flutter_zxing, so I would not recommend using stopImageStream().
|
gharchive/issue
| 2023-01-05T17:42:16 |
2025-04-01T04:34:45.818341
|
{
"authors": [
"kaboc",
"khoren93"
],
"repo": "khoren93/flutter_zxing",
"url": "https://github.com/khoren93/flutter_zxing/issues/71",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2493230641
|
[no-jira]:Translation strings ready
📲 What
All languages ready
:warning: Please install the to ensure uploads and comments are reliably processed by Codecov.
Codecov Report
Attention: Patch coverage is 0% with 4 lines in your changes missing coverage. Please review.
Project coverage is 67.66%. Comparing base (bd46335) to head (f9461cf).
Files with missing lines
Patch %
Lines
...i/activities/compose/projectpage/CheckoutScreen.kt
0.00%
2 Missing :warning:
.../activities/compose/projectpage/AddOnsContainer.kt
0.00%
1 Missing :warning:
...kstarter/ui/views/compose/checkout/BonusSupport.kt
0.00%
1 Missing :warning:
:exclamation: Your organization needs to install the Codecov GitHub app to enable full functionality.
Additional details and impacted files
@@ Coverage Diff @@
## master #2117 +/- ##
=========================================
Coverage 67.66% 67.66%
- Complexity 2189 2190 +1
=========================================
Files 367 367
Lines 22563 22563
Branches 3236 3236
=========================================
+ Hits 15267 15268 +1
Misses 5561 5561
+ Partials 1735 1734 -1
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
|
gharchive/pull-request
| 2024-08-29T00:42:59 |
2025-04-01T04:34:45.856097
|
{
"authors": [
"Arkariang",
"codecov-commenter"
],
"repo": "kickstarter/android-oss",
"url": "https://github.com/kickstarter/android-oss/pull/2117",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
506720598
|
Update to latest swiper 5.1.0 version
Is it possible to update the package to the latest version of swiper.
@kidjp85 🙏
please 5.1.0 now since it fixes errors with ios 13
As by documentation you install the latest version of swiper yourself npm install --save react-id-swiper@latest swiper@latest.
If there are any errors with the latest swiper version maybe you should post the error here.
Hi @groomain , sorry for the late response. I will try my best to update new version within this week.
Newest version was released!
|
gharchive/issue
| 2019-10-14T15:19:59 |
2025-04-01T04:34:45.858928
|
{
"authors": [
"fxOne",
"groomain",
"kidjp85",
"lhtdesignde"
],
"repo": "kidjp85/react-id-swiper",
"url": "https://github.com/kidjp85/react-id-swiper/issues/349",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1361693020
|
Fix compilation
Should solve https://eng-jenkins-csb-business-automation.apps.ocp-c1.prod.psi.redhat.com/blue/organizations/jenkins/KIE%2Fkogito%2Fmain%2Fnightly%2Fkogito-operator-deploy/detail/kogito-operator-deploy/111/pipeline/
Many thanks for submitting your Pull Request :heart:!
Please make sure your PR meets the following requirements:
[ ] You have read the contributors' guide
[ ] Pull Request title is properly formatted: [KOGITO-XYZ] Subject
[ ] Pull Request contains a link to the JIRA issue
[ ] Pull Request contains a description of the issue
[ ] Pull Request does not include fixes for issues other than the main ticket
[ ] Your feature/bug fix has a unit test that verifies it
[ ] You've ran make before-pr and everything is working accordingly
[ ] You've tested the new feature/bug fix in an actual OpenShift cluster
[ ] You've added a RELEASE_NOTES.md entry regarding this change
How to retest this PR or trigger a specific build:
Run operator BDD testing
Please add comment: /jenkins test
Run RHPAM operator BDD testing
Please add comment: /jenkins RHPAM test
/jenkins test
PR job #634 was: FAILURE
Possible explanation: Pipeline failure or project build failure
You can find test error logs here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/634//artifact/test/logs/error/
Please look here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/634/display/redirect
See console log:
Console Logs
[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/factory.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/data_index.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/explainability.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/completion/completion.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/shared/resource_checks.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/task_console.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/supportingservices.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/kogitobuildtype_converter.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/runtime_flag.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/project/use_project.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/test/common.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/config_converter.go[37mDEBU[0m[2022-09-05T10:33:02Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/project/factory.go[37mDEBU[0m[2022-09-05T10:33:02Z] found 0 boxes../../../../go/pkg/mod/github.com/googleapis/gnostic@v0.5.1/jsonschema/models.go:19:8: missing go.sum entry for module providing package gopkg.in/yaml.v3 (imported by github.com/googleapis/gnostic/openapiv2); to add: go get github.com/googleapis/gnostic/openapiv2@v0.5.1make: *** [Makefile:173: build-cli] Error 1[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Push Operator Image(s) to Openshift Registry)Stage "Push Operator Image(s) to Openshift Registry" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Run BDD tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] stage[Pipeline] { (Running smoke tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/634/api/json[Pipeline] readJSON[Pipeline] sh+ tail -n 50+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/634/consoleText
/jenkins test
PR job #635 was: FAILURE
Possible explanation: Pipeline failure or project build failure
You can find test error logs here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/635//artifact/test/logs/error/
Please look here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/635/display/redirect
See console log:
Console Logs
[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/project/new_project.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/artifact_converter.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/runtimetype_converter.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/deploy/factory.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/probe_flag.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/util/common_util.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/kogitoprobe_converter.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/install.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/supportingservices.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/gitsource_converter.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/runtime_flag.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/kogito_infra.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/explainability.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/config_flag.go[37mDEBU[0m[2022-09-05T13:14:18Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/image_flag.go[37mDEBU[0m[2022-09-05T13:14:18Z] found 0 boxes../../../../go/pkg/mod/github.com/googleapis/gnostic@v0.5.1/jsonschema/models.go:19:8: missing go.sum entry for module providing package gopkg.in/yaml.v3 (imported by github.com/googleapis/gnostic/openapiv2); to add: go get github.com/googleapis/gnostic/openapiv2@v0.5.1make: *** [Makefile:173: build-cli] Error 1[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Push Operator Image(s) to Openshift Registry)Stage "Push Operator Image(s) to Openshift Registry" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Run BDD tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] stage[Pipeline] { (Running smoke tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/635/api/json[Pipeline] readJSON[Pipeline] sh
I'll try it later locally.
@vaibhavjainwiz might be able to help as well.
/jenkins test
PR job #639 was: FAILURE
Possible explanation: Pipeline failure or project build failure
You can find test error logs here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/639//artifact/test/logs/error/
Please look here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/639/display/redirect
See console log:
Console Logs
[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/remove/factory.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/service/buildresource_service.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/service/runtime_service.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/shared/install_services.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/test/common.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/completion/factory.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/infraresource_flag.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/install.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/core/resource_manager.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/context/command_context.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/monitoring_converter.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/deploy/factory.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/kogito_infra.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/data_index.go[37mDEBU[0m[2022-09-06T13:03:27Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/project.go[37mDEBU[0m[2022-09-06T13:03:27Z] found 0 boxes../../../../go/pkg/mod/github.com/googleapis/gnostic@v0.5.1/jsonschema/models.go:19:8: missing go.sum entry for module providing package gopkg.in/yaml.v3 (imported by github.com/googleapis/gnostic/openapiv2); to add: go get github.com/googleapis/gnostic/openapiv2@v0.5.1make: *** [Makefile:173: build-cli] Error 1[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Push Operator Image(s) to Openshift Registry)Stage "Push Operator Image(s) to Openshift Registry" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Run BDD tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] stage[Pipeline] { (Running smoke tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/639/api/json[Pipeline] readJSON[Pipeline] sh
/jenkins test
PR job #640 was: FAILURE
Possible explanation: Pipeline failure or project build failure
You can find test error logs here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/640//artifact/test/logs/error/
Please look here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/640/display/redirect
See console log:
Console Logs
[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/context/logger.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/project/use_project.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/remove/supportingservices.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/service/runtime_service.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/main.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/trusty_ui.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/remove/factory.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/context/command_context.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/deploy/deploy_service.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/build_flag.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/infraresource_flag.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/flag/monitoring_flag.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/converter/gitsource_converter.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/install/kogito_infra.go[37mDEBU[0m[2022-09-06T13:37:42Z] *parser.Parser#Run [37mparsing[0m=/home/jenkins/agent/workspace/kogito/kogito-operator-PR-check/cmd/kogito/command/message/kubeconfig.go[37mDEBU[0m[2022-09-06T13:37:42Z] found 0 boxes../../../../go/pkg/mod/github.com/googleapis/gnostic@v0.5.1/jsonschema/models.go:19:8: missing go.sum entry for module providing package gopkg.in/yaml.v3 (imported by github.com/googleapis/gnostic/openapiv2); to add: go get github.com/googleapis/gnostic/openapiv2@v0.5.1make: *** [Makefile:173: build-cli] Error 1[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Push Operator Image(s) to Openshift Registry)Stage "Push Operator Image(s) to Openshift Registry" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Run BDD tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] stage[Pipeline] { (Running smoke tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/640/api/json[Pipeline] readJSON[Pipeline] sh
/jenkins test
PR job #642 was: ABORTED
Possible explanation: Most probably a timeout, please review
You can find test error logs here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/642//artifact/test/logs/error/
Please look here: https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/642/display/redirect
See console log:
Console Logs
go: downloading k8s.io/klog/v2 v2.9.0go: downloading golang.org/x/tools v0.1.2go: downloading golang.org/x/sys v0.0.0-20210616094352-59db8d763f22----> Generating client code [versioned]----> Generating client code [listers]----> Generating client code [informers]Aborted by IAM#tradisso@redhat.comSending interrupt signal to processsignal: terminatedFailed to generate informersmake: *** [Makefile:122: generate] Error 1script returned exit code 2[Pipeline] }[Pipeline] // withEnv[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Build Kogito CLI)Stage "Build Kogito CLI" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Push Operator Image(s) to Openshift Registry)Stage "Push Operator Image(s) to Openshift Registry" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Run BDD tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] stage[Pipeline] { (Running smoke tests)Stage "Run BDD tests" skipped due to earlier failure(s)[Pipeline] }[Pipeline] // stage[Pipeline] }[Pipeline] // stage[Pipeline] stage[Pipeline] { (Declarative: Post Actions)[Pipeline] script[Pipeline] {[Pipeline] }[Pipeline] // script[Pipeline] script[Pipeline] {[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/642/api/json[Pipeline] readJSON[Pipeline] sh+ wget --no-check-certificate -qO - https://jenkins-kogito-tools.kogito-cluster-0ad6762cc85bcef5745bb684498c2436-0000.us-south.containers.appdomain.cloud/job/kogito/job/kogito-operator-PR-check/642/consoleText+ tail -n 50
/jenkins test
|
gharchive/pull-request
| 2022-09-05T09:58:07 |
2025-04-01T04:34:45.916614
|
{
"authors": [
"kie-ci",
"radtriste",
"spolti"
],
"repo": "kiegroup/kogito-operator",
"url": "https://github.com/kiegroup/kogito-operator/pull/1272",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
172498864
|
option to create without destination manifest / just detination manifest
For DCDM's to save time can we have an option to option to create ~without destination manifest ~ just detination manifest?
Hey, thanks for raising this as well - we need to do this. I didn't think of adding that option, I think that's actually the best/easiest way to do it for now.
We can probably either alter move.py, or create a verify.py script to do the destionation manifest check if needs be.
Also, doing a manual check with md5deep is very simple if it comes to it
md5deep -ler drivename > manifest.md5 - then we can compare afterwards.
also lol I only fixed this after @AnjaMahler is not even with the IFI anymore! :(
|
gharchive/issue
| 2016-08-22T16:39:59 |
2025-04-01T04:34:45.926471
|
{
"authors": [
"AnjaMahler",
"kieranjol"
],
"repo": "kieranjol/IFIscripts",
"url": "https://github.com/kieranjol/IFIscripts/issues/70",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
177841373
|
moveit.py - add progress indicator
As we've moved from md5deep, we need a progress indicator. This or something similar might work @ecodonohoe ??
Yep, I'll start looking into the progress package and any other options I come across.
Awesome!
On 19 Sep 2016 6:35 p.m., "ecodonohoe" notifications@github.com wrote:
Yep, I'll start looking into the progress package and any other options I
come across.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/kieranjol/IFIscripts/issues/91#issuecomment-248063251,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ABEyvpjUmBd6PK540K0bV0tuQ6rZhlftks5qrsfmgaJpZM4KAufY
.
Thanks @ecodonohoe for adding this - closing.
|
gharchive/issue
| 2016-09-19T17:07:33 |
2025-04-01T04:34:45.930698
|
{
"authors": [
"ecodonohoe",
"kieranjol"
],
"repo": "kieranjol/IFIscripts",
"url": "https://github.com/kieranjol/IFIscripts/issues/91",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
910626601
|
AnkiConnect failed to fetch
I added https://killergerbah.github.io to ankiconnect but it doesn't work. Do you know how to fix it?
@yona19 Some questions:
What browser are you using? Do you have any ad blockers enabled?
Can you paste the logs that appear when trying the AnkiConnect URL in asbplayer?
Go to Chrome menu -> Other tools -> Developer console to see logs.
I changed the browser to Chrome and it worked. Thank you I really appreciate it
Glad it works now. Curious what browser you were using? I know with Brave you have to add the asbplayer site to a whitelist.
Sorry for the late response. I was using brave.
|
gharchive/issue
| 2021-06-03T15:47:37 |
2025-04-01T04:34:45.964557
|
{
"authors": [
"killergerbah",
"yona19"
],
"repo": "killergerbah/asbplayer",
"url": "https://github.com/killergerbah/asbplayer/issues/52",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2751247372
|
TagBot trigger issue
This issue is used to trigger TagBot; feel free to unsubscribe.
If you haven't already, you should update your TagBot.yml to include issue comment triggers.
Please see this post on Discourse for instructions and more details.
If you'd like for me to do this for you, comment TagBot fix on this issue.
I'll open a PR within a few hours, please be patient!
Triggering TagBot for merged registry pull request: https://github.com/JuliaRegistries/General/pull/121509
|
gharchive/issue
| 2024-12-19T19:25:48 |
2025-04-01T04:34:45.979639
|
{
"authors": [
"JuliaTagBot"
],
"repo": "kingaa/PartiallyObservedMarkovProcesses.jl",
"url": "https://github.com/kingaa/PartiallyObservedMarkovProcesses.jl/issues/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
373757752
|
whatanime.ga has moved to trace.moe
Hello, I'm (was) the owner of whatanime.ga
A few days ago, the .ga domain was taken down by freenom T_T,
so the site has moved to https://trace.moe
if you are still actively supporting this project, please update the URL, thank you m(_ _;)m
Thanks for let me know! 😃 I will update the URL.
|
gharchive/issue
| 2018-10-25T03:22:41 |
2025-04-01T04:34:45.981679
|
{
"authors": [
"kingdido999",
"soruly"
],
"repo": "kingdido999/atogatari",
"url": "https://github.com/kingdido999/atogatari/issues/20",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
163566269
|
Write Dockerfile that installs node & slackit
slackit already has a Dockerfile but it looks like it has a few issues. It might be easier to have a quick Dockerfile that installs node and then installs slackit, before running it.
Once this is up, we can continue with https://github.com/kingsreach/kingsreach-online/issues/23
Can't believe I accidentally called it slackit everywhere. Such a mungbeen 😧.
Done.
|
gharchive/issue
| 2016-07-03T15:14:17 |
2025-04-01T04:34:45.989486
|
{
"authors": [
"adamkdean"
],
"repo": "kingsreach/kingsreach-slackit",
"url": "https://github.com/kingsreach/kingsreach-slackit/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1627010371
|
Trouble with API calls - most certainly user error.
Hey mate,
First of all, thanks so much for this template and for all the work you've put in. I've got it mostly up and running on my machine. I'd say you can certainly preemptively mark this issue as 'user error' haha.
Basically, I'm just having trouble with my API calls. I apologize for reaching out, but I've been trying for a good few days and just cannot work it out!
Are you able to take a look at my screenshots below and let me know if I'm missing something obvious?
My env
My index.js
My getData.js
My console response:
and finally, my terminal response from my npm run dev (this all runs fine until the request happens).
Also, goes without saying, but if you can't be bothered looking at all this, just tell me to piss off and mark this as resolved haha! You've done enough just making the template.
Thank you for the kind words!
I just looked into this; It's partly my fault too that I ignored how the REST client actually works. I've pushed a new commit that fixes how REST client is created and it works now. The issue on your end was you weren't passing the path correctly. Here's the fix for that:
userRoutes.get("/api/rest-test", async (req, res) => {
const {client} = await clientProvider.restClient({ isOnline: false, req, res });
const data = await client.get({ path: "blogs" });
console.log(data.body.blogs)
res.status(200).send({ blogs: data.body.blogs });
});
Oh wow! Thanks so much for the quick response. I'm somewhat glad that is wasn't 100% user error haha.
I'll give this a go after work!
Thanks again for everything!
|
gharchive/issue
| 2023-03-16T09:07:01 |
2025-04-01T04:34:45.994752
|
{
"authors": [
"kinngh",
"ohermans1"
],
"repo": "kinngh/shopify-node-express-mongodb-app",
"url": "https://github.com/kinngh/shopify-node-express-mongodb-app/issues/66",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
520073628
|
packet: Use Route53 FQDN attribute instead of name
The name attribute only refers to the name used to create the DNS record. The values accepted are either the subdomain to be created on the dns_zone or the FQDN.
For example, if you do something like:
resource "aws_route53_record" "example" {
zone_id = "xxx" # let's say it is the zone id for awesome.lokomotive-k8s.com
name = "deleteme"
type = "A"
ttl = 300
# private IPv4 address for etcd
records = ["10.0.0.1"]
}
output "name" {
value = "${aws_route53_record.example.name}"
}
output "fqdn" {
value = "${aws_route53_record.example.fqdn}"
}
You will see:
fqdn = deleteme.awesome.lokomotive-k8s.com
name = deleteme
If you use the FQDN for the name attribute in the example resource, however, the name and fqdn will match. This is what we are currently doing, for example here:
https://github.com/kinvolk/lokomotive-kubernetes/blob/da1cdcfaec23acd727193f53b45ef40753960534/packet/flatcar-linux/kubernetes/controllers.tf#L8
However, the correct and reliable way to get the FQDN of a domain is to get the fqdn attribute, and not the name.
This patch just simply changes to that, so we use what we really want and not depend on how the record was created. Although in this case it is a no-op.
@invidian the name is whatever you used for the name, if you included the rest of the FQDN then it has it, if no, it doesn't.
For example:
resource "aws_route53_record" "example" {
zone_id = "xxx" # let's say it is the zone id for awesome.lokomotive-k8s.com
name = "deleteme"
type = "A"
ttl = 300
# private IPv4 address for etcd
records = ["10.0.0.1"]
}
output "name" {
value = "${aws_route53_record.example.name}"
}
output "fqdn" {
value = "${aws_route53_record.example.fqdn}"
}
You will see:
fqdn = deleteme.awesome.lokomotive-k8s.com
name = deleteme
Therefore, it will not be the FQDN. We happen to use the FQDN for the name (see https://github.com/kinvolk/lokomotive-kubernetes/blob/da1cdcfaec23acd727193f53b45ef40753960534/packet/flatcar-linux/kubernetes/controllers.tf#L8), therefore is not an issue in our case, but it is still the proper fix.
I'll update the commit message to be more clear about this
@invidian updated the PR comment. PTAL :)
It seems that AWS provider has a mess regarding this thing.
What is the mess, exactly?
The docs seem quite clear that the fqdn attr is the way to know the fqdn of the record, IMHO.
There is a lot of bug requests about the behavior of those 2 properties as far as I saw. I went trough them and trough documentation and honestly I'm still not sure if I should put the FQDN in name field or not without testing :smile:
haha, was not aware of that. It makes even more sense to use the guaranteed and documented way to get the fqdn, then :-)
|
gharchive/pull-request
| 2019-11-08T15:22:23 |
2025-04-01T04:34:46.001272
|
{
"authors": [
"invidian",
"rata"
],
"repo": "kinvolk/lokomotive-kubernetes",
"url": "https://github.com/kinvolk/lokomotive-kubernetes/pull/91",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
794370760
|
docs/quickstarts/aks: rephrase Kubernetes version note
AKS currently provides 1.18 (and other platforms 1.19). It's easy that
this note gets out of date so let's rephrase it so it's more generic.
fixed a typo in the commit message
fixed a typo in the commit message
|
gharchive/pull-request
| 2021-01-26T16:23:06 |
2025-04-01T04:34:46.002701
|
{
"authors": [
"iaguis"
],
"repo": "kinvolk/lokomotive",
"url": "https://github.com/kinvolk/lokomotive/pull/1344",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2461994848
|
Q: Custom App + Access Token
Hi,
how can I use this gem if I only have the x-access token from a custom install like described here:
https://shopify.dev/docs/apps/build/authentication-authorization/access-tokens/generate-app-access-tokens-admin
The shopify_app expects an whole oAuth that I cant seem to get working in case of a single user custom app.
You'll need to create custom app for the store and then grab access token and API credentials from it.
These credentials will be sufficient to set api_key and secret in ShopifyApp initializer. And then you can add the following method to Shop model:
class Shop < ActiveRecord::Base
include ShopifyApp::ShopSessionStorageWithScopes
def self.system
new(
shopify_domain: "MYSHOPIFY_DOMAIN",
shopify_token: "API_ACCESS_TOKEN_FOR_CUSTOM_APP"
)
end
end
Using this method, you should be able to make API calls like this:
Shop.system.with_shopify_session do
GetOrder.call(id: order.shopify_gid)
end
Thank you for answering my Question. I made a worarkound by using a custom class:
class MyShop
def self.with_session(&block)
ShopifyAPI::Auth::Session.temp(shop: "xxx.myshopify.com",
access_token: "xxxxx-token", &block)
end
end
So I can use it that way:
MyShop.with_session do
GetOrder.call(id: order.shopify_gid)
end
Is this any better or worse? Since I didnt even really install the shoppify_app gem after I couldnt get it to work in first case, but just pulled it in with your gem.
Any comment is highly appreciated as I'd like to enhance the documentation as very many questions in the official forums seem to come from Auth with token isntead of o-Auth as you need token auth for every custom app that is not to be sold/ deployed.
Documented both approaches. Thanks for sharing!
There seems to be no way to authenticate custom apps via shopify_app gem. I end up spinning my own authentication with devise. However, shopify_app gem still might be helpful for handling webhooks, for example.
|
gharchive/issue
| 2024-08-12T22:26:27 |
2025-04-01T04:34:46.033090
|
{
"authors": [
"kbachl",
"kirillplatonov"
],
"repo": "kirillplatonov/shopify_graphql",
"url": "https://github.com/kirillplatonov/shopify_graphql/issues/45",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
715332473
|
Python packages requiring rebuilds
Send a PR bumping the second field in the version file.
[x] asciinema @mmatongo
[x] py-sip @sdsddsd1
[x] pyqt5 @sdsddsd1
[x] pyqtwebengine @sdsddsd1
[x] qemu @jedavies-dev
[x] qutebrowser @sdsddsd1
[x] youtube-dl @Vouivre
These packages require rebuild for Python 3.9.
Done
Thanks everyone.
|
gharchive/issue
| 2020-10-06T04:40:02 |
2025-04-01T04:34:46.041112
|
{
"authors": [
"dylanaraps",
"mmatongo"
],
"repo": "kisslinux/community",
"url": "https://github.com/kisslinux/community/issues/1355",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
468803789
|
just --build=x86_64-pc-linux-musl
It can handle --host and --target automatically
Tested and works on my machine. Merged.
|
gharchive/pull-request
| 2019-07-16T18:35:59 |
2025-04-01T04:34:46.043275
|
{
"authors": [
"konimex",
"protonesso"
],
"repo": "kissx/packages",
"url": "https://github.com/kissx/packages/pull/29",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
100990618
|
Problem with new users
Hi,
So I got taskboard installed fine. Everything seems to work but when I created a user things went wrong.
When the new user logged in he doesn't see any taskboard. The user can't logout etc.
This is the console output:
GET http://----------.----.net/api/users/current 500 (Internal Server Error)(anonymous function) @ libs.min.js:507
q @ libs.min.js:503
g @ libs.min.js:500
(anonymous function) @ libs.min.js:532
q.$eval @ libs.min.js:546
q.$digest @ libs.min.js:543
q.$apply @ libs.min.js:547
g @ libs.min.js:501
u @ libs.min.js:505
w.onload @ libs.min.js:506
I'm running this on Apache, PHP/5.5.9, pdo_sqlite is loaded , AllowOverride is on etc.
I have read and done #3 and #7
The same problem is in your own demo version on your site http://taskboard.matthewross.me/demo/#/boards
When I create there a new user I got the same problem.
Thanks.
Christian
I just duplicated this, and it seems to be an issue when the user is created without a default board. I'll get it fixed soon!
|
gharchive/issue
| 2015-08-14T11:06:30 |
2025-04-01T04:34:46.047385
|
{
"authors": [
"CIJeu",
"kiswa"
],
"repo": "kiswa/TaskBoard",
"url": "https://github.com/kiswa/TaskBoard/issues/210",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1806865604
|
how to hot reload sql functions
when develop, i add a new function in the sql file
-- :name get-user-by-id :? :1
-- :doc returns a user object by id, or nil if not present
SELECT *
FROM user_profile
WHERE id = :id;
then, how can i hot reload the sql file, and exec the function?
everytime when i add a new sql function , i should re-run (def query-fn (:db.sql/query-fn state/system))
It would be better to use a (defn query-fn [] (:db.sql/query-fn state/system)), this way it'll grab the fresh state of the system when it's called.
Yes, that solves this problem,thanks
|
gharchive/issue
| 2023-07-17T03:06:07 |
2025-04-01T04:34:46.049230
|
{
"authors": [
"vinurs",
"yogthos"
],
"repo": "kit-clj/kit",
"url": "https://github.com/kit-clj/kit/issues/106",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
999442684
|
Revert "Add the enter key support in grid block to create the default text bl…"
Reverts kitconcept/volto-blocks-grid#25
@iFlameing I ran the acceptance tests twice and they still fail. Is it possible that we are missing another commit here to amend the Cypress tests?
|
gharchive/pull-request
| 2021-09-17T14:51:51 |
2025-04-01T04:34:46.055811
|
{
"authors": [
"tisto"
],
"repo": "kitconcept/volto-blocks-grid",
"url": "https://github.com/kitconcept/volto-blocks-grid/pull/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.