id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
379508292
|
Elaborate the diffs between build and cli
This fixes #137
It's a good start.
|
gharchive/pull-request
| 2018-11-11T10:09:01 |
2025-04-01T04:35:29.908555
|
{
"authors": [
"jessemonroy650",
"purplecabbage"
],
"repo": "phonegap/phonegap-docs",
"url": "https://github.com/phonegap/phonegap-docs/pull/305",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
276718236
|
Compiler warnings (mingw)
When compiling the code (version 0.20.0) using msys/mingw(32 bit) on windows 7 (64 bit pro) I get the following warnings:
string/convert.c: In function `pdbl_to_wcs':
string/convert.c:164: warning: passing arg 2 of `swprintf' makes pointer from integer without a cast
string/string.c: In function `pwcscatchar':
string/string.c:884: warning: passing arg 2 of `swprintf' makes pointer from integer without a cast
string/utf8.c:33: warning: large integer implicitly truncated to unsigned type
string/utf8.c:34: warning: large integer implicitly truncated to unsigned type
string/utf8.c:34: warning: large integer implicitly truncated to unsigned type
string/utf8.c:35: warning: large integer implicitly truncated to unsigned type
string/utf8.c: In function `u8_toutf8':
string/utf8.c:235: warning: comparison is always true due to limited range of data type
string/utf8.c:242: warning: comparison is always true due to limited range of data type
string/utf8.c: In function `u8_wc_toutf8':
string/utf8.c:268: warning: comparison is always true due to limited range of data type
string/utf8.c:274: warning: comparison is always true due to limited range of data type
string/utf8.c: In function `u8_escape_wchar':
string/utf8.c:468: warning: comparison is always false due to limited range of data type
string/utf8.c:470: warning: comparison is always true due to limited range of data type
libtool: warning: undefined symbols not allowed in i686-pc-mingw32 shared libraries; building static only
The swprintf problem can be resolved by adding the following snippet to both string.c and convert.c
#if _WIN32
#define swprintf _snwprintf
#endif
The utf8.c related warnings have to do with the size of wchar_t on windows (unsigned 16 bit integer). On Linux wchar_t is defined as a 32 bit integer which means that utf8.c will most likely compile without a problem on Linux.
The author you got the utf8 - code from has this to say about his utf8 library:
I now use and recommend utf8proc instead of this library.
Hi @FreeBASIC-programmer, thank your for submitting this issue.
After setting up a test environment (Win 10, latest MSYS), I already failed on compiling any file because wctype.h is not available in the system. Did you got the same problem before?
Removing the include gets me to the same result as you reported, as it can be seen in the attached screenshot.
The binaries can't be linked due missing unicode functions.
It seems that the UNICODE/UTF8 detection does not work properly. Some more tests need to be done.
-- Jan
I did not get a problem when compiling the code other than the warnings as mentioned in my previous post (mingw created a static library called libphorward.a).
I am using the following (very old) version of mingw (which includes wctypes.h)
$ gcc --version
gcc.exe (GCC) 3.4.5 (mingw-vista special r3)
Copyright (C) 2004 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
The msys version I am using is msys 1.0.11.
Hey, I did some restructuring of the dependencies in configure.ac, and successfully tested it with the latest MSYS on a 32-Bit Windows 8.1. The mingw-gcc that I use is 6.3.0.
Can you please have a try on 9a236f323bd74519c204de3c044324b9d9df5536?
I compiled 9a236f3 and got the same warnings.
OK now I re-tested it with the same version that you have and was able to reproduce your warnings. I now cleared the warnings by doing some pre-processor switchting of code, but I am sure that there a still many bugs in the related functions on 32-bit platforms (or platforms with a 16-Bit wchar_t).
Maybe, the wide-character support will be removed in a future version due lacking of usage and portability. I think, writing tests would be a good option, but then for all cases and platforms.
The issue should be fixed by commit 3dd05c1defd26a2b0ca289bc1b7c953d880b53ad.
|
gharchive/issue
| 2017-11-25T02:01:08 |
2025-04-01T04:35:29.916654
|
{
"authors": [
"FreeBASIC-programmer",
"phorward"
],
"repo": "phorward/phorward",
"url": "https://github.com/phorward/phorward/issues/1",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
98634626
|
Changing text property of tinted BitmapText doesn't use correct glyphs
Add a BitmapText to the game and give it a tint:
var game = new Phaser.Game(320, 200, Phaser.AUTO, '', {preload: preload, create: create}),
text;
function preload() {
game.load.bitmapFont('desyrel', 'desyrel.png', 'desyrel.xml');
}
function create() {
text = game.add.bitmapText(0, 0, 'desyrel', 'bar', 64);
text.tint = 0x55ff55;
}
Once the page loads, type this in the console:
text.text = 'foobar';
Result:
As a workaround, you can type this in the console after setting the text:
text.removeChildren();
text.purgeGlyphs();
Correct result:
I'm using Phaser 2.4.2 and Chromium 43.0.2357.130 Ubuntu 14.04 (64-bit).
I think text is read only, and you should use text.setText('foobar') instead. http://phaser.io/docs/2.3.0/Phaser.BitmapText.html#setText
Same result with setText().
What about calling .updateText() after you set it?
.updateText() is called by the text property setter. In other words, nope.
Actually I think it's directly related to this: https://github.com/photonstorm/phaser/commit/9362a2b1f480ef570c2a5a05e2fceec03e169262
This is now fixed in a dev - and didn't require anything as intense as any of the suggestions above thankfully :)
Oh, hah. That's much simpler indeed. Thanks for the speedy fix!
I still get it in 2.6.1 albeit not as frequent. I first had the setText outside of the phaser game loop but even after moving it into the game loop, it happens from time to time.
|
gharchive/issue
| 2015-08-02T20:28:45 |
2025-04-01T04:35:29.923983
|
{
"authors": [
"Cristy94",
"DDowell",
"Pajamaman",
"photonstorm"
],
"repo": "photonstorm/phaser",
"url": "https://github.com/photonstorm/phaser/issues/1969",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
166574444
|
KeyCode class not listed in docs
This Issue is about
An error in the documentation
I do not find the KeyCode class in the Input namespace docs:
http://phaser.io/docs/2.6.1/index#input
Maybe it does not appear because the KeyCode definition is tagged as @namespace but not as @class. I can do a pull request with the change.
It's here http://phaser.io/docs/2.6.1/Phaser.KeyCode.html which is linked from the Key class: http://phaser.io/docs/2.6.1/Phaser.Key.html
It doesn't appear on the index page though. I could add it I guess? That page is generated manually, not from a script.
Yes sure, you can add it to the index manually.
Looking into the local docs (I guess those are generated by a script), in the namespace menu there is the items Phaser, KeyCode and PIXI (I think KeyCode should not appear there, out of Phaser.Input namespace). And in looking in the Input namespace menu, the KeyCode does not appear. So I guess to declare it as a class still a good solution.
KeyCode shouldn't appear in the namespace list, that is true. But the Input menu is manually created.
Manually added even in the local docs? Anyway... what's the options here then?
To change the docs of the phaser site manually (you only has access to that).
And change the local docs index manually too (I can do this).
Yes there are no automatic drop-downs generated in the local docs other than 'global' (and 'namespace' iirc), they are all hard-coded into the html template which is in the resources folder in the repo.
Ok thanks, let's go manually then.
We can change it to avoid it appearing in the namespace list. But it still needs manually correcting to solve those other 2 places too.
I see, well, I think there is some tooling behind it, docstrap or something like that. I can setup the environment or maybe it is more easy for you just to fix it next time you need to update the templates. I just do not want to take more of your time.
Ok I've updated it so it no longer appears in the Namespace list, and now appears in the Input drop-down in the generated docs.
|
gharchive/issue
| 2016-07-20T12:56:04 |
2025-04-01T04:35:29.930417
|
{
"authors": [
"boniatillo-com",
"photonstorm"
],
"repo": "photonstorm/phaser",
"url": "https://github.com/photonstorm/phaser/issues/2653",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1622470434
|
🛑 phpMyFAQ API is down
In f47265a, phpMyFAQ API (https://api.phpmyfaq.de/versions) was down:
HTTP code: 0
Response time: 0 ms
Resolved: phpMyFAQ API is back up in 579f4c5.
|
gharchive/issue
| 2023-03-13T23:45:16 |
2025-04-01T04:35:30.086362
|
{
"authors": [
"thorsten"
],
"repo": "phpMyFAQ/status.phpmyfaq.de",
"url": "https://github.com/phpMyFAQ/status.phpmyfaq.de/issues/4410",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1625432166
|
🛑 phpMyFAQ Homepage is down
In 04f4cba, phpMyFAQ Homepage (https://www.phpmyfaq.de) was down:
HTTP code: 0
Response time: 0 ms
Resolved: phpMyFAQ Homepage is back up in 7fb0187.
|
gharchive/issue
| 2023-03-15T12:27:55 |
2025-04-01T04:35:30.088700
|
{
"authors": [
"thorsten"
],
"repo": "phpMyFAQ/status.phpmyfaq.de",
"url": "https://github.com/phpMyFAQ/status.phpmyfaq.de/issues/4445",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1889307144
|
🛑 phpMyFAQ API is down
In ab27dce, phpMyFAQ API (https://api.phpmyfaq.de/versions) was down:
HTTP code: 0
Response time: 0 ms
Resolved: phpMyFAQ API is back up in 00a8ce0 after 11 minutes.
|
gharchive/issue
| 2023-09-10T20:06:16 |
2025-04-01T04:35:30.091011
|
{
"authors": [
"thorsten"
],
"repo": "phpMyFAQ/status.phpmyfaq.de",
"url": "https://github.com/phpMyFAQ/status.phpmyfaq.de/issues/5831",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2031432474
|
🛑 phpMyFAQ API is down
In c28a5d7, phpMyFAQ API (https://api.phpmyfaq.de/versions) was down:
HTTP code: 0
Response time: 0 ms
Resolved: phpMyFAQ API is back up in 3abb04b after 13 minutes.
|
gharchive/issue
| 2023-12-07T19:59:34 |
2025-04-01T04:35:30.093294
|
{
"authors": [
"thorsten"
],
"repo": "phpMyFAQ/status.phpmyfaq.de",
"url": "https://github.com/phpMyFAQ/status.phpmyfaq.de/issues/6600",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2069275710
|
🛑 phpMyFAQ API is down
In ca6f9d3, phpMyFAQ API (https://api.phpmyfaq.de/versions) was down:
HTTP code: 0
Response time: 0 ms
Resolved: phpMyFAQ API is back up in 7a306d0 after 5 minutes.
|
gharchive/issue
| 2024-01-07T20:26:42 |
2025-04-01T04:35:30.095980
|
{
"authors": [
"thorsten"
],
"repo": "phpMyFAQ/status.phpmyfaq.de",
"url": "https://github.com/phpMyFAQ/status.phpmyfaq.de/issues/6922",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
35950448
|
Raise test coverage to at least 80%
[x] 10%
[x] 20%
[x] 30%
[x] 40%
[x] 50%
[x] 60%
[ ] 70%
[ ] 80%
[ ] Make test suite more portable between php versions (use @group for specific versions)
A mayignore group was added to ignore redundant tests on travis.
|
gharchive/issue
| 2014-06-18T02:48:31 |
2025-04-01T04:35:30.101799
|
{
"authors": [
"c9s",
"marcioAlmada"
],
"repo": "phpbrew/phpbrew",
"url": "https://github.com/phpbrew/phpbrew/issues/253",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
225461597
|
Rounding error
To Reproduce:
bcround(0.000006, 5);
Expected: 0.00001
Actual: 0.00000
Confirmed that it is fixed. Huzzah!
|
gharchive/issue
| 2017-05-01T16:01:52 |
2025-04-01T04:35:30.105539
|
{
"authors": [
"hopeseekr"
],
"repo": "phpexpertsinc/MoneyType",
"url": "https://github.com/phpexpertsinc/MoneyType/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1596455768
|
Custom arrows > wrong condition cause a bug
I put the wrong condition here, we have declared 3 variables and assigned the default value for them ( adjustedBeginIndex, adjustedEndIndex, selectedCellIndex). But in the condition to set the disabled attribute for each arrow, I compared the dummy condition (slidePosition === 0 || selectedCellIndex === adjustedBeginIndex), and it always runs this code.
This PR will not be merged due to low code quality, lack of comments, and unclear changes.
|
gharchive/pull-request
| 2023-02-23T08:52:42 |
2025-04-01T04:35:30.206545
|
{
"authors": [
"phucbm",
"vuquangpham"
],
"repo": "phucbm/flickity-responsive",
"url": "https://github.com/phucbm/flickity-responsive/pull/18",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
80629425
|
fixed /proc/kmsg: permission denied in syslog-ng
I've just fixed default syslog-ng config and a minor sshd fix.
ping, I'm having this error too.
cc @FooBarWidget
@sokoow This PR has merge conflicts, please resolve those conflicts and let me know so I can merge your changes in! :)
Is there an existing issue that this PR is in reference to? If not, please kindly open a new issue which explains what you are trying to accomplish with these changes.
The syslog part of this PR seems already addressed by #382.
I would suggest to close this PR and open new PR with the sshd part of this PR (if still applicable - I do not use sshd).
@lucafavatella Thank you for the heads-up!
|
gharchive/pull-request
| 2015-05-25T18:47:22 |
2025-04-01T04:35:30.212920
|
{
"authors": [
"Theaxiom",
"lucafavatella",
"pragmaticivan",
"sokoow"
],
"repo": "phusion/baseimage-docker",
"url": "https://github.com/phusion/baseimage-docker/pull/223",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1650020987
|
add support for arrow key navigation / dpad navigation
By adding features that allow users to jump to the next focusable element with arrow keys or a TV dpad, and confirm with a single click of the enter button, would allow to use flutter_inappwebview on Android TV devices and improve usability on desktop devices
I also have this need
It seems it's impossible to focus inside webview programatically to use navigation #1974
Although when pass the keys you need it works fine. Like this https://github.com/gynzy/flutter-webview-arrow-keys/blob/main/lib/main.dart#L17
|
gharchive/issue
| 2023-03-31T20:59:16 |
2025-04-01T04:35:30.324852
|
{
"authors": [
"razfazz",
"wuchuixu",
"yamaha252"
],
"repo": "pichillilorenzo/flutter_inappwebview",
"url": "https://github.com/pichillilorenzo/flutter_inappwebview/issues/1607",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2334494940
|
[Snyk] Security upgrade typeorm from 0.2.24 to 0.2.25
This PR was automatically created by Snyk using the credentials of a real user.
Snyk has created this PR to fix 1 vulnerabilities in the npm dependencies of this project.
Snyk changed the following file(s):
package.json
package-lock.json
Vulnerabilities that will be fixed with an upgrade:
Issue
Score
Prototype Pollution SNYK-JS-TYPEORM-590152
422
[!IMPORTANT]
Check the changes in this PR to ensure they won't cause issues with your project.
Max score is 1000. Note that the real score may have changed since the PR was raised.
This PR was automatically created by Snyk using the credentials of a real user.
Note: You are seeing this because you or someone else with access to this repository has authorized Snyk to open fix PRs.
For more information:
🧐 View latest project report
📜 Customise PR templates
🛠 Adjust project settings
📚 Read about Snyk's upgrade logic
Learn how to fix vulnerabilities with free interactive lessons:
🦉 Prototype Pollution
🎉 Snyk hasn't found any issues so far.
✅ security/snyk check is completed. No issues were found. (View Details)
✅ license/snyk check is completed. No issues were found. (View Details)
|
gharchive/pull-request
| 2024-06-04T22:43:14 |
2025-04-01T04:35:30.338407
|
{
"authors": [
"pickford-snyk"
],
"repo": "pickford-snyk/nodejs-goof",
"url": "https://github.com/pickford-snyk/nodejs-goof/pull/10",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1422991687
|
Update scala-library to 2.13.10
Updates org.scala-lang:scala-library from 2.13.0 to 2.13.10.
GitHub Release Notes - Version Diff
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scala-lang", artifactId = "scala-library" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "@monthly" },
dependency = { groupId = "org.scala-lang", artifactId = "scala-library" }
}]
labels: library-update, early-semver-patch, semver-spec-patch, commit-count:1
Superseded by #182.
|
gharchive/pull-request
| 2022-10-25T19:53:29 |
2025-04-01T04:35:30.344189
|
{
"authors": [
"scala-steward"
],
"repo": "picnicml/doddle-model",
"url": "https://github.com/picnicml/doddle-model/pull/169",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2574202297
|
Make it work on windows and mac
See discussion here: https://github.com/pieper/MD_E15/issues/7
Given that Slicer is going to be start from an application icon, we need:
Ask the user to provide the path to gh executable
And possibly invoke the gh auth login inside the MorphoDepot (this will ensure that gh is working. And probably users will find this a bit more convenient).
Another option would be to install gh for the user so that we know the path.
Jc packaged up s5cmd as a pip-installable package this way and it's worked great for the idc-index project. FYI, once there are many scans with complex metadata it could make sense to have something akin to idc-index for searching and downloading.
And yes, I find the gh auth process a bit of a chore that may confuse new users but I'm not sure how much we should try to automate since they could change the procedure at any point. Invoking the command is probably good, but passwords 2fa, and an copying of codes will have to happen manually.
Another option would be to install gh for the user so that we know the path.
Does it work on all platforms?
We can worry less about gh auth if gh is available from get go...
Another option would be to install gh for the user so that we know the path.
Yes, that would be great. In MorphoCloud, I will have the gh CLI available from get go. So that will be less of a concern...
Yes, it would be available on all platforms. We can just copy what Jc did, or maybe just ask him to do it since he's already familiar with the details.
As a side comment, Jc and I discussed making a python package called something like 'slicer-runner' that would use pip to get Slicer or launch existing installations. We could do that also for SlicerMorph users could easily download and run and manage versions using pip. This would make python, which any computational scientist should have anyway, the key requirement instead of having different directions for all OS platforms.
I can't seem to create labels on your repo. I wanted to bump this up on the priority, because some people do indeed want to run MorphoDepotReview on their own computers, and most of them have Macs or Windows, not linux.
|
gharchive/issue
| 2024-10-08T20:59:27 |
2025-04-01T04:35:30.364413
|
{
"authors": [
"muratmaga",
"pieper"
],
"repo": "pieper/SlicerMorphoDepot",
"url": "https://github.com/pieper/SlicerMorphoDepot/issues/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1199440342
|
Cannot install torch 1.7.1 with cuda 11.1
When I tried following installation instructions in section 2.4, when i run pip install -U torch==1.7.1 torchvision==0.8.2 -f https://download.pytorch.org/whl/cu111/torch_stable.html it installs cuda 10.2 by default.
When I tried following installation instructions in section 2.4, when i run pip install -U torch==1.7.1 torchvision==0.8.2 -f https://download.pytorch.org/whl/cu111/torch_stable.html it installs cuda 10.2 by default.
Is CUDA 10.2 your default version on your computer? Quite honestly, the CUDA version does not really matter, as long as detectron2 builds from source and you can access your GPU as computing device in your virtual environment. See steps 2.6 and 2.7 in the installation instruction.
My cuda version is 11.1 by default. And it started training!. Thank you for your hardwork. This software is really amazing!
|
gharchive/issue
| 2022-04-11T06:07:58 |
2025-04-01T04:35:30.374530
|
{
"authors": [
"pieterblok",
"vinit13792"
],
"repo": "pieterblok/maskal",
"url": "https://github.com/pieterblok/maskal/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
462725855
|
Remove cluster resource from example list of resource types
The README mentions customresourcedefinition as example of namespaced resource to backup. This pull request removes it from that list.
:100: :+1:
|
gharchive/pull-request
| 2019-07-01T13:49:48 |
2025-04-01T04:35:30.375549
|
{
"authors": [
"Overv",
"pieterlange"
],
"repo": "pieterlange/kube-backup",
"url": "https://github.com/pieterlange/kube-backup/pull/54",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2066644805
|
🛑 Stripe Webhook URL (Test and Live) is down
In ad37f37, Stripe Webhook URL (Test and Live) ($STRIPEWEBHOOK) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Stripe Webhook URL (Test and Live) is back up in f38a2b4 after 12 minutes.
|
gharchive/issue
| 2024-01-05T02:46:34 |
2025-04-01T04:35:30.411319
|
{
"authors": [
"cybertheory"
],
"repo": "pikfarm/PikfarmStatus",
"url": "https://github.com/pikfarm/PikfarmStatus/issues/2940",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2087546801
|
🛑 www.pik.farm is down
In ce1f1d5, www.pik.farm (https://pik.farm) was down:
HTTP code: 0
Response time: 0 ms
Resolved: www.pik.farm is back up in 1c0fd89 after 12 minutes.
|
gharchive/issue
| 2024-01-18T05:41:57 |
2025-04-01T04:35:30.413779
|
{
"authors": [
"cybertheory"
],
"repo": "pikfarm/PikfarmStatus",
"url": "https://github.com/pikfarm/PikfarmStatus/issues/4739",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2089278171
|
🛑 www.pik.farm is down
In 97737b4, www.pik.farm (https://pik.farm) was down:
HTTP code: 0
Response time: 0 ms
Resolved: www.pik.farm is back up in 2bc162d after 55 minutes.
|
gharchive/issue
| 2024-01-18T23:36:56 |
2025-04-01T04:35:30.416209
|
{
"authors": [
"cybertheory"
],
"repo": "pikfarm/PikfarmStatus",
"url": "https://github.com/pikfarm/PikfarmStatus/issues/4844",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2104249094
|
🛑 Stripe Webhook URL (Test and Live) is down
In 9bc83b8, Stripe Webhook URL (Test and Live) ($STRIPEWEBHOOK) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Stripe Webhook URL (Test and Live) is back up in 85449c0 after 12 minutes.
|
gharchive/issue
| 2024-01-28T16:40:03 |
2025-04-01T04:35:30.418471
|
{
"authors": [
"cybertheory"
],
"repo": "pikfarm/PikfarmStatus",
"url": "https://github.com/pikfarm/PikfarmStatus/issues/6129",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
438255315
|
Forced Mode feature addition for #2
This adds the ability to run the bmp280 in "forced" mode, which requests a single conversion from the sensor after which it returns to sleep mode.
If you're using a large delay between temperature readings in your code, this should help to eliminate any temperature offset caused by the internal temperature of the sensor itself.
The setup method now also accepts temperature and pressure oversampling and temperature standby (the time between measurements in normal mode) settings.
Pull Request Test Coverage Report for Build 14
0 of 0 changed or added relevant lines in 0 files are covered.
9 unchanged lines in 1 file lost coverage.
Overall coverage decreased (-3.3%) to 91.743%
Files with Coverage Reduction
New Missed Lines
%
bmp280/init.py
9
91.74%
Totals
Change from base Build 11:
-3.3%
Covered Lines:
100
Relevant Lines:
109
💛 - Coveralls
|
gharchive/pull-request
| 2019-04-29T10:48:20 |
2025-04-01T04:35:30.468245
|
{
"authors": [
"Gadgetoid",
"coveralls"
],
"repo": "pimoroni/bmp280-python",
"url": "https://github.com/pimoroni/bmp280-python/pull/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
123384200
|
TypeError: cannot use a string pattern on a bytes-like object
When trying to run install.py, i get this error:
Traceback (most recent call last):
File "./install.py", line 109, in <module>
main()
File "./install.py", line 81, in main
verify_requirements()
File "./install.py", line 61, in verify_requirements
m = re.search('(\d).(\d)', version_text)
File "/usr/lib/python3.5/re.py", line 173, in search
return _compile(pattern, flags).search(string)
TypeError: cannot use a string pattern on a bytes-like object
On my system, when using pipes, i do sometimes get this weird error message (line 2 below). I haven't been able to figure out the cause, but i haven't seen any side effects from it before now, so i don't know if it is the problem here:
$ bash --version | head -1
bash: child setpgid (20301 to 20297): Operation not permitted
GNU bash, version 4.3.42(1)-release (x86_64-unknown-linux-gnu)
Here's what i get in the REPL:
$ python
Python 3.5.0 (default, Sep 20 2015, 11:28:25)
[GCC 5.2.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import subprocess
>>> subprocess.Popen("bash --version | head -1", shell=True, stdout=subprocess.PIPE).stdout.read()
b'GNU bash, version 4.3.42(1)-release (x86_64-unknown-linux-gnu)\n'
It works if i change line 60 to set version_text manually:
version_text = "GNU bash, version 4.3.42(1)-release (x86_64-unknown-linux-gnu)"
hmm, yes. the first line in bash --version is expected to be the version string. But in your case there is an error(or warning, I'm not sure):
bash: child setpgid (20301 to 20297): Operation not permitted
but that error message goes to stderr, not stdout:
$ bash --version | head -1 >/dev/null
bash: child setpgid (9003 to 8999): Operation not permitted
$ (bash --version | head -1) 2>/dev/null
GNU bash, version 4.3.42(1)-release (x86_64-unknown-linux-gnu)
so install.py shouldn't be reading in that line, right?
I got the same issue. It happens because you use a string pattern on a bytes object. Replace line 60 with following line will fix this issue:
m = re.search(b'(\d).(\d)', version_text)
I fixed this error by using Python 2.
I just got this issue too
|
gharchive/issue
| 2015-12-21T23:54:09 |
2025-04-01T04:35:30.497839
|
{
"authors": [
"NightMachinary",
"X1011",
"dessalines",
"et2010",
"pindexis"
],
"repo": "pindexis/marker",
"url": "https://github.com/pindexis/marker/issues/13",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
388195876
|
charset: make collation case insensitive (#66)
cherry-pick from #66
LGTM
|
gharchive/pull-request
| 2018-12-06T12:38:46 |
2025-04-01T04:35:30.544161
|
{
"authors": [
"winkyao",
"zz-jason"
],
"repo": "pingcap/parser",
"url": "https://github.com/pingcap/parser/pull/70",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
609648844
|
UCP: Reduce the unnecessary updateStoreStatusLocked call counts in heartbeat process
Description
Score
200
Mentor
@nolouch
Recommended Skills
go
Learning Materials
Ref #1846
/pick-up-challenge
thanks!
/pick-up-challenge
@mantuliu try again please~
/pick-up-challenge
Is this issue already finished? @mantuliu
@nolouch already finished.
|
gharchive/issue
| 2020-04-30T07:02:48 |
2025-04-01T04:35:30.547616
|
{
"authors": [
"mantuliu",
"nolouch",
"you06"
],
"repo": "pingcap/pd",
"url": "https://github.com/pingcap/pd/issues/2395",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2003657204
|
fips: tikv heap profiling failed after enable tls and set require_secure_transport = on
Bug Report
Please answer these questions before submitting your issue. Thanks!
What did you do?
What did you expect to see?
No error, works fine.
What did you see instead?
Profile Error
failed to fetch and write to temp file: failed to fetch profile with *.prof format: failed to fetch tikv heap profile: Use of uninitialized value $line in substitution (s///) at /dev/stdin line 3323. http://tikv1-peer:20180/debug/pprof/symbol doesn't exist
What version of TiDB Dashboard are you using (./tidb-dashboard --version)?
master
/assign @mornyx
/severity major
TiKV heap profile depends on jeprof which does not support mTLS yet.
https://github.com/pingcap/tidb-dashboard/blob/706489ff610dcccc238335b26e0bc97cbf486c6e/pkg/apiserver/profiling/fetcher.go#L78
|
gharchive/issue
| 2023-11-21T07:48:31 |
2025-04-01T04:35:30.554797
|
{
"authors": [
"mayjiang0203",
"overvenus"
],
"repo": "pingcap/tidb-dashboard",
"url": "https://github.com/pingcap/tidb-dashboard/issues/1613",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2654663265
|
feat: openai-like embedding model support
Support OpenAI-Like embedding models, one of them is ZhipuAI. https://github.com/pingcap/tidb.ai/issues/373
|
gharchive/pull-request
| 2024-11-13T08:32:10 |
2025-04-01T04:35:30.580804
|
{
"authors": [
"Icemap"
],
"repo": "pingcap/tidb.ai",
"url": "https://github.com/pingcap/tidb.ai/pull/375",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1320406808
|
Add benchmark for collation
What problem does this PR solve?
Issue Number: close #xxx
Problem Summary:
What is changed and how it works?
Check List
Tests
[ ] Unit test
[ ] Integration test
[ ] Manual test (add detailed scripts or steps below)
[ ] No code
Side effects
[ ] Performance regression: Consumes more CPU
[ ] Performance regression: Consumes more Memory
[ ] Breaking backward compatibility
Documentation
[ ] Affects user behaviors
[ ] Contains syntax changes
[ ] Contains variable changes
[ ] Contains experimental features
[ ] Changes MySQL compatibility
Release note
None
/cc @windtalker /cc @ywqzzy
Better to add a test for like
/merge
/run-all-tests
/merge
|
gharchive/pull-request
| 2022-07-28T05:18:55 |
2025-04-01T04:35:30.671681
|
{
"authors": [
"JaySon-Huang",
"windtalker",
"xzhangxian1008",
"ywqzzy"
],
"repo": "pingcap/tiflash",
"url": "https://github.com/pingcap/tiflash/pull/5491",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2435590332
|
Storages: Fix obtaining incorrect column information when there are virtual columns in the query (release-8.1)
This is an manual cherry-pick of #9189
What problem does this PR solve?
Issue Number: close #9188
Problem Summary:
When parsing some column expressions, it may need to use column index to obtain column information, but since virtual columns are filtered out before being sent to the storage layer, the original columns and the columns for storage to read are
inconsistency.
What is changed and how it works?
1. Use the original columns in `query_info.dag_query` instead of `columns_to_read` when building `RSOperator`.
2. For runtime filters, creating the `DM::Attr` object in `StorageDeltaMerge::read`, so it doesn't need to rely on `column_to_read`.
Currently, if a query contains virtual columns, hash join will not be pushed down to TiFlash. So in fact, runtime filters will not occur when a query contains virtual columns. But in order to keep the relevant codes consistent, make them not rely on columns_to_read, the way that runtime filters creating DM::Attr object is modified.
Check List
Tests
[ ] Unit test
[x] Integration test
[ ] Manual test (add detailed scripts or steps below)
[ ] No code
Side effects
[ ] Performance regression: Consumes more CPU
[ ] Performance regression: Consumes more Memory
[ ] Breaking backward compatibility
Documentation
[ ] Affects user behaviors
[ ] Contains syntax changes
[ ] Contains variable changes
[ ] Contains experimental features
[ ] Changes MySQL compatibility
Release note
Fix obtaining incorrect column information when there are virtual columns in the query.
/run-all-tests
|
gharchive/pull-request
| 2024-07-29T14:30:59 |
2025-04-01T04:35:30.677369
|
{
"authors": [
"JaySon-Huang",
"JinheLin"
],
"repo": "pingcap/tiflash",
"url": "https://github.com/pingcap/tiflash/pull/9265",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1721579482
|
[DNM]: test ci on release-6.1
What problem does this PR solve?
Issue Number: close #xxx
What is changed and how it works?
Check List
Tests
Unit test
Integration test
Manual test (add detailed scripts or steps below)
No code
Questions
Will it cause performance regression or break compatibility?
Do you need to update user documentation, design documentation or monitoring documentation?
Release note
Please refer to [Release Notes Language Style Guide](https://pingcap.github.io/tidb-dev-guide/contribute-to-tidb/release-notes-style-guide.html) to write a quality release note.
If you don't think this PR needs a release note then fill it with `None`.
/run-all-tests
/debug cdc-integration-mysql-test
/debug cdc-integration-kafka-test
/debug cdc-integration-kafka-test
/debug cdc-integration-kafka-test
/debug cdc-integration-mysql-test
/debug all
/debug all
/run-dm-compatibility-tests
/run-dm-integration-tests
/debug dm-integration-test
/debug dm-integration-test
/debug dm-integration-test
/debug dm-integration-test
/debug all
/retest
/debug cdc-integration-kafka-test
/debug cdc-integration-mysql-test
/debug dm-integration-test
/debug cdc-integration-kafka-test
|
gharchive/pull-request
| 2023-05-23T09:03:14 |
2025-04-01T04:35:30.684738
|
{
"authors": [
"purelind"
],
"repo": "pingcap/tiflow",
"url": "https://github.com/pingcap/tiflow/pull/9025",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1741080293
|
metrics(ticdc): fix inaccurate kafka v2 batch time metrics.
What problem does this PR solve?
Issue Number: close #9129
What is changed and how it works?
Check List
Tests
Integration test
Questions
Will it cause performance regression or break compatibility?
Do you need to update user documentation, design documentation or monitoring documentation?
Release note
`None`
/test all
/test verify
/test dm-integration-test
/test dm-integration-test
|
gharchive/pull-request
| 2023-06-05T06:35:18 |
2025-04-01T04:35:30.688075
|
{
"authors": [
"3AceShowHand"
],
"repo": "pingcap/tiflow",
"url": "https://github.com/pingcap/tiflow/pull/9128",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
121662204
|
关于Browsersync
web跑起来之后,右上角一直会出现connect to browsersync和disconnect to browsersync
并且有时候会一直不断地自我刷新
这个是hexo-browsersync插件导致的吗,请问如何解决
这是一个文件更改后同步刷新的 Hexo 插件,通过 npm uninstall --save hexo-browsersync 卸载。
项目介绍:https://github.com/hexojs/hexo-browsersync
OK.
|
gharchive/issue
| 2015-12-11T08:34:12 |
2025-04-01T04:35:30.729412
|
{
"authors": [
"Lwxiang",
"pinggod"
],
"repo": "pinggod/hexo-theme-apollo",
"url": "https://github.com/pinggod/hexo-theme-apollo/issues/8",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1884319510
|
Updating for v9
Hi everyone!
Updating dependencies: I updated every dependency to the latest version, and tests are up and running. As @jsumners was mentioning, we could have a problem with deps switching to ESM-only, but to be honest I did not find a way of programmatically understanding if they are or not, if you have any suggestions please share!
Dropping Node <= 16: As this module is pretty lean, there were no flags like "for Node X support", but I am open to comments if you find something that you know that is not supported in 18 ( or better if there is a tool that does that )
Updating CI configuration: Here I improvised a little bit, as I never worked with GH Actions before. So probably there is a better and more correct way of writing the .yaml, or if there are additional steps to do.
Testing suite configuration (tap): I managed to switch to tap. Here again, suggestions and comments are more than welcomed
As it's my first PR I expect quite some things to refactor, so don't go easy on me!
Thank you for your time!
Nice! Should I close this and open another one, or can I use this for the CI change?
Close this one and open a new one with only the CI changes.
|
gharchive/pull-request
| 2023-09-06T16:00:44 |
2025-04-01T04:35:30.737425
|
{
"authors": [
"Ranieri93",
"jsumners"
],
"repo": "pinojs/json-colorizer",
"url": "https://github.com/pinojs/json-colorizer/pull/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
712834226
|
Using import aliases on static Java methods with reserved Kotlin keyword names trigger the Unused Import rule
Expected Behavior
Mockito contains a static Java method names when, which will be imported in Kotlin like:
import org.mockito.Mockito.`when`
Clearly, if you try to use an import alias to get rid of the when, for example like:
import org.mockito.Mockito.`when` as _when
Observed Behavior
When using ktlint version >= 0.38.0, the above import triggers the no-unused-import rule. For ktlint versions up to 0.37.2 this was working properly.
Steps to Reproduce
Simply do an import alias on a Java static method (or even a variable) which is named as a reserved Kotlin keyword (has been reproduced with Mockito.when and org.hamcrest.CoreMatchers.is) and run ktlint. Note that the actual alias name given does not seem to matter.
Your Environment
Version of ktlint used: 0.38.0, 0.39.0
Name and version (or code for custom task) of integration used (Gradle plugin, Maven plugin, command line, custom Gradle task): Gradle plugin: org.jlleitschuh.gradle:ktlint-gradle:9.4.0
Version of Gradle used (if applicable): 5.6.4
Operating System and version: Debian Linux Buster
Link to your project (if it's a public repository):
Should be fixed by #911, you can try a snapshot
I tried the latest snapshot (ktlint-0.39.0-SNAPSHOT) and it appears that the issue is still there. I have created a sample app that demonstrates the issue here.
oh, interesting, thanks for the sample 👍
This appears not to just be a static import of java methods but using import aliases in general. I was able to easily reproduce this.
package test.ktlint
import something as `It`
class Test {
fun test() {
It()
}
}
Problem seems not be reproducable in at least version 0.43.2. If I recall correctly we solve d a bug a while ago in which import statements were removed when sorting the imports. Please reopen, if problem pop ups again.
|
gharchive/issue
| 2020-10-01T13:15:53 |
2025-04-01T04:35:30.755047
|
{
"authors": [
"AndrewReitz",
"gmetal",
"paul-dingemans",
"romtsn"
],
"repo": "pinterest/ktlint",
"url": "https://github.com/pinterest/ktlint/issues/937",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
230861126
|
Depend on SwiftCheck when running tests
Add SwiftCheck as a test dependency so that when toJson lands, we can make a SwiftCheck test that does toJson(fromJson(arbitrarySchema)) == arbitrarySchema.
SwiftPM doesn't have a legit test dependency solution right now, but the strategy most SwiftPM authors are following right now is to define an environment variable of the form SWIFTPM_TEST_libname and conditioning on that in the Package.swift (see ReactiveSwift and Algebra)
🚫 CI failed with log
@bkase - Can you fix the lint errors?
⛔️ Line 7: Variable name should be between 3 and 40 characters long: 'x'
This lint rule is 👎
🚫 CI failed with log
🚫 CI failed with log
It works locally for me with make test, not sure why CI is failing?
Closing this for now, lets re-open when we get something passing with an updated version of this lib
|
gharchive/pull-request
| 2017-05-23T22:03:44 |
2025-04-01T04:35:30.759674
|
{
"authors": [
"bkase",
"pinterest-ios",
"rahul-malik"
],
"repo": "pinterest/plank",
"url": "https://github.com/pinterest/plank/pull/52",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1104553099
|
How to deploy ION to kubernetes?
Your environment.
Version: v1.10.0
Browser: latest chrome
Other Information -
What did you do?
I have deployed ION to Kubernetes, everything is working except for SFU unable to create peer connection, here is my deployment and service
apiVersion: apps/v1
kind: Deployment
metadata:
name: ion-sfu
spec:
selector:
matchLabels:
app: ion-sfu
replicas: 1
template:
metadata:
labels:
app: ion-sfu
spec:
imagePullSecrets:
- name: github-registry-key
containers:
- name: ion-sfu
image: pionwebrtc/ion:latest-sfu
imagePullPolicy: {{ .Values.image.pullPolicy }}
ports:
- containerPort: 3478
---
apiVersion: v1
kind: Service
metadata:
name: ion-sfu
spec:
type: ClusterIP
ports:
- protocol: TCP
port: 3478
targetPort: 3478
selector:
app: ion-sfu
---
once trying to get video streaming working, i'm getting the below in browser
and the below in ION-SFU pod logs
[2022-01-15 06:22:51.406] [INFO] [peer.go:199] => PeerLocal got offer logger=ion-sfu-node/sfu peer_id=ahmed v=0
[2022-01-15 06:22:51.407] [DEBUG] [publisher.go:117] => ice connection status logger=ion-sfu-node/sfu state=checking v=1
[2022-01-15 06:22:51.407] [INFO] [peer.go:210] => PeerLocal send answer logger=ion-sfu-node/sfu peer_id=ahmed v=0
[2022-01-15 06:22:51.407] [DEBUG] [peer.go:165] => on publisher ice candidate called for peer logger=ion-sfu-node/sfu peer_id=ahmed v=1
[2022-01-15 06:22:51.407] DEBUG default: [service.go:145] [sfu.(*SFUService).Signal.func2] [S=>C] peer.OnIceCandidate: target = 0, candidate = candidate:706165708 1 udp 2130706431 45.79.247.128 5000 typ host
[2022-01-15 06:22:51.407] DEBUG default: [service.go:238] [sfu.(*SFUService).Signal] [S=>C] join.description: answer v=0
I also updated the sfu.toml config to have my cluster public ip
nat1to1 = ["public ip"]
is there a way to know what's missing in the config?
I found the solution, I just needed port 5000 also to be open, so if anyone else facing same issue, here is the correct deployment and service config:
apiVersion: apps/v1
kind: Deployment
metadata:
name: ion-sfu
spec:
selector:
matchLabels:
app: ion-sfu
replicas: 1
template:
metadata:
labels:
app: ion-sfu
spec:
imagePullSecrets:
- name: github-registry-key
containers:
- name: ion-sfu
image: pionwebrtc/ion:latest-sfu
imagePullPolicy: {{ .Values.image.pullPolicy }}
ports:
- containerPort: 3478
- containerPort: 5000
protocol: UDP
---
apiVersion: v1
kind: Service
metadata:
name: ion-sfu
spec:
type: ClusterIP
ports:
- protocol: TCP
name: tcp
port: 3478
targetPort: 3478
- protocol: UDP
name: udp
port: 5000
targetPort: 5000
selector:
app: ion-sfu
---
@leewardbound would you be able to have a look? Thank you so much
hey, so this is a familiar issue, i think we used to have some docs/examples about this but i'm not sure where they are right now --
the issue is related to how webrtc works, basically UDP ports need to be open (but not load balanced or mapped to different host ports). this is tricky on kubernetes. in config.toml, you'll see that we have some UDP port range configured, usually 5000-5200 is the default (which allows about 200 participants i believe?). when a webrtc client (like pion or ion-sfu) sends an offer/answer packet, it includes a list of suggested IPs and ports for the connection to be opened. you need more than just nat1to1={public_ip} in k8s, because a) this doesn't open the proper ports, and b) the public IP will actually depend on which node the pod is scheduled to run on, which you can't anticipate when you are setting up config.toml.
my preferred choice to workaround this is to use a DaemonSet instead of a Deployment (assuring 1 ion-sfu is scheduled on each node), and to use HostNetwork: true so that the pod is able to a) discover its own IP address, and b) open the 5000-5200 UDP ports, without listing them 1-by-1 in the Ports section as would normally be necessary in k8s, which doesn't support opening port ranges.
i've never deployed ion (the whole app) on k8s, just ion-sfu, but here's the config i've used as recently as ion-sfu:1.9, i haven't really updated my ion-sfu projects in a while:
---
apiVersion: v1
kind: Service
metadata:
name: sfu
labels:
name: sfu
annotations:
prometheus.io/scrape: "true"
prometheus.io/port: "9123"
spec:
type: ClusterIP
ports:
- name: grpc
port: 50051
targetPort: grpc
- name: jsonrpc
port: 7000
targetPort: jsonrpc
selector:
app: sfu
---
apiVersion: apps/v1
kind: DaemonSet
metadata:
name: sfu
spec:
selector:
matchLabels:
name: sfu
template:
metadata:
labels:
name: sfu
app: sfu
spec:
hostNetwork: true
dnsPolicy: ClusterFirstWithHostNet
terminationGracePeriodSeconds: 30
containers:
- name: sfu
image: pionwebrtc/ion-sfu:v1.9.4-allrpc
command: ["/usr/local/bin/sfu", "-c", "/configs/sfu.toml", "-gaddr", ":50051", "-jaddr", ":7000", "-maddr", ":9123"]
ports:
- name: grpc
containerPort: 50051
- name: jsonrpc
containerPort: 7000
- name: metrics
containerPort: 9123
@leewardbound Thank you so much, i tried deploying ion-sfu beside the rest of ion pods and didn't get much luck, maybe because both sfu app/container inside ion has different deplyoment requirements from the independent ion-sfu.
currently i'm getting the below 2 lines in both signal and sfu pods and trying to wrap my head around, hopefully tonight.
[2022-01-16 21:54:48.548] INFO nats-discovery.Client: [client.go:91] [client.(*Client).handleNatsMsg] node.delete: dc1.sfu-EhR9Kq
[2022-01-16 21:54:48.548] INFO default: [node.go:138] [ion.(*Node).handleNeighborNodes] Service down: rtc node id => [sfu-EhR9Kq]
one observation that the docker version that works locally has the following ports setup
sfu:
image: pionwebrtc/ion:latest-sfu
build:
dockerfile: ./docker/sfu.Dockerfile
context: .
volumes:
- "./configs/docker/sfu.toml:/configs/sfu.toml"
ports:
- "5000:5000/udp"
- 3478:3478
which means, theoretically at least, it can work with only 3478 and 5000 if configured correctly. That's why i'm trying to get the following service ports to work:
apiVersion: v1
kind: Service
metadata:
name: sfu
spec:
type: ClusterIP
ports:
- protocol: TCP
name: tcp
port: 3478
targetPort: 3478
- protocol: UDP
name: udp
port: 5000
targetPort: 5000
selector:
app: sfu
I will continue digging and post my findings.
@Sean-Der @adwpc by chance do you have ideas around here?
Thank you everyone so much
@leewardbound BTW i'm running ION inside K8s but the client UI that trying to stream on localhost, do you know if there is a requirement to have the client UI also on a public IP in order for the ICE connection to work?
looking at the following offer and answer, what i can see both has 127.0.0.1 which i anticipate to be the issue, still tring to find ways around that
{
"type": "offer",
"sdp": "v=0\r\no=- 2711940938522023088 3 IN IP4 127.0.0.1\r\ns=-\r\nt=0 0\r\na=group:BUNDLE 0 1\r\na=extmap-allow-mixed\r\na=msid-semantic: WMS 3a8fffb3-cf06-44f8-a10c-144e209991bb\r\nm=application 57934 UDP/DTLS/SCTP webrtc-datachannel\r\nc=IN IP4 24.150.74.40\r\na=candidate:3194784207 1 udp 2122262783 2001:1970:5692:c200:bc83:7b6d:313e:f371 61453 typ host generation 0 network-id 2 network-cost 10\r\na=candidate:1511920713 1 udp 2122194687 192.168.0.2 57934 typ host generation 0 network-id 1 network-cost 10\r\na=candidate:4081163164 1 udp 1685987071 24.150.74.40 57934 typ srflx raddr 192.168.0.2 rport 57934 generation 0 network-id 1 network-cost 10\r\na=candidate:4042224959 1 tcp 1518283007 2001:1970:5692:c200:bc83:7b6d:313e:f371 9 typ host tcptype active generation 0 network-id 2 network-cost 10\r\na=candidate:345893049 1 tcp 1518214911 192.168.0.2 9 typ host tcptype active generation 0 network-id 1 network-cost 10\r\na=ice-ufrag:0/zJ\r\na=ice-pwd:+9L5W9JtXWOfH+NRU7kzFlHr\r\na=ice-options:trickle\r\na=fingerprint:sha-256 F6:D4:47:2E:60:11:71:0A:47:E4:66:B0:8E:AB:9B:67:03:BD:FB:56:0F:F7:93:A9:8E:8F:50:3B:A6:DB:64:36\r\na=setup:actpass\r\na=mid:0\r\na=sctp-port:5000\r\na=max-message-size:262144\r\nm=video 9 UDP/TLS/RTP/SAVPF 96\r\nc=IN IP4 0.0.0.0\r\na=rtcp:9 IN IP4 0.0.0.0\r\na=ice-ufrag:0/zJ\r\na=ice-pwd:+9L5W9JtXWOfH+NRU7kzFlHr\r\na=ice-options:trickle\r\na=fingerprint:sha-256 F6:D4:47:2E:60:11:71:0A:47:E4:66:B0:8E:AB:9B:67:03:BD:FB:56:0F:F7:93:A9:8E:8F:50:3B:A6:DB:64:36\r\na=setup:actpass\r\na=mid:1\r\na=extmap:1 urn:ietf:params:rtp-hdrext:toffset\r\na=extmap:2 http://www.webrtc.org/experiments/rtp-hdrext/abs-send-time\r\na=extmap:3 urn:3gpp:video-orientation\r\na=extmap:4 http://www.ietf.org/id/draft-holmer-rmcat-transport-wide-cc-extensions-01\r\na=extmap:5 http://www.webrtc.org/experiments/rtp-hdrext/playout-delay\r\na=extmap:6 http://www.webrtc.org/experiments/rtp-hdrext/video-content-type\r\na=extmap:7 http://www.webrtc.org/experiments/rtp-hdrext/video-timing\r\na=extmap:8 http://www.webrtc.org/experiments/rtp-hdrext/color-space\r\na=extmap:9 urn:ietf:params:rtp-hdrext:sdes:mid\r\na=extmap:10 urn:ietf:params:rtp-hdrext:sdes:rtp-stream-id\r\na=extmap:11 urn:ietf:params:rtp-hdrext:sdes:repaired-rtp-stream-id\r\na=sendonly\r\na=msid:3a8fffb3-cf06-44f8-a10c-144e209991bb 2dd8c802-f7be-4801-a6aa-df98e21c13c8\r\na=rtcp-mux\r\na=rtcp-rsize\r\na=rtpmap:96 VP8/90000\r\na=rtcp-fb:96 goog-remb\r\na=rtcp-fb:96 transport-cc\r\na=rtcp-fb:96 ccm fir\r\na=rtcp-fb:96 nack\r\na=rtcp-fb:96 nack pli\r\na=ssrc:2659409037 cname:TiUOWZgap2jBYNbL\r\na=ssrc:2659409037 msid:3a8fffb3-cf06-44f8-a10c-144e209991bb 2dd8c802-f7be-4801-a6aa-df98e21c13c8\r\na=ssrc:2659409037 mslabel:3a8fffb3-cf06-44f8-a10c-144e209991bb\r\na=ssrc:2659409037 label:2dd8c802-f7be-4801-a6aa-df98e21c13c8\r\n"
}
{
"type": "answer",
"sdp": "v=0\r\no=- 3940562069726591883 1642473100 IN IP4 127.0.0.1\r\ns=-\r\nt=0 0\r\na=group:BUNDLE 0 1\r\na=msid-semantic: WMS\r\na=ice-lite\r\nm=application 5000 UDP/DTLS/SCTP webrtc-datachannel\r\nc=IN IP4 45.79.247.128\r\na=candidate:706165708 1 udp 2130706431 45.79.247.128 5000 typ host generation 0\r\na=candidate:706165708 2 udp 2130706431 45.79.247.128 5000 typ host generation 0\r\na=candidate:706165708 1 udp 2130706431 45.79.247.128 5000 typ host generation 0\r\na=ice-ufrag:NGrczxsIXdECNyjl\r\na=ice-pwd:dEenZietQjixEhwubLiIGYBsXjueWNHA\r\na=fingerprint:sha-256 90:EA:F3:AC:B3:3E:B7:30:D4:EB:DC:86:F5:55:0B:CC:F8:1C:E7:2C:B3:5B:2E:B4:F4:EF:3C:89:DF:44:80:A7\r\na=setup:active\r\na=mid:0\r\na=sctp-port:5000\r\nm=video 9 UDP/TLS/RTP/SAVPF 96\r\nc=IN IP4 0.0.0.0\r\na=rtcp:9 IN IP4 0.0.0.0\r\na=ice-ufrag:NGrczxsIXdECNyjl\r\na=ice-pwd:dEenZietQjixEhwubLiIGYBsXjueWNHA\r\na=fingerprint:sha-256 90:EA:F3:AC:B3:3E:B7:30:D4:EB:DC:86:F5:55:0B:CC:F8:1C:E7:2C:B3:5B:2E:B4:F4:EF:3C:89:DF:44:80:A7\r\na=setup:active\r\na=mid:1\r\na=extmap:10 urn:ietf:params:rtp-hdrext:sdes:rtp-stream-id\r\na=extmap:9 urn:ietf:params:rtp-hdrext:sdes:mid\r\na=extmap:4 http://www.ietf.org/id/draft-holmer-rmcat-transport-wide-cc-extensions-01\r\na=recvonly\r\na=rtcp-mux\r\na=rtcp-rsize\r\na=rtpmap:96 VP8/90000\r\na=rtcp-fb:96 goog-remb\r\na=rtcp-fb:96 transport-cc\r\na=rtcp-fb:96 ccm fir\r\na=rtcp-fb:96 nack\r\na=rtcp-fb:96 nack pli\r\n"
}
|
gharchive/issue
| 2022-01-15T06:49:17 |
2025-04-01T04:35:30.775563
|
{
"authors": [
"ahmed-adly-khalil",
"leewardbound"
],
"repo": "pion/ion",
"url": "https://github.com/pion/ion/issues/597",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1659034601
|
Unable to run the script
When I open tampermonkey, it shows that chatgpt-exporter is enabled, but I dont see any screen to export the chat.
How do I get the export screen?
Hi, can you update to 2.5.4 first and open F12 - console, then take a screenshot so that I can check the error message.
When I open tampermonkey, it shows that chatgpt-exporter is enabled, but I dont see any screen to export the chat.
How do I get the export screen?
me too, please fix it
@pdsing @tinhthaison please provide the following information:
Browser
OS
Script Version (Open it up and check the version in the script manager)
Error Message (Press F12, go to console, take a screenshot)
thks: Browser: Edge Version 112.0.1722.34 (Official build) (64-bit); OS win 10; Script:https://raw.githubusercontent.com/pionxzh/chatgpt-exporter/master/dist/chatgpt.user.js; Error:
How to export, i can 't see how to export
@tinhthaison thanks
For the script version, you can click on Tampermonkey -> Dashboard
For the error message, I will need the "complete" screenshot. If your screen is not big enough, you can try to copy the text directly. 🙏
I'm also encountering this issue. The full error message I'm encountering is:
Script error in [Greasemonkey script pionxzh/ChatGPT Exporter; version 2.6.1]:
TypeError: GM_getValue is not a function ChatGPT Exporter:18661:19
<anonymous> user-script:pionxzh/ChatGPT Exporter:18661
Hi @weaversam8 @pdsing @tinhthaison
Please help to verify if 2.6.2 fixed the problem. Thanks.
|
gharchive/issue
| 2023-04-07T17:38:27 |
2025-04-01T04:35:30.819282
|
{
"authors": [
"pdsing",
"pionxzh",
"tinhthaison",
"weaversam8"
],
"repo": "pionxzh/chatgpt-exporter",
"url": "https://github.com/pionxzh/chatgpt-exporter/issues/112",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
830330177
|
Rename MySQL error code const
What this PR does / why we need it:
Rename mysqlErrorCodeDuplicate to mysqlErrorCodeDuplicateEntry to clarify it from other duplicate errors.
ref: https://fromdual.com/mysql-error-codes-and-messages-1050-1099#error_er_dup_entry
Which issue(s) this PR fixes:
Fixes #
Does this PR introduce a user-facing change?:
NONE
/approve
|
gharchive/pull-request
| 2021-03-12T17:48:46 |
2025-04-01T04:35:30.824299
|
{
"authors": [
"khanhtc1202",
"nghialv"
],
"repo": "pipe-cd/pipe",
"url": "https://github.com/pipe-cd/pipe/pull/1701",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1123974552
|
Fix orphan command cleaner can not stop on done
What this PR does / why we need it:
Which issue(s) this PR fixes:
Fixes #
Does this PR introduce a user-facing change?:
NONE
Nice catch!
/approve
|
gharchive/pull-request
| 2022-02-04T09:12:56 |
2025-04-01T04:35:30.826110
|
{
"authors": [
"khanhtc1202",
"nghialv"
],
"repo": "pipe-cd/pipecd",
"url": "https://github.com/pipe-cd/pipecd/pull/3203",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1257686441
|
Revert "Add feature to show piped config on web console (#3673)"
This reverts commit 6f722eb0a32521e75036439c36dc0275cdb1bbe5.
What this PR does / why we need it:
Let me revert this temporally to release the next patch version.
Which issue(s) this PR fixes:
Fixes #
Does this PR introduce a user-facing change?:
NONE
/changelog
|
gharchive/pull-request
| 2022-06-02T04:58:15 |
2025-04-01T04:35:30.828085
|
{
"authors": [
"knanao"
],
"repo": "pipe-cd/pipecd",
"url": "https://github.com/pipe-cd/pipecd/pull/3716",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2684581200
|
screen recording
https://github.com/user-attachments/assets/74cf39c5-2d64-4c98-b3ee-cf7308753da6
https://github.com/user-attachments/assets/7585647b-4e8d-4bcf-bed9-0985fe94de2b
https://github.com/user-attachments/assets/8947dfd1-5d4a-43b9-b0c7-22cb52f72dc3
|
gharchive/issue
| 2024-11-22T21:11:19 |
2025-04-01T04:35:30.829908
|
{
"authors": [
"basnijholt"
],
"repo": "pipefunc/graphviz-anywidget",
"url": "https://github.com/pipefunc/graphviz-anywidget/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
200417367
|
web3.eth.contract does not perform validation on address parameter.
What was wrong?
This code does not throw any errors or warnings.
abi = [...] # the contract ABI
code = "0x..." # the contract code
web3.eth.contract(abi, code)
The above code is actually passing in the code as the address parameter since the function signature of web3.eth.contract is def contract(self, abi, address=None, **kwargs):.
Validating the address is well formed (and potentially normalizing it) is trivial to do and would keep this subtle error from happening.
Cute Animal Picture
Recommend using the utility functions from eth_utils for address validation and normalization.
Done via #199
|
gharchive/issue
| 2017-01-12T16:40:09 |
2025-04-01T04:35:30.832689
|
{
"authors": [
"pipermerriam"
],
"repo": "pipermerriam/web3.py",
"url": "https://github.com/pipermerriam/web3.py/issues/150",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1022962099
|
jsp-file broken
The <jsp-file> element inside a servlet element in web.xml is broken. It used to work.
Hi @folker-kuhn. I can take a look. Do you have any examples of the issue?
This is a bit tricky...
We need to set the WaspJspManager as the JspManager to allow calls to JspManager::addJspFile, as this method is called when parsing the web.xml file.
However, we can't do it too early because it would assume by default that the*.jsp mapping is handled by WaSP, but the spec allows overriding the *.jsp mapping to allow swap implementations (I guess).
Probably we'll need to split this process in two step.
@mnriem Thoughts?
You should still be able to map *.jsp to your own Servlet right, as is indeed allowed?
I just checked, the TCK also contains jsp-file entries in some xml files, which are indeed broken now too. So the TCK is broken one way or the other at the moment.
@arjantijms We need to fix both issues and not swap the one for the other. Hence as it currently stands I would not be in favor of merging this PR as it currently stands. @Thihup Please continue your exploration. @folker-kuhn Thank you for pointing this out and working on this. Between you and @Thihup I am convinced you will be able to find a way out.
@Thihup That is maybe bit verbose and not so easy to remember. I added simple -D option to existing pom. You can copy paste failing test from the log with this and run in one go.
I just asked, and it appears to be relevant. I also checked the TCK and there is no test for fully replacing a JSP implementation at such a level that things like using that implementation for deeply integrated things like jsp-file should work just based on a mapping.
The only thing that needs to work is mapping random servlets to the .jsp pattern.
@arjantijms @Thihup @mnriem why on earth are you reverting and re-revering my PR all the time???
Did I do something something wrong? :/
Sorry for the confusion. What is needed is to support overriding the *.jsp servlet-mapping only, not to swap full implementation
Yes, my idea is to go with your suggestion and do split. But now you made it difficult for me to test this, and I don't know why?
@mnriem sorry, did I miss having to create an issue? Yes, please add to the PR template so we know in the future.
@folker-kuhn Not for this one. I wanted to make sure you saw the reasoning for what was happening in any of the issues you were working on. And there was also some confusion here between me and @arjantijms. Rest assured it was not your contributions! Please keep them coming. Note I might as for some clarification and/or want to discuss some more than others.
@mnriem I partially understand, you want a discussion about the need to not have the annoying warnings? It's a bit strange, but let's give it a shot.
This means; find the existing mapping only, remove that mapping (if the mapping was stored in a map directly that would have happened automatically), and then add the new mapping. WDYT?
@folker-kuhn That seems a very good idea! Could you try to implement it?
|
gharchive/issue
| 2021-10-11T17:37:38 |
2025-04-01T04:35:30.844836
|
{
"authors": [
"Thihup",
"arjantijms",
"folker-kuhn",
"mnriem"
],
"repo": "piranhacloud/piranha",
"url": "https://github.com/piranhacloud/piranha/issues/1959",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1654049404
|
change schedule frequency to timeUnit
We need to set timeUnit instead of frequency to have a proper flow schedule set up
will need to properly design the endpoint later, close the pr for now. We will be utilizing the free format of obj in week2.
|
gharchive/pull-request
| 2023-04-04T14:53:07 |
2025-04-01T04:35:30.862844
|
{
"authors": [
"liqunc"
],
"repo": "pitchmuc/aepp",
"url": "https://github.com/pitchmuc/aepp/pull/44",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
242929825
|
Disable floating stemcells before tile upgrade
Hello,
Before starting a tile upgrade in a large scale environment, I would like to make sure the floating stemcell feature is disabled. It would be nice to have this checked as part of the upgrade-tile pipeline.
See:
https://docs.pivotal.io/pivotalcf/1-7/customizing/understanding-stemcells.html
https://discuss.pivotal.io/hc/en-us/articles/224704307-How-to-disable-floating-stemcells-in-Pivotal-Cloud-Foundry
Thanks!
@benfab thanks for using our pipelines! I'm not familiar with the reason(s) why anyone wants to disable floating stemcells, although it does seem like this is not uncommon if Pivotal has an article explaining how to disable this. Can you please let me know what the reason(s)?
When activated during a tile upgrade requiring a new stemcell, the floating stemcells feature will force all VMs from all deployments to be recreated with the new stemcell. In very large scale setup, this behavior will cause a single tile update to take significantly longer wait time, which in production needs to be carefully handled.
Hi Ryan,
We also have a requirement to disable floating stemcells, at least for running production upgrades, it would be great to see this as a feature of pcf-pipelines.
Hi @benfab, thank you for opening this issue - and sorry it's taken so long to add further remarks. I plan on closing this issue because we do not have plans to add this to our backlog - and also Ops Manager has plans to allow you to opt out of using floating stemcells thereby allowing you to control when a tile's stemcell is upgraded. Please do not hesitate to get in touch should you have any questions.
|
gharchive/issue
| 2017-07-14T08:22:06 |
2025-04-01T04:35:30.866371
|
{
"authors": [
"RMeharg",
"abbyachau",
"benfab",
"ryanpei"
],
"repo": "pivotal-cf/pcf-pipelines",
"url": "https://github.com/pivotal-cf/pcf-pipelines/issues/151",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
184153439
|
cities microservice deployment in cf dev - EndpointMBeanExportAutoConfiguration Bean Error
Hi,
I am trying to deploy the microservices set into my cf dev. We got the sample from the location "https://github.com/Pivotal-Field-Engineering/pcf-workspace-microservices/tree/master/micro" and when we try to deploy this, we are getting the below error. How can we customize this?
2016-10-20T12:51:13.49+0530 [APP/0] OUT 2016-10-20 07:21:13.495 ERROR 12 --- [ main] o.s.b.d.LoggingFailureAnalysisReporter :
2016-10-20T12:51:13.49+0530 [APP/0] OUT ***************************
2016-10-20T12:51:13.49+0530 [APP/0] OUT APPLICATION FAILED TO START
2016-10-20T12:51:13.49+0530 [APP/0] OUT ***************************
2016-10-20T12:51:13.49+0530 [APP/0] OUT Description:
2016-10-20T12:51:13.49+0530 [APP/0] OUT Constructor in org.springframework.boot.actuate.autoconfigure.EndpointMBeanExportAutoConfiguration required a single
bean, but 3 were found:
2016-10-20T12:51:13.49+0530 [APP/0] OUT - halObjectMapper: defined by method 'halObjectMapper' in org.springframework.data.rest.webmvc.config.RepositoryR
estMvcConfiguration
2016-10-20T12:51:13.49+0530 [APP/0] OUT - objectMapper: defined by method 'objectMapper' in org.springframework.data.rest.webmvc.config.RepositoryRestMvc
Configuration
2016-10-20T12:51:13.49+0530 [APP/0] OUT - _halObjectMapper: defined in null
2016-10-20T12:51:13.49+0530 [APP/0] OUT Action:
2016-10-20T12:51:13.49+0530 [APP/0] OUT Consider marking one of the beans as @Primary, updating the consumer to accept multiple beans, or using @Qualifier to identify the bean that should be consumed
2016-10-20T12:51:13.58+0530 [APP/0] OUT Exit status 1
2016-10-20T12:51:13.59+0530 [CELL/0] OUT Exit status 0
2016-10-20T12:51:13.60+0530 [CELL/0] OUT Destroying container
2016-10-20T12:51:13.66+0530 [API/0] OUT App instance exited with guid 2ca83c70-b801-48e6-be55-ed13816fe475 payload: {"instance"=>"", "index"=>0, "reason"=>"
CRASHED", "exit_description"=>"2 error(s) occurred:\n\n* 2 error(s) occurred:\n\n* Exited with status 1\n* cancelled\n* cancelled", "crash_count"=>3, "crash_time
stamp"=>1476948073600725767, "version"=>"d7282888-0c34-498d-93b5-71e6bebc42f3"}"
@pcmgit007 could you please provide the commands that you ran? We were able to successfully push the app by following the instructions in the README.
Hi @aemengo , thanks for the reply.
I am trying to run this application with cf version 1.8 on winddows. The default gradle for the sample build has the Spring Boot as 1.1.6.RELEASE and Gradle 1.12. When I tried with these I am getting the below error:
_:cities-client:compileJava UP-TO-DATE
:cities-client:processResources UP-TO-DATE
:cities-client:classes UP-TO-DATE
:cities-client:jar UP-TO-DATE
:cities-client:assemble UP-TO-DATE
:cities-service:compileJava
FAILURE: Build failed with an exception.
What went wrong:
java.lang.UnsupportedOperationException (no error message)_
I have changed it to the latest versions as I have in my local installation (Spring Boot - 1.4.1 and Gradle 3.1), the application assemble (gradle assemble) was success, but during the push I started getting aforementioned error.
It looks like an error from Spring Boot version to me as it was showing the error with multiple beans.
Perhaps cleaning your gradle environment is in order.
Again we were able to successfully push the applications following the README instructions and with Gradle 1.12. Below is our gradle version output
$ gradle -v
------------------------------------------------------------
Gradle 1.12
------------------------------------------------------------
Build time: 2014-04-29 09:24:31 UTC
Build number: none
Revision: a831fa866d46cbee94e61a09af15f9dd95987421
Groovy: 1.8.6
Ant: Apache Ant(TM) version 1.9.3 compiled on December 23 2013
Ivy: 2.2.0
JVM: 1.8.0_91 (Oracle Corporation 25.91-b14)
OS: Mac OS X 10.10.3 x86_64
Closing due to inactivity and inability to reproduce. Feel free to let us know if we need to re-open this issue.
|
gharchive/issue
| 2016-10-20T07:24:05 |
2025-04-01T04:35:30.872936
|
{
"authors": [
"aemengo",
"pcmgit007",
"sclevine"
],
"repo": "pivotal-cf/pcfdev",
"url": "https://github.com/pivotal-cf/pcfdev/issues/158",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
256810431
|
PCF Dev Fails to Provision
Machine Info:
MacOS Sierra
cf version 6.28.0+9e024bdbd.2017-06-27
PCF Dev version 0.27.0 (CLI: 01c0c2b, OVA: 0.529.0)
Running behind Corporate Proxy servers
Here are the proxy variables I have set:
declare -x http_proxy="http://"
declare -x https_proxy="http://"
declare -x no_proxy="localhost,127.0.0.1,192.168.11.1,192.168.11.11,local.pcfdev.io,.local.pcfdev.io"
I get the following error when I run: cf dev start
Error: failed to provision VM: Process exited with: 127. Reason was: ().
When I tail the provision.log file, I see:
Setting quota apps-manager-quota to org system as admin...
OK
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0curl: (6) Could not resolve host: raw.githubusercontent.com
/var/pcfdev/deploy-apps-manager.sh: line 84: nvm: command not found
It looks like curl is not honoring the proxy variables.
Duplicate of #255, about to cut a new release that fixes this.
Can someone please tell me why I'm timing out here. I've tried 0.24.0, 0.29.0, and tried setting http_proxyies as shown (after if failed with 0.29.0) - still no luck:
declare -x http_proxy="http://"
declare -x https_proxy="http://"
declare -x no_proxy="localhost,127.0.0.1,192.168.11.1,192.168.11.11,local.pcfdev.io,.local.pcfdev.io"
My system details:
macOS Sierra. Version 10.12.6
VirtualBox 5.1.30 r118389 (Qt5.6.3)
PCF Dev version 0.29.0 (CLI: 49405ec, OVA: 0.548.0)
pcfdev-debug.tgz.zip
cf dev destroy
PCF Dev VM has been destroyed.
cf dev start -s all
Using existing image.
Allocating 8192 MB out of 16384 MB total system memory (8954 MB free).
Importing VM...
Starting VM...
Provisioning VM...
Waiting for services to start...
Timed out after 3600 seconds.
FAILED
Error: failed to provision VM: Process exited with: 1. Reason was: ().
Follow #324
|
gharchive/issue
| 2017-09-11T19:09:04 |
2025-04-01T04:35:30.881815
|
{
"authors": [
"cassidd",
"echurchm",
"sclevine"
],
"repo": "pivotal-cf/pcfdev",
"url": "https://github.com/pivotal-cf/pcfdev/issues/259",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
390942676
|
Migrated pivpn setup allows old clients to connect, but cannot generate new client keys
Issue
After migration, newly generated client configurations do not work. Old configurations work fine; they can connect.
I copied the contents of /etc/openvpn from my old setup to my new, and the old clients are still able to connect normally and everything is fine. But when using pivpn add nopass, a new client configuration is successfully generated but it does not connect and always gives the same error:
Dec 13 21:54:43 piserver ovpn-server[428]: Authenticate/Decrypt packet error: packet HMAC authentication failed
Dec 13 21:54:43 piserver ovpn-server[428]: TLS Error: incoming packet authentication failed from [AF_INET]71.176.238.134:1024
Console output of pivpn debug
::: Generating Debug Output
::: :::
:: PiVPN Debug ::
::: :::
:: Latest Commit ::
::: :::
commit 95f78aa861761bfe9faa6e34e4b7965266e48ef3
Merge: 375faa6 0b4464b
Author: redfast00 <redfast00@users.noreply.github.com>
Date: Sat Oct 20 16:13:34 2018 +0100
Merge pull request #632 from pivpn/test
Merge test branch
::: :::
:: Recursive list of files in ::
:: /etc/openvpn/easy-rsa/pki ::
::: :::
/etc/openvpn/easy-rsa/pki/:
ca.crt
crl.pem
Default.txt
dh2048.pem
index.txt
index.txt.attr
index.txt.attr.old
index.txt.old
issued
<list of ovpns>
/etc/openvpn/easy-rsa/pki/issued:
<list of certs>
/etc/openvpn/easy-rsa/pki/private:
<list of keys>
::: :::
:: Output of /etc/pivpn/* ::
::: :::
:: START /etc/pivpn/DET_PLATFORM ::
Raspbian
:: END /etc/pivpn/DET_PLATFORM ::
:: START /etc/pivpn/INSTALL_PORT ::
1194
:: END /etc/pivpn/INSTALL_PORT ::
:: START /etc/pivpn/INSTALL_PROTO ::
udp
:: END /etc/pivpn/INSTALL_PROTO ::
:: START /etc/pivpn/INSTALL_USER ::
pi
:: END /etc/pivpn/INSTALL_USER ::
:: START /etc/pivpn/NO_UFW ::
1
:: END /etc/pivpn/NO_UFW ::
:: START /etc/pivpn/pivpnINTERFACE ::
eth1
:: END /etc/pivpn/pivpnINTERFACE ::
:: START /etc/pivpn/setupVars.conf ::
pivpnUser=pi
UNATTUPG=unattended-upgrades
pivpnInterface=eth1
IPv4dns=127.0.0.1
IPv4addr=192.168.1.4
IPv4gw=192.168.1.1
pivpnProto=udp
PORT=1194
ENCRYPT=256
APPLY_TWO_POINT_FOUR=true
DOWNLOAD_DH_PARAM=false
PUBLICDNS=<public address>
OVPNDNS1=8.8.8.8
OVPNDNS2=8.8.4.4
:: END /etc/pivpn/setupVars.conf ::
:: START /etc/pivpn/setupVars.conf.update.bak ::
pivpnUser=pi
UNATTUPG=unattended-upgrades
pivpnInterface=eth1
IPv4dns=127.0.0.1
IPv4addr=192.168.1.241
IPv4gw=192.168.1.2
pivpnProto=udp
PORT=1194
ENCRYPT=256
APPLY_TWO_POINT_FOUR=true
DOWNLOAD_DH_PARAM=false
PUBLICDNS=
OVPNDNS1=8.8.8.8
OVPNDNS2=8.8.4.4
:: END /etc/pivpn/setupVars.conf.update.bak ::
:: START /etc/pivpn/TWO_POINT_FOUR ::
:: END /etc/pivpn/TWO_POINT_FOUR ::
::: :::
:: /etc/openvpn/easy-rsa/pki/Default.txt ::
::: :::
client
dev tun
proto udp
remote <publicip> 1194
resolv-retry infinite
nobind
persist-key
persist-tun
key-direction 1
remote-cert-tls server
tls-version-min 1.2
verify-x509-name server_f25KhqmvUDxKqemN name
cipher AES-256-CBC
auth SHA256
comp-lzo
verb 3
::: :::
:: Debug Output Complete ::
::: :::
:::
::: Debug output completed above.
::: Copy saved to /tmp/debug.txt
:::
Console Output of `pvpn add'
pi@piserver:~ $ pivpn add nopass
Enter a Name for the Client: test
spawn ./easyrsa build-client-full test nopass
Note: using Easy-RSA configuration from: ./vars
Extra arguments given.
rand: Use -help for summary.
Generating a RSA private key
....................+++++
..........+++++
writing new private key to '/etc/openvpn/easy-rsa/pki/private/test.key.qpsiuPtOWw'
-----
Using configuration from /etc/openvpn/easy-rsa/openssl-1.0.cnf
Check that the request matches the signature
Signature ok
The Subject's Distinguished Name is as follows
commonName :ASN.1 12:'test'
Certificate is to be certified until Dec 10 21:57:38 2028 GMT (3650 days)
Write out database with 1 new entries
Data Base Updated
Client's cert found: test.crt
Client's Private Key found: test.key
CA public Key found: ca.crt
tls-auth Private Key found: ta.key
========================================================
Done! test.ovpn successfully created!
test.ovpn was copied to:
/home/pi/ovpns
for easy transfer. Please use this profile only on one
device and create additional profiles for other devices.
========================================================
Check the content of the file /etc/openvpn/easy-rsa/pki/ta.key against the content of the text between and in a new and old .ovpn file. Do they differ? Do not post the content here, just compare. Instead, post the output of grep ta.key /etc/openvpn/server.conf.
I'm trying to under what changed in the process, I thought tls-auth key was different but it's fine.
Can you post the output of /etc/openvpn/server.conf, /etc/openvpn/easy-rsa/pki/Default.txt and the settings in the new and old ovpn file (lines before <ca> tag). Censor the public ip in the last two outputs f you want.
Ok so, actually you have <tls-crypt> inside NEW ovpns and <tls-auth> inside OLD ovpns. From your first answer I though you had <tls-auth> on both. Fix those new clients by changing tls-crypt to tls-auth (both opening and closing tag). Fix future clients by running sudo rm /etc/pivpn/TWO_POINT_FOUR
Awesome, that fixed it. Thanks!
|
gharchive/issue
| 2018-12-14T02:40:47 |
2025-04-01T04:35:30.889678
|
{
"authors": [
"orazioedoardo",
"valentinAlkan"
],
"repo": "pivpn/pivpn",
"url": "https://github.com/pivpn/pivpn/issues/651",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
152448601
|
Failed Submission Set Reload at div id
Is there anyway to stop the form from reloading at the top of the page and instead have the page reloaded at a particular id location (e.g. ) when errors are detected? Depending on where the form is located, a user may have no idea submission failed unless they scroll down to see errors.
Usage questions should be posted over on craftcms.stackexchange.com or on the Craft Slack group https://craftcms.com/community#slack
|
gharchive/issue
| 2016-05-01T21:29:57 |
2025-04-01T04:35:30.906183
|
{
"authors": [
"aaronsransom",
"takobell"
],
"repo": "pixelandtonic/ContactForm",
"url": "https://github.com/pixelandtonic/ContactForm/issues/52",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1322936431
|
gRPC support for C based gRPC libraries
Is your feature request related to a problem? Please describe.
As of now Pixie only supports go-gRPC tracing. Since the gRPC protocol itself is supported by many widely used languages such as Python and Node, there is big gap in pixie's observability capabilities when it comes to no go languages.
Describe the solution you'd like
As mentioned on Pixie's slack channel, C++, Python, Ruby, Objective-C, PHP and C#, all relies on c-gRPC library. So covering this library with UProbes would give a huge boost to pixie's observability capabilities.
Duplicate of #546.
|
gharchive/issue
| 2022-07-30T08:29:12 |
2025-04-01T04:35:30.913393
|
{
"authors": [
"MrSupiri",
"htroisi"
],
"repo": "pixie-io/pixie",
"url": "https://github.com/pixie-io/pixie/issues/545",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
176733653
|
PIXI.extras.MovieClip.onUpdate / onFrameChanged
I have to execute certain code when movie clip is updated.
Rasing the issue because i could find only onComplete callback
Of course I can use setInterval for now, but I think other ppl may also like the idea.
Just so I am clear, you are asking for events at key moments in the lifetime of a MovieClip that you can hook into?
yep, exactly: the moment a new texture from the MovieClip sequence is being rendered
Created a PR with what I believe you are after
@englercj You can close this after https://github.com/pixijs/pixi.js/pull/2948 was merged
|
gharchive/issue
| 2016-09-13T19:53:07 |
2025-04-01T04:35:30.941576
|
{
"authors": [
"englercj",
"themoonrat",
"thesunlover"
],
"repo": "pixijs/pixi.js",
"url": "https://github.com/pixijs/pixi.js/issues/2947",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
447003620
|
Regression in v5 - interactive graphics object not firing events with alpha 0
Running v5.0.3
Events not firing on graphics with alpha 0.
Example:
const app = new PIXI.Application({ antialias: true });
document.body.appendChild(app.view);
const graphics = new PIXI.Graphics();
graphics.beginFill(0xFFF, 1) // click works
// graphics.beginFill(0xFFF, 0) // click DOES NOT work
graphics.drawRect(0, 0, 500, 500)
graphics.interactive = true
graphics.on('click', () => console.log('CLICKED'))
app.stage.addChild(graphics);
Try it here: https://codepen.io/anon/pen/wbpaaV
A continuation of https://github.com/pixijs/pixi.js/issues/5614 ?
Didn't notice that issue. Closing this one then, the workaround mentioned in that issue even though annoying, works.
|
gharchive/issue
| 2019-05-22T08:35:37 |
2025-04-01T04:35:30.943826
|
{
"authors": [
"Mykybo",
"themoonrat"
],
"repo": "pixijs/pixi.js",
"url": "https://github.com/pixijs/pixi.js/issues/5709",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
485967143
|
Sprite size inconsistency depending on parent.
Expected Behavior
Two sprites that are set up with identical sizes should be rendered as the same size.
Current Behavior
The 2 identical sprites are rendered at different sizes.
Steps to Reproduce
https://codepen.io/Lancer611/pen/YzKVEEd
The green squares are set up with identical values. However, s3 is added as child of s, while t3 is added as child of t2. Why does the parent affect the rendered size of the child? Shouldn't the child's size be unaffected by its parent?
Environment
pixi.js version: 5.1.1
Browser & Version: Chrome 75.0.3770.142
OS & Version: Windows 10
Running Example: https://codepen.io/Lancer611/pen/YzKVEEd
Because its same in Flash.
There are many hacks that can help with that, I dont know even which one to recommend ...
My attempt at making frame size independent from scale was rejected 4 years ago.
|
gharchive/issue
| 2019-08-27T18:33:53 |
2025-04-01T04:35:30.947737
|
{
"authors": [
"ivanpopelyshev",
"lancewellspring"
],
"repo": "pixijs/pixi.js",
"url": "https://github.com/pixijs/pixi.js/issues/6046",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1013945037
|
Added reverse integer solution in C
I have added the reverse integer solution in C language and also updated the readme file.
I hope you will like it.
Thank you.
Repo excluded
@lucy2512 Please read the pinned issue to know more!
|
gharchive/pull-request
| 2021-10-02T06:32:16 |
2025-04-01T04:35:30.956327
|
{
"authors": [
"lucy2512",
"piyushsharma220699"
],
"repo": "piyushsharma220699/LeetCode-Problems-Solution-Book",
"url": "https://github.com/piyushsharma220699/LeetCode-Problems-Solution-Book/pull/154",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1965848699
|
[BUG] Roblox Game Client has stopped working after 2 - 15 minutes of playing
Acknowledgement of preliminary instructions
[x] I have read the preliminary instructions, and I am certain that my problem has not already been addressed.
What problem did you encounter?
Roblox Game Client has stopped working is the message I get after playing for around 2 - 15 minutes, I've changed bloxstrap settings and nothing works, please help
Are you using Vulkan?
Are you using Vulkan? What did you change in the settings?
I am using Direct3D 11 and I modified some settings that make the menu go back to the 2015 one, 2013 cursor, custom font, old oof sound, disable desktop app, and alternate graphics quality selector
I've just realized that you're talking about the error message that shows when Windows detects a frozen process. Are you sure this doesn't happen without Bloxstrap? https://github.com/pizzaboxer/bloxstrap/wiki/Switching-between-Roblox-and-Bloxstrap
|
gharchive/issue
| 2023-10-27T16:34:32 |
2025-04-01T04:35:30.959873
|
{
"authors": [
"AsherHeist",
"pizzaboxer"
],
"repo": "pizzaboxer/bloxstrap",
"url": "https://github.com/pizzaboxer/bloxstrap/issues/852",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
975291457
|
Can I unmount a zfs dataset from a pot?
Is your feature request related to a problem? Please describe.
I have a running pot with a zfs dataset mounted somewhere on its rootfs, which was mounted with mount-in command. Now, I need to move the mountpoint to another path but I'm failing to find any info on how to unmount a zfs dataset from a pot.
Describe the feature you'd like to have
mount-out or unmount command for unmounting a dataset/fscomp/directory would be great. Or a documentation entry that explains how to remove a dataset from a pot would work too.
Describe potential alternatives or workaround you've considered (if any)
None I can think of.
Hi.
Unmount (or mount-out) is a missing feature. I recently implemented copy-out, but I guess it's time to implement mount-out as well.
Did you use mount-in with the -w option (re-mount)?
Do you need to umount it while pot is still running?
The -w option will change the mountpoint of the dataset. Without it, it will use nullfs
|
gharchive/issue
| 2021-08-20T05:56:44 |
2025-04-01T04:35:30.963024
|
{
"authors": [
"f8122dac91",
"pizzamig"
],
"repo": "pizzamig/pot",
"url": "https://github.com/pizzamig/pot/issues/170",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2616000148
|
Update logback-classic to 1.5.12
About this PR
📦 Updates ch.qos.logback:logback-classic from 1.3.14 to 1.5.12
Usage
✅ Please merge!
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
⚙ Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "ch.qos.logback", artifactId = "logback-classic" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "30 days" },
dependency = { groupId = "ch.qos.logback", artifactId = "logback-classic" }
}]
labels: test-library-update, early-semver-minor, semver-spec-minor, commit-count:1
Superseded by #108.
|
gharchive/pull-request
| 2024-10-26T16:52:37 |
2025-04-01T04:35:30.972237
|
{
"authors": [
"scala-steward"
],
"repo": "pjfanning/micrometer-pekko",
"url": "https://github.com/pjfanning/micrometer-pekko/pull/98",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2278576207
|
🛑 ✅JrebelLicenseServer is down
In 11612ee, ✅JrebelLicenseServer (https://jrebel.wayok.cn) was down:
HTTP code: 0
Response time: 0 ms
Resolved: ✅JrebelLicenseServer is back up in 50c5c31 after 8 minutes.
|
gharchive/issue
| 2024-05-03T23:36:38 |
2025-04-01T04:35:30.989421
|
{
"authors": [
"pkl1024"
],
"repo": "pkl1024/status",
"url": "https://github.com/pkl1024/status/issues/1200",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2241675825
|
🛑 WAY👌 Blog is down
In e557206, WAY👌 Blog (https://wayok.cn) was down:
HTTP code: 0
Response time: 0 ms
Resolved: WAY👌 Blog is back up in f1ce6a0 after 8 minutes.
|
gharchive/issue
| 2024-04-13T17:49:16 |
2025-04-01T04:35:30.991750
|
{
"authors": [
"pkl1024"
],
"repo": "pkl1024/status",
"url": "https://github.com/pkl1024/status/issues/774",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
311439843
|
cloaning process not successful
after branch, there is a colon (:) sign at the end of the folder 04 _02b. then, unable to further after this not even type anything.
Yes the same happening with me, when I re-start the Terminal, it was not finding gulp so I did a local install of gulp. When I run gulp I get assert.js:350
throw err;
^
AssertionError [ERR_ASSERTION]: Task function must be specified
This course needs an update, not just because of the gulp issue, but because of changes within the framework. It's on my schedule to re-do, but I don't have a due date to give you right now.
|
gharchive/issue
| 2018-04-05T01:00:15 |
2025-04-01T04:35:31.089270
|
{
"authors": [
"ayagirma",
"jfambrini",
"planetoftheweb"
],
"repo": "planetoftheweb/reactinterface",
"url": "https://github.com/planetoftheweb/reactinterface/issues/4",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1706394150
|
PlanetScale migrate
Is there a way to use the Drizzle ORM migrate API with PlanetScale?
I tried:
import { drizzle } from "drizzle-orm/mysql2";
import { migrate } from "drizzle-orm/mysql2/migrator";
import mysql from "mysql2/promise";
async function main() {
// create the connection
const poolConnection = mysql.createPool({
host: process.env.MYSQL_HOSTNAME,
user: process.env.MYSQL_USERNAME,
database: process.env.MYSQL_DATABASE,
multipleStatements: true,
});
const db = drizzle(poolConnection);
// this will automatically run needed migrations on the database
await migrate(db, { migrationsFolder: "./migrations" });
}
main();
But, I got code: 'ECONNREFUSED'. I have confirmed that the stuff in my env is all good.
Thanks!
I'm not sure the error based on this information alone, but none of what I'm seeing here is using this database-js driver.
Can you either ask in https://github.com/planetscale/discussion/discussions or open a support ticket? Seems like there's something else going on.
|
gharchive/issue
| 2023-05-11T19:08:55 |
2025-04-01T04:35:31.091592
|
{
"authors": [
"gederer",
"mattrobenolt"
],
"repo": "planetscale/database-js",
"url": "https://github.com/planetscale/database-js/issues/113",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
135841062
|
chore(): change rules
Configuración para CSSComb: https://gist.github.com/gmq/c255863cd7377ef3912b
Si lo dejan en alguna parte se puede configurar así en atom-css-comb:
En general CSSComb soluciona los problemas de formato que pide el mono excepto el de espacio entre las reglas (EmptyLineBetweenBlocks) que está en la branch dev todavía. De todas maneras la regla ya está en el gist para cuando hagan un release.
genial, gracias!
|
gharchive/pull-request
| 2016-02-23T19:38:14 |
2025-04-01T04:35:31.134633
|
{
"authors": [
"blackjid",
"gmq"
],
"repo": "platanus/hound-scss",
"url": "https://github.com/platanus/hound-scss/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2686599017
|
sugerencia al README.md
una line a de sugerencia
me parece bien
|
gharchive/pull-request
| 2024-11-23T20:56:59 |
2025-04-01T04:35:31.152297
|
{
"authors": [
"hsjm-sistemas"
],
"repo": "platzi/git-github",
"url": "https://github.com/platzi/git-github/pull/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2329760219
|
Update refined to 0.11.2
About this PR
📦 Updates eu.timepit:refined from 0.11.0 to 0.11.2
📜 GitHub Release Notes - Version Diff
Usage
✅ Please merge!
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
🔍 Files still referring to the old version number
The following files still refer to the old version number (0.11.0).
You might want to review and update them manually.
example/project/plugins.sbt
project/plugins.sbt
⚙ Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "eu.timepit", artifactId = "refined" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "30 days" },
dependency = { groupId = "eu.timepit", artifactId = "refined" }
}]
labels: test-library-update, early-semver-minor, semver-spec-patch, old-version-remains, commit-count:1
Superseded by #126.
|
gharchive/pull-request
| 2024-06-02T16:54:16 |
2025-04-01T04:35:31.157519
|
{
"authors": [
"scala-steward"
],
"repo": "play-swagger/play-swagger",
"url": "https://github.com/play-swagger/play-swagger/pull/91",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2544975519
|
[RFC] Async/Lazy Component Systems
Before you can add a component to an entity, you currently need to register the component system with the app.
opts.componentSystems = [RenderComponentSystem]
const app = new AppBase(canvas);
app.init(opts);
app.start();
new Entity('cube').addComponent('render');
This has a couple of friction points:
Creates unnecessary boilerplate. Engine users are required to manage a list of component system. This either means that systems are included when not needed, or worse, components are added without systems, causing errors and friction for developers
In editor projects, every component systems is added by default, regardless of whether they're used or not.
In both cases neither can be tree-shaken, as they're instantiated by the App
An optimal solution would;
Only instantiate components systems when necessary
Not require manually maintaining a list of component systems.
Support tree-shaking
Proposal
Update the addComponent() method to resolve to a lazily loaded component system:
const createComponentSystem = async path => {
const class = await import(path);
return new class();
}
addComponent(type, data) {
let system = this._app.systems[type];
if (!system) {
switch(type) {
case 'camera' : await createComponentSystem('./CameraComponentSystem.js');
case 'render' : await createComponentSystem('./RenderComponentSystem.js');
}
}
// ...
}
Benefits of this approach
Maintains similar api addComponent('camera')
Component Systems are lazily loaded.
Simplifies engine only startup by deprecating the need for specifying systems array.
Better supports tree-shaking.
Editor projects only load necessary components.
Cons
Modifies the addComponent() to become async. Strictly speaking this is a breaking change as it requires await addComponent() which means a semantic major version bump.
Other options
An alternative non breaking change would be to use statically imported components which are not async
import CameraComponentSystem from './CameraComponentSystem'
addComponent(type) {
// ...
switch(type) {
case 'camera' : CameraComponentSystem;
....
}
}
This still has the benefit of a not requiring specifying a component system array, but is harder to tree-shake.
An additional note is to ensure that the order of component systems and their update execution is deterministic and consistent between runs and through the development while using various numbers of systems.
I think the order of execution is something the engine would specify which would be independent of the order in which they're added.
I like this.
3. In both cases neither can be tree-shaken, as they're instantiated by the App
Three-shaking works in the engine only project at the moment. The AppBase does not import any components, and so the components the user does not provide when the AppBase is created can be tree-shaken.
I'm not convinced by the dynamic import (though its a really cool idea) but I like the idea of the auto system instanciation
An alternative solution is to extend parameters addComponent accepts to allow Component Classes too ie addComponent(CameraComponent). Then introduce a static system = CameraComponentSystem in the component which is used to instantiate and register the system.
class MySystem {}
class MyComponent { static system = MySystem }
entity.addComponent(MyComponent)
The string syntax addComponent('camera') would still be supported, so this is a non-breaking change. Strictly speaking this is a tight coupling between Component and System, but this already exist in the component lookup in the entity anyway.
An alternative solution is to extend parameters addComponent accepts to allow Component Classes too ie addComponent(CameraComponent). Then introduce a static system = CameraComponentSystem in the component which is used to instantiate and register the system.
class MySystem {}
class MyComponent { static system = MySystem }
entity.addComponent(MyComponent)
The string syntax addComponent('camera') would still be supported, so this is a non-breaking change. Strictly speaking this is a tight coupling between Component and System, but this already exist in the component lookup in the entity anyway.
That would require an additional import when using ESM?
The string syntax addComponent('camera') would still be supported
if the old way is still supported (and has to be for loading the scene from json), than this won't help with tree-shaking? What's the advantage then.
Editor project can be tree-shaken as we won't need to include the static list of every component systems in the start up script. Also for engine only it will simplify the boilerplate
Editor project can be tree-shaken as we won't need to include the static list of every component systems in the start up script. Also for engine only it will simplify the boilerplate
How you will ensure that the right component systems are in the engine, and avoid extra network file requests?
Editor projects are bundled, so unused modules can be excluded from the
build.
On Thu, Sep 26, 2024 at 12:22 PM mrmaxm @.***> wrote:
Editor project can be tree-shaken as we won't need to include the static
list of every component systems in the start up script. Also for engine
only it will simplify the boilerplate
How you will ensure that the right component systems are in the engine,
and avoid extra network file requests?
—
Reply to this email directly, view it on GitHub
https://github.com/playcanvas/engine/issues/6986#issuecomment-2376666187,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AADJFLHQ66NVNNICYM4LWVLZYPU5PAVCNFSM6AAAAABOX6LRVKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGNZWGY3DMMJYG4
.
You are receiving this because you authored the thread.Message ID:
@.***>
but components can be added by string from scripts. Also the scene format has their names, and creates them by the string. How can you detect what is not used.
In order to do that for Editor there are few complexities:
Identify all component systems that are used across all scenes.
Identify all component systems that are used across all template assets which are not "excluded".
Identify all component systems from scripts but only if ESM scripts are used, and only during bundling, as it wont work for string based addComponent.
It has to do it for each Launch, otherwise the engine differences between launcher and published projects can lead to unintended differences and bugs.
The save from tree-shaking will be neglectable, complexity is huge, and if developers really need to save their tiny KBs, they can compile engine themselves excluding unnecessary component systems. Another branching in API with achieving same thing but a different way, leading to reduced learning curve long-term.
This will introduce complexity, branching API, more failure points, and potentially bad UX if things don't go smoothly in Editor, with extra delays when using Launcher. With seemingly tiny KBs savings, which are easily eclipsed by a single 512 PNG texture.
The original plan we had (and I don't see a better solution at the moment) is that somewhere in project settings, we'd have a list of components with checkboxes. By default all are ticked. User can untick some they don't need and the bundler will not import those components to three-shake them. In a way equivalent to engine only project where we list those. Manual, but simple. Most people don't need to touch it, because as you said, the saving are minimal for majority of projects.
|
gharchive/issue
| 2024-09-24T10:16:03 |
2025-04-01T04:35:31.176492
|
{
"authors": [
"LeXXik",
"MAG-AdrianMeredith",
"Maksims",
"marklundin",
"mvaligursky"
],
"repo": "playcanvas/engine",
"url": "https://github.com/playcanvas/engine/issues/6986",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2265767295
|
Update package.json
Fixes #
PR to address this issue
I confirm I have read the contributing guidelines and signed the Contributor License Agreement.
See comments on #6292
|
gharchive/pull-request
| 2024-04-26T13:14:01 |
2025-04-01T04:35:31.178969
|
{
"authors": [
"gVguy",
"kpal81xd"
],
"repo": "playcanvas/engine",
"url": "https://github.com/playcanvas/engine/pull/6298",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
189713392
|
Top nav gets overlapped
Move to https://play.google.com/store/apps/dev?id=6720847872553662727&hl=en_GB
scroll the page till the "arrow icon" reaches the top nav. You can see that icon is overlapping the nav.
Tried the following and looks like it works. But I am not sure where all .cover for featured apps is used. There should be a reason why its z-index is 2002.
.action-bar-container.sticky .action-bar-inner {
position: fixed;
top: 0;
z-index: 1052;
}
.developer-featured-apps-content .cover {
height: 64px;
margin: 0;
width: 64px;
z-index: 100;
}
You should post this type of question to the play console help directly. This project is focused on the use of Google Play Game Services SDK in the game.
|
gharchive/issue
| 2016-11-16T14:04:52 |
2025-04-01T04:35:31.197442
|
{
"authors": [
"claywilkinson",
"imashankar"
],
"repo": "playgameservices/android-basic-samples",
"url": "https://github.com/playgameservices/android-basic-samples/issues/226",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2325319123
|
🛑 Biosyslit is down
In 433dd78, Biosyslit (https://zenodo.org/communities/biosyslit/) was down:
HTTP code: 503
Response time: 670 ms
Resolved: Biosyslit is back up in 2df5fb4 after 38 minutes.
|
gharchive/issue
| 2024-05-30T10:33:11 |
2025-04-01T04:35:31.212445
|
{
"authors": [
"retog"
],
"repo": "plazi/monitoring",
"url": "https://github.com/plazi/monitoring/issues/1026",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2358305783
|
🛑 eBioDiv Matching Service is down
In b877ed5, eBioDiv Matching Service (https://prod.ebiodiv.org/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: eBioDiv Matching Service is back up in 3347b66 after 5 minutes.
|
gharchive/issue
| 2024-06-17T21:36:50 |
2025-04-01T04:35:31.215150
|
{
"authors": [
"retog"
],
"repo": "plazi/monitoring",
"url": "https://github.com/plazi/monitoring/issues/1063",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1638738174
|
Can you recomment which layer to freeze?
Hi, plemeri
I'm trying transfer learing from ckpt_base(Plus_Ultra) with my custom dataset
My dataset include about 100 samples until now, planned to be 10,000
To avoid overfitting, Can you recommend whith layer to be frozen?
It would be nice to have some additional advice on hyper-parameters, but think it's better to do it after learning
You can freeze backbone network if there are only 100 samples, while you don't need to freeze any other layers if you have more than 10000 samples. Freezing layers is especially unnecessary if they are domain specific, such as human only samples.
Also, try using lower learning rate, different settings for scheduler might be useful.
Thanks.
Got it. Thanks.
|
gharchive/issue
| 2023-03-24T05:20:27 |
2025-04-01T04:35:31.217325
|
{
"authors": [
"plemeri",
"sanghyun0927"
],
"repo": "plemeri/InSPyReNet",
"url": "https://github.com/plemeri/InSPyReNet/issues/23",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2165249327
|
Oracle resource detector
Added OracleResourceDetector for:
Ruby (scheduler)
JS (backend, discordgateway2http)
Python (voice)
please also increment versions for scheduler and voice
actually, bump version everywhere
|
gharchive/pull-request
| 2024-03-03T08:24:46 |
2025-04-01T04:35:31.218808
|
{
"authors": [
"moflwi",
"plengauer"
],
"repo": "plengauer/Philbot",
"url": "https://github.com/plengauer/Philbot/pull/17",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1665375226
|
Support encrypted .intunewin Uploads
Add functionality to decrypt .intunewin & extract data for win32 app uploads
@santese happy to help you on this issue and send a PR. please let me know if you have an approach in mind for decrypting SHA256 encrypted intunewin files.
|
gharchive/issue
| 2023-04-12T22:56:57 |
2025-04-01T04:35:31.288261
|
{
"authors": [
"santese",
"ttfreeman"
],
"repo": "pliancy/intune-node",
"url": "https://github.com/pliancy/intune-node/issues/125",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
579108403
|
zero f1
Hi, plkmo !
I have managed to run the main_task script using follow command:
python main_task.py --train_data ./data/finetune/SemEval2010_task8_all_data/SemEval2010_task8_training/TRAIN_FILE.TXT --test_data ./data/finetune/SemEval2010_task8_all_data/SemEval2010_task8_testing_keys/TEST_FILE_FULL.TXT --use_pretrained_blanks 0 --model_no 0
but always got zero f1 in about 20 ecpos, you can see the detail in following:
Epoch finished, took 29.90 seconds. Losses at Epoch 20: 2.6830234 Train accuracy at Epoch 20: 0.1756250 Test f1 at Epoch 20: 0.0000000
What is the other arguments used for main_task.py? Can you try running with default arguments?
|
gharchive/issue
| 2020-03-11T09:05:38 |
2025-04-01T04:35:31.306732
|
{
"authors": [
"plkmo",
"xiongjun19"
],
"repo": "plkmo/BERT-Relation-Extraction",
"url": "https://github.com/plkmo/BERT-Relation-Extraction/issues/7",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
793858371
|
add .env-example
closes #2
@plocket I wasn't sure if this was what you wanted or how to test it
@plocket I wasn't sure if this was what you wanted or how to test it
@plocket updated the defaults mostly to your requests. Also added some comments
@plocket updated the defaults mostly to your requests. Also added some comments
Did not know that comments in .env are now a Thing. Will wonders never cease...
Did not know that comments in .env are now a Thing. Will wonders never cease...
Need to remember to add this to CHANGELOG on next version bump or something. And maybe to documentation.
Need to remember to add this to CHANGELOG on next version bump or something. And maybe to documentation.
|
gharchive/pull-request
| 2021-01-26T01:49:11 |
2025-04-01T04:35:31.312142
|
{
"authors": [
"berit",
"plocket"
],
"repo": "plocket/docassemble-cucumber",
"url": "https://github.com/plocket/docassemble-cucumber/pull/143",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
118464752
|
Fix directions which led to duplicate resourcess being delivered
(closes #174)
Thank you.
|
gharchive/pull-request
| 2015-11-23T20:24:42 |
2025-04-01T04:35:31.337333
|
{
"authors": [
"davilima6"
],
"repo": "plone/training",
"url": "https://github.com/plone/training/pull/177",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1374378076
|
Makefile: support Mac?
The following code is of course helpful, but lets Mac users with shell zsh stumble. Who is expert in make and conditionalize it?
Defensive settings for make:
https://tech.davis-hansson.com/p/make/
SHELL:=bash
.ONESHELL:
.SHELLFLAGS:=-xeu -o pipefail -O inherit_errexit -c
.SILENT:
.DELETE_ON_ERROR:
MAKEFLAGS+=--warn-undefined-variables
MAKEFLAGS+=--no-builtin-rules
@ksuess What was the problem you ran into? We already have these lines in the Makefile, and SHELL:=bash ensures that the Makefile commands will be run using bash, even if the user running the command has their shell set to zsh.
With an upgrade of make to 4.x I got the following error:
% make docs-livehtml
bash: line 0: inherit_errexit: invalid shell option name
make: *** [Makefile:123: docs-livehtml] Error 2
Now I upgraded bash from 3.x to 5.x. which solved by problem.
Thank you @davisagli for caring!
|
gharchive/issue
| 2022-09-15T11:30:49 |
2025-04-01T04:35:31.340564
|
{
"authors": [
"davisagli",
"ksuess"
],
"repo": "plone/volto",
"url": "https://github.com/plone/volto/issues/3642",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2043639554
|
A small typo in contributing-core.md file
the pnp typo command does not work . Should be chnaged to pnpm
Can I be assigned this issue?
I already submiited a PR for it
Ok.
@aarishshahmohsin please read and follow Contributing to Plone and First-time contributors. We do not assign issues.
@ujjwaleee26 see https://6.docs.plone.org/contributing/first-time.html#write-code for how to avoid the problem you had with your PR.
|
gharchive/issue
| 2023-12-15T12:25:37 |
2025-04-01T04:35:31.343796
|
{
"authors": [
"aarishshahmohsin",
"stevepiercy",
"ujjwaleee26"
],
"repo": "plone/volto",
"url": "https://github.com/plone/volto/issues/5537",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1600776739
|
#3250 Fixed responsive error in login
Edited login.css to make login page responsive. Just a small contribution ❤️
Duplicate of #4305 and #3319 and #4309
Already fixed
|
gharchive/pull-request
| 2023-02-27T09:13:09 |
2025-04-01T04:35:31.344860
|
{
"authors": [
"HyperDanisH",
"davisagli",
"sneridagh"
],
"repo": "plone/volto",
"url": "https://github.com/plone/volto/pull/4437",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1848096622
|
replaced Widget with type in Hardcoded object_list issue#4878
Changed "Widget" to "type" in Hardcoded object_list.
@ThisisMS11 you need to sign the Plone Contributor Agreement to merge this pull request.
Learn about the Plone Contributor Agreement: https://plone.org/foundation/contributors-agreement
This is not sufficient for fixing issue 4878. There are other places in the code that use object_list as a widget, and even if we add support for it as a type instead of a widget, we need to remain backward compatible with existing addons that use it as a widget.
Also, please read https://6.docs.plone.org/contributing/index.html and https://6.docs.plone.org/contributing/first-time.html -- there are some requirements for contributing to Plone that you haven't followed here.
Thanks a lot @davisagli . Will keep those things in mind. can beginners contribute to this repo?
@ThisisMS11 Yes, anyone can contribute as long as they sign the contributor agreement and follow the other guidelines in the contributor documentation. Please note that in general, the core group of people who review contributions is small, busy, and has trouble providing timely feedback to all contributions. So please expect that it may take some time for contributions to be reviewed.
|
gharchive/pull-request
| 2023-08-12T16:07:06 |
2025-04-01T04:35:31.348225
|
{
"authors": [
"ThisisMS11",
"davisagli",
"mister-roboto"
],
"repo": "plone/volto",
"url": "https://github.com/plone/volto/pull/5082",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2317013071
|
🛑 KCG is down
In fb4f708, KCG (https://kunstrijclubgroningen.nl) was down:
HTTP code: 0
Response time: 0 ms
Resolved: KCG is back up in cc5a689 after 20 minutes.
|
gharchive/issue
| 2024-05-25T13:35:38 |
2025-04-01T04:35:31.444652
|
{
"authors": [
"pluim003"
],
"repo": "pluim003/upptime",
"url": "https://github.com/pluim003/upptime/issues/1517",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2427449014
|
🛑 KCG is down
In 6b8c605, KCG (https://kunstrijclubgroningen.nl) was down:
HTTP code: 0
Response time: 0 ms
Resolved: KCG is back up in ed4de57 after 27 minutes.
|
gharchive/issue
| 2024-07-24T12:22:28 |
2025-04-01T04:35:31.447101
|
{
"authors": [
"pluim003"
],
"repo": "pluim003/upptime",
"url": "https://github.com/pluim003/upptime/issues/3327",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2526754932
|
🛑 Pierced Arrows is down
In 87a137f, Pierced Arrows (https://www.piercedarrows.nl) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Pierced Arrows is back up in 5b72a3d after 29 minutes.
|
gharchive/issue
| 2024-09-15T05:57:56 |
2025-04-01T04:35:31.449393
|
{
"authors": [
"pluim003"
],
"repo": "pluim003/upptime",
"url": "https://github.com/pluim003/upptime/issues/4769",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2731455741
|
🛑 KCG is down
In e5fd02f, KCG (https://kunstrijclubgroningen.nl) was down:
HTTP code: 0
Response time: 0 ms
Resolved: KCG is back up in 070a5a6 after 10 minutes.
|
gharchive/issue
| 2024-12-10T23:31:50 |
2025-04-01T04:35:31.451587
|
{
"authors": [
"pluim003"
],
"repo": "pluim003/upptime",
"url": "https://github.com/pluim003/upptime/issues/6946",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2267539026
|
🛑 Dick Pluim.com is down
In b310896, Dick Pluim.com (https://dickpluim.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Dick Pluim.com is back up in 8262baa after 20 minutes.
|
gharchive/issue
| 2024-04-28T12:29:02 |
2025-04-01T04:35:31.454492
|
{
"authors": [
"pluim003"
],
"repo": "pluim003/upptime",
"url": "https://github.com/pluim003/upptime/issues/709",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1110392342
|
Overlapping subtitles
the extension just overlaps any lines that run at the same time.
I also noticed that behaviour. I believe starting with 6.1 version. I use a 6.0.13 version of extension and dont have that problem with overlaping lines. And also broken transparency. Just see an example below :
6.0.13
6.1.6
Release 6.2.x will fix that issue.
A new checkbox is added to the customizing tab, which enables snapToLines, which prevents overlapping subtitles.
If you want to show the subtitle on the bottom, you have to select a high position value, like 94%
do not show overlapping hint for firefox
|
gharchive/issue
| 2022-01-21T12:02:57 |
2025-04-01T04:35:31.484644
|
{
"authors": [
"Phoenomenal",
"ste-xx"
],
"repo": "plussub/plussub",
"url": "https://github.com/plussub/plussub/issues/34",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2646623016
|
Fix errors in extracting session content for script 02/02
When running the script src/open_discourse/02_preprocessing/02_split_xml_electoral_term_1_and_2.py the log reveils a problem with the following files:
Period 1
No session content found in 01014.
No session content found in 01001.
No session content found in 01042.
No session content found in 01280.
No session content found in 01041.
No session content found in 01281.
No session content found in 01224.
No session content found in 01019.
No session content found in 01223.
No session content found in 01183.
Period2
No session content found in 02078.
No session content found in 02188.
We need to investigate the problem. This was already a problem in the old script as the same files are missing there
Directly related to https://github.com/open-discourse/open-discourse/issues/115, just have to check if it is the same error for all listed files
After a short investigation the problem is definitely NOT only related to problems in the start statement and line breaks. The problem are breaks in the session protocol. A session can be interrupted and later continued. When this happens we have something like:
Ich unterbreche infolgedessen die Sitzung und berufe den Ältestenrat zu einer Sitzung zusammen.
(Große Unruhe und Pultdeckelklappen.)
(Unterbrechung der Sitzung:
17 Uhr 55 Minuten.)
Die Sitzung wird um 19 Uhr 8 Minuten wieder aufgenommen.
which seems totally fine to me. The only problem is, that our current get_session_content function that is looking for the beginning and end of a session, only works for two cases:
There are more beginnings than ends and exactly ONE end -> take the first beginning until the only end
exactly N beginnings and N ends -> take all the sessions in between
But in the above mentioned cases, thanks to the interruptions, we have N beginnings but M endings with N > M, meaning we need to detect interruptions and basically ignore them.
|
gharchive/issue
| 2024-11-09T22:40:24 |
2025-04-01T04:35:31.523271
|
{
"authors": [
"pmayd"
],
"repo": "pmayd/open-discourse",
"url": "https://github.com/pmayd/open-discourse/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2280761342
|
[BUG] Recursive models generate invalid typescript
We have the following python model:
class FilterRule(BaseModel):
value: str | list[str]
operator: Literal["eq", "ne", "in"]
field: str
field_type: Literal["list", "string"] = "string"
def __str__(self) -> str:
if self.field_type == "list":
if isinstance(self.value, list):
joined_value = ", ".join([v for v in self.value])
else:
joined_value = self.value
return f"{self.field}/any(g: search.in(g, '{joined_value}'))"
else:
if self.operator == "in":
if isinstance(self.value, list):
predicate = " or ".join(
[f"{self.field} eq '{v}' " for v in self.value]
)
else:
predicate = f"{self.field} eq '{self.value}'"
return f"({predicate})" if predicate else ""
return f"{self.field} {self.operator} '{self.value}'"
def __bool__(self) -> bool:
return bool(self.value) and bool(self.field)
class Filters(BaseModel):
logical_op: Literal["and", "or", "not"] = "and"
rules: list[Union[FilterRule, "Filters"]]
def __str__(self) -> str:
filter_str = f" {self.logical_op} ".join(
[
str(rule) if isinstance(rule, FilterRule) else f"({rule})"
for rule in self.rules
if bool(rule)
]
)
return filter_str.strip()
Which generates the following OpenAPI schema type:
"Filters":{
"properties":{
"logical_op":{
"type":"string",
"enum":[
"and",
"or",
"not"
],
"title":"Logical Op",
"default":"and"
},
"rules":{
"items":{
"oneOf":[
{
"$ref":"#/components/schemas/FilterRule"
},
{
"$ref":"#/components/schemas/Filters"
}
]
},
"type":"array",
"title":"Rules"
}
},
"type":"object",
"required":[
"rules"
],
"title":"Filters"
},
Which is correct. But the generated typescript file is invalid:
import type { FilterRule } from "./FilterRule.js";
import type { Filters2 } from "./Filters.js";
export type Filters = {
logical_op?: "and" | "or" | "not";
rules: Array<FilterRule | Filters2>;
};
There exists not "Filters2". It should simply be:
import type { FilterRule } from "./FilterRule.js";
export type Filters = {
logical_op?: "and" | "or" | "not";
rules: Array<FilterRule | Filters>;
};
How can this be fixed? If we manually adjust it, every time we run the client it is overwritten.
Thanks for letting me know about this! I will look at this and get it fixed ASAP.
Unfortunately, there's not yet a way to exclude a type file from being automatically recreated. That's coming soon (#886). You've also inspired me add another feature to prevent an individual file from being overwritten (#891). That will be trivial to implement so worst case I'll have a release out by the end of the day that allows you to keep your manual fix.
It's probably a matter of removing the !this.exports.has(name) part in this line.
https://github.com/pmcelhaney/counterfact/blob/84a421d7e14e31a13f7ad660fe0142fb92e74efa/src/typescript-generator/script.js#L30
Fixed in 0.40.2. Thank you for the bug report! If you have any other feedback, please don't hesitate to share.
|
gharchive/issue
| 2024-05-06T12:24:44 |
2025-04-01T04:35:31.528540
|
{
"authors": [
"kaikun213",
"pmcelhaney"
],
"repo": "pmcelhaney/counterfact",
"url": "https://github.com/pmcelhaney/counterfact/issues/890",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1915936037
|
no patch package after install
Version 0.26.0 was not working because it's trying to run patch-package after install, and the command patch-package is "not found". The only package that's being patched is a dev dependency so it's not needed anyway.
Pull Request Test Coverage Report for Build 6328844608
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage increased (+0.1%) to 87.853%
Totals
Change from base Build 6328656972:
0.1%
Covered Lines:
2409
Relevant Lines:
2776
💛 - Coveralls
|
gharchive/pull-request
| 2023-09-27T16:26:04 |
2025-04-01T04:35:31.533996
|
{
"authors": [
"coveralls",
"pmcelhaney"
],
"repo": "pmcelhaney/counterfact",
"url": "https://github.com/pmcelhaney/counterfact/pull/583",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1125089533
|
Suggestion: Advanced List indexing options
It has been suggested to add more indexing options for lists.
Here an example from Python: Indexing in Python
I looked into it a bit ; some of these features (like reverse indexing) are not easily possible without breaking the current List model in microScript which is very close to JavaScript arrays (and based on them). I think the other features (like slicing) are already covered by List functions in microScript (or let me know and they will!). I do not see a great benefit in adding more syntax options for it.
But wait! Here is something that should interest you: microScript 2.0 allows you to add your own custom functions or operators for lists, strings and functions:
You have a List prototype on which you can attach any useful function you want to apply to all your lists
You can override or define common operators on the List prototype too, that is to say you can set what + - * / % & | do when applied to lists.
Example:
List.get = function(index)
this[if index<0 then length+index else index end]
end
List."%" = function(list,length)
list.slice( 0 , length )
end
With that defined, you can now do:
[1,2,3,4,5,6,7,8].get(-1) // index elements from the end
8
[1,2,3,4,5,6,7,8] % 4 // get a sub-list of the list
[1,2,3,4]
While we are here, let's do something crazy :-) Python has this my_list[start_index:end_index:step_size], in microScript we could mimic it with this code:
List."|" = function(list,arg)
local start_index = arg[0]
local end_index = arg[1]
local step = arg[2]
local res = []
for i=start_index to end_index by step
res += list[i]
end
res
end
List.range = function(num) // let's also create our version of range(100)
local res = []
for i=0 to num-1 by 1
res += i
end
res
end
Now that this is set, we can do:
List.range(100)|[12,84,3]
[12,15,18,21,24,27,30,33,36,39,42,45,48,51,54,57,60,63,66,69,72,75,78,81,84]
That surpasses the original request by miles :)
Loving it ❤️
|
gharchive/issue
| 2022-02-06T03:01:41 |
2025-04-01T04:35:31.576651
|
{
"authors": [
"TinkerStudio",
"pmgl"
],
"repo": "pmgl/microstudio",
"url": "https://github.com/pmgl/microstudio/issues/83",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1254314330
|
⚠️ Slack has degraded performance
In 88779b9, Slack (https://status.slack.com/api/v2.0.0/current) experienced degraded performance:
HTTP code: 200
Response time: 131 ms
Resolved: Slack performance has improved in 8ecff92.
|
gharchive/issue
| 2022-05-31T19:14:47 |
2025-04-01T04:35:31.580139
|
{
"authors": [
"pmmmwh"
],
"repo": "pmmmwh/upptime",
"url": "https://github.com/pmmmwh/upptime/issues/140",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1333915141
|
⚠️ GitHub has degraded performance
In 077d967, GitHub (https://www.githubstatus.com/api/v2/status.json) experienced degraded performance:
HTTP code: 200
Response time: 237 ms
Resolved: GitHub performance has improved in c114ccb.
|
gharchive/issue
| 2022-08-10T00:24:22 |
2025-04-01T04:35:31.582693
|
{
"authors": [
"pmmmwh"
],
"repo": "pmmmwh/upptime",
"url": "https://github.com/pmmmwh/upptime/issues/182",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1917702919
|
🛑 Google is down
In f33b643, Google (https://www.google.com) was down:
HTTP code: 429
Response time: 352 ms
Resolved: Google is back up in 88c667b after 7 minutes.
|
gharchive/issue
| 2023-09-28T14:43:21 |
2025-04-01T04:35:31.585128
|
{
"authors": [
"pmmmwh"
],
"repo": "pmmmwh/upptime",
"url": "https://github.com/pmmmwh/upptime/issues/528",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2056715771
|
XR integration
Do you have any hints How can I integrate with @react-three/xr ? I would like to use it as it works by default with additional support of VR, when when enter VR is clicked. Any hints are welcome.
I'm also interested in integrating it with XR, but I haven't tried @react-three/xr yet. I'll give it a try when I get back to this project.
You would have to manipulate the camera returned from https://threejs.org/docs/#api/en/renderers/webxr/WebXRManager.getCamera rather than the default camera. No other changes are needed.
Thanks, @CodyJasonBennett! Good to know!
@ErdongChen-Andrew in my project I got pretty stuck on this and the solution was to make sure you do
gl.xr.enabled = true; after you do the getCamera suggestion above or the headset wont follow movement or do any xr things
Thank you, @antpb ! I'll definitely keep that in mind while I'm working on XR!
Yo, glad this is being worked on, been trying to make it work myself using @react-three/xr but I'm not able to get the movement / rotation right.
I was able to get the camera connected, but I'd like to be able to use the vr controllers.
left controller joystick to move around relative to where im facing (click in to sprint)
right controller joystick for rotating left and right
a button to jump
I noticed there was controller support added but doesn't seem to work with the xr controls.
anyways big thanks for the library @ErdongChen-Andrew hope you get around to this soon :)
Bumping this thread for priority
PR welcome, and reminder open-source is voluntary community work.
PR welcome, and reminder open-source is voluntary community work.
Hey thanks for the bumping this thread again, I love open source, I can contribute what I can but I know there are those who are better at doing this than me and I wish they can do what they can as soon as possible. Will keep you updated in the PR i plan to create.
|
gharchive/issue
| 2023-12-26T21:23:46 |
2025-04-01T04:35:31.595182
|
{
"authors": [
"CodyJasonBennett",
"ErdongChen-Andrew",
"antpb",
"jamesrez",
"krzysztofpniak"
],
"repo": "pmndrs/ecctrl",
"url": "https://github.com/pmndrs/ecctrl/issues/21",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1144973487
|
Bug: Disabled input values can be mutated via keyboard
The disabled input can be focused using tab or shift+tab, and then mutated using keyboard shortcuts and keys.
Demo GIF:
Demo Code:
@Sec-ant thanks for reporting this, it should now be fixed!
@Sec-ant thanks for reporting this, it should now be fixed!
Much obliged. Thank you, sir!
|
gharchive/issue
| 2022-02-20T10:01:10 |
2025-04-01T04:35:31.598096
|
{
"authors": [
"Sec-ant",
"dbismut"
],
"repo": "pmndrs/leva",
"url": "https://github.com/pmndrs/leva/issues/309",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
893145348
|
Able to drag Leva window under text
It can't be dragged back again as the text eats the click event. Is it possible to disable dragging on Leva entirely?
https://user-images.githubusercontent.com/25351731/118467883-1d07db00-b6fc-11eb-9ed4-eadf71a1c1e1.mov
Hey!
Thank you so much for catching that! Just removed the ability to drag here https://github.com/pmndrs/market/commit/df2c3fb17fc65507cb832c33dcb7861026395f35
Anything else feel free to make issues :)
Awesome 🚀
|
gharchive/issue
| 2021-05-17T09:39:58 |
2025-04-01T04:35:31.600651
|
{
"authors": [
"SaraVieira",
"alii"
],
"repo": "pmndrs/market",
"url": "https://github.com/pmndrs/market/issues/20",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1551969247
|
other: update sbt-native-packager from 1.9.7 to 1.9.13
Updates com.github.sbt:sbt-native-packager from 1.9.7 to 1.9.13.
GitHub Release Notes - Changelog - Version Diff
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "com.github.sbt", artifactId = "sbt-native-packager" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequests = { frequency = "30 days" },
dependency = { groupId = "com.github.sbt", artifactId = "sbt-native-packager" }
}]
labels: sbt-plugin-update, early-semver-patch, semver-spec-patch, commit-count:1
Codecov Report
Base: 100.00% // Head: 100.00% // No change to project coverage :thumbsup:
Coverage data is based on head (46accb3) compared to base (631a91b).
Patch has no changes to coverable lines.
Additional details and impacted files
@@ Coverage Diff @@
## master #79 +/- ##
===========================================
Coverage 100.00% 100.00%
Complexity 29 29
===========================================
Files 1 1
Lines 114 114
Branches 8 8
===========================================
Hits 114 114
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
:umbrella: View full report at Codecov.
:loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
Superseded by #81.
|
gharchive/pull-request
| 2023-01-22T01:11:41 |
2025-04-01T04:35:31.647286
|
{
"authors": [
"codecov-commenter",
"pnerg"
],
"repo": "pnerg/jmx-runtime-json",
"url": "https://github.com/pnerg/jmx-runtime-json/pull/79",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
278993577
|
Genome binning
@brwnj I' started implementing concoct and am hesitating to use anvi'o. which has a lot of additional tools (may be too much).. what do you think?
If you decide you really want to dive into it, I would make it an optional protocol and a separate conda environment -- like you're starting to do with concoct. That way we have a primary conda env for the default and optional environments for alternate processing paths.
It's been a while since I used anvio. It can't be as bad as CheckM's install, right?
Hehe, Anvi'o can be inst via bioconda now.
On Dec 4, 2017, 18:29, at 18:29, Joe Brown notifications@github.com wrote:
If you decide you really want to dive into it, I would make it an
optional protocol and a separate conda environment -- like you're
starting to do with concoct. That way we have a primary conda env for
the default and optional environments for alternate processing paths.
It's been a while since I used anvio. It can't be as bad as CheckM's
install, right?
--
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub:
https://github.com/pnnl/atlas/issues/57#issuecomment-349037819
|
gharchive/issue
| 2017-12-04T13:21:34 |
2025-04-01T04:35:31.656814
|
{
"authors": [
"SilasK",
"brwnj"
],
"repo": "pnnl/atlas",
"url": "https://github.com/pnnl/atlas/issues/57",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
487886171
|
New command: create custom connector
Implemented as paconn create in PowerPlatform Connectors CLI (https://github.com/microsoft/PowerPlatformConnectors)
Command
paconn create : Creates a new custom connector from the given directory.
Arguments
--api-def : Location of the Open API definition JSON document.
--api-prop : Location of the API properties JSON document.
--env -e : Power Platform environment ID.
--icon : Location for the icon file.
--pau -u : Power Platform URL.
--pav -v : Power Platform api version.
--secret -r : The OAuth2 client secret for the connector.
--settings -s : A settings file containing required parameters. When a settings file is
specified some commandline parameters are ignored.
Global Arguments
--debug : Increase logging verbosity to show all debug logs.
--help -h : Show this help message and exit.
--output -o : Output format. Allowed values: json, jsonc, table, tsv. Default: json.
--query : JMESPath query string. See http://jmespath.org/ for more information and
examples.
--verbose : Increase logging verbosity. Use --debug for full debug logs.
Examples
Create connector
paconn create
Closing due to lack of activity
|
gharchive/issue
| 2019-09-01T13:26:17 |
2025-04-01T04:35:31.675808
|
{
"authors": [
"garrytrinder",
"waldekmastykarz"
],
"repo": "pnp/cli-microsoft365",
"url": "https://github.com/pnp/cli-microsoft365/issues/1083",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.