id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
39804785
|
hi,PaulGuo
hi~,你的 juicer 非常好用,但在使用过程中,我发现的这种链式调用会比较实用,比如: ${text|func1|func2|func3} 请问最近有实现此功能的计划吗?
这个,我觉得模板还是应该专注模板该做的事情,真要那么多function 的话,应该再用一个function 去包装他们
话说已经2016年了,链式调用到底啥时候能支持
|
gharchive/issue
| 2014-08-08T09:03:53 |
2025-04-01T04:55:30.974918
|
{
"authors": [
"Riant",
"imloki",
"judasnow"
],
"repo": "PaulGuo/Juicer",
"url": "https://github.com/PaulGuo/Juicer/issues/54",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1276556459
|
feat(automatic): Automatically convert urls on block commit
When a person enters a URL and saves the block automatically get the title
Will need to have a user preference to set the default.
Hi, I don’t know how to automatically convert the URL, but I submitted a PR https://github.com/PaulKinlan/logseq-webpage-title/pull/25 which can set the shortcut key to achieve this conversion.
|
gharchive/issue
| 2022-06-20T08:25:42 |
2025-04-01T04:55:30.976681
|
{
"authors": [
"PaulKinlan",
"Theigrams"
],
"repo": "PaulKinlan/logseq-webpage-title",
"url": "https://github.com/PaulKinlan/logseq-webpage-title/issues/22",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
314569507
|
SMTP Address
Changed
SMTP address in example config (156645819)
Delivers #156645819
Skipped scrutiniser check as it had hung and requeueing also hung. Four ticks from other checks and known untested file (.env)
|
gharchive/pull-request
| 2018-04-16T09:30:19 |
2025-04-01T04:55:31.054847
|
{
"authors": [
"srleach"
],
"repo": "PayBreak/basket",
"url": "https://github.com/PayBreak/basket/pull/261",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1758143082
|
🛑 PbtServers (Minecraft) is down
In cbbb826, PbtServers (Minecraft) (mc.pbtservers.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: PbtServers (Minecraft) is back up in ad08b76.
|
gharchive/issue
| 2023-06-15T06:39:23 |
2025-04-01T04:55:31.091235
|
{
"authors": [
"PbtServers"
],
"repo": "PbtServers/PbtServers-Estado",
"url": "https://github.com/PbtServers/PbtServers-Estado/issues/232",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
736049027
|
Autoplacing of the dropdown menu in the Select
When the dropdown menu fails to place it automatically.
@MarikTar Try to use lib-react-components@1.2.1-alpha.0.
Changes:
Added Select dropdown auto-placement.
Added Autocomplete dropdown auto-placement.
Please test
@donskov, thanks, i tried it and can suggest the following:
placement : 'top' | 'bottom' | 'auto' = 'bottom' not all dropdown need apply automatically orientation for example:
bottom
auto
or in mobile version.
it also prevents that if someone uses this library will face the problem that the "placement" is no longer available.
on mobile i have next problem
if page have fixed "footer" happened next (can use z-index but this not good solution):
no footer
have footer
P.S. I do not know if it will be appropriate, but I would add an auto-flip of the arrow.
@MarikTar Published lib-react-components@1.2.1-alpha.2
Fixed dropdown left position.
Added props:
/**
* Component dropdown start opened direction.
*/
placement: PropTypes.oneOf(['top', 'bottom']),
/**
* If `true`, component will automatically calc possible dropdown opened direction.
*/
flip: PropTypes.bool,
Please try to use
Added auto-flip for the arrow
Added auto-flip for the arrow
|
gharchive/issue
| 2020-11-04T11:45:09 |
2025-04-01T04:55:31.152956
|
{
"authors": [
"MarikTar",
"donskov"
],
"repo": "PeculiarVentures/peculiar-react-components",
"url": "https://github.com/PeculiarVentures/peculiar-react-components/issues/82",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
961239789
|
gorc config gas prices take
Lands #125
This bit has been reviewed several times so going to squash and merge shortly.
|
gharchive/pull-request
| 2021-08-04T23:59:30 |
2025-04-01T04:55:31.157781
|
{
"authors": [
"levicook"
],
"repo": "PeggyJV/gravity-bridge",
"url": "https://github.com/PeggyJV/gravity-bridge/pull/126",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2580355031
|
Finished_assignment_2
What changes are you trying to make? (e.g. Adding or removing code, refactoring existing code, adding reports)
Finished and submitted assignment_2.
What did you learn from the changes you have made?
I've practiced functions of Regressions 1,2 what we learned.
Was there another approach you were thinking about making? If so, what approach(es) were you thinking of?
Maybe, some functions/ models could be more consice.
Were there any challenges? If so, what issue(s) did you face? How did you overcome it?
I had challenges in modeling. And some information in assignment was new for me.
How were these changes tested?
On my side, all functions / models work.
A reference to a related issue in your repository (if applicable)
N/A
Checklist
[ Yes] I can confirm that my changes are working as intended
A2: Complete
Great work
Q2 part(ii) More specifically it represents the line of best fit in linear regression, which aims to minimize the squared distance between the observations and the fitted line.
|
gharchive/pull-request
| 2024-10-11T03:41:10 |
2025-04-01T04:55:31.180526
|
{
"authors": [
"Melokazz",
"PeleshkoV"
],
"repo": "PeleshkoV/applying_statistical_concepts",
"url": "https://github.com/PeleshkoV/applying_statistical_concepts/pull/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
555911871
|
[BUG] "Write About Me" does not mention required LANDING_PAGE_TITLE
Description
Describe the bug
The documentation does not mention that it is necessary to set LANDING_PAGE_TITLE before the landing page will be included.
To Reproduce
Steps to reproduce the behavior:
Set THEME='elegant'
Create a simple page with slug set to landing-page-about-hidden and status set to hidden containing "XYZZY"
Generate site
See that index.html does not contain "XYZZY"
Expected behavior
Documentation will specify all steps necessary to achieve desired outcome.
Screenshots
N/A
Desktop (please complete the following information):
N/A
Smartphone (please complete the following information):
N/A
Additional context
Add any other context about the problem here.
:tada: This issue has been resolved in version 5.2.0 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
|
gharchive/issue
| 2020-01-28T00:18:39 |
2025-04-01T04:55:31.186267
|
{
"authors": [
"neirbowj",
"talha131"
],
"repo": "Pelican-Elegant/elegant",
"url": "https://github.com/Pelican-Elegant/elegant/issues/561",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2573675959
|
Fix remove outdated api calls
fix bug causing duplicate api calls to /packages endpoint while trying to fetch dataset files
Get Packages
https://github.com/user-attachments/assets/bb317613-9f8b-44da-9cd9-224b6a5dfac1
@vijaysparc This looks great on first pass. Can you validate in dev that the switch organization functionality is not impacted by this change? Thanks!
@vijaysparc This looks great on first pass. Can you validate in dev that the switch organization functionality is not impacted by this change? Thanks!
Sure, I can verify that. This change affects internal navigation in dataset files component. Switching organization seems to be out of scope for this change, will still verify and do smoke test though.
|
gharchive/pull-request
| 2024-10-08T16:24:14 |
2025-04-01T04:55:31.188745
|
{
"authors": [
"annaleeswitek",
"vijaysparc"
],
"repo": "Pennsieve/pennsieve-app",
"url": "https://github.com/Pennsieve/pennsieve-app/pull/183",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1357843794
|
Add Support for CEN-IO devices
Updated CEN-IO class inheritance of EssentialsDevice to CrestronGenericBaseDevice
DIGIN-104
RY-104
IR-104
Tested in the field with CEN-IO-IR-104 and CEN-IO-RY-104. Merged hotfix/huddle-room-interfaces branch.
This PR also addresses a missing interface on the HuddleRoom.
|
gharchive/pull-request
| 2022-08-31T20:37:25 |
2025-04-01T04:55:31.198822
|
{
"authors": [
"andrew-welker",
"jkdevito"
],
"repo": "PepperDash/Essentials",
"url": "https://github.com/PepperDash/Essentials/pull/988",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2049333929
|
Flight duration is resetting
Example: While I craft one stack of Flight duration 3 rockets, after using one rocket, the whole stack is resetting the Flight duration.
Do you think it's possible to fix?
Fixed in the latest version, thanks for reporting!
Thank you!!
|
gharchive/issue
| 2023-12-19T19:47:09 |
2025-04-01T04:55:31.201487
|
{
"authors": [
"Pepperoni-Jabroni",
"sternschnaube"
],
"repo": "Pepperoni-Jabroni/RandomizedDefaultFireworks",
"url": "https://github.com/Pepperoni-Jabroni/RandomizedDefaultFireworks/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1508518000
|
Add a license
Very cool project! Thanks for releasing it. I took the same approach of running a tRPC binary as a sidecar for my Tauri app.
I noticed this project doesn’t seem to have a license, so I wonder if you would consider adding one.
Thanks!
Thanks for the heads-up. I overlooked this!
MIT license added
|
gharchive/issue
| 2022-12-22T20:41:37 |
2025-04-01T04:55:31.210810
|
{
"authors": [
"Perfect7M",
"techpeace"
],
"repo": "Perfect7M/taurine",
"url": "https://github.com/Perfect7M/taurine/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1445978838
|
portal problem
i made the 2x2 portal with the dirt and flowers around it, but it did not turn into the portal entrance, is this a glitch or is there an alternate way in?
also most of the commands dont work
Yep, it's a bug!
I have not worked on this project in a while, and I've lacked the motivation to continue developing Minecraft plugins.
Thanks for taking interest in this project though, I really appreciate it!
|
gharchive/issue
| 2022-11-11T20:34:39 |
2025-04-01T04:55:31.214822
|
{
"authors": [
"EclipseGamer61",
"PeriodicSeizures"
],
"repo": "PeriodicSeizures/TwilightForestPlugin",
"url": "https://github.com/PeriodicSeizures/TwilightForestPlugin/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1142136436
|
feat(shadowing): bch-wallet-starter shadowing
PR related to this trello task --> https://trello.com/c/jwYJfoSR
:tada: This PR is included in version 1.4.0 :tada:
Your semantic-release bot :package::rocket:
|
gharchive/pull-request
| 2022-02-18T01:00:56 |
2025-04-01T04:55:31.242849
|
{
"authors": [
"christroutner",
"danielhumgon"
],
"repo": "Permissionless-Software-Foundation/bch-wallet-starter",
"url": "https://github.com/Permissionless-Software-Foundation/bch-wallet-starter/pull/39",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
420538568
|
Principal Eigen vectors inversion, based on the principal eigen vector
Thank you for open sourcing your code. I have a question regarding the two lines of code below.
Why you inversed the eigen vectors based on the principal eigen vector's values?
# Make sure vectors are pointing the right way
if V[0, 0] < 0: V[:, 0] *= -1
if V[0, 1] < 0: V[:, 1] *= -1
The eigen vectors physically correspond to color vectors in optical density space so they should be elementwise positive. The math doesn't know about this (if x is an eigenvector so is -x). So we have to see if we got the negative one and if so invert. Does that make sense?
Yes @Peter554. But we are flipping the vectors based on the direction of the principal vector. My question is, is there something else that i miss?
For example, why we are not taking the absolute value of the whole V array?
The eigenvectors are columns of that matrix. There is no special treatment for the principle eigenvector. I only test one element of the vector as that's enough to check if it points the right way.
@Peter554 Thanks for your help. Could you please provide the macenko stain normalized samples to cross check them with mine? I saw only the original on your drive.
@Peter554
I would be obliged if you can send me by email the normalized samples you have on your website.
My email is: dimimallios@gmail.com
Kind regards,
Dimitris
Given issue #32 I'm having trouble even using the package at the moment but when this resolves I'll try to remember to send you my result.
|
gharchive/issue
| 2019-03-13T14:42:34 |
2025-04-01T04:55:31.248285
|
{
"authors": [
"Peter554",
"dimimal"
],
"repo": "Peter554/StainTools",
"url": "https://github.com/Peter554/StainTools/issues/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
296862510
|
Extra Argument
I have been using this library for a while however I cannot for the life of me figure out how to pass an EXTRA argument. Any arguments I append to the end of the install command are dropped.
The Service Path Should be: C:\MyService.exe action:run arg1:MyArg
Please Help
Sorry, I am kind of swamped right now so I wasn't able to look into this. what I suggest is cloning the repo and giving it a go locally. It's fairly simple, if there's a big you should be able to spot it no probs. I myself never needed this functionality therefore I just added it "blindly" so it might not be working properly. It would be weird though cause there's quite a few people using it and I would be surprised if you were the first one to try this. But you never know. Worst case I'll give it a go in few days/weeks as time allows
I was able to get it working, it looks like there is a minor bug... there is some code that drops the 1st argument if the App is a standalone exe (opposed to .dll). I believe this was mistakenly dropping my extra argument. I was able to work around the issue by passing 2 extra arguments, the 1st is dropped and the 2nd makes it through. Not pretty but it works.
ah yeah, I remember that, the first argument normally is the path to the dll. Or at least was when I've written it. Hence I dropped it it wasn't necessary. Ok thanks for looking into it, I'll fix it when I get to it. Good to know that you can at least work for now with that hack.
Fixed by https://github.com/PeterKottas/DotNetCore.WindowsService/releases/tag/v2.0.7. Feel free to reopen if you find any issues with the current solution.
Hello, I'm using this versión of lib:
I launch this command to register the service:
"SePro TS OPCUASwitch Client Service.exe" "name:SeP..." action:install uno dos tres
The service has two parameters and "one" parameter is lost:
Thank you for this library :-)
|
gharchive/issue
| 2018-02-13T19:46:38 |
2025-04-01T04:55:31.253937
|
{
"authors": [
"PeterKottas",
"dgfmdrv",
"rydergillen-compacSort"
],
"repo": "PeterKottas/DotNetCore.WindowsService",
"url": "https://github.com/PeterKottas/DotNetCore.WindowsService/issues/74",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
563994494
|
UnhandledPromiseRejectionWarning:
Hi,
I tried today to use your plugin you created but it doesn't work. I'm getting following error:
(node:21584) UnhandledPromiseRejectionWarning: Error: Request failed with status code 400
at createError (D:\Dev\codeceptJS\node_modules\axios\lib\core\createError.js:16:15)
at settle (D:\Dev\codeceptJS\node_modules\axios\lib\core\settle.js:17:12)
at IncomingMessage.handleStreamEnd (D:\Dev\codeceptJS\node_modules\axios\lib\adapters\http.js:236:11)
at IncomingMessage.emit (events.js:333:22)
at IncomingMessage.EventEmitter.emit (domain.js:485:12)
at endReadableNT (_stream_readable.js:1201:12)
at processTicksAndRejections (internal/process/task_queues.js:84:21)
(node:21584) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag --unhandled-rejections=strict (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 2)
(node:21584) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
do you know what could be the issue?
Thanks
@xgene34 what is your config file?
This is inside
testrail: {
require: 'codeceptjs-testrail',
host: 'https://URL',
user: 'USERNAME',
password: '1DwsVf09t43JMm3CKfWd-4HHFGrh4OPjAa2Yw5FAx',
suiteId: 1569,
projectId: 117058,
runName: 'Automated tests run ${dd/mm/yyy H:M}',
enabled: true
}
@xgene34 nothing is wrong it seemed, could you copy the logs when running with --verbose flag?
@PeterNgTr
here are the logs:
npx codeceptjs run --grep "Refund" --verbose
(node:27676) UnhandledPromiseRejectionWarning: Error: Request failed with status code 400
at createError (D:\Dev\codeceptJS\node_modules\axios\lib\core\createError.js:16:15)
at settle (D:\Dev\codeceptJS\node_modules\axios\lib\core\settle.js:17:12)
at IncomingMessage.handleStreamEnd (D:\Dev\codeceptJS\node_modules\axios\lib\adapters\http.js:236:11)
at IncomingMessage.emit (events.js:333:22)
at IncomingMessage.EventEmitter.emit (domain.js:485:12)
at endReadableNT (_stream_readable.js:1201:12)
at processTicksAndRejections (internal/process/task_queues.js:84:21)
(node:27676) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag --unhandled-rejections=strict (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 2)
(node:27676) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
CodeceptJS v2.4.3
Using test root "D:\Dev\codeceptJS"
Helpers: WebDriver, ChaiWrapper, Mochawesome
Plugins: screenshotOnFail, retryFailedStep, testrail
After execution:
OK | 1 passed, 11 skipped // 20s
Emitted | global.result ([object Object])
Emitted | global.after ([object Object])
(node:27676) UnhandledPromiseRejectionWarning: Error: Request failed with status code 400
at createError (D:\Dev\codeceptJS\node_modules\axios\lib\core\createError.js:16:15)
at settle (D:\Dev\codeceptJS\node_modules\axios\lib\core\settle.js:17:12)
at IncomingMessage.handleStreamEnd (D:\Dev\codeceptJS\node_modules\axios\lib\adapters\http.js:236:11)
at IncomingMessage.emit (events.js:333:22)
at IncomingMessage.EventEmitter.emit (domain.js:485:12)
at endReadableNT (_stream_readable.js:1201:12)
at processTicksAndRejections (internal/process/task_queues.js:84:21)
(node:27676) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag --unhandled-rejections=strict (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 4)
@xgene34 well, I'm into this.
@PeterNgTr thx. Let me know if you figure something out or u need something
oh, the issue comes from here -> https://github.com/PeterNgTr/codeceptjs-testrail/blob/6ddb0e62af7f1968510e46b5ce1de1fcfafb5a49/index.js#L205
runId is undefined on that call, hence, it cannot update testrun and leads to the next error. (cannot update test result on that testrun while testrun is null)
@PeterNgTr but shouldn't run be created by plugin?
@xgene34 it's now fixed with 1.4.0. Closing this for now.
@PeterNgTr - I noticed today when having more then 2 tests cases to report I get an error in console:
{ id: '720084', elapsed: '7s' }
{ id: '529330', elapsed: '7s' }
(node:28592) UnhandledPromiseRejectionWarning: Error: Cannot attach file due to Error: read ECONNRESET
at D:\Dev\PayTestFramework\node_modules\codeceptjs-testrail\index.js:156:10
at processTicksAndRejections (internal/process/task_queues.js:97:5)
(node:28592) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag --unhandled-rejections=strict (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 2)
(node:28592) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
(node:28592) UnhandledPromiseRejectionWarning: Error: Cannot attach file due to Error: Request failed with status code 500
at D:\Dev\PayTestFramework\node_modules\codeceptjs-testrail\index.js:156:10
at processTicksAndRejections (internal/process/task_queues.js:97:5)
(node:28592) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag --unhandled-rejections=strict (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 4)
npm ERR! Test failed. See above for more details.
Although error appear all tests are reported in testrail.
Can you help here?
Thanks
@PeterNgTr - I noticed today when having more then 2 tests cases to report I get an error in console:
{ id: '720084', elapsed: '7s' }
{ id: '529330', elapsed: '7s' }
(node:28592) UnhandledPromiseRejectionWarning: Error: Cannot attach file due to Error: read ECONNRESET
at D:\Dev\PayTestFramework\node_modules\codeceptjs-testrail\index.js:156:10
at processTicksAndRejections (internal/process/task_queues.js:97:5)
(node:28592) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag --unhandled-rejections=strict (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 2)
(node:28592) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code.
(node:28592) UnhandledPromiseRejectionWarning: Error: Cannot attach file due to Error: Request failed with status code 500
at D:\Dev\PayTestFramework\node_modules\codeceptjs-testrail\index.js:156:10
at processTicksAndRejections (internal/process/task_queues.js:97:5)
(node:28592) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag --unhandled-rejections=strict (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 4)
npm ERR! Test failed. See above for more details.
Although error appear all tests are reported in testrail.
Can you help here?
Thanks
Hey @xgene34 are you using the latest version of this lib?
@PeterNgTr ahh it looks like it not :-) I'll try with 1.4.2 I had 1.4.0
@PeterNgTr - so don't get stacktrace now but still I get an error:
Cannot attach file due to Error: read ECONNRESET
Cannot attach file due to Error: Request failed with status code 500
npm ERR! Test failed. See above for more details.
I also noticed that testrail itself get's unresponsive when sending results to it
@xgene34 could you do me a favor? Run with --verbose mode to get more detailed logs.
@PeterNgTr hmm so I updated also codeceptjs now and webdriver and I runned the tests with --verbose mode
here is the result :
FAIL | 6 passed, 2 failed, 4 skipped // 3m
Emitted | global.result ([object Object])
Emitted | global.after ([object Object])
Cannot attach file due to Error: Request failed with status code 500
Cannot attach file due to Error: Request failed with status code 500
@xgene34 are you using self-hosted testrail or cloud testrail? also, please make sure your account has proper permissions for attachments.
http://docs.gurock.com/testrail-userguide/howto-permissions
@PeterNgTr it's self-hosted. Yeah could be that I don't have permission. Have to check with testrail admin
@xgene34 maybe you should take a look at this. https://discuss.gurock.com/t/cannot-write-to-reports-dir/1617
@PeterNgTr thanks a lot for the help
|
gharchive/issue
| 2020-02-12T13:18:49 |
2025-04-01T04:55:31.283461
|
{
"authors": [
"PeterNgTr",
"xgene34"
],
"repo": "PeterNgTr/codeceptjs-testrail",
"url": "https://github.com/PeterNgTr/codeceptjs-testrail/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
957138410
|
"undefined" as both title and message for all Toasts created
<head lang="en">
<link href="node_modules/bootstrap/dist/css/bootstrap.min.css" rel="stylesheet">
<link href="node_modules/bootstrap-icons/font/bootstrap-icons.css" rel="stylesheet">
<link rel="stylesheet" href="node_modules/bootstrap-toaster/css/bootstrap-toaster.min.css" />
<link rel="stylesheet" href="./style.css">
<script src="node_modules/jquery/dist/jquery.min.js" type="application/javascript"></script>
<script src="node_modules/bootstrap/dist/js/bootstrap.min.js" type="application/javascript"></script>
<script src="node_modules/socket.io-client/dist/socket.io.min.js" type="application/javascript"></script>
</head>
<body>
...
<script src="node_modules/bootstrap-toaster/js/bootstrap-toaster.min.js"></script>
<script> window.onload = function() {
Toast.create("Wow, that was easy!", "Just like that, this toast will appear on the page", TOAST_STATUS.SUCCESS, 5000);}
</script>
</body>
I won't have my computer that I develop on with me for a while, so I can't be a great help, but what versions of Bootstrap and Bootstrap Toaster do you have installed?
Make sure to use 5.x if you are running Bootstrap 5, and if you're using Bootstrap 4, use 4.x.
Latest for both. Here is my package.json:
{
"dependencies": {
"bootstrap": "^5.0.2",
"bootstrap-toaster": "^5.1.0-rc1",
"jquery": "^3.6.0",
"resumablejs": "^1.1.0",
"socket.io-client": "^4.1.3"
}
}
Changed bootstrap-toaster to an older version in my package.json, and reinstalled dependencies. Everything works now!
"bootstrap-toaster": "^5.0.0"
Glad you got it working! Yeah I’ve made a lot of changes in 5.1.0 so I haven’t officially released it yet
|
gharchive/issue
| 2021-07-31T00:56:09 |
2025-04-01T04:55:31.291118
|
{
"authors": [
"PeytonRG",
"eamonniknafs"
],
"repo": "PeytonRG/BootstrapToaster",
"url": "https://github.com/PeytonRG/BootstrapToaster/issues/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
741825025
|
Cannot create sets with circular reference
We cannot create a set and add it to itself. The following code when evaluated in JavaScript answers 1 instead of 2.
s := Set new.
s add: 'A'.
s add: s.
s size.
The problem is caused by includes: method defined in PjTArray. This caused another bug. The following code answers true instead of false, when evaluated in JavaScript.
array := #('X').
array includes: array.
|
gharchive/issue
| 2020-11-12T18:23:33 |
2025-04-01T04:55:31.301376
|
{
"authors": [
"bouraqadi"
],
"repo": "PharoJS/PharoJS",
"url": "https://github.com/PharoJS/PharoJS/issues/56",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2129153895
|
Inventory drag and drop
It's very unintuitive to me to click to pick up and click to drop, though many games support this. Can a click and drag to move items be added?
Currently no plans for me to add this, as gamepad and mouse interaction use the same input map actions to keep things simple and flexible.
Would be happy to add/include if someone else wants to tackle this.
|
gharchive/issue
| 2024-02-11T20:23:32 |
2025-04-01T04:55:31.302780
|
{
"authors": [
"Phazorknight",
"ac-arcana"
],
"repo": "Phazorknight/Cogito",
"url": "https://github.com/Phazorknight/Cogito/issues/42",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1571295402
|
advances in code formatting
The documentation for formatting is a little outdated, so here are some advances for it.
Still no custom indentation configuration but it starts applying some formatting in certain circumstances, probably we can discuss on several ways to indent the code based on the tokens you have defined.
Is it possible to introduce an indentation configuration based on this PR? maybe there'd be a global configuration hosted in the IDEA platform that we can read from it and apply.
that depends totally in the tokens you have declared.
I personally tend to use two indentation styles:
my_rule( Param1, Param2 ) :-
number( Param1 ),
between( 1, 10, Param2 ).
That's the one I've seen in many books.
And the other I use is:
my_rule(
Param1
, Param2
) :-
number( Param1 )
, between( 1, 10, Param2 )
.
This is more SQL inspired and is more practical when you need to comment a whole line yet less readable.
According to documentation the options that we have are:
spacing
indent
none
regular
continuation
wrap
never
always
if exceeds border
alignment
Strangely enough, in my local testing environment, the code formatting hardly does anything on most of the code I've provided (I've also tried to add blank space manually and see if the formatting eliminates it, but nothing happened). Do you have any example that works with the formatting function introduced by this PR?
Strangely enough, in my local testing environment, the code formatting hardly does anything on most of the code I've provided (I've also tried to add blank space manually and see if the formatting eliminates it, but nothing happened). Do you have any example that works with the formatting function introduced by this PR?
@javatlacati have you got any working example for me to test on? The behavior of the formatter seems wrong to me.
not yet, I need to get trough all of your tokens to see what can be done
|
gharchive/pull-request
| 2023-02-05T07:13:38 |
2025-04-01T04:55:31.398151
|
{
"authors": [
"Phosphorus15",
"javatlacati"
],
"repo": "Phosphorus15/intellij-prolog",
"url": "https://github.com/Phosphorus15/intellij-prolog/pull/122",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2015002845
|
[Feature] Separate prompts between "BREAK" / "AND" keywords
Is this feature related to existing issues?
No response
What feature do you want or what suggestions do you have?
I typically separate prompts using BREAK and AND with newlines to make them easier to find/differentiate. I'd like to be able to see the same thing in the All-in-One prompt field too.
Are there any similar competitors that can be referenced?
No response
Additional information
No response
"BREAK" currently supports display in separate lines.
|
gharchive/issue
| 2023-11-28T17:59:47 |
2025-04-01T04:55:31.401573
|
{
"authors": [
"Physton",
"ghostsquad"
],
"repo": "Physton/sd-webui-prompt-all-in-one",
"url": "https://github.com/Physton/sd-webui-prompt-all-in-one/issues/278",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1105059955
|
[BUG] Unique index or primary key violation
Describe the bug
A clear and concise description of what the bug is.
Screenshots and Logs
[15:52:31 WARN]: Error connecting to the storage. The plugin will not work correctly.
[15:52:31 WARN]: org.h2.jdbc.JdbcSQLIntegrityConstraintViolationException: Unique index or primary key violation: "PUBLIC.PRIMARY_KEY_3 ON PUBLIC.JOBPLAYERS(UUID) VALUES 2"; SQL statement:
[15:52:31 WARN]: INSERT INTO jobplayers (uuid) VALUES (?) [23505-200]
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.message.DbException.getJdbcSQLException(DbException.java:459)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.message.DbException.getJdbcSQLException(DbException.java:429)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.message.DbException.get(DbException.java:205)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.message.DbException.get(DbException.java:181)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.index.BaseIndex.getDuplicateKeyException(BaseIndex.java:103)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.mvstore.db.MVSecondaryIndex.checkUnique(MVSecondaryIndex.java:221)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.mvstore.db.MVSecondaryIndex.add(MVSecondaryIndex.java:196)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.mvstore.db.MVTable.addRow(MVTable.java:531)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.command.dml.Insert.insertRows(Insert.java:195)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.command.dml.Insert.update(Insert.java:151)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.command.CommandContainer.update(CommandContainer.java:198)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.command.Command.executeUpdate(Command.java:251)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.jdbc.JdbcPreparedStatement.executeUpdateInternal(JdbcPreparedStatement.java:191)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//org.h2.jdbc.JdbcPreparedStatement.executeUpdate(JdbcPreparedStatement.java:152)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//com.gmail.picono435.picojobs.storage.sql.file.FlatfileStorageFactory.createPlayer(FlatfileStorageFactory.java:54)
[15:52:31 WARN]: at PicoJobs-1.0-beta-5.jar//com.gmail.picono435.picojobs.api.managers.PlayersManager.lambda$getJobPlayerFromStorage$2(PlayersManager.java:102)
[15:52:31 WARN]: at org.bukkit.craftbukkit.v1_18_R1.scheduler.CraftTask.run(CraftTask.java:101)
[15:52:31 WARN]: at org.bukkit.craftbukkit.v1_18_R1.scheduler.CraftAsyncTask.run(CraftAsyncTask.java:57)
[15:52:31 WARN]: at com.destroystokyo.paper.ServerSchedulerReportingWrapper.run(ServerSchedulerReportingWrapper.java:22)
[15:52:31 WARN]: at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136)
[15:52:31 WARN]: at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635)
[15:52:31 WARN]: at java.base/java.lang.Thread.run(Thread.java:833)
Software Details:
Server Software: Paper
Server Software Version: 1.18.1-RO.1-SNAPSHOT
Plugin Version: 1.0-beta-5
Java Version: 1.8.0_311
Additional context
Add any other context about the problem here.
Hello when does this happen?
each time I start the server, to each player who enters
Please test in the latest development build by clicking in Actions.
Fixed issue in the lastest commit, thank you for reporting it. Be free to create another issue or re-open this one if the issue persists.
|
gharchive/issue
| 2022-01-16T14:00:26 |
2025-04-01T04:55:31.425962
|
{
"authors": [
"LeonMaximal1",
"Picono435"
],
"repo": "Picono435/PicoJobs",
"url": "https://github.com/Picono435/PicoJobs/issues/150",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1334216245
|
Chat command to reselect teams
This probably just needs to invoke the gm_showteam ConCommand.
Resolved via commit f561f7cb583760eb7da0d0f36da1eabc08bed6ab.
|
gharchive/issue
| 2022-08-10T07:44:44 |
2025-04-01T04:55:31.431543
|
{
"authors": [
"Piengineer12"
],
"repo": "Piengineer12/public-addons",
"url": "https://github.com/Piengineer12/public-addons/issues/54",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
527937341
|
why MIoU can not reach 77 in cityscapes val dataset?
Thanks for your great work;but when I use your pretrained model to validate,the MIoU is 70.19536 (hardnet70_cityscapes_model.pkl)and 71.179(hardnet70_cityscapes_model_2.pkl) respectively.So I want to ask is this the best model? Or do I need to do anything else? thanks!
Hello, thank you very much for the report. It seems like there is an accidentally deleted line for converting RGB->BGR in the last pull request. I have fixed it, please git pull and try again. Thanks!
|
gharchive/issue
| 2019-11-25T08:51:23 |
2025-04-01T04:55:31.444693
|
{
"authors": [
"GitHubLiaoYong",
"PingoLH"
],
"repo": "PingoLH/FCHarDNet",
"url": "https://github.com/PingoLH/FCHarDNet/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
124909793
|
use vec! macro for ImageBuffer::new
Change ImageBuffer default constructor (for Vec container) to use vec! macro.
before:
test buffer::test::bench_conversion ... bench: 7,130,486 ns/iter (+/- 106,108) = 420 MB/s
test dynimage::bench::bench_conversion ... bench: 7,123,419 ns/iter (+/- 34,759) = 420 MB/s
test imageops::sample::tests::bench_resize ... bench: 8,605,023 ns/iter (+/- 50,953) = 236 MB/s
after:
test buffer::test::bench_conversion ... bench: 5,204,968 ns/iter (+/- 25,449) = 576 MB/s
test dynimage::bench::bench_conversion ... bench: 5,204,117 ns/iter (+/- 20,338) = 576 MB/s
test imageops::sample::tests::bench_resize ... bench: 7,673,833 ns/iter (+/- 99,008) = 265 MB/s
Changes impact impageproc too:
test affine::test::bench_rotate_bilinear ... bench: 723,508 ns/iter (+/- 43,365)
test affine::test::bench_rotate_nearest ... bench: 417,903 ns/iter (+/- 20,282)
test affine::test::bench_translate ... bench: 938,693 ns/iter (+/- 6,434)
test contrast::test::bench_equalize_histogram ... bench: 3,933,365 ns/iter (+/- 27,685)
test corners::test::bench_is_corner_fast12_12_noncontiguous ... bench: 27 ns/iter (+/- 0)
test corners::test::bench_is_corner_fast9_9_contiguous_lighter_pixels ... bench: 29 ns/iter (+/- 0)
test filter::test::bench_box_filter ... bench: 4,973,883 ns/iter (+/- 62,633)
test filter::test::bench_filter3x3_i32_filter ... bench: 3,381,504 ns/iter (+/- 21,686)
test filter::test::bench_horizontal_filter ... bench: 9,714,388 ns/iter (+/- 30,177)
test filter::test::bench_separable_filter ... bench: 6,983,245 ns/iter (+/- 26,106)
test filter::test::bench_vertical_filter ... bench: 9,856,291 ns/iter (+/- 48,827)
test integralimage::test::bench_column_running_sum ... bench: 3,698 ns/iter (+/- 51)
test integralimage::test::bench_integral_image ... bench: 1,608,849 ns/iter (+/- 10,780)
test integralimage::test::bench_row_running_sum ... bench: 3,483 ns/iter (+/- 17)
test suppress::test::bench_local_maxima_dense ... bench: 42,237 ns/iter (+/- 9,673)
test suppress::test::bench_local_maxima_sparse ... bench: 44,919 ns/iter (+/- 330)
test suppress::test::bench_suppress_non_maximum_decreasing_gradient ... bench: 209,578 ns/iter (+/- 1,372)
test suppress::test::bench_suppress_non_maximum_increasing_gradient ... bench: 229,177 ns/iter (+/- 1,350)
test suppress::test::bench_suppress_non_maximum_noise ... bench: 258,725 ns/iter (+/- 1,678)
after
test affine::test::bench_rotate_bilinear ... bench: 665,395 ns/iter (+/- 7,632)
test affine::test::bench_rotate_nearest ... bench: 356,090 ns/iter (+/- 4,673)
test affine::test::bench_translate ... bench: 507,069 ns/iter (+/- 3,498)
test contrast::test::bench_equalize_histogram ... bench: 3,520,825 ns/iter (+/- 45,802)
test corners::test::bench_is_corner_fast12_12_noncontiguous ... bench: 27 ns/iter (+/- 1)
test corners::test::bench_is_corner_fast9_9_contiguous_lighter_pixels ... bench: 29 ns/iter (+/- 0)
test filter::test::bench_box_filter ... bench: 4,603,757 ns/iter (+/- 311,278)
test filter::test::bench_filter3x3_i32_filter ... bench: 2,564,050 ns/iter (+/- 112,980)
test filter::test::bench_horizontal_filter ... bench: 9,290,717 ns/iter (+/- 142,560)
test filter::test::bench_separable_filter ... bench: 6,668,658 ns/iter (+/- 180,966)
test filter::test::bench_vertical_filter ... bench: 9,458,380 ns/iter (+/- 183,229)
test integralimage::test::bench_column_running_sum ... bench: 3,820 ns/iter (+/- 142)
test integralimage::test::bench_integral_image ... bench: 1,452,976 ns/iter (+/- 78,702)
test integralimage::test::bench_row_running_sum ... bench: 3,485 ns/iter (+/- 19)
test suppress::test::bench_local_maxima_dense ... bench: 38,114 ns/iter (+/- 426)
test suppress::test::bench_local_maxima_sparse ... bench: 45,574 ns/iter (+/- 546)
test suppress::test::bench_suppress_non_maximum_decreasing_gradient ... bench: 211,604 ns/iter (+/- 2,357)
test suppress::test::bench_suppress_non_maximum_increasing_gradient ... bench: 231,685 ns/iter (+/- 3,345)
test suppress::test::bench_suppress_non_maximum_noise ... bench: 261,134 ns/iter (+/- 4,973)
Thanks!
|
gharchive/pull-request
| 2016-01-05T07:06:57 |
2025-04-01T04:55:31.473862
|
{
"authors": [
"bvssvni",
"tafia"
],
"repo": "PistonDevelopers/image",
"url": "https://github.com/PistonDevelopers/image/pull/500",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
285230453
|
Import tool doesn't support tilemap.json or cache files
Need to come up with a solution for importing from projects that have cached sprites and tilemap.json files.
This is fixed.
|
gharchive/issue
| 2017-12-30T23:41:03 |
2025-04-01T04:55:31.521240
|
{
"authors": [
"PixelVision8",
"jessefreeman"
],
"repo": "PixelVision8/GameCreator",
"url": "https://github.com/PixelVision8/GameCreator/issues/575",
"license": "MS-PL",
"license_type": "permissive",
"license_source": "github-api"
}
|
245742785
|
Add launch and import back into edit tool
When selecting an asset or tool archive a contextual button should appear in the action bar to offer extra functionality.
System and Artpacks should be imported into a project and tools should be launched.
This has been fixed.
|
gharchive/issue
| 2017-07-26T14:33:18 |
2025-04-01T04:55:31.522563
|
{
"authors": [
"PixelVision8"
],
"repo": "PixelVision8/GameCreator",
"url": "https://github.com/PixelVision8/GameCreator/issues/84",
"license": "MS-PL",
"license_type": "permissive",
"license_source": "github-api"
}
|
157387169
|
Can't downgrade 9.2 2DS with 11.0 emu
Using sysUpdater: 004001000022000.cia install failed with exception "0xD8E08027"
Using PlaiSysUpdater: downgrade failed for nfirm ("An error occured, press button power...." message)
Already tried 12 times to restore emunand and retry
In the luma folder put the firmware.bin correspondent to your console
Really reammy weird.
I already had firmware.bin. I formated everything and tried again from scratch, unlinked nands and downgraded, the same "an error occured...". Rebooted to OTPHelper, used the one click setup, didn't worked, reinjected FIRM0 and FIRM1, retried, it worked.
And... AND.... Bricked. I'll install a hardmod next weekend
You used One Click Setup, everything verified properly, and you can't boot without the SD card in?
Exactly
You are bricked and will need a hardmod, but you should contact @d0k3 about this odd issue.
And what about the failed downgrade for nfirm? Should I make an issue in PlaiSysUpdater repo?
I have an assumption that the system settings (that's the app that did not install properly at first) still did not install correctly. Possible reasons? Some earlier modifications maybe?
Anyways @Amadren, get back to me once you have that hardmod. Best to join us on freenode #Cakey then.
@dok3, Nope no previous modifications and I don't think of any possible reason, it was my little brother's (4yo) 2DS, I even used a whole new SD Card. And ok, I'll join you on #Cakey when i'll have my hardmod. I'll install it sometimes during next week.
|
gharchive/issue
| 2016-05-29T13:50:00 |
2025-04-01T04:55:31.554377
|
{
"authors": [
"Amadren",
"Plailect",
"d0k3",
"willkof2"
],
"repo": "Plailect/Guide",
"url": "https://github.com/Plailect/Guide/issues/164",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2730509351
|
chore(tooling): set node alpine to node:20-alpine3.20
https://pins-ds.atlassian.net/browse/APPLICS-1189
we don't specify an alpine version in our dockerfiles and things have broken specifically for prisma.
node alpine has been updated, and openssl path has changed, setting node alpine to node:20-alpine3.20
Useful information to review or test
Type of change 🧩
[ ] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Documentation update
[ ] Other (please explain in the description section above)
Checklist before requesting a review
[ ] I have performed a self-review of my own code
[ ] I have double checked this work does not include any hardcoded secrets or passwords
[ ] I have made corresponding changes to the documentation
[ ] I have provided details on how I have tested my code
[ ] I have referenced the ticket number above
[ ] I have provided a description of the ticket
[ ] I have included unit tests to cover any testable code changes
:tada: This PR is included in version 2.33.0 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
:tada: This PR is included in version 2.41.0 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
:tada: This PR is included in version 1.4.0 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
|
gharchive/pull-request
| 2024-12-10T15:50:51 |
2025-04-01T04:55:31.569782
|
{
"authors": [
"AnthonyJWalsh",
"emil-placheta"
],
"repo": "Planning-Inspectorate/applications-service",
"url": "https://github.com/Planning-Inspectorate/applications-service/pull/1398",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2181078323
|
PermissionError: [Errno 13] Permission denied: '/usr/lib/python3/dist-packages/guidata/tests/data/genreqs/requirements.rst'
Hello, while preparing the Debian package of guidata, I got this failure.
110s =================================== FAILURES ===================================
110s _______________________ test_generate_requirement_tables _______________________
110s
110s def test_generate_requirement_tables():
110s """Test generate_requirement_tables."""
110s > genreqs.gen_path_req_rst(GR_PATH, "guidata", ["Python>=3.8", "PyQt>=5.11"], GR_PATH)
110s
110s tests/unit/test_genreqs.py:23:
110s _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
110s
110s path = '/usr/lib/python3/dist-packages/guidata/tests/data/genreqs'
110s modname = 'guidata', additional_reqs = ['Python>=3.8', 'PyQt>=5.11']
110s destpath = '/usr/lib/python3/dist-packages/guidata/tests/data/genreqs'
110s
110s def gen_path_req_rst(
110s path: str, modname: str, additional_reqs: list[str], destpath: str | None = None
110s ) -> None:
110s """Generate install 'requirements.rst' reStructuredText text.
110s This reStructuredText text is written in a file which is by default located in
110s the `doc` folder of the module.
110s
110s Args:
110s path (str): Path to folder containing pyproject.toml or setup.cfg file
110s modname (str): Module name
110s additional_reqs (list[str]): Additional requirements
110s destpath (str): Destination path for requirements.rst file (optional).
110s """
110s requirements = extract_requirements_from_toml(path)
110s if requirements is None:
110s requirements = extract_requirements_from_cfg(path)
110s if requirements is None:
110s raise RuntimeError(
110s "Could not find pyproject.toml or setup.cfg file in %s" % path
110s )
110s requirements = extract_requirements_from_toml(path)
110s if requirements is None:
110s requirements = extract_requirements_from_cfg(path)
110s if requirements is None:
110s raise RuntimeError(
110s "Could not find pyproject.toml or setup.cfg file in %s" % path
110s )
110s text = f"""The :mod:`{modname}` package requires the following Python modules:
110s
110s {reqlist_to_table(requirements["main"]+additional_reqs)}"""
110s for category, title in (
110s ("dev", "development"),
110s ("doc", "building the documentation"),
110s ("test", "running test suite"),
110s ):
110s if category in requirements:
110s text += f"""
110s
110s Optional modules for {title}:
110s
110s {reqlist_to_table(requirements[category])}"""
110s if destpath is None:
110s destpath = osp.join(path, "doc")
110s > with open(osp.join(destpath, "requirements.rst"), "w") as fdesc:
110s E PermissionError: [Errno 13] Permission denied: '/usr/lib/python3/dist-packages/guidata/tests/data/genreqs/requirements.rst'
110s
110s /usr/lib/python3/dist-packages/guidata/utils/genreqs.py:192: PermissionError
During the test it tries to write some file in a non writable location.
here the command executed for the tests.
Test-Command: set -efu
; export GUIDATA_UNATTENDED_TESTS=1
; export HOME="$AUTOPKGTEST_TMP"
; cp -r guidata/tests "$AUTOPKGTEST_TMP"
; for py in $(py3versions -r 2>/dev/null)
; do cd "$AUTOPKGTEST_TMP"
; echo "Testing with $py:"
; $py -m pytest tests
; done
Thanks for the feedback Fred ;-)
I was about to release a critical bug fix (V3.4.1).
So I'm also fixing this right now: in fact this test should not be executed within the test suite (it should be executed manually for development purpose).
@picca : FYI, just release V3.4.1
thanks :))
----- Le 12 Mar 24, à 14:01, Pierre Raybaut @.*** a écrit :
@picca : FYI, just released V3.4.1
--
Reply to this email directly or view it on GitHub:
https://github.com/PlotPyStack/guidata/issues/70#issuecomment-1991601342
You are receiving this because you were mentioned.
Message ID: @.***>
|
gharchive/issue
| 2024-03-12T09:03:03 |
2025-04-01T04:55:31.599951
|
{
"authors": [
"PierreRaybaut",
"picca"
],
"repo": "PlotPyStack/guidata",
"url": "https://github.com/PlotPyStack/guidata/issues/70",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2288671545
|
Improve switch time complexity for constant cases
First version of an improvement to the switch statement to change the time complexity from O(n) to O(log n) (see https://github.com/PlutoLang/Pluto/issues/860).
This does currently not pass the debug tests because of the extra code generated in the main chunk.
Some notes for the current implementation.
The generated bytecode is not stable due to the use of lua tables. I do not know how important this point is.
It would also be possible to move the lookup table initialization into the chunk. This would remove like two instructions from the switch for the initialization test. However, this would iniliatize all switches, even the ones never used.
Rebased this branch because we treat Git as a tree; not a directed graph.
The switchimpl function is now incredibly long, so I'm assuming you're taking ownership and I can ping you if there's any issues with it? :P
The generated bytecode is not stable due to the use of lua tables. I do not know how important this point is.
Could you elaborate on this some more? Would this PR cause issues with the upgrade to Lua 5.5?
You can ping me when there are issues with the switch implementation.
With not stable I mean that string.dump will give out different results on different runs for functions with switch statements due to different order of the table initialization.
I use a lua table to during the compiling to gather all the constants, but a lua table traversal order is not defined, so can result in different orderings.
I see. That does sound a bit ugly, but I also can't imagine a better solution than putting a std::stack<std::map<...>> on LexState.
I could use a std::vector in the switchimpl function, but that would just be there for the ordering. So if a stable output is important, this should be simple to add.
You cannot stack-allocate such a thing because RAII is not guaranteed here.
You are right, then I need to look into case_pc too as it does this currently.
Yes, with the table lookup, check for initialization, and other overhead the binary search will only be used from around 16 cases upwards.
Try with the switch speed test:
local N = 2000
local code = "return function(val)\nswitch val do"
for i=1, N do
code = code .. "\ncase " .. i .. ".5: return " .. i
end
code = code .. "\ndefault: return -1\nend\nend"
local f = load(code)()
-- Initialize, this might take more time
assert(f(0) == -1)
for i=1, N do
assert(f(i+0.5) == i)
end
assert(f(0) == -1)
The problem is that I think >95% of switch blocks are gonna have <16 cases.
And are the other stack allocated std::vector in the parser then not also buggy?
That depends. If there is a chance a Lua error could be thrown in their scope, then it would not be deallocated when a longjump is used.
The problem is that I think >95% of switch blocks are gonna have <16 cases.
Yes, but they are hard to get better. When multiple case values map to the same case then it can happen earlier.
int average_time_linear = num_consts/2;
int average_time_table = luaO_ceillog2((unsigned int)case_pc.size()) + 4; /* Load and check cache, load value, check nil case, then do the binary search */
if (average_time_linear <= average_time_table) {
/* Not worth the effort, just do a linear scan */
That depends. If there is a chance a Lua error could be thrown in their scope, then it would not be deallocated when a longjump is used.
Yes, that could happen in almost all the cases I saw, so I assumed that Pluto is using exceptions and not longjump.
I fixed the memory leak issues with std::vector in the switch statement, need to check std::function.
Will also look in the speed decrease of the example you mentioned.
I would advise just using a void* for user-data and casting it as before. Also nothing wrong with std::vector, as long as it's somewhere on LexState, because that does actually get cleaned up in all cases.
When putting the vector on LexState it would need to handle reentrancy, And the tables also use the correct allocation function from the Lua state.
I guess the speed decrease for the simple case is due to the switch always starting with a jump and not with a conditional jump like before. This results in one instruction more per loop. I will look to emit the first case in place and jump on false to the other logic.
Still a pessimisation, which really shouldn't happen since you seem to have logic to "not bother" for small switches? Maybe this patch would've been better if it focused on trying to add the new functionality without effectively rewriting the whole thing.
When just looking at
local value = 3
switch value do
case 1:
case 2:
case 3:
cc = cc + 1
break
case 4:
case 5:
break
end
I get the same code from main & this branch now. So I have no clue where the slowdown should come from.
You are able to reproduce it, tho, right?
No, I am not able to reproduce the slowdown. Over multiple runs sometimes the main branch is faster, sometimes this one.
Well, it seems to only be on MSVC, and I guess "within margin of error," but now I've tried a benchmark where your implementation should undoubtably be faster, and I can also not detect a difference:
local start = os.clock()
local cc = 0
for i = 1, 100000000 do
local value = 3
switch value do
case 1: break
case 2: break
case 3: break
case 4: break
case 5: break
case 6: break
case 7: break
case 8: break
case 9: break
case 10: break
case 11: break
case 12: break
case 13: break
case 14: break
case 15: break
case 16: break
case 17: break
case 18: break
case 19: break
case 20: cc = cc + 1 break
end
end
print(os.clock() - start)
return print(cc)
With the 20 cases both average_time_linear and average_time_table are both 9, so it will still use the linear lookup. Either add a case 21 or remove the break from all cases but 19.
Okay, now we have ~3.5 on main, and ~2.4 on this branch. That is cool, but...
It's an optimisation targeted at a very small minority of switch blocks.
That is slower on MSVC in the remaining (overwhelming majority) of switch blocks.
At the cost of increasing the severely increasing the code complexity of the parser's switch functions.
I don't know why the same bytecode and only changes to the parser make the interpreter slower on MSVC. Maybe something gets aligned differenty which reduces the speed. But that could happen with any change.
As for the complexity vs benefit is something you need to decide. I have no problem with closing this. I had my fun implementing this and have no problem when I do not need to maintain in and further.
I think this is a very impressive optimization, however I share some of the concerns already raised in this thread. The most concerning point goes to maintainability of the switch statement. The complexity increase is great enough to say that the lessons learned over time with our switch statement — the bugs fixed, the overall polish applied, the time-tested factor — would be reset. While the performance increase is quite good, the scope of when that performance increase will apply is too limited to justify the aforementioned costs.
|
gharchive/pull-request
| 2024-05-10T00:23:18 |
2025-04-01T04:55:31.632097
|
{
"authors": [
"Sainan",
"XmiliaH",
"well-in-that-case"
],
"repo": "PlutoLang/Pluto",
"url": "https://github.com/PlutoLang/Pluto/pull/862",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1399684759
|
Proposal: podcast:episodeType
I'm preparing an episode of The Audacity to Podcast to teach about the "full," "bonus," and "trailer" episode types and when to use them. And I thought it would be good for use to include a drop-in replacement tag for <itunes:episodeType> since it pairs very well with episode numbers and seasons.
Thus, the <podcast:episodeType> tag, but with some improvements.
Apple explains their tag as follows:
The episode type.
If an episode is a trailer or bonus content, use this tag.
Where the episodeType value can be one of the following:
Full (default). Specify full when you are submitting the complete content of your show.
Trailer. Specify trailer when you are submitting a short, promotional piece of content that represents a preview of your current show.
Bonus. Specify bonus when you are submitting extra content for your show (for example, behind the scenes information or interviews with the cast) or cross-promotional content for another show.
The rules for using trailer and bonus tags depend on whether the itunes:season and itunes:episode tags have values:
Trailer:
No season or episode number: a show trailer
A season number and no episode number: a season trailer. (Note: an episode trailer should have a different than the actual episode)
Episode number and optionally a season number: an episode trailer/teaser, later replaced with the actual episode
Bonus:
No season or episode number: a show bonus
A season number: a season bonus
Episode number and optionally a season number: a bonus episode related to a specific episode
The resulting tag would look like <itunes:episodeType>trailer</itunes:episodeType>
The most important takeaway is that these bonus and trailer types are connected to something: the whole show, a season, or an episode.
I think our tag should function the same with this information, and add some additional uses.
Extra or special: This would be the type for any kind of extra episode that's a break from the norm, but is not a bonus for existing content. For example, if my show is a weekly solo show, but I publish a bunch of 5-minute interviews from an event, those interviews would not be the bonus type, because they're not attached to any existing content. They're standalone, but extra, and thus the "extra" type would be more appropriate than "bonus." This would give podcasters the episode type I often see them incorrectly thinking "bonus" is for.
Rerun: This would indicate that the episode is republishing old content—either in whole or multiple segments.
Sequel or miniseries: That might not be the right word, but here's what I envision. This type indicates that an episode is a part of a miniseries regardless of the whole show type. Consider some TV shows, for example. The show is an overall serial format, but then there's a part 1 and part 2 miniseries. In other words, a small series within a series. But with out bonus and trailer types are attached to an episode/season number, I think this tag would stand alone, allowing "part 1" to be episode 10 and "part 2" to be episode 11. What happens when there are multiple miniseries and how to group them? That could probably be handled with our season tag. But the following is another possible solution.
Additional label attribute?
While the Apple tag allows only a single value, I think our tag should have an optional "label" attribute, which can display to the audience as well as help group miniseries together.
For example, <podcast:episodeType label="Foundations">miniseries</podcast:episodeType>
This could be used when a particular season is already necessary, but the miniseries spreads across seasons or is independent of them.
But I'll admit this is probably an extremely rare scenario and nearly every situation I consider could be handled by <podcast:season> instead.
For example, many pastors will preach sermon series, but sometimes take a break from that series for holiday themes or summer. So a "Foundations" sermon series could be indicated in the season tag, while the episodes are assigned to it as normal episodes. But the "miniseries" episode type could be for a series too small to warrant its own season.
Support multiple types?
I can foresee the need for an episode to be multiple types. Like rerunning a past bonus episode. Or maybe something is a trailer for a miniseries. So it might be necessary to allow multiple instances of the episodeType tag.
Added a "rerun" type, thanks to inspiration from thejavakat.
It was pointed out to me that we're allowing decimals in episode numbers to allow for the things I think episode types can solve better.
I thought of an additional episode type: promo or promotion. This can be used for an episode that serves only to promote something, like a special event, another podcast, a sale, and such.
This kind of content shouldn't be a trailer or bonus because it might not be tied to the rest of the show (it could be promoting a different show or something that's not even a podcast). And it's not a "special episode" either.
Promos would be the kind of content a podcaster might want to remove after the event has happened. But they might also opt to leave the promo in the feed indefinitely, such as when crosspromoting another podcast (of their own or someone else's).
This would pair perfectly with my new proposal for a <podcast:expire> tag (#448) to make episodes (of any type) automatically dim or hide from podcast apps and directories after a time is past.
|
gharchive/issue
| 2022-10-06T13:55:28 |
2025-04-01T04:55:31.655517
|
{
"authors": [
"theDanielJLewis"
],
"repo": "Podcastindex-org/podcast-namespace",
"url": "https://github.com/Podcastindex-org/podcast-namespace/issues/398",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2318314292
|
I output gradient values during training and observed that gradient explosion has always existed, but as long as nan values do not appear, it does not seem to affect the training effect. However, more and more gradient disappearance problems have made it impossible for my network to learn useful information.
Hello, I trained on my two-category semantic segmentation task. I found that the first few pieces of data trained well, but the performance will be very poor later. I output the gradient information to view it. I found that as the network trains, More and more layers are experiencing the vanishing gradient problem, but I cannot add a bn layer to the network. Do you have any suggestions?
Maybe you can try to add a normalization in QKV project just as PTv2?
I also notice gradient explosions giving me nans and infs, I suspect some additional normalization might help, but I'm not sure where and for which layer. For now I am using gradient clipping at a value of 1.0 (torch.nn.utils.clip_grad_norm_(model.parameters(), grad_clip_val))
It seems to be related more to the 3D sparse convolutions for my use case rather than the attention stuff. I'm curious if you have tried using regular batch norm here instead of layer norm? In my mind the layer norm makes more sense, but I do see examples in spconv using 3d convolutions followed by batch norm. I'll give it a shot and report back at some point.
Maybe you can try to add a normalization in QKV project just as PTv2?
do you mean at the start or at the end of the attention layer ?
I also notice gradient explosions giving me nans and infs, I suspect some additional normalization might help, but I'm not sure where and for which layer. For now I am using gradient clipping at a value of 1.0 (torch.nn.utils.clip_grad_norm_(model.parameters(), grad_clip_val))
It seems to be related more to the 3D sparse convolutions for my use case rather than the attention stuff. I'm curious if you have tried using regular batch norm here instead of layer norm? In my mind the layer norm makes more sense, but I do see examples in spconv using 3d convolutions followed by batch norm. I'll give it a shot and report back at some point.
We used to apply BN in v1 and v2; (Check our ablation in paper) we want to enable LN as it is independent of batch size. Nice try, waiting for your response.
I also notice gradient explosions giving me nans and infs, I suspect some additional normalization might help, but I'm not sure where and for which layer. For now I am using gradient clipping at a value of 1.0 (torch.nn.utils.clip_grad_norm_(model.parameters(), grad_clip_val))
It seems to be related more to the 3D sparse convolutions for my use case rather than the attention stuff. I'm curious if you have tried using regular batch norm here instead of layer norm? In my mind the layer norm makes more sense, but I do see examples in spconv using 3d convolutions followed by batch norm. I'll give it a shot and report back at some point.
Any update on this ?
Has anyone found the origin of the gradient explosion and any way to fix it without gradient clipping ?
@AxelCole
If you look at the spconv library, they have 1D batch norm layers on the feature channels of the points, try replacing the 3D submanifold convolution layers with a sparse sequential using 3d submanifold convolution, batch norm and relu, might do the trick and add some stability.
Clip gradient is very easy to apply (I will update it in the next version), like below:
(pointcept.engine.train.py)
def run_step(self):
input_dict = self.comm_info["input_dict"]
for key in input_dict.keys():
if isinstance(input_dict[key], torch.Tensor):
input_dict[key] = input_dict[key].cuda(non_blocking=True)
with torch.cuda.amp.autocast(enabled=self.cfg.enable_amp):
output_dict = self.model(input_dict)
loss = output_dict["loss"]
self.optimizer.zero_grad()
if self.cfg.enable_amp:
self.scaler.scale(loss).backward()
self.scaler.unscale_(self.optimizer)
if self.cfg.clip_grad is not None:
torch.nn.utils.clip_grad_norm_(
self.model.parameters(), self.cfg.clip_grad
)
self.scaler.step(self.optimizer)
# When enable amp, optimizer.step call are skipped if the loss scaling factor is too large.
# Fix torch warning scheduler step before optimizer step.
scaler = self.scaler.get_scale()
self.scaler.update()
if scaler <= self.scaler.get_scale():
self.scheduler.step()
else:
loss.backward()
if self.cfg.clip_grad is not None:
torch.nn.utils.clip_grad_norm_(
self.model.parameters(), self.cfg.clip_grad
)
self.optimizer.step()
self.scheduler.step()
if self.cfg.empty_cache:
torch.cuda.empty_cache()
self.comm_info["model_output_dict"] = output_dict
|
gharchive/issue
| 2024-05-27T06:08:57 |
2025-04-01T04:55:31.698109
|
{
"authors": [
"AxelCole",
"Gofinge",
"JamesMcCullochDickens",
"SpeedyGonzales949",
"yueyangwen"
],
"repo": "Pointcept/PointTransformerV3",
"url": "https://github.com/Pointcept/PointTransformerV3/issues/54",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
855164062
|
Typo on Zamazenta's French name
Hi! It seems that the French name for Zamazenta is wrong on https://pokeapi.co/api/v2/pokemon-species/889
It should be "Zamazenta", not "Zamazent"!
Have a nice day!
(English is not my native language, sorry in advance!)
Should be fixed, can you check?
Yep, that's perfect! Thank you!
|
gharchive/issue
| 2021-04-10T21:56:28 |
2025-04-01T04:55:31.706253
|
{
"authors": [
"Naramsim",
"polidano10"
],
"repo": "PokeAPI/pokedex",
"url": "https://github.com/PokeAPI/pokedex/issues/42",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
169001051
|
Allow for 3-7 decimal points for coordinates
Some exports only provide three decimal point accuracy, and sometimes 4.440000 turns into 4.44, need to adjust the regex.
:+1:
|
gharchive/pull-request
| 2016-08-02T22:09:54 |
2025-04-01T04:55:31.716901
|
{
"authors": [
"L422Y",
"TheSavior"
],
"repo": "PokemonGoF/PokemonGo-Bot",
"url": "https://github.com/PokemonGoF/PokemonGo-Bot/pull/2402",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
172688909
|
Made docker file a whole lot smaller! (67% to be exact)
Short Description:
Rebased the docker to use alpine python build since it's a lot lighter and easier to download and handle on low connections speed
please wait for merge trying to optimize it even more
@TheSavior save to merge now :)
if someone else has an idea on how to make it even smaller let me know.
Please keep the following ENTRYPOINT code
ENTRYPOINT ["python", "pokecli.py"]
In order to execute docker image and pass parameters, such as ptc, username, password
@nivong isn't this a bit odd? why would one run setup.sh -i during docker build. why would i save my username and password in the docker image?
@mjmadsen: this creates some security concerns for me if the username and password are in the docker image.
Sorry it's not the file that is smaller it's the installation. Now it's 800 something MB. This puts it down to 299 MB by using alpine Linux. But I'm new to docker so if there is a better way of streamlining the whole thing please do!
I used setup.sh because installation instructions might be updated in the feature to not have to edit everything I used the setup.sh
And yes you won't be promoted ;) tested that out ofc
Good point about the entry point will change that asap when I'm back in the office (sorry about the stupid close fat fingers)
@nivong kudos for using alpine that part i totally agree and support
my 2 cents were regarding the readability and transparency of what the Dockerfile does
Thanks will clean it up a bit. Could you perhaps contribute too? You can pr to the dev branch in my fork to update it here. Like I said I'm totally new at docker and not Linux. So if you have a better idea I'm all ears and eyes!
@th3w4y updated, is this more like it?
@nivong
looks good
but check below version for the ADD directly from github
that way if user has any file modified they don't get pulled in by docker build.
FROM python:2.7.12-alpine
RUN apk add --update --no-cache alpine-sdk bash wget git
ADD https://github.com/PokemonGoF/PokemonGo-Bot/archive/dev.tar.gz .
RUN tar -zxvf dev.tar.gz && mv PokemonGo-Bot-dev /usr/src/app && rm dev.tar.gz
WORKDIR /usr/src/app
VOLUME ["/usr/src/app/configs", "/usr/src/app/web"]
ARG timezone=Etc/UTC
RUN ln -snf /usr/share/zoneinfo/$TZ /etc/localtime && echo $TZ > /etc/timezone
#setup the bot
RUN cd /tmp && wget http://pgoapi.com/pgoencrypt.tar.gz \
&& tar zxvf pgoencrypt.tar.gz \
&& cd pgoencrypt/src \
&& make \
&& cp libencrypt.so /usr/src/app/encrypt.so \
&& cd /tmp \
&& rm -rf /tmp/pgoencrypt*
ENV LD_LIBRARY_PATH /usr/src/app
RUN ln -s /usr/include/locale.h /usr/include/xlocale.h
RUN pip install --no-cache-dir -r requirements.txt
#remove unused stuff
RUN apk del alpine-sdk\
&& apk cache clean\
&& rm -rf /var/cache/apk/*
ENTRYPOINT ["python", "pokecli.py"]
no idea what it does but you have more knowledge than me ;) added it!
@nivong check https://github.com/nivong/PokemonGo-Bot/pull/4
@nivong thanks
@nivong: before you take out wget you would have to modify the Dockerfile like this:
-RUN cd /tmp && wget http://pgoapi.com/pgoencrypt.tar.gz \
- && tar zxvf pgoencrypt.tar.gz \
+ADD http://pgoapi.com/pgoencrypt.tar.gz /tmp/
+RUN cd /tmp && tar zxvf pgoencrypt.tar.gz \
as for git i don't know if is used
can you do a PR to my fork :) ? but swampt today
@nivong https://github.com/nivong/PokemonGo-Bot/pull/6
@nivong please pull in also https://github.com/nivong/PokemonGo-Bot/pull/7 otherwise the suggestion i have with pulling the archive from github.com and the fact that i hardcoded dev will not be compatible with the master branch
check!
@th3w4y this comes up on testing the docker build:
Step 21 : RUN apk del alpine-sdk && rm -rf /var/cache/apk/*
---> Running in 2ba527e2474e
WARNING: Ignoring APKINDEX.167438ca.tar.gz: No such file or directory
WARNING: Ignoring APKINDEX.a2e6dac0.tar.gz: No such file or directory
Removing intermediate container 2ba527e2474e
The command '/bin/sh -c apk del alpine-sdk && rm -rf /var/cache/apk/*' returned a non-zero code: 1
Why is this a problem out of the blue? Worked before....
@nivong checking now...
docker build https://raw.githubusercontent.com/nivong/PokemonGo-Bot/691037348b7d40649f8c507e73a12334854b5526/Dockerfile
... still building
@nivong
I found the cullprit that step si ruined twice....
#remove unused stuff
RUN apk del alpine-sdk\
&& rm -rf /var/cache/apk/*
ENTRYPOINT ["python", "pokecli.py"]
#remove unused stuff
RUN apk del alpine-sdk \
&& rm -rf /var/cache/apk/*
ENTRYPOINT ["python", "pokecli.py"]
lol how did that happen, changed it thanks for checking, if you have more please do a PR i am out of the office again
@nivong sorry was my bad in commit 215832021111bd08038b337487b09d68e73787ce i copy paste from my file over the text in the github.com editor and probably some lines remain unselected when pasting over.
@nivong now all looks good 👍
14:06 $ docker build https://raw.githubusercontent.com/nivong/PokemonGo-Bot/8832b16f610bbc7ea7cbce10b8c682450a6d05da/Dockerfile
Downloading build context from remote url: https://raw.githubusercontent.com/nivong/PokemonGo-Bot/8832b16f610bbc7ea7cbce10b8c682450a6d05da/Dockerfile 1.56 kB
Sending build context to Docker daemon 3.584 kB
Step 1 : FROM python:2.7.12-alpine
---> 67fa4e31547b
Step 2 : RUN apk add --update --no-cache alpine-idk git
---> Using cache
---> a4ae9dd908be
Step 3 : ARG BUILD_BRANCH
---> Using cache
---> 321d0cbbdaf6
Step 4 : ENV BUILD_BRANCH ${BUILD_BRANCH:-master}
---> Using cache
---> 2dcad1163c8d
Step 5 : ARG BUILD_REPO
---> Using cache
---> b4eaf959943c
Step 6 : ENV BUILD_REPO ${BUILD_REPO:-PokemonGoF/PokemonGo-Bot}
---> Using cache
---> 12cc3e912e5d
Step 7 : LABEL build_repo $BUILD_REPO build_branch $BUILD_BRANCH
---> Using cache
---> 06c9d4154b8c
Step 8 : ADD https://github.com/$BUILD_REPO/archive/$BUILD_BRANCH.tar.gz .
Downloading [==================================================>] 3.521 MB/3.521 MB
---> Using cache
---> ec957c9739f9
Step 9 : RUN tar -zxvf $BUILD_BRANCH.tar.gz && mv PokemonGo-Bot-* /usr/src/app && rm $BUILD_BRANCH.tar.gz
---> Using cache
---> 047661f5c49e
Step 10 : WORKDIR /usr/src/app
---> Using cache
---> c39fa7fe8bbc
Step 11 : VOLUME /usr/src/app/configs /usr/src/app/web
---> Using cache
---> 47815483d5f0
Step 12 : ARG timezone=Etc/UTC
---> Using cache
---> 9ca574da35f9
Step 13 : RUN ln -snf /usr/share/zoneinfo/$TZ /etc/localtime && echo $TZ > /etc/timezone
---> Using cache
---> 2bc5d76191de
Step 14 : ADD http://pgoapi.com/pgoencrypt.tar.gz /tmp/
Downloading [==================================================>] 243.6 kB/243.6 kB
---> Using cache
---> dc84fb88af8b
Step 15 : RUN cd /tmp && tar zxvf pgoencrypt.tar.gz && cd pgoencrypt/src && make && cp libencrypt.so /usr/src/app/encrypt.so && cd /tmp && rm -rf /tmp/pgoencrypt*
---> Using cache
---> 861c0b5a4784
Step 16 : ENV LD_LIBRARY_PATH /usr/src/app
---> Using cache
---> eb78d51f0398
Step 17 : RUN ln -s /usr/include/locale.h /usr/include/xlocale.h
---> Using cache
---> fa24a7864b5c
Step 18 : RUN pip install --no-cache-dir -r requirements.txt
---> Using cache
---> 0e858ca5141c
Step 19 : RUN apk del alpine-sdk && rm -rf /var/cache/apk/*
---> Using cache
---> 7f5393c52c4e
Step 20 : ENTRYPOINT python pokecli.py
---> Using cache
---> d483306518cb
Successfully built d483306518cb
(anaconda3) (root) ✔ ~
14:06 $ echo $?
0
(anaconda3) (root) ✔ ~
14:06 $
Thanks for the merge @solderzzc and thanks @th3w4y for the help :)
|
gharchive/pull-request
| 2016-08-23T12:43:01 |
2025-04-01T04:55:31.731735
|
{
"authors": [
"helloj",
"nivong",
"th3w4y"
],
"repo": "PokemonGoF/PokemonGo-Bot",
"url": "https://github.com/PokemonGoF/PokemonGo-Bot/pull/4618",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
917350612
|
Implement callback for create_order
Will close #88
@haerdib I need your help with RPC stuff ... I mentioned your name in the comments
|
gharchive/pull-request
| 2021-06-10T13:31:15 |
2025-04-01T04:55:31.800298
|
{
"authors": [
"ksr30"
],
"repo": "Polkadex-Substrate/polkadexTEE-worker",
"url": "https://github.com/Polkadex-Substrate/polkadexTEE-worker/pull/100",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
367552677
|
lit-html, lit-element and redux : _stateChanged ignore template attributes update
Also I came to this specific use-case where a LitElement updates one of its attributes from a redux state property and an attribute in a lit-html template (the template returned from the render method).
Here's the reproduction
I tried to simplify as much as I could.
Simply put, the main view (person-view) html is embedding another custom element (person-card) and set one of its attributes using the "dot notation" (.personId="${this._personId}").
When the state changes (clicking the button) the main view updates its property _personId. Naturally I was expecting the view to update the html attribute of person-card as well so when person-card _stateChanged method is called its property personId value is coherent to the application global state.
I found that was natural because person-card is one leaf of the tree which is the application and should rely on its parents' state.
I think this behavior is caused from the methods _stateChanged called in an asynchronous way or technically not awaiting upper elements dom full update (I can't really tell as my understand of how the core libraries work is limited).
As for now I am trying to find a workaround but If you think this case deserves special attention I'd be gladly waiting for an update.
Love you Polymer Team !
Just a quick addition, this is the workaround I found and I think this is the syntax some lit-element showcases are using.
However in my project I really need to do the treatment in the _stateChanged method because the computed property is coherent to the element and is used subsequently in other elements' methods.
What I mean is that the property is not used for rendering purpose only. So I maintain the state of this issue.
It looks like lit-element will rerender the template of the view asynchronously, so it turns out the personId gets updated in the child component after the _stateChanged method has been called. In your example the parent already has the correct state, but this state hasn't been propagated to the child yet.
I'm not sure what the best approach is to solve the issue (add a synchronous variant of _invalidate to LitElement?), but in any case this looks more like an issue for Polymer/lit-element than lit-html. Could you open a ticket there?
@bgotink Ok I will, thanks for your answer.
|
gharchive/issue
| 2018-10-07T13:02:48 |
2025-04-01T04:55:31.821277
|
{
"authors": [
"bgotink",
"vdegenne"
],
"repo": "Polymer/lit-html",
"url": "https://github.com/Polymer/lit-html/issues/545",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
733593603
|
Remove a test.only in the lit-html tests.
I haven't looked into the cause for the failures this exposes, but I believe they're all legitimately failing tests so we shouldn't (necessarily) wait to make them pass in this PR before merging.
Sorry about this; I'm working on fixes for this, I think it's better to fix them than to have master red for some amount of time.
Ok, that was a simple merging mistake when reverting one of the code factorings late in the lit-ssr PR: https://github.com/Polymer/lit-html/commit/961e95c95ef3a2dd36fbe00d660ab927012d923e#diff-bead45bb43cda474d916704550a1b26322667aa42b814c48300b05be79b9a7ebR1181
Will push a fix to this branch.
|
gharchive/pull-request
| 2020-10-31T00:52:49 |
2025-04-01T04:55:31.823567
|
{
"authors": [
"bicknellr",
"kevinpschaaf"
],
"repo": "Polymer/lit-html",
"url": "https://github.com/Polymer/lit-html/pull/1402",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
85705600
|
paper-input docs not showing up in catalog
They are there when I run polyserve. Instead, I get paper-textarea and friends.
Closed by 4f6b270a8
|
gharchive/issue
| 2015-06-06T05:56:31 |
2025-04-01T04:55:31.827477
|
{
"authors": [
"atotic",
"mbleigh"
],
"repo": "Polymer/polymer-element-catalog",
"url": "https://github.com/Polymer/polymer-element-catalog/issues/112",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
76071841
|
[0.8.0-rc.5] observers as an array
Property observers work when custom element is instantiated, but never after that. No errors in console. Here is the javascript from the custom element definition used in my test:
var ChatUser = Polymer({
is: "chat-user",
properties: {
user: Object
},
constructor: function (user) {
this.user = user;
},
observers: [
"userStatusChanged(user.status)"
],
userStatusChanged: function (status) {
console.log("updating status", status);
this.notifyPath("user.status", status);
}
});
Testing the custom element's observer:
var userData = {username: "Bob", status: "Offline"};
var user = new ChatUser(userData); // console logs "updating status Offline"
// change status
userData.status = "Online"; // observer ignores this
You need to use the setPathValue method to update the property (or update it using
data binding). See:
https://www.polymer-project.org/0.8/docs/devguide/data-binding.html#path-binding
(Note: in 0.9, setPathValue changes its name to just set).
In this case, you're updating the status from outside, I think instead of:
userData.status = "Online";
In your example, you'd do something like:
user.set("user.status", "Online");
So I have to explicitly call a setter on every custom element using that object. I was under the impression that using the observers array would set up Object.observe. I was hoping I could have more than one element observe a single object, so that changes to that one object would essentially be broadcast to all the observing elements, updating their views.
A use case along the lines of this example would be a chat room with a user list of ChatUser elements, and a chat window of ChatMessage elements. If "Bob" were to change his nickname to "Robert", updating his userData object would change his nickname both in the user list, and on any chat messages he's posted in the chat window.
0.8/0.9 does not use Object.observe. This was one of the performance changes -- As I understand it, O.o and its polyfill were both quite unpredictable from a performance standpoint.
If you use data binding to pass values, you can get many of the same benefits, but how exactly you'd do that may depend on how your app is set up. I recommend putting together a code sample and either posting a question on Stack Overflow or joining the Polymer slack channel (bit.ly/polymerslack)... There are a lot of people on there who've been working with 0.8/0.9 and may have suggestions.
Thank you very much for your help. I'll do as you've suggested.
|
gharchive/issue
| 2015-05-13T18:06:03 |
2025-04-01T04:55:31.833069
|
{
"authors": [
"arthurevans",
"jasonpecor"
],
"repo": "Polymer/polymer",
"url": "https://github.com/Polymer/polymer/issues/1527",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
182035661
|
Deep property change from undefined to null is not propagated
Repro: http://jsbin.com/xatiqibofu/edit?html,output
The code changes a deep property and displays the object. Initially the object is empty. When I change the property to "x", the result becomes {property: "x"} which is expected. When I change property to null, it should become {property: null}, but instead it stays empty. The change is not even propagated at all.
I understand how undefined is handled strangely, but this is null which should work like any valid value. So hopefully this is just an oversight, ie. somewhere you're checking "if (oldvalue == newvalue)", and undefined happens to be == to null, but instead it should be ===, which would work. (I see you have a well thought out comparison here: https://github.com/Polymer/polymer/blob/ff6e884ef4f309d41491333860a8bc9c2f178696/src/standard/notify-path.html#L100)
Seems that your issue has been fixed, as the JSBin shows {"property":null} now.
|
gharchive/issue
| 2016-10-10T14:53:51 |
2025-04-01T04:55:31.836120
|
{
"authors": [
"TimvdLippe",
"fejesjoco"
],
"repo": "Polymer/polymer",
"url": "https://github.com/Polymer/polymer/issues/4053",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
138579128
|
Ability to configure a11ySuite rules, exceptions?
Is it possible to configure a11ySuite rules, or ignore specific instances of errors/warnings?
Yes you can pass an array of rule named to skip...
a11ySuite('default', ['duplicateId']);
The rulenames come from accessibility-developer-tools...
https://github.com/GoogleChrome/accessibility-developer-tools/tree/master/src/audits
Awesome! Thanks Rob!
There should be a way to configure these ignore rules from wct.conf.json file, so that they can be applied to all tests globally.
|
gharchive/issue
| 2016-03-04T20:31:30 |
2025-04-01T04:55:31.838760
|
{
"authors": [
"dbatiste",
"jaichandra",
"robrez"
],
"repo": "Polymer/web-component-tester",
"url": "https://github.com/Polymer/web-component-tester/issues/298",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
493532887
|
Upgrade to Typescript 3.5.3
Update dependencies on lint and tfjs
Disable new operator-linebreak lint rule
Fix Lint inssues in mock-slot-composer.ts and sequence.ts
Add skipLibCheck for tfjs as that is recommended in https://www.tensorflow.org/js/tutorials/setup#typescript
Okay, have approvals from two continents.
If this causes a problem I'd be happy to rollback.
|
gharchive/pull-request
| 2019-09-13T22:00:04 |
2025-04-01T04:55:31.845260
|
{
"authors": [
"lindner"
],
"repo": "PolymerLabs/arcs",
"url": "https://github.com/PolymerLabs/arcs/pull/3574",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
543288949
|
Break cycle between Runtime and Arc.
Moves cycle count from 12 to 11.
Introduces a StoreRegistry interface for arcs to use when registering
and unregistering their stores.
Part of https://github.com/PolymerLabs/arcs/issues/1878
Closing in favor of https://github.com/PolymerLabs/arcs/pull/4351.
|
gharchive/pull-request
| 2019-12-28T22:42:29 |
2025-04-01T04:55:31.846917
|
{
"authors": [
"shaper"
],
"repo": "PolymerLabs/arcs",
"url": "https://github.com/PolymerLabs/arcs/pull/4346",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
617810363
|
Fix memory leaks
(Root #1)
TaskHandle (from testapp) holds an EntityHandleManager
EntityHandleManager holds StorageProxies
StorageProxy holds a crdt field
crdt field holds a RawEntity
(Root #2)
(serviceConnection -> service.registerCallback -> onProxyMessage holds ServiceStore, also held by StorageProxy)
ServiceStore holds a Channel
channel has a Flow attached
channel can hold a lambda
lambda has in it { service.sendProxyMessage(message.toProto().toByteArray(), result) }
that lambda has in it a captured 'message' parameter from the function call
message contains ModelUpdate which holds RawEntity
Fixes to TestApp
Invoke EntityHandleManager.close() after tests finish
Fixed to Core
make StorageProxy.crdt a nullable field and make close() null it
change onLifecycleDestroyed() to invoke scope.cancel() (cancel the flow)
add channel.cancel() to ServiceStore.off() and reinitialize channel & flow with new fresh copies
/gcbrun
/gcbrun
I'm submitting this, only the HandleManager tests are flaky, but on G3 I was able to get TAP to pass with enough runs.
I'm submitting this, only the HandleManager tests are flaky, but on G3 I was able to get TAP to pass with enough runs.
I think this change caused additional flakyness, so we are unable to import it, see eg cl/313015371
|
gharchive/pull-request
| 2020-05-13T23:08:46 |
2025-04-01T04:55:31.851537
|
{
"authors": [
"cromwellian",
"galganif"
],
"repo": "PolymerLabs/arcs",
"url": "https://github.com/PolymerLabs/arcs/pull/5301",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
247576088
|
Adding functions to events seems to be executed immediately
Hi! :100:
I'm trying this lib, and in first look it seems fantastic and works. Works, but almost. It's seems that when attach an event handler it is executed immediately. For example setting a function on onclick event is called on render, not when element is clicked.
const msg = 'Hello!'
const Main = () => html`<main>
<h1>${msg}!</h1>
<button onclick=${() => { console.log('woohoo') }}>+1</button>
</main>`
render(Main(), document.body)
http://jsbin.com/sogitix/28/edit?js,console,output
Because functions are used as extension points (function values are called with the Part object for an expression so that it can imperatively modify the Part), values that are intended to be functions, like event handlers must be returned from a wrapper function:
html`<main>
<h1>${msg}!</h1>
<button onclick=${_=> () => { console.log('woohoo') }}>+1</button>
</main>`
Also, make sure you're using lit-extended, because core lit-html only sets attributes, which means the function would be converted to a string.
This is talked about a little bit here, though could clearly use some more detail: https://github.com/PolymerLabs/lit-html#function-values
This is talked about a little bit here
Link don't work, the anchor :D nevermind, found it and probably make sense to show above example to clarify the things.
make sure you're using lit-extended
hearing it for the first time
values that are intended to be functions, like event handlers must be returned from a wrapper function:
Wouldn't it be better to switch the places. So if that function returns a function use it like now, so when someone don't care about parts and etc he will be able to pass normal function.
for example
html`<main>
<h1>${msg}!</h1>
<button onclick=${clicked}>+1</button>
</main>`
function clicked (evt) {
console.log('woo hoo, called when clicked'
// if it returns a function then use it as currently
return (part) => part.setValue((part.previousValue + 1) || 0)
}
Wouldn't it be better to switch the places. So if that function returns a function use it like now, so when someone don't care about parts and etc he will be able to pass normal function.
Yeah, and I filed #29 for it. I'll try to get it done asap
Hooo!! :tada: Awesome! :clap:
So i believe it will fix this issue too, or I must use lit-extended? And if i must, should i include and the core too?
Well, to write to properties instead of attributes you need to use lit-extended - because that's a pretty opinionated default
I'm going to close this as working-as-intended for now, but I'll change the behavior in #29. Thanks for the report!
Also, you need to import html from lit-html.js and render from lit-extended.js for now... I'll revisit re-exporting html from lit-extended soon.
Great, great, thanks for the tip. :wink:
|
gharchive/issue
| 2017-08-03T02:39:30 |
2025-04-01T04:55:31.859132
|
{
"authors": [
"charlike",
"justinfagnani"
],
"repo": "PolymerLabs/lit-html",
"url": "https://github.com/PolymerLabs/lit-html/issues/26",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
170719438
|
Issues with data:image?
It appears that when running polymerCssBuild on our project, data:image css declarations seem to get formatted incorrectly.
for example...
background: url("data:image/svg+xml;base64,PD94bW....");
becomes
background: url("data:image/svg+xml;)
Does that only happen with image/svg+xml? I just checked our project, and our image/png data: URLs survived the processing properly.
|
gharchive/issue
| 2016-08-11T18:56:06 |
2025-04-01T04:55:31.861386
|
{
"authors": [
"ankon",
"barnomics"
],
"repo": "PolymerLabs/polymer-css-build",
"url": "https://github.com/PolymerLabs/polymer-css-build/issues/2",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1609569346
|
crackmapexec /usr/lib/python3/dist-packages/pywerview/requester.py:144: SyntaxWarning: "is not" with a literal. Did you mean "!="? if result['type'] is not 'searchResEntry':
Describe the bug
I can not use crackmapexec due this issue
To Reproduce
Steps to reproduce the behavior:
Go to 'terminal'
Write crackmapexec smb IP
See error
Expected behavior
I need to use crackmapexec but I can not, i would like to fix this error
Screenshots
Crackmapexec info
OS: [e.g. Kali]
Version of CME [e.g. v5.0.2]
Installed from apt or using latest release ? Please try with latest release before openning an issue
Additional context
I tried to fix it by seeing the code, go to the file, and you have this code
I tried writing != instead "is not", and then I tried to do it, nothing happens,I have no the error, but crackmapexec do nothing now, whitout any request
Please update crackmapexec, your version is way out of date. This Syntax version is fixed for several versions
|
gharchive/issue
| 2023-03-04T03:28:37 |
2025-04-01T04:55:31.903362
|
{
"authors": [
"1Freezer",
"NeffIsBack"
],
"repo": "Porchetta-Industries/CrackMapExec",
"url": "https://github.com/Porchetta-Industries/CrackMapExec/issues/756",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2561529731
|
chore: allow manual initialization of the SDK
:bulb: Motivation and Context
Closes https://github.com/PostHog/posthog-flutter/issues/22
Closes https://github.com/PostHog/posthog-flutter/pull/113
Required for https://github.com/PostHog/posthog-flutter/pull/116
Automatic mode is still default so not a breaking change.
:green_heart: How did you test it?
Running auto init and automatic mode on android and ios
:pencil: Checklist
[X] I reviewed the submitted code.
[ ] I added tests to verify the changes.
[ ] I updated the docs if needed.
[X] No breaking change or entry added to the changelog.
@thisames if you'd like to review as well :)
@thisames if you'd like to review as well :)
for me, everything is ok
|
gharchive/pull-request
| 2024-10-02T12:42:27 |
2025-04-01T04:55:31.926148
|
{
"authors": [
"marandaneto",
"thisames"
],
"repo": "PostHog/posthog-flutter",
"url": "https://github.com/PostHog/posthog-flutter/pull/117",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1457429247
|
[Bug]: Connecting to environment not working
What happened?
Connecting via username/password still not possible. After the version update the error message changed, but the problem still exists, that I can't connect to any environment.
The log output was added through creating the bug via the error message. Don't know if that helps.
What version of the tool are you using?
v2.1.1
What version of Visual Studio Code are you running?
1.73.1
Any relevant session id from VSCode?
acfaf457-e549-442d-85bf-574c00bd8e231669014766144
Relevant log output
Error occured in connectDataverse command.
Error code: undefined.
Error message: undefined.
Error stack: undefined.
I have exactly same issue. Please help
Can you check with newer versions released? Thanks.
|
gharchive/issue
| 2022-11-21T07:18:58 |
2025-04-01T04:55:32.013834
|
{
"authors": [
"Power-Maverick",
"neb-miletic",
"wd21sos"
],
"repo": "Power-Maverick/DataverseDevTools-VSCode",
"url": "https://github.com/Power-Maverick/DataverseDevTools-VSCode/issues/92",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
496916626
|
ADReplicationSiteLink Add Integration tests, add setting 'options'
Pull Request (PR) description
Add integration testing ability. Adds ability to enable options such as Change Notification Replication
This Pull Request (PR) fixes the following issues
Fixes #356
Fixes #504
Task list
[ ] Added an entry under the Unreleased section in the CHANGELOG.md.
Entry should say what was changed, and how that affects users (if applicable).
[ ] Resource documentation added/updated in resource directory README.md.
[ ] Resource parameter descriptions added/updated in schema.mof
and comment-based help.
[ ] Comment-based help added/updated.
[ ] Conceptual help topic added/updated (cultureFolder\about_ResourceName.help.txt).
[ ] Localization strings added/updated in all localization files as appropriate.
[ ] Examples appropriately added/updated.
[ ] Unit tests added/updated. See DSC Resource Testing Guidelines.
[ ] Integration tests added/updated (where possible). See DSC Resource Testing Guidelines.
[ ] New/changed code adheres to DSC Resource Style Guidelines and Best Practices.
This change is
Codecov Report
Merging #509 into dev will decrease coverage by <1%.
The diff coverage is 80%.
@@ Coverage Diff @@
## dev #509 +/- ##
===================================
- Coverage 98% 98% -1%
===================================
Files 23 23
Lines 3108 3159 +51
Branches 10 10
===================================
+ Hits 3063 3098 +35
- Misses 35 51 +16
Partials 10 10
Hi @johlju , are you able to review this and #510 at all please? I know there will be a conflict on the changelog.md when one is merged. Thank you
@1800Zeta Thank you for this! The integration tests worked perfectly! 😃
Thank you so much @johlju
|
gharchive/pull-request
| 2019-09-23T06:31:05 |
2025-04-01T04:55:32.032750
|
{
"authors": [
"1800Zeta",
"codecov-io",
"johlju"
],
"repo": "PowerShell/ActiveDirectoryDsc",
"url": "https://github.com/PowerShell/ActiveDirectoryDsc/pull/509",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2042382516
|
Find-PSResource with multiple tags not working as expected
Prerequisites
[X] Write a descriptive title.
[X] Make sure you are able to repro it on the latest released version
[X] Search the existing issues.
Steps to reproduce
I can run these commands to search for resources by tag.
Find-PSResource -Tag yml -type module
Find-PSResource -Tag json-type module
The -Tag parameter will accept an array, so it is reasonable to assume I can run:
Find-PSResource -Tag yml,json -type module
Expected behavior
I would expect to see the results from the two separate commands.
Actual behavior
I get the error: Package with Tags 'yml, json' could not be found in any registered repositories
The parameter will work only when all tags are in the same module.
PS C:\> Find-PSResource -Tag demo,teaching -Type module
Name Version Prerelease Repository Description
---- ------- ---------- ---------- -----------
PSTeachingTools 4.2.0 PSGallery A set of commands and tools for teaching PowerShell.
Error details
Exception :
Type : Microsoft.PowerShell.PSResourceGet.UtilClasses.ResourceNotFoundException
Message : Package with Tags 'yml, json' could not be found in any registered repositories.
HResult : -2146233088
TargetObject : Microsoft.PowerShell.PSResourceGet.Cmdlets.FindHelper
CategoryInfo : ObjectNotFound: (Microsoft.PowerShel….Cmdlets.FindHelper:FindHelper) [Find-PSResource], ResourceNotFoundException
FullyQualifiedErrorId : PackageWithTagsNotFound,Microsoft.PowerShell.PSResourceGet.Cmdlets.FindPSResource
InvocationInfo :
MyCommand : Find-PSResource
ScriptLineNumber : 1
OffsetInLine : 1
HistoryId : 26
Line : Find-PSResource -Tag yml,json -type module
Statement : Find-PSResource -Tag yml,json -type module
PositionMessage : At line:1 char:1
+ Find-PSResource -Tag yml,json -type module
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
InvocationName : Find-PSResource
CommandOrigin : Internal
ScriptStackTrace : at <ScriptBlock>, <No file>: line 1
PipelineIterationInfo :
Environment data
PS C:\> Get-module Microsoft.PowerShell.PSResourceGet;$PSVersionTable | format-Table
ModuleType Version PreRelease Name ExportedCommands
---------- ------- ---------- ---- ----------------
Binary 1.0.0 Microsoft.PowerShell.PSResourceGet {Find-PSResource, Get-InstalledPSResource, Get-P…
Name Value
---- -----
PSVersion 7.4.0
PSEdition Core
GitCommitId 7.4.0
OS Microsoft Windows 10.0.22631
Platform Win32NT
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…}
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
WSManStackVersion 3.0
Same behavior in Windows PowerShell.
Visuals
No response
Indeed, it looks like Find-Module's OR logic with respect to multiple tags mistakenly became AND logic in Find-PSResource.
@jdhitsolutions thanks for creating this issue! We intentionally changed our -Tag logic to reflect AND logic with Find-PSResource instead of OR, so that the search is less computationally expensive and the search result is narrower. If a user wishes to see packages with tag1ORtag2` they would need to provide 2 separate commands.
Good catch that the docs do not yet reflect this, we will update this on our end, thanks!
@jdhitsolutions in the docs for Find-PSResource for the -Tag parameter it states, "If multiple tags are specified, the cmdlet only returns resources that include all the tags provided." Just wanted to point to that, but please feel free to re-open if you feel we should add more clarity thanks :)
|
gharchive/issue
| 2023-12-14T19:57:33 |
2025-04-01T04:55:32.042809
|
{
"authors": [
"anamnavi",
"jdhitsolutions",
"mklement0"
],
"repo": "PowerShell/PSResourceGet",
"url": "https://github.com/PowerShell/PSResourceGet/issues/1492",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1015443708
|
Fixes #1720 - move rule docs and update tests
PR Summary
PR Checklist
Moved ./RulesDocumentation to ./docs/Rules
Moved ./docs/markdown to ./docs/Cmdlets
Update Pester test for documentation
[x] PR has a meaningful title
Use the present tense and imperative mood when describing your changes
[x] Summarized changes
[x] Change is not breaking
[ ] Make sure all .cs, .ps1 and .psm1 files have the correct copyright header
[x] Make sure you've added a new test if existing tests do not effectively test the code changed and/or updated documentation
[x] This PR is ready to merge and is not Work in Progress.
If the PR is work in progress, please add the prefix WIP: to the beginning of the title and remove the prefix when the PR is ready.
Looks like platyPS invocations in the build script also need to be updated
@rjmholt OK, I missed that. Will update.
|
gharchive/pull-request
| 2021-10-04T17:33:28 |
2025-04-01T04:55:32.048768
|
{
"authors": [
"rjmholt",
"sdwheeler"
],
"repo": "PowerShell/PSScriptAnalyzer",
"url": "https://github.com/PowerShell/PSScriptAnalyzer/pull/1724",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
583232965
|
Having to click twice to close
Steps to reproduce:
Open powershell in any folder
Write to powershell "code ." //open the folder with visual studio code
Click to powershell window close button one time. Window will not close.
Click a second time, windows will close.
I think it should close when I first click. This problem is just "code . " Occurs when I run the command.
I guess when powershell is unfocus when it opens another window it can't detect the first shutdown.
That's an interesting find (in the future, please provide details about your environment, as suggested in the issue template; in the case at hand, it would have been helpful to know that the problem only occurs (in regular console windows) on Windows).
However, the problem is not related to PowerShell, it is the behavior of regular console windows (conhost.exe) on Windows, as evidenced by the fact that a regular cmd.exe console window exhibits the same behavior.
Therefore, please close this issue and report it at https://github.com/microsoft/terminal/issues.
Also interesting to note that Windows Terminal is not affected.
|
gharchive/issue
| 2020-03-17T19:02:47 |
2025-04-01T04:55:32.064622
|
{
"authors": [
"mklement0",
"sefacan"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/12144",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
592843679
|
Get-ChildItem does unnecessary work if path isn't found
I have tested the behaviour with PS5 and PS7.
Steps to reproduce
$files = Get-ChildItem -Path 'c:\mysepecialfolders*' -Filter '*.dll' -Recurse
foreach ($file in $files)
{
$file.FullName
}
Expected behavior
Seach for all folder which start with mysepecialfolders and search inside this folders for all .dll files.
Actual behavior
Depends.
IF there is a folder like c:\mysepecialfolders1 or there are more folders that match,
then Get-ChildItem is fast and searches only inside this folders for all *.dll files.
If there is no folder that matches, then Get-ChildItem searches the whole drive.
Behavior request
If there is no folder that matches, then Get-ChildItem should return no files immediately without searching the whole drive.
Environment data
Name Value
---- -----
PSVersion 7.0.0
PSEdition Core
GitCommitId 7.0.0
OS Microsoft Windows 10.0.18363
Platform Win32NT
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0…}
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
WSManStackVersion 3.0
I have checked the behavior with Process Monitor.
|
gharchive/issue
| 2020-04-02T18:34:51 |
2025-04-01T04:55:32.067167
|
{
"authors": [
"Tragen"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/12245",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1780036563
|
Powershell can't run on macos 14 beta2
Prerequisites
[X] Write a descriptive title.
[X] Make sure you are able to repro it on the latest released version
[X] Search the existing issues.
[X] Refer to the FAQ.
[X] Refer to Differences between Windows PowerShell 5.1 and PowerShell.
Steps to reproduce
install powershell-7.3.5-osx-arm64.pkg ok
run powershell Application PowerShell cannot be opened
Expected behavior
-
Actual behavior
-
Error details
No response
Environment data
-
Visuals
No response
Do you have Rosetta installed? There was a problem where MacOS thought that PowerShell.app required Rosetta...
https://github.com/PowerShell/PowerShell/issues/18548
|
gharchive/issue
| 2023-06-29T03:00:04 |
2025-04-01T04:55:32.072149
|
{
"authors": [
"rhubarb-geek-nz",
"teslachina"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/19869",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
223135430
|
ConvertFrom-Yaml, ConvertTo-Yaml
Would be great to support Yaml natively.
This was also mentioned by @fabiendibot on #3046
It would also be nice if the CMDLets had the goal of cleanly handling conversion of objects that came from XML as it seems like it would be a frequent use case. Maybe some good tests around this conversion?
We had a similar discussion from DSC aspect,
allowing us to change json based configuration files, we wanted to have options for modifying xml based files, YAML based files ,INI based files supporting RegEx swaps from within Text Manipulation cmdlets.
Lack of existing support in PS means we have to work hard to get such ability.
It has been on hold pending community contribution, but if it was baked into PS, it would make it much easier for the DSC part as well.
When you say natively, do you mean like XML or JSON?
The current thinking is that YAML should not be baked into PowerShell at all, instead it should be a separate module that you can update without picking up a new version of PowerShell.
If YAML were baked into PowerShell like XML, that would be impossible (think [xml]"b")
If we went the JSON route, you'd have cmdlets to work with YAML - so not really baked into PowerShell, but you'd still have the drawbacks of needing to update PowerShell to get YAML updates.
@lzybkr I know we said we didn't want to bring in a new library, but I think this is something we might need to reassess. Ideally, we should also ship the module on the Gallery, but I think a TON of modern scenarios require YAML now.
Maybe not in 6.0 timeframe, but we should talk about it.
@ArieHein - I have some simple functions that save and retrieve a hash array to the registry. Only handle REG_SZ - but for a simple set of settings it is sufficient - let me know if you want a copy.
I mispoke when I said "native" - I primarily meant "built-in" - it wouldn't bother me if they were shipped-in script modules that could be updated.
Our first discussion #2109
@iSazonov - ah yes I see!
I noticed the reference to AWS support of YAML on the thread - I have been converting some templates and have found this to helpful: https://github.com/awslabs/aws-cfn-template-flip
@iSazonov thanks for the pointer, I couldn't find it for some reason. Remember it well, though.
In re-reading that original thread, I think we should definitely implement the cmdlets at some point in the future, and ship them in the Gallery. Based on their quality, and people's perceived usefulness (along with some refactoring work we hope to do after 6.0.0), we can make the in-box vs. Gallery-only call.
We have Yaml in https://github.com/PowerShell/platyPS
yeah this would be awesome to have, ended up using https://github.com/awslabs/aws-cfn-template-flip to convert
@MattTunny Welcome to contribute! :-)
There's a Windows Server user voice openned to vote for it :-)
https://windowsserver.uservoice.com/forums/301869-powershell/suggestions/11088495-out-of-the-box-support-for-yaml-like-csv-xml-j
This should definitely be part of the native PS 6.1 library. So many things these days are in YAML.
There are now psyaml and powershell-yaml modules on the PSGallery but both are not even able to round-trip a YAML file from a VSTS build definition. I don't mind if the module is baked into PowerShell or is a module from the PSGallery.
I wonder if the core problem here is the clunky way we deploy modules. Today, you have to find, trust and install a module before you can use it. Compare this with the (apparently) slick way that Javascript does var m = require('mymodule'). Maybe we should have some way to do what DSC does but for native PowerShell. In DSC, when a module is referenced in a configuration, it's automatically downloaded and installed on the target node with no manual effort. Making critical but non-core modules available that way should eliminate the "it should be part of core" arguments. And for nodes that are disconnected from the net, we could have a tool that bundled the dependencies in a script into an archive which is then deployed to the target. This is how the Azure DSC resource extension works - there is a tool that scans a script to figure out the required modules then builds a zip file containing everything that is needed and publishes it to a blob. The Azure resource extension then pulls this blob, installs the modules and runs the script.
@bgshacklett From what I've heard from the Puppet guys, there are just no good YAML parsers :-)
Is platyPS parser good enough?
@vors Is there simple way to reuse platyPS YAML parser in PowerShell Core repo?
I prefer the idea of a separate official module in PowerShell Gallery like @lzybkr says because it would be possible to use it in older powershell versions and it could have its own releases. That would be like the sqlserver module. @BrucePay if it were a page in PowerShell Gallery with Microsoft own modules, it would be easier to find and everybody would know that they can trust them.
But I would understand if it were backed into Powershell as XML and JSON.
The important thing is that it exists ConvertFrom-YAML and ConvertFrom-YAML official functions because YAML is a widely used format for configuration files and it shouldn't be a third-party module, as @bgshacklett point out.
I made a blog entry testing and comparing the two modules I've found to work with YAML files: PSYaml and powershell-yaml.
They have different behaviours because internally they're using different objects:
module
mappings
sequences
PSYaml
OrderedDictionary
Array
powershell-yaml
Hastable
List
I think we need a standard ConvertFrom-YAML and ConvertFrom-YAML.
Actually, ConvertFrom-Yaml in powershell-yaml uses OrderedDictionary when converting with the -ordered parameter.
I've been using this module successfully for a while (in my Datum module for DSC Configuration data, and with kitchen yamls), but don't have a vsts build definition to test with.
Bear in mind that the right way to call it is: get-content -Raw MyFile.yml | ConvertFrom-Yaml -Ordered (people often miss the -Raw).
I wonder why we'd need a Microsoft official module, putting even more overhead on MSFT and reinventing the wheel... Maybe trying to contribute to an existing one first, add tests to avoid regression, open issues to make sure the owner knows the problems is a better approach...
You know what happens when you're trying to create a standard out of the 99 existing implementations...
And yes it would be better outside the language, I agree that the dependency management could be better, bundling everything in PS is not a solution though.
The broad npm issue is also a failure in process. Fork and re-publish fixed it in no time, building apps out of latest version of the internet was the reason it broke so many live apps.
I agree with @gaelcolas I think this is better with everyone working with the owners of an existing community module to raise and ensure quality.
I'll just add that tests for such a project should include working with a large number of real-world YAML files for things like AppVeyor, Travis CI, VSTS, AWS CloudFormation, etc. For my own experience with YAML deserilization, I have had little success with one solution working universally and have ultimately had to reinvent the wheel several times. In that sense, I agree with @BrucePay "there are just no good YAML parsers".
We are talking about this platyPS module because it is already actively used in PowerShell Help environment. I guess no one from MSFT can tell how good this module is because of Code of Conduct. They can either silently reject it or improve it.
And although we've been talking about this a long time ago I don't see how we could use the components of this module here in a simple way.
Maybe @adityapatwardhan and @SteveL-MSFT will open their plans and timeline especially as the new Help RFC is already in the experiment stage.
My personal view is that I would rather see more community modules succeed and become de facto standard than requiring "official" modules from Msft.
@iSazonov It is one thing to have a solution that works for serializing/deserializing a well defined schema. It is quite another to have a solution that works in general with all schemas that are compliant YAML.
I understand the desire of MSFT to reuse community projects to cut costs. But the situation is, in fact, that MSFT may not make use of so many community projects:
many have bad code, have no trust
many projects are one person
MSFT has published Powershell specifications more 10 years ago, but nobody ported it yet until MSFT did.
The OpenSSL Project has existed for many years but nobody ported it to Windows while MSFT has not done this.
MSFT revealed many thousands of API interfaces, but how many of them were ported to Unix?
The interesting thing about why the company launched its project .Net Core rather than reuses Mono?
PowerShell is already a year and a half is an open source project, but I see that in this repository only one person from the community makes systematic contribution in the code @markekraus and only one person makes systematic analysis @mklement0.
I don't think that if we divide the project into parts, then we got more contributions.
I don't think the situation will change tomorrow. I wouldn't count on it.
@markekraus I very hope on http://yaml.org/spec/1.2/spec.html#id2802346 :-)
@iSazonov makes important points about support, trust and maintenance of 3rd party modules. Some 3rd party module can become a success and mature like e.g. Pester.
However, one should not assume that a great YAML module will evolve on its own over the next years. The reality is that most modules are published by authors who solved a particular problem and did the good deed of publishing their generic base code. This is how we ended up with 2 modules that aim to solve the same problem. Ideally one would need to merge them to focus efforts, otherwise they are going drift apart further in the future or just become stale and soon there will be more modules published by other people.
The underlying problem of having a proper parser indicates that basic (and substantial in terms of effort) ground work is needed and required to have a good YAML module.
I am not a YAML expert, but is this just a problem of the loose language specification itself or specific interpretation by various systems like VSTS or AppVeyor or is this only the lack of a good parser?
I found it frustrating to write YAML in VSCode and only when running it in VSTS to get an error that the VSTS parser does not like it...
To me this conversation is a case in point with open source's "code curation / architecture" problem.
Open source provides good seeding ideas and code bases - but if a serious architecture eye is not given to it when adopted as the most general solution - then it's 10 years of bug fixes for items that could have been taken care of in a decent design review.
In the true cases of @bergmeister "mature successes" it is often an active maintainer that has taken on the mission of generalizing the code base. But that can't be guaranteed to happen.
I think some of us our saying "YAML support is like support for writing files - it's core - it should be architected in the same way => with intention to the be the gold standard for that functionality"
Very valid points on software maturity. I haven't looked closely at the two modules listed here, nor at yamldotnet to make any opinion. Something we can look at as we start planning for 6.2.0
Don't get me wrong, I do value the experience and systematic approach of the PowerShell team and MSFT developers, I just think it's wrong for them to try to fill all the gaps with a module of their own stamped MSFT... It does not scale (and we've seen the problem with DSC resources already).
Increasing the reliance on MSFT provided modules is fragile, and does not help grow the community, nor the diversity of the ecosystem.
I'm in favour of MSFT contributing to open source projects to share their experience and help improve practices and quality, while not creating a dependence on them (because you know, squirrels...!).
The MSFT as unique provider of approved things is an old model that they struggle already to educate on, and it is not helping the community to encourage this approach (i.e. I'll wait, or moan, at Microsoft for not solving the problem I have kind of attitude in the OSS ecosystem).
I agree YAML support is core, instead of the PS team re-writing from scratch, why not help existing maintainers of projects to improve, and give them an opportunity to merge projects and hear from them what it would take. A bit like an apprenticeship/mentorship from PS team on core functionality modules.
Just re-writing a new module sounds like an engineer's reaction to solve a problem which is not an engineering problem. Re-writing a YAML module is an easy engineering task for the PS Team, but would not (help to) fix the community maturity problem, nor give the right incentive.
Whether Yaml is the strategic item to tackle this is MSFT's call though :)
@bergmeister
I'll preface this with myself not being a YAML expert. I happened to do some research on this when I wanted to bake some AppVeyor like yaml configs into my own franken-pipeline. I looked at how a dozen or so C# projects were consuming YAML. Since the PowerShell projects use YamlDotNet, I can only assume it's no easier. Though i have at leats toyed around with both PSYaml and powershell-yaml and look less closely at a few projects which use them.
I am not a YAML expert, but is this just a problem of the loose language specification itself or specific interpretation by various systems like VSTS or AppVeyor or is this only the lack of a good parser?
I suspect it's the nature of YAML being readable by humans at the possible expense of being more easily readable by machines. This readability-first paradigm extends into the way YAML authors write their YAML files. Though the resulting YAML is compliant under YAML spec, it is parsed in such away as to be unusable in code without using the deserialized object as an intermediary to an actually useful object.
That is to say, that 90% of the time the deserilization from YAML to an Object is not the issue, but the data design/architecture is. The other 10% of the time is parsing issues for which I can only chalk up to "YAML is hard to parse, man." However, the deserialized objects are often only slightly more useful than regexing what you are looking for....
As an example, the secure strings in AppVeyor.yml
environment:
my_var1: value1
my_var2: value2
my_secure_var1:
secure: FW3tJ3fMncxvs58/ifSP7w==
powershell-yaml and YamlDotNet do convert this to an object, but good luck using it without a bunch of logic. Once you have that logic, good for this schema, but what about another?
Some of these same data design problems plague JSON, but it is (in my experience and opinion) much easier to make models that can work around those shortcomings due to the more rigid nature of JSON. Trying to make models for any of the yaml deserializers mentioned in this thread is a nightmare if and where it is possible.
Granted, models are not a feature currently available in the JSON cmdlets, though I would really like to add it. If I had a say in the "official" YAML module/cmdlets I would put it down as a "must have" feature. It is a missed opportunity especially with the addition of PowerShell classes in v5.
IMO, Just getting YAML strings into an Object isn't good enough. That appears to be easy (90% of the time at least). The trick is getting YAML strings into useful objects. That requires some flexibility from the solution. But that flexibility must also be somewhat approachable and not require @IISResetMe and @lzybkr there to give you serialization advice....
To that affect, I haven't seen anything that works on general scope. Projects adopt the solutions available, and then use their output as intermediaries for actually useful objects (leading to a bunch of wheel reinventing that probably should be baked in upstream). Or, the projects compromise YAML readability for ease of parsing from YAML to objects.
@gaelcolas
I agree YAML support is core, instead of the PS team re-writing from scratch, why not help existing maintainers of projects to improve, and give them an opportunity to merge projects and hear from them what it would take
Ask yourself why MSFT started .Net Core project instead of continuing Mono many years later.
MSFT is a community too. And as any community has the same problems of interaction with other communities.
For context, I am not implying any work be done from scratch - code could be adopted - but should then be scrutinized from a Systems Development architecture perspective before being improved. It could even be open source after that review and re-release.
My point is to have a significant architectural review and remediation from a team that thoroughly understands the nuances of core code that will be leveraged virtually everywhere.
Another model always worth considering is acquire/contract/second. On this basis an effort is made to reach commercial terms with one or more community members/firms to recruit their services for a MSFT-led/facilitated development cycle to re-vamp and (in some fashion) integrate/connect the product(s). This was done successfully with Xamarin, which kicked the project to the Net Foundation, licensed it under MIT, and recruited/contracted/involved key resources such as Miguel de Icaza and Nat Friedman via Xamarin. Some whine that this is open source treason. But it does create positive incentives for folks and small firms to conceive and develop projects that later could be fit for widespread adoption and integration into at least one major ecosystem. Certainly it's preferred to jumping straight to a blank slate in-house redo that copies the whole concept and functionality and many of the ideas but jettisons the creators and (ostensibly) the code.
@iSazonov sorry for a late reply, no the platyPS yaml parser is no good: it only supports key value pairs. We also use YamlDotNet to generate yaml there.
Regarding the significant sentiment towards keeping this out of the core feature set: there's a very significant difference in how PowerShell handles dependencies compared to, say, Ruby, Python or Node.js.
Each of these languages has dependency management tools (bundler, pip, npm/yarn) which make the management of external dependencies easy and, more importantly, reproducible. Having something like a Gemfile/Gemfile.lock or package.json/package-lock.json [,yarn.lock] which makes for easy installation of all required packages and ensures that you are staying at a very specific patch level is a very significant distinction which is, in my opinion, what makes third-party libraries for something this fundamental feasible.
Perhaps there's something that could be done with Nuget to solve this issue, but I've never seen any articles describing dependency management strategies/patterns for PowerShell. Having the gallery is great, but if you've got to install all required packages manually it becomes unfeasible for any significant deployment.
What I do no each new pwsh system? I do Install-Module -Name powershell-yaml.
Mongo, Kubernetes, Istio, Ansible - I use. All these are YAML and I do have YAML templates and transfomations.
@dzmitry-lahoda Issue #5681 proposed to have a rich' version of PowerShell that ships with a set of common modules such as e.g. Pester`, etc. Please post in this issue but given that there seems to be no clear winner between the currently 2 available yaml modules and they are clobbering each other, it might be a hard decision to pick a favourite.
I see only one YAML :(
Pester, yeh. Too heavy to ship BDD framework into mainline, unlike YAML reader for my pwsh container applications.
Has this thread been concluded. What is the recommended(or suggested) module to use by Microsoft?
DevOps pipelines uses yaml. All my deployment automation are build with powershell. Seems yaml and powershell do not play nice. Is powershell a bad choice for Azure DevOps automation?
Need to think carefully of my future use/innovation and would appreciate some direction.
Thanks in advance!
@dirkslab You could use https://github.com/cloudbase/powershell-yaml
Thanks @iSazonov, that is the solution I am testing with at the moment. The solution so far seems to work fine. There is probably nothing wrong using the solution.
Note that using powershell-yaml you need to ok an untrusted module. This is the part that i am battling to understand. Microsoft recommends using yaml pipelines. Microsoft (or at least this thread) suggest using 3rd party module so you can integrate yaml config with powershell, but do not endorse or recommend any. How do you explain that logically to enterprise.
My experience so far has always been that if you do not use Microsoft endorsed solutions, that would mute any support or understanding from Microsoft for your solution issues(this does not matter if the unsupported part touches anything causing problems). The mere fact that you have unsupported part usually results in no support/responsibility .
Maybe things have changed in this OpenSource era. Simple official response and guidance from Microsoft would set me at ease and help me understand.
Appreciated your response. Regards.
@dirkslab I think your MSFT account manager is right person to ask about support policy.
CoreFX team asks about use cases
Appart from obvious benefits that yaml is all around us in CI/CD today and number of configuration systems, the added benefit of ConvertTo-Yaml is representing nasted HashTable in human readable format, unlike ConvertTo-Json that we have to use now which makes output not very readable.
I use Write-HashTable in the meantime, but this would be great to have OTB.
I hate yaml, I really hate it. However there are a couple of facets worth the MS team considering:
It's become the defacto language of CI: docker-compose.yaml, ansible, kuber, k8s, github, circle, azure, ... And it seems to crawl out of CI into the projects that use it.
$config = Invoke-WebRequest https://$ciserver/api/projects/$project/config.yaml | ConvertFrom-Yaml
$config['started'] = Get-Date
$config['options'] = $options
Invoke-WebRequest -Method Post https://$ciserver/api/projects/$project/build -Body ($config | ConvertTo-Yaml)
Having this ship with powershell would be transformative in evangelising to CI groups.
"If we switch to ms powershell we can automatically" -> "Tell me more?"
vs
"If we switch to ms powershell and download some scripts from the gallery" -> "no"
Really, this is by-the-by, but yaml is a superset of json, such that json is an abbreviated form of yaml, an efficient yaml parser is an efficient json parser,
Can this be reconsidered for 7.1? I'm also having issues with using an untrusted module and something so DevsOpsy should really be native to PowerShell.
IMHO, YAML is as popular as JSON and CSV, and not having inbox converters for YAML in PowerShell is kind of sad. Having inbox YAML converters will also ensure that their behavior is on par with JSON converters, which is not the case with community modules.
Don't get me wrong - I appreciate that folks create modules for community, but in the current state of the world, YAML conversion is table stakes - we don't expect folks to download 3rd-party modules for JSON conversion.
@dirkslab You could use https://github.com/cloudbase/powershell-yaml
GitHub**cloudbase/powershell-yaml**PowerShell CmdLets for YAML format manipulation. Contribute to cloudbase/powershell-yaml development by creating an account on GitHub.
The problem is that unlike ConvertFrom-Json, this module deserializes to Hashtable instead of a proper object (or PSCustomObject), which doesn't allow using the pipeline in the following way:
$obj | ? Categories -gt 0 | % Name
My current workaround:
'...' | ConvertFrom-Yaml | ConvertTo-Yaml -JsonCompatible | ConvertFrom-Json
Suboptimal, to say the least.
@marektoman, your example should work with a hashtable:
PS> @{ one = 1; two = 2 } | ? one -eq 1 | % two
2
Also note that you can cast hashtables to [pscustomobject] (albeit only one by one - see #13838):
PS> @{ one = 1; two = 2 }, @{ one = 10; two = 20 } | % { [pscustomobject] $_ }
one two
--- ---
1 2
10 20
@marektoman, your example should work with a hashtable:
PS> @{ one = 1; two = 2 } | ? one -eq 1 | % two
2
Also note that you can cast hashtables to [pscustomobject] (albeit only one by one - see #13838):
PS> @{ one = 1; two = 2 }, @{ one = 10; two = 20 } | % { [pscustomobject] $_ }
one two
--- ---
1 2
10 20
Good point, the behavior is unrelated to hashtable, but it is likely related to the way each module does output.
When I run
$yamlArray | ConvertFrom-Yaml | ? Name -eq 'All'
, it doesn't work. However, this does:
($yamlArray | ConvertFrom-Yaml) | ? Name -eq 'All'
Thanks @marektoman , my solution was to remove the extra yaml module overhead and settle somewhere in between yaml and json.
I did something similar as you when requiring an pscustomobject:
$obj = Get-Content -Path .\deployment-config.yml -raw | ConvertFrom-Yaml | convertto-json -Depth 100 | ConvertFrom-json -Depth 100
The direction I have chosen to go with:
I have replaced all my yaml config with .psd1 and using Import-PowerShellDataFile
Not as clean as yaml, more than sufficient and easy to read though. Also eliminated any and all module issue.
Examples:
Sharing project with other devs that are domain policy restricted on installing random and unsigned modules, not an issue anymore.
I can now . browse path before I convert it to a string
adding comments are easy
My personal goals were to need little to no 3rd party modules for my ps project. I am now back on track and it's hassle free.
@marektoman : The (...) forces enumeration, which in turn suggests that when ConvertFrom-Yaml outputs an array of hashtables that it does so as a single object, which contravenes the expected streaming (one-by-one) pipeline behavior. It is what '[ 1, 2 ]' | ConvertFrom-Json used to do (see #3424), and still does in Windows PowerShell).
@dirkslab Good idea for that use case. My use case is a shell independent tool that produces single output, which is both human and machine readable.
@mklement0 I see, thanks for all the info.
It is absolutely CRAZY that there is not first class support for yaml in PowerShell. I am just stunned.
The Kubernetes movement alone should be more than enough to get this prioritized.
We should admit that YAML is used practically in many/most of cloud, CI, configuration, orchestration tools that IT pros, devs and DevOps need to use. So if PowerShell want to stay (most :D) valuable in that world for each it was designed, created and tailored specifically it should support it out of the box.
can i donate 10 USD for the feature? I know there are services where donation is done until some dev implements
This page talks a lot about the powershell-yaml module as an option to get YAML support.
Does this library work on Linux with Powershell Core? or it just a Windows only thing. (I noticed that its dependencies include .Net Framework, so I am worried.)
@Vaccano it works well on Linux.
My initial thinking is that YAML support should be standard in PowerShell 7.x. YAML is such an integral language to today's enterprises that its omission is glaring considering support for XML and JSON. I have no problem with a separate module that can also be installed from the gallery, but it should be included by default in a new PowerShell 7.x installation. I know there are organizations that won't consider anything open source, and certainly nothing that doesn't originate or is supported by Microsoft. Given the value of PowerShell as a cloud language, native YAML support seems like the right decision.
I am also less concerned about making the module downlevel compatible. If someone needs YAML conversion support for Windows PowerShell, there appear to be a number of options in the PowerShell Gallery.
As a normal user, I've installed PowerShell Core and use it in my work.
Our projects' ps1 scripts are not compatible with PowerShell Core. I have switched to the legacy PowerShell to run them.
And it's difficult to install a new module in this constrained work environment.
I don't know how to write PowerShell scripts, and now I have to live with the old PowerShell.
I feel sad because the new shiny PowerShell Core that I'm in love, is not usable in even a simple real-world usage.
Maybe, I have to rewrite those *.ps1 files by using a serious programming language.
@PowerShell/wg-powershell-cmdlets has been discussing this issue, and despite the extremely high demand (and my own personal desire to have strong, canonical YAML cmdlets), we're unfortunately going to have to close it (for now) as Won't Fix. The reasoning is two-fold:
While the PowerShell package is already fairly large, we're doing our best not to add new dependencies to the package so that we don't continue to grow our size unfettered, and because we believe that pulling modules from the PowerShell Gallery (or an internal trusted Gallery, having been vetted appropriately by an enterprise) should be a standard part of a PS7 workflow.
As a team, we simply don't have the capacity to build a .NET YAML parser from scratch, nor do we have the capacity to vet, support, and service a 3rd party library for which the security implications are high-stakes.
If .NET were to officially support a YAML library, we (as Microsoft) could potentially look into supporting a YAML module on the Gallery. But until that point, the existing community modules will have to suffice.
Thanks, Joey. I'd like to file a Feature Request in dotnet but don't understand it well enough to know which repo to add the issue. Can you please suggest which repo would be most appropriate within https://github.com/dotnet ?
(or could a Microsoft person/PowerShell team member place to the request as it may get a higher priority?)
See https://github.com/dotnet/runtime/issues/28414
Thank you @iSazonov. That doesn't look promising :/ Seems that we do need an advocate from the PowerShell team to create the request, eh?
I guess Microsoft has internal conclusion that it is low priority and small area.
You can ask PowerShell team online today on Community Call https://github.com/PowerShell/PowerShell-RFC/issues/299
This makes me sad.
Powershell says ".Net has to do it for us to do it"
.Net says "It is up to Powershell if they want to do it"
I wonder if some of the higher ups at Microsoft realize the runaround that is going on here.
YAML is important. Having it from a signed Trusted Source is critical for most companies. Powershell has built in restrictions making it hard to use things not signed by trusted sources.
I give up. It seems that the enough dust has been kicked up to obscure this issue. I imagine at some point someone high up at Microsoft will say "What? No YAML support in PowerShell?" Fingers will be pointed, blame laid, excuses made (etc). Then it will get done. But no amount of requests by the community will get this through.
Why has this been closed as WONTFIX? Seriously, YAML is so common these days, all the cloud-native stuff uses it. One can do so much with just vanilla PowerShell before they need to import a module, and not being to deal with YAML shouldn't be one of them. The built-in JSON cmdlets is one of the reasons why JSON is to popular in PowerShell: it's there, it's standardized, and it works. I am sure I am not alone defaulting to JSON just for that reason, and thinking twice about using YAML if I don't really need it, as it makes my scripts less portable or standalone.
@awakecoding joey's explanation is here: https://github.com/PowerShell/PowerShell/issues/3607#issuecomment-901288121
The Json library, even though it was Newtonsoft, was "blessed" by Microsoft as the official JSON library till system.text.json came around.
It's a bummer for sure, I would love native YAML support as well, but I personally think the reasoning is sound. I think energies should be focused on more easily pulling down dependencies quickly for a script/module.
Seriously surprised this isn't already included. It is sorely needed in the community and would seem like a obvious inclusion.
It's a shame.
We pay many billions in royalties and still have to beg M$ for many years so that not Amazon or Google don't step in again to solve such "problems".
Yes, it's a shame.
This makes me sad.
Powershell says ".Net has to do it for us to do it" .Net says "It is up to Powershell if they want to do it"
I wonder if some of the higher ups at Microsoft realize the runaround that is going on here.
YAML is important. Having it from a signed Trusted Source is critical for most companies. Powershell has built in restrictions making it hard to use things not signed by trusted sources.
I give up. It seems that the enough dust has been kicked up to obscure this issue. I imagine at some point someone high up at Microsoft will say "What? No YAML support in PowerShell?" Fingers will be pointed, blame laid, excuses made (etc). Then it will get done. But no amount of requests by the community will get this through.
After 5 years, it's time to ask M$$ again to do its homework, especially because we need YAML on azure.
Please vote:
https://github.com/PowerShell/PowerShell/issues/16785
@mubed Probably won't happen until MS releases a first party YAML library:
https://github.com/PowerShell/PowerShell/issues/3607#issuecomment-901288121
Several YAML modules available on the gallery though:
https://www.powershellgallery.com/packages?q=yaml
Native YAML support in PowerShell is extremely necessary nowadays. We have to remember that PowerShell was primary designed for IT professionals and system administrators which most of cases have to deploy vanilla PowerShell scripts over the production environment because they cannot deploy modules (no internet connection, security constraints, not signed binaries etc). Also the module footprint is a concern. A single 2KB script that connects to SQL Server requires a 50 MB Sql Server module to work. If an script requires more than one module, I cannot imagine how much disk space would be required. NPM packages feeling? I really does not understand this concern of PowerShell team of not changing the standard library. Are PowerShell going to the same Javascript path?
@ronascentes it's pretty simple. There is no first-party Microsoft YAML library (there is a JSON one and hence why its there), so having a first-party YAML support brings all kinds of licensing and whatnot issues until that happens.
Fortunately there's lots of great community modules, my favorite being YaYaML.
https://github.com/jborean93/PowerShell-Yayaml
Just run install-module yayaml and you're good to go.
I really does not understand this concern of PowerShell team of not changing the standard library. Are PowerShell going to the same Javascript path?
@ronascentes - It is two parts of Microsoft pointing the finger at the other. The PowerShell team says that they need a YAML parser built into .NET before they can add native YAML support to PowerShell. The PowerShell team says it does not have the manpower to build and maintain a .NET YAML parser, and thus cannot add native YAML support PowerShell. (They are unwilling add native YAML support based on a community YAML parser. This is due to licensing and support issues.)
The .NET team says there is not a need for a native .NET YAML parser. They site that there is a fine community YAML parser already out there. They say that if PowerShell needs a YAML parser, they need to build and maintain their own parser.
So basically, it is one company (Microsoft) acting as two companies (PowerShell and .NET). The two parts of Microsoft are unwilling to get together and make a plan to fix this situation.
The .NET team says there is not a need for a native .NET YAML parser. They site that there is a fine community YAML parser already out there. Because of this, they are unwilling to build a YAML parser into .NET. They say that if PowerShell needs a YAML parser, they need to build and maintain their own parser. (Which the PowerShell team says they cannot do.)
Same thing was said about json, yet we have System.Text.Json today 🤷
@StephenHodgson agreed, and as soon as the dotnet team recognizes that and ships System.Text.Yaml in Dotnet 11 (or whatever), I'm sure a first party PS module will be soon to follow :)
It will take someone thinking that the existing libraries are not "good enough" in the same way it was determined for NewtonSoft. Honestly surprised as much of a Not-Invented-Here culture that exists over there that it hasn't happened sooner.
Related Discussion FYI: https://github.com/dotnet/runtime/issues/83313
The .NET team says there is not a need for a native .NET YAML parser. They site that there is a fine community YAML parser already out there. Because of this, they are unwilling to build a YAML parser into .NET. They say that if PowerShell needs a YAML parser, they need to build and maintain their own parser. (Which the PowerShell team says they cannot do.)
Same thing was said about json, yet we have System.Text.Json today 🤷
Without JSON patch support...
You could make a module for that using this library.
https://github.com/weichch/system-text-json-jsondiffpatch
|
gharchive/issue
| 2017-04-20T17:04:45 |
2025-04-01T04:55:32.153042
|
{
"authors": [
"ArieHein",
"BrucePay",
"DarwinJS",
"DavidChennells",
"JustinGrote",
"MattTunny",
"Satak",
"StephenHodgson",
"SteveL-MSFT",
"Vaccano",
"autosvet",
"awakecoding",
"bergmeister",
"bgshacklett",
"dirkslab",
"dzmitry-lahoda",
"gaelcolas",
"iSazonov",
"jdhitsolutions",
"joeyaiello",
"josepmv",
"kfsone",
"lzybkr",
"m3nax",
"majkinetor",
"marektoman",
"markekraus",
"mklement0",
"potatoqualitee",
"redstrike",
"ronascentes",
"schittli",
"vors",
"wasker"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/3607",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
223200981
|
Need a way to bring back legacy Windows aliases for consistent experience
To address https://github.com/PowerShell/PowerShell/issues/929 we removed aliases that collided with native tools. However, some customers may want the legacy Windows PowerShell aliases because they are consistent with what they are used to. Proposal that has been discussed is to have a module that only contains aliases that users can import to get back that previous experience.
Where should this new module go?
It's not about "what we're used to." I'm a Mac user that has barely used PowerShell at all on windows. If I switched to powershell, it would be for the object-oriented model. My expectation is that ls gives me a stream of objects, because that is powershell's killer feature. Without that, PS is nothing.
I still remember curl and Invoke-WebRequest quarrel.
Powershell thought about introducing this grammar?
{Module name}. Command name
or
(Module name). Command name
or others
In this case, add a linux local command module, the inside of the ls command, link to / bin / ls can.
It is simple and does not change the old ones.
You can even make modules based on some linux release.eg:
(Centos).ls <----- module name.command name
{Mac}.ls
[Ubuntu1604].ls
@Gaelan for your use case, I would suggest using Get-ChildItem or gci as the short form, meant for interactive usage. I understand that it can be difficult to retrain muscle memory (I'm having trouble myself), but being explicit with what your asking the shell to give you is almost always going to give you a better experience than relying on aliases which were not originally intended for your platform. The decision to remove these aliases has not removed any functionality in PowerShell, but only changed the way the commands need to be invoked.
ls certainly falls into a grey area, as it's a built-in in both PowerShell and Bash, but in the case of commands such as curl, wget, etc. which mask functionality that already exists on the system, I don't think there was really any question of what the "correct" behavior is, only the least disruptive. There was no solution that was going to be ideal for everyone. At least with this solution you will have an easy option to re-enable the aliases you wish to have in your environment.
@dchristian3188 If your question was about a module place I believe it is https://github.com/PowerShell/PowerShell/tree/master/src/Modules/Shared
Putting them in a module is only one proposal. The point is that there needs to be an interactive (and possibly also parse-time) mechanism for turning GNU/POSIX colliding aliases on and off. (Side note: it's not just *nix users who may want to turn them on, Windows users may also want to turn these aliases off.)
Also, current thinking is that this might be blocked on PowerShell/PowerShell-RFC#48, which is why I'm removing up-for-grabs and personally prioritizing driving that RFC to completion (it's blocking other usability issues as well).
@chuanjiao10 for cmdlet name collision, we already support that today:
Microsoft.PowerShell.Management\Get-ChildItem
3q @SteveL-MSFT .
Hello everyone:
I found this linux local command return error in powershell, but normal in BASH:
PS /root> /usr/bin/ls -l /root/*.xml
#return /usr/bin/ls: can not access /root/*.xml: no that file or directory
PS /root> ls -l /root/*.xml
#return /usr/bin/ls: can not access /root/*.xml: no that file or directory
So I suggest: linux local command in THE module name,THE module link to linux path. Powershell explain linux local command, different treatment.
Linux\ls
Or
Ubuntu1604\ls
@chuanjiao10: That might be the behavior you want, but you received far too much feedback on day one that people want things like ls -la to just work out of the box. (Side note: the full path to the binary will always work, no matter what path we take here.)
The fact that you have a preference is why we're proposing that there be an opt-in mechanism for exactly the behavior you're describing.
Sound reasonable?
Also, the reason your example is failing is because we need to implement globbing with #954
Is here a way to explicitly start external (OS) command?
This is start cmdlet:
&"where"
Can we do somethink like:
$nativecommand:where
@iSazonov - I believe that can be done with Get-Command:
$where = Get-Command -Name where -CommandType Application
& $where
@chuanjiao10 the globbing issue you have above should be fixed in beta.1
yes i saw. thanks @daxian-dbw goodjob!
|
gharchive/issue
| 2017-04-20T21:08:34 |
2025-04-01T04:55:32.168937
|
{
"authors": [
"Gaelan",
"SteveL-MSFT",
"bgshacklett",
"bielawb",
"chuanjiao10",
"dchristian3188",
"iSazonov",
"joeyaiello"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/3610",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
346410465
|
Windows PowerShell Event Provider GUID still in use
The Windows PowerShell guid is used for tracing here:
https://github.com/PowerShell/PowerShell/blob/c1c5344a8897262433141ecbc13bb06ac2c4bbef/src/System.Management.Automation/utils/tracing/TracingGen.cs#L38
It should use the PowerShell Guid from here:
https://github.com/PowerShell/PowerShell/blob/master/src/PowerShell.Core.Instrumentation/PowerShell.Core.Instrumentation.man#L34
Actually, I think this is WorkFlow code and can be removed.
|
gharchive/issue
| 2018-08-01T01:01:36 |
2025-04-01T04:55:32.171240
|
{
"authors": [
"TravisEz13"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/issues/7421",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
911116756
|
Thread-safe HttpKnownHeaderNames initialization
PR Summary
Fix #15513
Move HttpKnownHeaderNames initialization to static constructor.
PR Context
See #15513.
PR Checklist
[x] PR has a meaningful title
Use the present tense and imperative mood when describing your changes
[x] Summarized changes
[x] Make sure all .h, .cpp, .cs, .ps1 and .psm1 files have the correct copyright header
[x] This PR is ready to merge and is not Work in Progress.
If the PR is work in progress, please add the prefix WIP: or [ WIP ] to the beginning of the title (the WIP bot will keep its status check at Pending while the prefix is present) and remove the prefix when the PR is ready.
Breaking changes
[x] None
OR
[ ] Experimental feature(s) needed
[ ] Experimental feature name(s):
User-facing changes
[x] Not Applicable
OR
[ ] Documentation needed
[ ] Issue filed:
Testing - New and feature
[x] N/A or can only be tested interactively
OR
[ ] Make sure you've added a new test if existing tests do not effectively test the code changed
Tooling
[ ] I have considered the user experience from a tooling perspective and don't believe tooling will be impacted.
OR
[ ] I have considered the user experience from a tooling perspective and opened an issue in the relevant tool repository. This may include:
[ ] Impact on PowerShell Editor Services which is used in the PowerShell extension for VSCode
(which runs in a different PS Host).
[ ] Issue filed:
[ ] Impact on Completions (both in the console and in editors) - one of PowerShell's most powerful features.
[ ] Issue filed:
[ ] Impact on PSScriptAnalyzer (which provides linting & formatting in the editor extensions).
[ ] Issue filed:
[ ] Impact on EditorSyntax (which provides syntax highlighting with in VSCode, GitHub, and many other editors).
[ ] Issue filed:
Even with this, the hashset will still be accessed from multiple threads concurrently in the cases given in #15513, won't it? We should probably add a test using the code samples in the issue to verify that this is sufficient to resolve it. Since the user reporting it mentioned that it was a bit intermittent, we should repeat their test a few times in our own tests.
Even with this, the hashset will still be accessed from multiple threads concurrently in the cases given in #15513, won't it?
A root of the issue was that second thread read the HashSet while first thread still initializes the HashSet.
The HashSet is used only in one place for reading so no need to concern more about race conditions.
We should probably add a test
It's not worth it. We will not be able to make such tests stable and defensible. Please use compiled artifact to check locally.
Even with this, the hashset will still be accessed from multiple threads concurrently in the cases given in #15513, won't it?
A root of the issue was that second thread read the HashSet while first thread still initializes the HashSet.
The HashSet is used only in one place for reading so no need to concern more about race conditions.
We should probably add a test
It's not worth it. We will not be able to make such tests stable and defensible. Please use compiled artifact to check locally.
If the only writes are the ones when the object is generated :
s_contentHeaderSet.Add(HttpKnownHeaderNames.Allow);
s_contentHeaderSet.Add(HttpKnownHeaderNames.LastModified);
I think the write race condition is eliminated with this change.
I am a C# dev, but I have zero knowledge on the Powershell source code. Help I could help you somehow.
I am a C# dev, but I have zero knowledge on the Powershell source code. Help I could help you somehow.
PowerShell is great project to improve skills. It's too big to learn it well in a short time but you can pick an area you like best and work there with some workgroup https://devblogs.microsoft.com/powershell/powershell-working-groups/
After monitoring issues and PRs for about a month you will be able to make good contributions.
@daxian-dbw Could you please review?
@SteveL-MSFT Could you please approve? It is impossible to merge without green flag :-)
I see that there's a pending review from @daxian-dbw
Is there anything else you need from me?
Thank you
@TravisEz13 Please review/approve.
|
gharchive/pull-request
| 2021-06-04T04:54:17 |
2025-04-01T04:55:32.190825
|
{
"authors": [
"dperez83",
"iSazonov",
"vexx32"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/pull/15519",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1837546752
|
Attempt to fix the issue https://github.com/PowerShell/PowerShell/issues/20055
Attempt to fix the issue of Powershell's magnet icon being too large on the Windows 10 start menu and not displaying the application name
Original issuse: https://github.com/PowerShell/PowerShell/issues/20055
Expected fix:
fixed tile displaying application name 'PowerShell' on Windows 10 22h2 Start menu
https://answers.microsoft.com/zh-hans/windows/forum/all/有些microsoft/ee1b391c-5ce6-4304-903e-94be1fa29bb7
Please use the pull request template and give the PR a descriptive name
请使用拉取请求模板并为 PR 指定一个描述性名称
finnished
|
gharchive/pull-request
| 2023-08-05T03:14:36 |
2025-04-01T04:55:32.194945
|
{
"authors": [
"Zhoneym",
"dkaszews"
],
"repo": "PowerShell/PowerShell",
"url": "https://github.com/PowerShell/PowerShell/pull/20080",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
230142919
|
xSQLServerNetwork: When TcpDynamicPorts is set to '0' then Test-TargetResource always fail
Details of the scenario you tried and the problem that is occurring:
When parameter TcpDynamicPorts is set to '0' then Test-TargetResource will always fail because network protocol property TcpDynamicPort is always set to the actual listening port when SQL Server is restarted.
If parameter TcpDynamicPort is set to '0' then Test-TargetResource should only return false if network protocol property TcpDynamicPort is not set to a value.
The DSC configuration that is using the resource (as detailed as possible):
xSQLServerNetwork "$($currentSqlInstance)-InstanceNetworkConfiguration"
{
SQLServer = $Node.NodeName
InstanceName = $Node."$($currentSqlInstance)InstanceName"
ProtocolName = 'Tcp'
IsEnabled = $true
TcpDynamicPorts = '0'
RestartService = $true
}
Version of the Operating System, SQL Server and PowerShell the DSC Target Node is running:
All
What module (SqlServer or SQLPS) and which version of the module the DSC Target Node is running:
n/a
Version of the DSC module you're using, or 'dev' if you're using current dev branch:
Dev
This should have been solved with a boolean parameter instead, but hold off on that until another PR. See issue https://github.com/PowerShell/xSQLServer/issues/534.
|
gharchive/issue
| 2017-05-20T09:46:33 |
2025-04-01T04:55:32.255959
|
{
"authors": [
"johlju"
],
"repo": "PowerShell/xSQLServer",
"url": "https://github.com/PowerShell/xSQLServer/issues/564",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
198461899
|
README.md: Small fixes to sections and fixed wrong product name
Changes to README.md
Now the correct product name is written in the installation section, and a typo was also fixed.
Fixed a typo in the Requirements section.
Added link to Examples folder in the Examples section.
This Pull Request (PR) fixes the following issues:
Fixes #283
[x] Change details added to Unreleased section of CHANGELOG.md?
[ ] Added/updated documentation, comment-based help and descriptions in .schema.mof files where appropriate?
[ ] Examples appropriately updated?
[ ] New/changed code adheres to Style Guidelines?
[ ] Unit and (optional) Integration tests created/updated where possible?
This change is
Reviewed 2 of 2 files at r1.
Review status: all files reviewed at latest revision, all discussions resolved.
Comments from Reviewable
|
gharchive/pull-request
| 2017-01-03T12:09:41 |
2025-04-01T04:55:32.261787
|
{
"authors": [
"johlju"
],
"repo": "PowerShell/xSQLServer",
"url": "https://github.com/PowerShell/xSQLServer/pull/284",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
207039828
|
xSQLServerSetup: Support feature CONN and BC
It now detects that feature Client Connectivity Tools (CONN) and Client Connectivity Backwards Compatibility Tools (BC) is installed. This is based on the work of @joshan1120 in PR #163. This resolved issue #159 and might also resolve #142.
This Pull Request (PR) fixes the following issues:
Fixes #159
[x] Change details added to Unreleased section of CHANGELOG.md?
[ ] Added/updated documentation, comment-based help and descriptions in .schema.mof files where appropriate?
[ ] Examples appropriately updated?
[x] New/changed code adheres to Style Guidelines?
[x] Unit and (optional) Integration tests created/updated where possible?
This change is
Codecov Report
Merging #383 into dev will increase coverage by <1%.
The diff coverage is 87%.
@@ Coverage Diff @@
## dev #383 +/- ##
====================================
+ Coverage 63% 64% +<1%
====================================
Files 31 31
Lines 3092 3104 +12
====================================
+ Hits 1978 1988 +10
- Misses 1114 1116 +2
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 4304e00...233296e. Read the comment docs.
Reviewed 2 of 3 files at r1, 1 of 1 files at r2.
Review status: all files reviewed at latest revision, all discussions resolved, some commit checks failed.
Comments from Reviewable
Reviewed 3 of 3 files at r3.
Review status: all files reviewed at latest revision, all discussions resolved.
Comments from Reviewable
|
gharchive/pull-request
| 2017-02-12T09:43:12 |
2025-04-01T04:55:32.271387
|
{
"authors": [
"codecov-io",
"johlju"
],
"repo": "PowerShell/xSQLServer",
"url": "https://github.com/PowerShell/xSQLServer/pull/383",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2458879638
|
🛑 HDArea is down
In b1502d6, HDArea ($HDAREA) was down:
HTTP code: 502
Response time: 342 ms
Resolved: HDArea is back up in 44ad395 after 6 minutes.
|
gharchive/issue
| 2024-08-10T03:49:23 |
2025-04-01T04:55:32.274193
|
{
"authors": [
"PowerfulRichard"
],
"repo": "PowerfulRichard/PT-Status",
"url": "https://github.com/PowerfulRichard/PT-Status/issues/85",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
947177091
|
[WIP] Random weight and favored
This PR is basically a combination of random favored idea https://github.com/Practical-Formal-Methods/AFL-public/pull/4 and weight random selection https://github.com/Practical-Formal-Methods/AFL-public/pull/5.
@wuestholz, I am setting up the fuzzbench experiment with the following fuzzers.
vanilla AFL
the latest setting of the random favored PR (increase_boost_inputs && base_weight_fac == 4.0 && max_weight_fac_decr == 3.75)
combination of random favored and weight random selection (this PR with the customzed values same as fuzzer no.2 — increase_boost_inputs && base_weight_fac == 4.0 && max_weight_fac_decr == 3.75)
combination of random favored and weight random selection (this PR with default parameters)
@wuestholz, could you please help to confirm that I have correct fuzzers for the evaluation?
@jiradeto Thanks! This will be a local experiment, right?
Did you already run an experiment to compare increase_boost_inputs with new_increase_boost_fast_seqs (from https://github.com/Practical-Formal-Methods/AFL-public/pull/4#issuecomment-879207228)? If not we should also include just the increase_boost_inputs configuration.
@jiradeto Thanks! This will be a local experiment, right?
@wuestholz, yes I mean the local experiment.
So, based on your your comment in another PR (https://github.com/Practical-Formal-Methods/AFL-public/pull/4#issuecomment-882443156), I assume the current setting means the default value of increase_boost_inputs which is:
default && max_weight_fac_incr == 15.0
If this is the case, I will run fuzzbench with the following fuzzers:
vanilla AFL
the increase_boost_inputs setting of the random favored PR (max_weight_fac_incr == 15.0)
combination of random favored and weight random selection (this PR with max_weight_fac_incr == 15.0)
combination of random favored and weight random selection (this PR with default parameters)
Could you please tell me if I am missing something?
@jiradeto Great! Thanks!
I think we don't need the fourth configuration.
Could you please also update this PR to make 15.0 the default value?
Let's add the following flags:
DISABLE_WRS: disables weighted random selection
DISABLE_RF: disables random favorites
ENABLE_UF: enables uniformly random favorites (i.e., weight always 1)
DISABLE_RF: disables random favorites
@wuestholz, does this mean that we disable both random weighted selection (PR https://github.com/Practical-Formal-Methods/AFL-public/pull/4) and uniformly random favorites (PR https://github.com/Practical-Formal-Methods/AFL-public/pull/5)?
DISABLE_RF: disables random favorites
@wuestholz, does this mean that we disable both random weighted selection (PR #5) and uniformly random favorites (PR #4)?
@jiradeto No, just uniformly random favorites (PR https://github.com/Practical-Formal-Methods/AFL-public/pull/4).
DISABLE_RF: disables random favorites
@wuestholz, does this mean that we disable both random weighted selection (PR #5) and uniformly random favorites (PR #4)?
@jiradeto No, just uniformly random favorites (PR #4).
@wuestholz OK, that makes sense.
But I am unsure when to use ENABLE_UF since the default weight is already set to 1.0 if all additional parameters enable_throttle_inputs, enable_boost_fast_seqs, enable_boost_inputs are not enabled.
DISABLE_RF: disables random favorites
@wuestholz, does this mean that we disable both random weighted selection (PR #5) and uniformly random favorites (PR #4)?
@jiradeto No, just uniformly random favorites (PR #4).
@wuestholz OK, that makes sense.
But I am unsure when to use ENABLE_UF since the default weight is already set to 1.0 and untouched if additional parameters enable_throttle_inputs, enable_boost_fast_seqs, enable_boost_inputs are not enabled.
@jiradeto I see. I suggest we replace enable_throttle_inputs, enable_boost_fast_seqs, and enable_boost_inputs with a single flag ENABLE_UF. When it is not set we will use enable_boost_fast_seqs and enable_boost_inputs.
We can probably remove the enable_throttle_inputs flag and the corresponding code.
@jiradeto Great! Thanks for the latest changes. There's only a small typo I noticed. I think we could try to set up an fuzzbench experiment with the following fuzzers:
afl (vanilla AFL)
afl_no_favs: DISABLE_WRS && DISABLE_RF && !ENABLE_UF && DISABLE_FAVS
afl_wrs_rf: !DISABLE_WRS && !DISABLE_RF && !ENABLE_UF && !DISABLE_FAVS
afl_rf: DISABLE_WRS && !DISABLE_RF && !ENABLE_UF && !DISABLE_FAVS
afl_rf_u: DISABLE_WRS && !DISABLE_RF && ENABLE_UF && !DISABLE_FAVS
afl_wrs: !DISABLE_WRS && DISABLE_RF && !ENABLE_UF && !DISABLE_FAVS
Can you think of another setup that would be interesting?
afl_rf_u: !DISABLE_WRS && !DISABLE_RF && ENABLE_UF && !DISABLE_FAVS
@jiradeto Sure, why not. :)
I was thinking a public one.
@wuestholz, I feel that I'd rather at least have a local experiment result before I can go with the public one. Maybe we should start with the local experiment to see whether the above configurations behave properly?
@jiradeto Sure, that's also fine with me. 👍
@jiradeto As discussed, I would first do the following:
close #8
merge #9
merge #7
close #3, #4, and #5
Afterwards, it would be great if you could run a local experiment that compares the following:
afl (vanilla AFL)
afl_wrs_rf: !DISABLE_WRS && !DISABLE_RF && !ENABLE_UF && !DISABLE_FAVS && DISABLE_RP
afl_wrs_rf_u: !DISABLE_WRS && !DISABLE_RF && ENABLE_UF && !DISABLE_FAVS && DISABLE_RP
afl_wrs_rf_t: !DISABLE_WRS && !DISABLE_RF && !ENABLE_UF && !DISABLE_FAVS && DISABLE_RP && changes below
The changes for the last configuration are:
if (!enable_uniformly_random_favorites) {
// enable_boost_inputs
double base_weight_fac_boost_inputs = 1.0;
double max_weight_fac_incr = 7.0;
double scale_fac_boost_inputs = 0.001;
double num_selections = (double)q->num_fuzzed;
weight *= base_weight_fac_boost_inputs + max_weight_fac_incr / (scale_fac_boost_inputs * num_selections + 1.0);
// enable_boost_fast_seqs
double base_weight_fac_boost_fast = 8.0;
double max_weight_fac_decr = 7.0;
double scale_fac_boost_fast = 0.005;
double execs_per_sec = 1000000.0 / (double) q->exec_us;
weight *= base_weight_fac_boost_fast - max_weight_fac_decr / (scale_fac_boost_fast*execs_per_sec + 1.0);
}
@wuestholz thank you for your detailed guidance. The fuzzbench evaluation for the changes you requested is available and can access using the following links:
bloaty
sqlite
libjpeg
libxml
freetype
@jiradeto Thanks a lot! Looks like it's slightly better. If you agree I suggest we update the code.
|
gharchive/pull-request
| 2021-07-19T01:41:30 |
2025-04-01T04:55:32.305848
|
{
"authors": [
"jiradeto",
"wuestholz"
],
"repo": "Practical-Formal-Methods/AFL-public",
"url": "https://github.com/Practical-Formal-Methods/AFL-public/pull/6",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
226378320
|
Cannot Use SVGs that are not from FontAwesome
Submit an Issue
Hi,
Thanks for helping us improve the Predix UI platform by submitting an issue.
Expected behavior and actual behavior:
Looking at the APIs for px-map-marker-symbol (part of px-map) and the only way to feed in input is fa- where is substituted for one of the several icons in FontAwesome. It is not clear how to feed in own SVGs from a local copy
Thanks for this report. We looked into this a bit to figure out the best approach.
Our early thoughts were allowing developers to pick the icon CSS class and define new icons by creating a new CSS class. But we decided that wouldn't be feasible because (1) there's no good way for developers to get their new icon CSS class definition into the shadow root of the px-map component, and (2) we're looking to move away from webfonts for icons in the long term in favor of SVGs.
I think we've landed on changing the way <px-map-marker-symbol> works by allowing the developer to define and use their own SVGs. Two ways we could approach this:
Expose a new attribute that accepts a string of SVG markup, and render that markup in the marker (e.g. <px-map-marker-symbol svg="<svg>...</svg>"></px-map-marker-symbol)
Follow the path of px-polymer-font-awesome and use iron-iconsets defined by the developer, or default iconsets defined by the component, to render SVG in the marker. This option would be a bit more work for our team to implement, and a bit more work for the developer, but has the added benefit of allowing developers to use their defined SVG icons in a marker or to use then in <iron-icon> tags elsewhere in the app.
My initial estimate is we can release a minor update within 1-2 weeks to accomplish this, but we'll comment back here with updates if the timeline changes. We'd also accept a contribution (pull request) that implemented this functionality.
cc @Menkhaus-ge @talimarcus
Hi @icicle167,
As of v2.0.0, px-map has been updated to take svg icons (rather than an icon font, such as Font Awesome) everywhere icons are used:
px-map-marker-symbol
px-map-marker-group
px-map-control-zoom
px-map-control-locate
|
gharchive/issue
| 2017-05-04T19:04:47 |
2025-04-01T04:55:32.322963
|
{
"authors": [
"davidrleonard",
"icicle167",
"talimarcus"
],
"repo": "PredixDev/px-map",
"url": "https://github.com/PredixDev/px-map/issues/88",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
760309326
|
Fixing the bug that appeared when anonymize customer last name option has been added.
Questions
Answers
Description?
Fixing the bug that appeared when anonymize customer last name option has been added.
Type?
bug fix
BC breaks?
no
Deprecations?
no
Fixed ticket?
Fixes PrestaShop/Prestashop#22222.
How to test?
Issue describes it enough.
CSFixer is red, please run it :)
@PierreRambaud I did it against latest dev branch with your CS fixer changes.
@Oksydan I don't know how you rebased your PR but something's wrong 🤔, you shouldn't see someone else's commits on your PR.
@atomiix
branch fixed
Hi @Oksydan,
It is ok :heavy_check_mark:
Thanks!
Thank you @Oksydan
|
gharchive/pull-request
| 2020-12-09T13:01:11 |
2025-04-01T04:55:32.438920
|
{
"authors": [
"NeOMakinG",
"Oksydan",
"PierreRambaud",
"atomiix",
"khouloudbelguith"
],
"repo": "PrestaShop/productcomments",
"url": "https://github.com/PrestaShop/productcomments/pull/89",
"license": "AFL-3.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
392101553
|
Merge pull request #11 from PrestaShop/dev
error 500 when enable the module.
just change in the ps_viewedproduct.php at the start of the file the line
use PrestaShop\PrestaShop\Adapter\Presenter\Product\ProductListingPresenter;
to
use PrestaShop\PrestaShop\Core\Product\ProductListingPresenter;
Thanks to @Matt75 we have a Pull Request that follows these guidelines 😉 : https://github.com/PrestaShop/ps_viewedproduct/pull/13
So I close this one
|
gharchive/pull-request
| 2018-12-18T10:56:56 |
2025-04-01T04:55:32.441196
|
{
"authors": [
"matks",
"shacker2"
],
"repo": "PrestaShop/ps_viewedproduct",
"url": "https://github.com/PrestaShop/ps_viewedproduct/pull/12",
"license": "AFL-3.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2212172758
|
[Request] Add Labels on Endscreen
Adds Labels on Endscreen
Examples.
Noclip Deaths/Accuracy, From%, Best runs.
Pls, This is cool.
Nice idea
|
gharchive/issue
| 2024-03-28T01:33:46 |
2025-04-01T04:55:32.445877
|
{
"authors": [
"LouemPH",
"PotatoTriesIt"
],
"repo": "Prevter/OpenHack",
"url": "https://github.com/Prevter/OpenHack/issues/139",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
433315307
|
As an admin, I want changes made to books, cards, and people in the admin interface to automatically update the public search, so that content in the search and admin interface stay in sync.
Dev notes
This will require signal handling to be implemented for Django in parasolr.
duplicates #339
|
gharchive/issue
| 2019-04-15T14:31:29 |
2025-04-01T04:55:32.452043
|
{
"authors": [
"bwhicks",
"thatbudakguy"
],
"repo": "Princeton-CDH/mep-django",
"url": "https://github.com/Princeton-CDH/mep-django/issues/272",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
264735442
|
As a global admin, I want addresses automatically associated with accounts in the logbook where possible to minimize the amount of manual cleanup required.
Encoder will reference personography to determine which account addresses are NOT the account-holder's address.
@elspethgreen Ben filled me in on the logic you all discussed for associating addresses with accounts on import. I'm converting this issue to handle address logic on import (as an extension of #9 ) and have asked @bwhicks to document the logic he intends to implement and what pieces of information will nee to be cleaned up manually (e.g., care of addresses, people with multiple addresses). If I understand correctly, it sounds like this should handle the majority of simple cases and leave the complicated ones to be sorted out manually by the team after import.
@elspethgreen after discussing a bit more with @bwhicks, I'm wondering if we should hold off on any logic to automatically associate addresses with accounts until we import the lending cards. What do you think?
If it's easier to do it on the next import, that's fine! But when encoders start actually working with the system, we'd want those addresses automatically associated.
@elspethgreen we're going to push this back to the next release. It requires some conversation among me, @bwhicks and @jabauer to figure out how to model addresses properly in the database (we'll probably be back to the team with questions to refine this soon).
We absolutely will do what we can to clean those up automatically before you do any manual clean up or association. It may actually be a good thing to look at this as we're working on the card import in the next stage.
@clmahoney @i-davis could you play with this a bit? I've taken a look and I think it's working as described in Rebecca's testing notes, but since it's complicated I'd like some other eyes on it.
I think all this looks good to me! I should say I'm a bit lost by testing notes, generally, & this last batch has felt esp a bit beyond me; but I think I get this one!
|
gharchive/issue
| 2017-10-11T21:04:50 |
2025-04-01T04:55:32.456298
|
{
"authors": [
"elspethgreen",
"i-davis",
"rlskoeser"
],
"repo": "Princeton-CDH/mep-django",
"url": "https://github.com/Princeton-CDH/mep-django/issues/44",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
345687483
|
生成成功后GIF为空GIF
`[root@VM_0_13_centos ~]# ffmpeg
ffmpeg version 3.4.2 Copyright (c) 2000-2018 the FFmpeg developers
built with gcc 4.4.7 (GCC) 20120313 (Red Hat 4.4.7-23)
configuration:
libavutil 55. 78.100 / 55. 78.100
libavcodec 57.107.100 / 57.107.100
libavformat 57. 83.100 / 57. 83.100
libavdevice 57. 10.100 / 57. 10.100
libavfilter 6.107.100 / 6.107.100
libswscale 4. 8.100 / 4. 8.100
libswresample 2. 9.100 / 2. 9.100
Hyper fast Audio and Video encoder
usage: ffmpeg [options] [[infile options] -i infile]... {[outfile options] outfile}...
Use -h to get full help or, even better, run 'man ffmpeg'
`
以上是检查FFMPEG时的代码,我以前装了最新版(v4.0)但是不行,又装了教程里的版本还是不行,php.ini的system函数也开了。网站:www.make233.cn,求大佬帮忙解决,Q:2403393945,谢谢了!
使用宝塔面板 PHP5.6 求帮忙解决
你试试这个新的 https://github.com/PrintNow/Make-Funny-Image ,需要开启 exec 函数
好的,我试试。
|
gharchive/issue
| 2018-07-30T09:50:49 |
2025-04-01T04:55:32.468140
|
{
"authors": [
"PrintNow",
"Redbeanw44602"
],
"repo": "PrintNow/php-sorry-gif",
"url": "https://github.com/PrintNow/php-sorry-gif/issues/18",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
381638511
|
"language-git" don't show any coloration on my navigator (FireFox & Chrome).
Hi,
I use prismjs with « Okaidia » theme,
pro coloring code of my learning « HTML » page's of web‑dev note's.
And strangely even if I have downloaded the version on the official page,
with all language's & some extra feature's,
"language-git" doesn't show any coloration, except pro :
comment's starting with « # ??? » in dark‑grey,
phrase starting with « - » shown in magenta,
string using « " ??? " » or « ' ??? ' » in green.
Can you help please ?
This is because the git uses mostly non-standard token names which are not supported by any Prism themes.
You can edit the Okaidia theme to highlight the different parts of git commands and their outputs to your liking. You can find all tokens names in the language definition along with comments explaining every single one of them.
Thank pro your help,
but it is strange that "prismjs" propose support of this language without any result ?
Well, there are results for comment, string, and parameter.
The rest of the language is for the outputs of git commands which are pretty hard to fit into the current style system of Prism themes.
So leaving the choice of color to the user doesn't seem like a bad idea IMO.
You are right in that it is a little confusing that the themes of Prism do not cover/highlight every (little) feature Prism supports. But as these themes are intended to be modified by the user, I think that it is alright for them to be minimalistic.
I'm not very good in JavaScript (beginner) to start to configure by my self this language-git "js" & "css", I'm still happy to get good result with HTML, CSS, PHP & JS language, but really I don't comprehend why each user's should make this modification by himself, it's such a waste of time, if we addition all the time we use all of us to make them. It would be surely better if the standard commands had directly specific syntax colors according to theme chose pro install.
It would be surely better if the standard commands had directly specific syntax colors according to theme chosen pro install.
That's true but doing so for every little feature of every theme would blow up the file size of the themes and it makes them hard to maintain. So for simplicity's sake, only the most common features are supported by Prism's themes.
I comprehend the problem of file size, but « prism » could be proposed in two version light & heavy, at least we will have the choice…
|
gharchive/issue
| 2018-11-16T15:10:54 |
2025-04-01T04:55:32.474836
|
{
"authors": [
"Nemzag",
"RunDevelopment"
],
"repo": "PrismJS/prism",
"url": "https://github.com/PrismJS/prism/issues/1615",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1930612741
|
Update Minecraft wiki link to new domain
The Minecraft Fandom wiki has been forked to a new domain: minecraft.wiki. Learn more here: https://minecraft.wiki/w/Minecraft_Wiki:Moving_from_Fandom. This PR updates the old wiki link accordingly.
Thanks!
|
gharchive/pull-request
| 2023-10-06T16:52:16 |
2025-04-01T04:55:32.481367
|
{
"authors": [
"Spongecade",
"extremeheat"
],
"repo": "PrismarineJS/bedrock-protocol",
"url": "https://github.com/PrismarineJS/bedrock-protocol/pull/455",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1338302228
|
Added several glass/tranparent navbar effect
added three different transparent navbar effect, choose the one that is most suitable
Hey @hassana123 can you make this boxes go away ?? or do I have to tell you which code you should remove ?
on it
Hey @hassana123 can you make it more like this ?
Don't change the color. Instead use this piece of code
.navbar {
position: fixed;
top: 0;
background: rgba(0,0,0,0.3);
backdrop-filter: saturate(180%) blur(10px);
left: 0;
display: flex;
width: 100%;
justify-content: space-between;
padding: 10px;
z-index: 10;
color: whitesmoke;
}
done
Thank you @hassana123 🙌
make sure to check other issues
Sure
|
gharchive/pull-request
| 2022-08-14T17:05:34 |
2025-04-01T04:55:32.486301
|
{
"authors": [
"PritamSarbajna",
"hassana123"
],
"repo": "PritamSarbajna/tourism-website",
"url": "https://github.com/PritamSarbajna/tourism-website/pull/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2610328831
|
Update security.md
As per the issue #1441
Fixes: #1441
Description
I changes the file with thw symbols and tried to make it interactive. So that it becomes easy to read it fastly and easily.
If any further changes required please let me know.
Type of PR
[ ] Bug fix
[ ] Feature enhancement
[ x] Documentation update
[ ] Other (specify): _______________
Screenshots / videos (if applicable)
Checklist:
[ x] I have made this change from my own.
[ x] I have taken help from some online resources.
[ x] My code follows the style guidelines of this project.
[x ] I have performed a self-review of my own code.
[ x] I have commented my code, particularly in hard-to-understand areas.
[ x] I have made corresponding changes to the documentation.
[ x] My changes generate no new warnings.
[ x] I have tested the changes thoroughly before submitting this pull request.
[ x] I have provided relevant issue numbers and screenshots after making the changes.
If possible please add the labels of a level 2 or level 3 , it will be a great help from your side.
Thank you for giving me an opportunity.
|
gharchive/pull-request
| 2024-10-24T03:49:18 |
2025-04-01T04:55:32.493742
|
{
"authors": [
"Nikhil0-3"
],
"repo": "PriyaGhosal/BuddyTrail",
"url": "https://github.com/PriyaGhosal/BuddyTrail/pull/1442",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2584598170
|
Change the color of text for better visible
Related Issue
Feat:Text Not Visible #640
Fixes: #issue no.640
Description
The current text is not visible properly. I've changed the font-size, weight, and color. So that it can properly visible.
Issue no: #640
Type of PR
[ ] Feature enhancement
Screenshots / videos (if applicable)
Before:
After:
Checklist:
[ x] I have made this change from my own.
[ x] My code follows the style guidelines of this project.
[x ] I have performed a self-review of my own code.
[ x] My changes generate no new warnings.
[ x] I have tested the changes thoroughly before submitting this pull request.
[x ] I have provided relevant issue numbers and screenshots after making the changes.
@PriyaGhosal @samarth-6 Please review my code and also add label of hacktoberfest'24, hacktoberfest-accepted, level-2.
@Sudhanshu248
Your pull request has been successfully merged! Don't stop here—continue contributing by addressing more issues and aiming to become a top contributor. Exclusive swag awaits for those who go the extra mile!
Great contribution to GSSOC'24 EXTENDED! 🎉
|
gharchive/pull-request
| 2024-10-14T02:54:30 |
2025-04-01T04:55:32.498790
|
{
"authors": [
"PriyaGhosal",
"Sudhanshu248"
],
"repo": "PriyaGhosal/BuddyTrail",
"url": "https://github.com/PriyaGhosal/BuddyTrail/pull/642",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1200470111
|
Cometchat stuck on loading
Downloaded the clone and it runs without issue the only problem I am seeing is with CometChat.
When I click the chat with seller button
it just hangs on this loading screen. I have verified that the .env file is setup correct with the IDs
Hi @Mattw11486. Could you please share your console output?
There is probably additional information there on what seems to be wrong when loading CometChat
I am a noob when it comes to this but in Terminal on Visual Studio code it just shows a few warnings about links not working which is fine. Nothing about cometchat. Output is blank.
No problem @Mattw11486. The console I was looking for is the browser console.
So on Chrome (if that is what you are using) you can press F12 and a new window will appear with several tabs. One of which is the browser console. This is where the application will show runtime errors.
Thank you
Looks like that was the key to my fix. I realize cometchat 2.0 is installed on the project while on my dashboard in cometchat I had to select the 2.0 version. Hopefully that won't be an issue later if cometchat forces to update version? Anyway it seems to be working now!
|
gharchive/issue
| 2022-04-11T20:02:36 |
2025-04-01T04:55:32.503185
|
{
"authors": [
"Mattw11486",
"igorasilveira"
],
"repo": "PriyankaGurnaniCometChat/marketplace-cometchat",
"url": "https://github.com/PriyankaGurnaniCometChat/marketplace-cometchat/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
327914886
|
ForceDelete model
$storedFile = Storedfile::findOrFail($id);
$storedFile->forceDelete();
Outputs error:
Call to a member function toArray() on array
exception | Symfony\Component\Debug\Exception\FatalThrowableError
file | \vendor\proai\eloquent-versioning\src\BuilderTrait.php
I have the same issue with normal delete:
$storedFile = Storedfile::findOrFail($id);
$storedFile->delete();
The fix from netcom does not help.
Are there some news about this?
I'm using a custom BuilderTrait (to patch few bugs). Try use this in your case @10bias :
/**
* Get affected records.
*
* @return array
*/
protected function getAffectedRecords()
{
// model only
if ($this->model->getKey()) {
$records = [$this->model];
}
// mass assignment
else {
$records = $this->query->get()->all();
}
return $records;
}
/**
* Run the default delete function on the builder.
*
* @return mixed
*/
public function forceDelete()
{
// get records
$affectedRecords = $this->getAffectedRecords();
$ids = array_map(function($record) {
return $record->{$this->model->getKeyName()};
}, $affectedRecords);
// delete main table records
if (! $this->query->delete()) {
return false;
}
// delete version table records
$db = $this->model->getConnection();
return $db->table($this->model->getVersionTable())
->whereIn($this->model->getVersionKeyName(), $ids)
->delete();
}
|
gharchive/issue
| 2018-05-30T22:00:47 |
2025-04-01T04:55:32.526173
|
{
"authors": [
"10bias",
"AceWT",
"Mortred974"
],
"repo": "ProAI/eloquent-versioning",
"url": "https://github.com/ProAI/eloquent-versioning/issues/21",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
822535945
|
Adding the Growth Chamber MK2
It's an improved version of the current growth chamber.
It yields 4x the amounts of the growth chamber while using only 2x as much power.
It has 9 input slots and 27 output slots.
As you see I made an entire new "AMachineMK2" file, the reason for that is because I wanted the new machine to have more input and output slots but didn't see another way to do it without making a new file and changing the registered slots. I also made the "GrowthChamberMK2" file because I also didn't see a way to change the amounts it produces without affecting the already existing machine. I haven't tested it yet to see that nothing breaks, so if you could take a quick peak at the code and tell me if you see anything that could easily break I'd thank that a lot.
Please use the Antigravity Bubble as a reference make a new ui instead of making a seperate abstract machine.
AMachineMK2 is not needed at all. That is currently that I all see
Got it
Why did you close it?
You an just add a commit
I'm making the file, will reopen when done
oh ok 👍
|
gharchive/pull-request
| 2021-03-04T21:48:19 |
2025-04-01T04:55:32.539264
|
{
"authors": [
"ProfElements",
"waleks647"
],
"repo": "ProfElements/DynaTech",
"url": "https://github.com/ProfElements/DynaTech/pull/33",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1406501439
|
Eprice555 patch 1
@manishapriya94 testing
@manishapriya94 testing 2
@manishapriya94 Testing again.
|
gharchive/pull-request
| 2022-10-12T16:33:34 |
2025-04-01T04:55:32.549036
|
{
"authors": [
"eprice555",
"paramsiddharth"
],
"repo": "ProgramEquity/amplify",
"url": "https://github.com/ProgramEquity/amplify/pull/335",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2069105624
|
Password Strength helper
Existing helper class created, PasswordStrength, which should use configuration to determine minimum password strength and return friendly results as required.
This is handled by ASP.NET Identity
|
gharchive/issue
| 2024-01-07T12:11:54 |
2025-04-01T04:55:32.550198
|
{
"authors": [
"ProgramX-NPledger"
],
"repo": "ProgramX-NPledger/Taggloo4",
"url": "https://github.com/ProgramX-NPledger/Taggloo4/issues/36",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1496342234
|
[API] Funktionsschnittstelle
Über die DSL werden die Entitäten angelegt, also mit Components versehen.
Components werden ggf. auch über die DSL konfiguriert (Parameter).
Interessanter wird der Fall, wenn Verhaltensweisen in der DSL definiert werden.
Anmerkung: In der DSL werden die Funktionen nicht richtig selbst programmiert, sondern eher aus vorhandenen zusammengesteckt.
Ein simples Beispiel, wäre eine Falle.
Eine Falle ist eine Entity mit einem PositionComponent, AnimationComponent und HitBoxComponent (so grob).
Jetzt muss der Falle aber noch mitgeteilt werden (über die DSL), was sie bei einer Kollision macht.
Ganz Stumpf:
onCollision: hitPlayer(5); deactivate();
hitPlayer und deactivate wären über die API vordefinierte Funktionen.
Die Frage ist mehr: Wo kommt onCollision hin. Für mich logisch wäre im HitBoxComponent aber das sind eigentlich keine Daten. Aber in ein System wäre das zu übergreifend, oder?
Außerdem: Woher weiß die Falle, dass sie mit dem Player kollidiert (aber vermutlich eher was für #87)
Dieses Problem ist repräsentativ für weitere Probleme. Wir wollen Teilweise noch komplexere Verhalten definieren (vgl #91)
@malt-r Habe ich das so richtig/gut beschrieben?
Yes, passt passt aus meiner Perspektive gut :)
Du hast Recht, dass das onCollision nicht zu den Daten gehört, allerdings müssen die Components ja irgendwie in der Lage sein, sich mitzuteilen. "Sauberer" (also mehr dem Gedanken ECS folgend) wäre es sicherlich, das über ein System zu regeln. Allerdings wäre meine Einschätzung, dass das eine gute pragmatische Aufweichung vom ECS Gedanken ist, wenn wir neben reinen Daten auch Callback-Funktionen in den Components erlauben.
Allerdings steht man dann vor der Herausforderung, dass man ja in der DSL irgendeine "Reaktion" in die Callback-Funktion einhängt und da dann wieder andere Komponenten des gleichen GameObjects manipulieren will, ggfs. über Methodenaufrufe. Ob die Components dann die Logik dieser Methoden selbst implementieren oder die Implementierung in ein System ausgelagert werden sollte, kann ich grade irgendwie nicht gut abschätzen.
Ein Beispiel wäre hier hitPlayer(5) als Reaktion auf onCollision: wird da dann direkt eine Methode aufgerufen, die den health-Wert in einer Komponente im Player abzieht oder merkt sich das die Komponente bis zum nächsten Update Zyklus und lässt dann den health-Wert durch ein System abziehen?
#Time: 16 WiMi Std.?
@malt-r Könnte ungefähr hinkommen. Aber mit Tests etc... tendiere zu 24
|
gharchive/issue
| 2022-12-14T10:54:45 |
2025-04-01T04:55:32.556117
|
{
"authors": [
"AMatutat",
"malt-r"
],
"repo": "Programmiermethoden/Dungeon",
"url": "https://github.com/Programmiermethoden/Dungeon/issues/97",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2478268374
|
Update luna detection
Try to use ApplyTransformToPointsd to update luna detection pipeline
Checks
[ ] Avoid including large-size files in the PR.
[ ] Clean up long text outputs from code cells in the notebook.
[ ] For security purposes, please check the contents and remove any sensitive info such as user names and private key.
[ ] Ensure (1) hyperlinks and markdown anchors are working (2) use relative paths for tutorial repo files (3) put figure and graphs in the ./figure folder
[ ] Notebook runs automatically ./runner.sh -t <path to .ipynb file>
I didn't modify the generate_detection_val_transform and generate_detection_inference_transform because it would offer no significant benefit. Making these changes would require converting boxes to points, applying the transform, and then converting the points back to boxes.
Looks good to me overall. Hi @Can-Zhao , we introduced some early features about point operations. Would you mind taking a look at this PR? Thanks!
|
gharchive/pull-request
| 2024-08-21T15:00:05 |
2025-04-01T04:55:32.574359
|
{
"authors": [
"KumoLiu",
"mingxin-zheng"
],
"repo": "Project-MONAI/tutorials",
"url": "https://github.com/Project-MONAI/tutorials/pull/1795",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1227112083
|
02 05 authorization
Добавлена папка приложения Comment, изменен дизайн публичной части сайта
Рейтинг
|
gharchive/pull-request
| 2022-05-05T19:51:30 |
2025-04-01T04:55:32.600803
|
{
"authors": [
"dr2moscow"
],
"repo": "Project-room2/habr",
"url": "https://github.com/Project-room2/habr/pull/22",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
}
|
1163252001
|
Primeira issue do projeto
Neste ticket, todos vocês vão ser notificados
Salve salve galerinha do zap
OK, TEAtcher
Boa noite !
Obrigada pela notificação! Grata!
ótima tarefa, irei iniciar em breve
hey
conta Catarina!!
chore não rei delax
bom dia
|
gharchive/issue
| 2022-03-08T22:57:55 |
2025-04-01T04:55:32.648474
|
{
"authors": [
"Athsxt",
"Emillysouzz",
"LUKASRIB15",
"MatheuzsFer",
"Tiago-Ramon",
"felipebastos",
"vitoriajuli"
],
"repo": "ProjetoIntegrador2021-2/backend",
"url": "https://github.com/ProjetoIntegrador2021-2/backend/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1260615282
|
export NodeViewConstructor and MarkViewConstructor
Would be nice to expose these types for libraries that build abstractions on top of these two.
We'll also need doc comments on these if they are going to appear in the reference docs.
|
gharchive/pull-request
| 2022-06-04T04:00:52 |
2025-04-01T04:55:32.680348
|
{
"authors": [
"BrianHung",
"marijnh"
],
"repo": "ProseMirror/prosemirror-view",
"url": "https://github.com/ProseMirror/prosemirror-view/pull/127",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1923185887
|
片源可播,能保障多久呢?
片源可播,能持续修复吗?
怕上个站,就挂了
正常情况下应该不至于挂,毕竟选的几个采集站都还算稳定长久,而且播放源也是一对多,同时挂的可能性很小,就算挂了重新换主站从零采集也才3-10分钟左右,时间也不算很长
|
gharchive/issue
| 2023-10-03T02:52:18 |
2025-04-01T04:55:32.700387
|
{
"authors": [
"ProudMuBai",
"yoorxee"
],
"repo": "ProudMuBai/GoFilm",
"url": "https://github.com/ProudMuBai/GoFilm/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1709149715
|
🛑 Service - Help Center is down
In bb80639, Service - Help Center (https://help.proxeuse.com) was down:
HTTP code: 500
Response time: 10588 ms
Resolved: Service - Help Center is back up in 84b802e.
|
gharchive/issue
| 2023-05-15T01:10:50 |
2025-04-01T04:55:32.707653
|
{
"authors": [
"roelreijneveld"
],
"repo": "Proxeuse/status",
"url": "https://github.com/Proxeuse/status/issues/142",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1712948393
|
🛑 Service - Help Center is down
In f194763, Service - Help Center (https://help.proxeuse.com) was down:
HTTP code: 500
Response time: 5435 ms
Resolved: Service - Help Center is back up in 726f176.
|
gharchive/issue
| 2023-05-17T01:24:33 |
2025-04-01T04:55:32.709987
|
{
"authors": [
"roelreijneveld"
],
"repo": "Proxeuse/status",
"url": "https://github.com/Proxeuse/status/issues/149",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2524532255
|
Pub 2941 fixing missing logging of invalid bids
Ticket: PUB-2941
This ticket addresses an issue in which an auction containing only invalid bids is not recorded.
Invalid bids are mapped to bidtype 1. Timed-out bids are mapped to bidtype 3. If any auction contains only bidtypes 1 or 3, it is now sent to pubx (assuming the relevant downsampling permits this).
There was a query as to whether or not auctions. with no bids at all were recorded. using the javascript array.prototype.every function on an empty array returns true, so this condition is satisfied. An additional test already verified this in prebid, but the test was quite obscure, so i tried to tidy it up a little.
Additional tests were added to confirm that auctions with timed out bids or invalid bids (as far as i can tell this only occurs due to pricing issues) are also recorded.
@nathan-pubx It's showing all other non-related commits in this PR. Could you please rebase the upstream master before adding this commit to avoid duplicate commits?
@nathan-pubx It's showing all other non-related commits in this PR. Could you please rebase the upstream master before adding this commit to avoid duplicate commits?
Sorry @pnhegde , that's done now. The branch was based on the prebid.js master branch, but now is based on the pubx master
|
gharchive/pull-request
| 2024-09-13T10:49:56 |
2025-04-01T04:55:32.734032
|
{
"authors": [
"nathan-pubx",
"pnhegde"
],
"repo": "Pubx-ai/Prebid.js",
"url": "https://github.com/Pubx-ai/Prebid.js/pull/26",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2167905196
|
Eva Weekly Notes -- 2024-03-04
RA Projects
AwardFindR / award-pynder
5/16 data sources ported to Python
handed off to Nic + Isaac
Research Software
EAGER / Software Sustainability
[ ] Create dataset and basic EDA for "follow up sample" -- specifically:
Select only CISE and BIO from 2013 - 2023 (ask Nic if he wants to 2023 because need grants that have concluded)
Remove no code and survey responses from the set
Remove EAGER, Workshop, REU, and coop grants
[ ] Build on top of this dataset by:
Retrieve project reports
Identify Publications and Software from those reports
Put these in a dataset linked by award (award-software and award-publication)
Find multiple publications under the same award using WoS
[ ] Come up with clear questions to answer using the existing survey data
[x] Timeline plot / think through other ways to visualize "typical development patterns as they relate to grant funding timelines"
RS-Graph / Database for RS Research
[ ] Review and talk through planning doc with Nic
[x] Put together a plan for Akhil and Terra to work on the CRediT work
[ ] Ask Isaac for help writing a Llama/Mistral/etc. fine-tuning script for the Dev - Author annotation data
[x] ICSSI: minor plan for abstract proposal
[ ] Write WoS script for getting papers who mention software, tool, pipeline, or package in title from set of journals
Civic
Public Comment Segmentation
[ ] Write evaluation script for application of model for determining the segmentation quality (i.e. are we missing parts before, after, etc.)
[ ] Train and eval models for Seattle and Oakland
[ ] Choose next city council for data annotation and prep data
ML for PIT
No progress.
PhD Activities
Writing
[ ] Review and contribute to the Gig Driver Deactivation paper
[ ] Review and contribute to the CZI Software Dependency Graph paper
Reading
Read this week:
https://doi.org/10.1007/s11192-023-04806-2
Plan to read this week:
https://www.semanticscholar.org/paper/Ecosystem-level-determinants-of-sustained-activity-Valiev-Vasilescu/e8eb4dc83c7f74848c331005aa7c124dc13cf6a9?utm_source=direct_link
https://www.semanticscholar.org/paper/Beyond-the-Repository-Casari-Ferraioli/21dd9f046fc807a498840e19f6c81e69ffad8651?utm_source=direct_link
https://www.semanticscholar.org/paper/Fairness-and-Abstraction-in-Sociotechnical-Systems-Selbst-Boyd/9dae5041ed97ab21f637fb282d758fd9cb771476?utm_source=direct_link
https://doi.org/10.1162/qss_a_00275
https://doi.org/10.1162/qss_a_00277
https://doi.org/10.1162/qss_a_00264
https://doi.org/10.1162/qss_a_00266
Specific tasks 2024-03-07:
finish constructing dataset for ankil and terra
put together script for evaluating whole public comment segmentation model
choose next city council and prep data for annotation
keep working on icssi
Specific tasks 2024-03-08
put together script for evaluating whole public comment segmentation model (train fine-tune-mpnet-three-sentences for real and then apply)
read two papers from above list
start JOSS review
choose next city council and prep data
check in on nic for new email dump
check in on akhil and terra for practice annotations
|
gharchive/issue
| 2024-03-04T22:31:43 |
2025-04-01T04:55:32.750912
|
{
"authors": [
"evamaxfield"
],
"repo": "PugetSoundClinic-PIT/ProjectTracking",
"url": "https://github.com/PugetSoundClinic-PIT/ProjectTracking/issues/134",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2743355733
|
Feature request: skipmissing keyword
Now it's hard to implement skipmissing
listingtable(
df,
:Conc => "Concentration (ng/mL)",
rows = [:Formulation => "Formulation", :Subject => "ID"],
cols = :Time => "Time (hr)",
summarize_rows = :Formulation => [length => "N",
Pair(x->mean(skipmissing(x)), "Mean"),
Pair(x->std(skipmissing(x)),"SD"),
Pair(x->median(skipmissing(x)) , "Median"),
Pair(x->maximum(skipmissing(x)) , "Maximum"),
Pair(x->minimum(skipmissing(x)) , "Minimum")]
)
also it lead to error when only missing data found in group.
You can use f ∘ skipmissing (written \circ + tab) for a short version of what you're doing, but if you want robustness against all-missing data, you can define a helper function like
function robust_skipmissing(f)
return function(vec)
all(ismissing, vec) && return Annotated("NC", "NC - Not computable", label = nothing)
return (f ∘ skipmissing)(vec)
end
end
data = DataFrame(
concentration = [missing, 4.5, 2.0, 1.5, 0.1, missing, 3.2, 1.8, 1.2, 0.2],
id = repeat([1, 2], inner = 5),
time = repeat([0, 0.5, 1, 2, 3], 2)
)
listingtable(
data,
:concentration => "Concentration (ng/mL)",
rows = :id,
cols = :time => "Time (hr)",
summarize_rows = [
length => "N",
robust_skipmissing(mean) => "Mean",
robust_skipmissing(std) => "SD",
robust_skipmissing(median) => "Median",
robust_skipmissing(maximum) => "Maximum",
robust_skipmissing(minimum) => "Minimum",
]
)
|
gharchive/issue
| 2024-12-16T20:39:50 |
2025-04-01T04:55:32.761014
|
{
"authors": [
"PharmCat",
"jkrumbiegel"
],
"repo": "PumasAI/SummaryTables.jl",
"url": "https://github.com/PumasAI/SummaryTables.jl/issues/61",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2615621901
|
added terms and conditions
@Puskar-Roy
I have added the terms and conditions page and made it responsive for the mobile view as well
Sorry I have deleted the previous PR by mistake, that why I created this new PR
And also in the previous PR, you have added all the labels and considered it as level3, please dont forget to add the labels here
The issue number is #123
Kindly check it out
https://github.com/user-attachments/assets/82c5652e-ae83-4c63-a5aa-78d1b4b7f9a9
Hello @AswaniBolisetti, thank you for your pull request! We will review it shortly. 😇❣️
Your contributions help us grow and improve our project. Please make sure you have read and followed our CONTRIBUTING GUIDELINES.
Your efforts are appreciated, and we value your participation in our community. Don't forget to ⭐ our repository!
Happy coding!✨
|
gharchive/pull-request
| 2024-10-26T07:23:27 |
2025-04-01T04:55:32.774291
|
{
"authors": [
"AswaniBolisetti",
"Puskar-Roy"
],
"repo": "Puskar-Roy/create-my-api",
"url": "https://github.com/Puskar-Roy/create-my-api/pull/147",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2110651131
|
Closes #33 make the black format workflow apply to all code
went for pre-commit
Hmm, I was hoping for fewer complaints. One step closer ...
|
gharchive/pull-request
| 2024-01-31T17:49:47 |
2025-04-01T04:55:32.776727
|
{
"authors": [
"JohannesFinsveen"
],
"repo": "PxTools/PxBuild",
"url": "https://github.com/PxTools/PxBuild/pull/73",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1053643019
|
🛑 Overseerr is down
In d64d788, Overseerr (https://request.pxlbuzzard.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Overseerr is back up in ba8ea2c.
|
gharchive/issue
| 2021-11-15T13:04:34 |
2025-04-01T04:55:32.779003
|
{
"authors": [
"PxlBuzzard"
],
"repo": "PxlBuzzard/upptime",
"url": "https://github.com/PxlBuzzard/upptime/issues/134",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1434040668
|
🛑 ArchiveBox is down
In 769c389, ArchiveBox (https://archivebox.pxlbuzzard.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: ArchiveBox is back up in bd9a55a.
|
gharchive/issue
| 2022-11-03T03:10:33 |
2025-04-01T04:55:32.781312
|
{
"authors": [
"PxlBuzzard"
],
"repo": "PxlBuzzard/upptime",
"url": "https://github.com/PxlBuzzard/upptime/issues/730",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.