id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
178360561
|
java.lang.ClassNotFoundException: org.postgresql.Driver
使用postgresql,无法运行
感谢你的使用与反馈,目前bin中的jar仅import了mariadb的驱动jar。
所以默认支持mariadb与mysql, 当在使用其它类型的数据库运行时,需要import相应的数据库驱动jar。
主要操作方法有两个:
1、fork源码在pom.xml中增加数据库驱动依赖,重新打包再以jar -jar GGCode-xxx.jar运行
2、将驱动包下载到GGCode-0.0.1.jar同级目录,使用java -cp GGCode-0.0.1.jar:[driver.jar] App运行(windows则使用java -cp GGCode-0.0.1.jar;[driver.jar] App运行)
|
gharchive/issue
| 2016-09-21T14:18:24 |
2025-04-01T06:40:29.868671
|
{
"authors": [
"printfcoder",
"stotem"
],
"repo": "stotem/GGCode",
"url": "https://github.com/stotem/GGCode/issues/1",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
125920771
|
and another package version not initialised
before a repository can be released, all package.xml need to have the same version or prepare-release is doomed to fail :-(
I'll merge it myself and re-release afterwards
now the re-release worked: https://lcas.lincoln.ac.uk/jenkins/job/prepare-release/384/
|
gharchive/pull-request
| 2016-01-11T11:09:23 |
2025-04-01T06:40:29.870729
|
{
"authors": [
"marc-hanheide"
],
"repo": "strands-project/v4r_ros_wrappers",
"url": "https://github.com/strands-project/v4r_ros_wrappers/pull/24",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2316551808
|
StatefulJob can't easily tar node data
The StatefulJob docs state this, and I'm attempting to use it for the same function.
Strangelove uses it to compress and upload snapshots of chain data.
I'm having a problem achieving this though. On the Provenance chain our nonpruned nodes contain about 1TB of data. Our PVCs are setup for 1.25TB and to grow when they are at 90% used. With about 25% space overhead, there's not enough space to compress the data and store the tar.gz on the same volume.
The two ways I could think that the StatefulJob could support this would be the following:
Allow configuration to specify additional PV/PVCs that are created/cleaned up
Allow setting additional snapshot PV size. Then the PV is restored from the snapshot, it is then edited to increase its size further based on this config
I can try tackling this if we settle on the solution.
We currently handle this by doing a streamed compress and upload so that storage is not necessary for the compressed file prior to upload.
For resumable uploads though, it would be great to have this feature so that the file is compressed once, and retries can be retried.
We could add an additional parameter that would allow creating the StatefulJobs PVC with something like twice the size so that additional room was available for these kinds of operations.
|
gharchive/issue
| 2024-05-25T01:30:57 |
2025-04-01T06:40:29.873610
|
{
"authors": [
"agouin",
"scirner22"
],
"repo": "strangelove-ventures/cosmos-operator",
"url": "https://github.com/strangelove-ventures/cosmos-operator/issues/420",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2490451932
|
fix alpine version crahes for cosmwasm based networks
reference: https://github.com/CosmWasm/wasmvm/issues/523
Currently Osmosis builds cause a cgo panic. Also affected wormhole w/ testing for Joel.
I have found a solution to this problem if necessary CosmWasm/wasmvm#576
I think we should submit a PR on the Heighliner project to systematically use an Alpine 3.18 image, at least for projects requiring cosmwasmvm.
@vimystic @0xPuncker
Considering generi-sizing all of heighliner to have choice of alpine base and install golang version from the mod file of the respective chains .Similar to what you suggest but for everything.
Alternatively , if the desired golang alpine combo does not exist , we then pass in an ARG to the docker file to do the needful
Will run it by people internally and keep you posted on this issue.
|
gharchive/issue
| 2024-08-27T22:06:06 |
2025-04-01T06:40:29.876715
|
{
"authors": [
"Reecepbcups",
"mrzigha",
"nourspace",
"vimystic"
],
"repo": "strangelove-ventures/heighliner",
"url": "https://github.com/strangelove-ventures/heighliner/issues/273",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
332569040
|
Quickstart is not working on my mac
api_1 | [2018-06-14T21:18:40.844Z] info Creating your application... It might take a few seconds.
api_1 | [2018-06-14T21:18:40.928Z] error $ strapi new can only be called in an empty directory.
api_1 | [2018-06-14T21:18:41.177Z] error This command can only be used inside a Strapi project.
strapi-docker_api_1 exited with code 0
docker pull node:9.11.1-alpine
and redo quickstart. Now it works.
|
gharchive/issue
| 2018-06-14T21:21:49 |
2025-04-01T06:40:29.886884
|
{
"authors": [
"demobo-com"
],
"repo": "strapi/strapi-docker",
"url": "https://github.com/strapi/strapi-docker/issues/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
302732519
|
tmpop: add tendermint evidence
At last, we're able to compute full tendermint evidence for our segments!
We can't really use the lite package as it's meant to be used by clients and is quite complex to fit in our infrastructure (it acts as a proxy node) but I think the data we're able to access in the votes is enough to produce a correct proof.
It's a bit complicated to do so because we can do it only at the beginning of block N+3 so we need to be careful with off-by-one errors in many places.
I've tested on a filetmpop that the evidence produced looks correct and is correctly validated by the Verify() method, but this is a tricky piece of code so I'm counting on you to find potential issues during the PR ;).
Don't hesitate to drop by my desk to draw diagrams of what object signs what part if you're unsure.
This change is
Reviewed 10 of 10 files at r1.
Review status: all files reviewed at latest revision, all discussions resolved, some commit checks failed.
cs/evidence_tendermint_test.go, line 189 at r1 (raw file):
}
assert.True(t, e.Verify(linkHash), "Proof should be verified")
supernit: shouldn't this assert be in its own test case?
tmpop/tmpop.go, line 376 at r1 (raw file):
linkHashes, err := t.getCommitLinkHashes(evidenceHeight)
if err != nil {
log.Warn("Could not get link hashes for this block. Evidence will not be generated.")
we should add more information about the block in the warning message
tmpop/tmpop.go, line 386 at r1 (raw file):
validatorHash, err := t.getValidatorHash(evidenceHeight)
if err != nil {
log.Warn("Could not get validator hash for this block. Evidence will not be generated.")
here as well
Comments from Reviewable
at last we have proper evidence! good job!
Review status: all files reviewed at latest revision, 3 unresolved discussions, some commit checks failed.
Comments from Reviewable
Review status: all files reviewed at latest revision, 3 unresolved discussions, some commit checks failed.
cs/evidence_tendermint_test.go, line 189 at r1 (raw file):
Previously, such (Adrien Montfort) wrote…
supernit: shouldn't this assert be in its own test case?
I liked making sure that the proof was valid before I was passing it on to the tests that might modify it.
But now that it works, maybe it's not necessary, I'll have a second look at it.
tmpop/tmpop.go, line 376 at r1 (raw file):
Previously, such (Adrien Montfort) wrote…
we should add more information about the block in the warning message
Good idea.
tmpop/tmpop.go, line 386 at r1 (raw file):
Previously, such (Adrien Montfort) wrote…
here as well
Good idea as well :).
Comments from Reviewable
Note that I'd like to work in the next few weeks on setting up a real metrics/monitoring stack for Indigo, this will be the occasion to provide good analytics on what happens in TMPoP.
Review status: 8 of 10 files reviewed at latest revision, 3 unresolved discussions.
Comments from Reviewable
And a first brick for IndigoEntreprise. Nice!
Review status: 8 of 10 files reviewed at latest revision, all discussions resolved.
Comments from Reviewable
Reviewed 9 of 10 files at r1, 1 of 2 files at r2.
Review status: 9 of 10 files reviewed at latest revision, 2 unresolved discussions, some commit checks failed.
tmpop/tmpoptestcases/evidence.go, line 150 at r1 (raw file):
}
tmClientMock.EXPECT().Block(int64(5)).Return(blocks[5], nil).AnyTimes()
nit: can factorize all block initialization
tmpop/tmpoptestcases/evidence.go, line 310 at r1 (raw file):
err := makeQuery(h, tmpop.GetSegment, linkHash4, got)
assert.NoError(t, err)
assert.Len(
nit: assert.Empty()
Comments from Reviewable
Yeah !
Review status: 9 of 10 files reviewed at latest revision, 2 unresolved discussions, some commit checks failed.
tmpop/tmpoptestcases/evidence.go, line 150 at r1 (raw file):
Previously, alexppxela (Alexandre Thibault) wrote…
nit: can factorize all block initialization
Not really, because a few blocks aren't built exactly like the others (some don't have votes).
This E2E test is big and a bit hard to follow I'll admit :)
Comments from Reviewable
Review status: 8 of 10 files reviewed at latest revision, 2 unresolved discussions.
tmpop/tmpoptestcases/evidence.go, line 310 at r1 (raw file):
Previously, alexppxela (Alexandre Thibault) wrote…
nit: assert.Empty()
Done.
Comments from Reviewable
Reviewed 1 of 2 files at r2, 1 of 1 files at r3.
Review status: all files reviewed at latest revision, all discussions resolved, some commit checks failed.
Comments from Reviewable
looking great !
Reviewed 7 of 10 files at r1, 1 of 2 files at r2, 1 of 1 files at r3.
Review status: all files reviewed at latest revision, 4 unresolved discussions, some commit checks failed.
cs/evidence_tendermint_test.go, line 171 at r3 (raw file):
// generates a valid block and its proof, and returns the link
// and the evidence.
func CreateTendermintProof(t *testing.T, linksCount int) (*types.Bytes32, *evidences.TendermintProof) {
nit: does this need to be exported ?
cs/evidences/evidences.go, line 192 at r2 (raw file):
// We validate that nodes signed the header.
if !p.validateVotes(p.Header, p.HeaderVotes) {
as discussed IRL: are the (potential) byzantine votes included in p.HeaderVotes by tendermint ? if yes, then we should make sure than more than 1/3 of the votes are valid, not all of them.
tmpop/tmClient.go, line 86 at r3 (raw file):
for _, tx := range tmBlock.Block.Txs {
tmTx, err := unmarshallTx(tx)
if !err.IsOK() || tmTx.TxType != CreateLink {
why do you check if the transaction was a CreateLink (even though we only have this type) ?
tmpop/tmpoptestcases/evidence.go, line 337 at r3 (raw file):
// vote creates a valid vote for a given header.
// It simulates nodes signing a header and is crucial for the proof.
func vote(header *tmtypes.Header) []*evidences.TendermintVote {
is this the same function as above (cs/evidence_tendermint_test.go) ? if yes, is there a way to factorize ?
Comments from Reviewable
The TendermintProof verification should be updated in JS now! ;)
Reviewed 1 of 2 files at r2.
Review status: all files reviewed at latest revision, 4 unresolved discussions, some commit checks failed.
Comments from Reviewable
The developer named Bastien is not available anymore. Please try another developer.
Review status: all files reviewed at latest revision, 4 unresolved discussions, some commit checks failed.
cs/evidence_tendermint_test.go, line 171 at r3 (raw file):
Previously, simonvadee (Simon Vadée) wrote…
nit: does this need to be exported ?
I think it's useful yes
cs/evidences/evidences.go, line 192 at r2 (raw file):
Previously, simonvadee (Simon Vadée) wrote…
as discussed IRL: are the (potential) byzantine votes included in p.HeaderVotes by tendermint ? if yes, then we should make sure than more than 1/3 of the votes are valid, not all of them.
Yes good point! I'll dive more into simulating byzantine nodes next.
tmpop/tmClient.go, line 86 at r3 (raw file):
Previously, simonvadee (Simon Vadée) wrote…
why do you check if the transaction was a CreateLink (even though we only have this type) ?
It feels more future-proof when/if we add other TxTypes :)
We most likely only want to create evidence for CreateLink operations (even though I admit it depends on what other operations we add in the future).
But the main usecase was that we might at some point have an AddEvidence operation to store external evidence, and we don't want to generate evidence on evidence...but that's still a bit blurry so we'll see later.
tmpop/tmpoptestcases/evidence.go, line 337 at r3 (raw file):
Previously, simonvadee (Simon Vadée) wrote…
is this the same function as above (cs/evidence_tendermint_test.go) ? if yes, is there a way to factorize ?
Yes it is, but I'm not a big fan of factorizing it yet, as Rob Pike says in Go sometimes a little duplication is better...if more functions need to be shared then we'll reevaluate :)
Comments from Reviewable
cs/evidences/evidences.go, line 192 at r2 (raw file):
Previously, t-bast (Bastien Teinturier) wrote…
Yes good point! I'll dive more into simulating byzantine nodes next.
That's a bit different than byzantine votes. A wrong signature is a bad way to try to mess with the system since it's so easily catchable. I'm pretty sure bad signature are not included in the block header by the Tendermint engine.
A byzantine vote would be voting twice on blocks with the same height for instance.
But yeah we should check that we have at least 2/3+ signatures.
Comments from Reviewable
^^
Comments from Reviewable
|
gharchive/pull-request
| 2018-03-06T14:54:52 |
2025-04-01T06:40:30.029194
|
{
"authors": [
"alexppxela",
"simonvadee",
"such",
"t-bast"
],
"repo": "stratumn/go-indigocore",
"url": "https://github.com/stratumn/go-indigocore/pull/356",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1920261385
|
Fix: DjangoOptimizerExtension corrupts nested objects' fields' prefetch objects
Description
This PR aims to resolve issue #379 .
I included a reproducible test and a suggested fix, where the Prefetch object gets deepcopied to avoid the side effects from add_prefix.
Considerations:
There were multiple ways to do this, including copying the OptimizerStore at a higher level in the execution stack, but this should be the least invasive one
I used deepcopy now as Prefetch didn't offer an easier way of copying the object that I know of, I'm happy about suggestions on improving this
I'm also very open for improving the test case included. If you know a better way on how to do it without the custom type setup, please let me know.
Types of Changes
[ ] Core
[x] Bugfix
[ ] New feature
[ ] Enhancement/optimization
[ ] Documentation
Issues Fixed or Closed by This PR
#379
Checklist
[x] My code follows the code style of this project.
[ ] My change requires a change to the documentation.
[ ] I have updated the documentation accordingly.
[x] I have read the CONTRIBUTING document.
[x] I have added tests to cover my changes.
[x] I have tested the changes and verified that they work and don't break anything (as well as I can manage).
I love the work you do here, thanks a lot for the really awesome work! ❤️
Codecov Report
All modified lines are covered by tests :white_check_mark:
Comparison is base (a17b51b) 87.98% compared to head (c03dc67) 87.99%.
Additional details and impacted files
@@ Coverage Diff @@
## main #380 +/- ##
=======================================
Coverage 87.98% 87.99%
=======================================
Files 33 33
Lines 2971 2973 +2
=======================================
+ Hits 2614 2616 +2
Misses 357 357
Files
Coverage Δ
strawberry_django/optimizer.py
89.13% <100.00%> (+0.06%)
:arrow_up:
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
|
gharchive/pull-request
| 2023-09-30T13:29:03 |
2025-04-01T06:40:30.040822
|
{
"authors": [
"aprams",
"codecov-commenter"
],
"repo": "strawberry-graphql/strawberry-graphql-django",
"url": "https://github.com/strawberry-graphql/strawberry-graphql-django/pull/380",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2097921153
|
Deprecate starlite
Now that we Litestar support I think we can deprecate Starlite 😊
### Tasks
- [ ] Add notice in the docs
- [ ] Use typing_extensions.deprecated to mark class as deprecated (see https://peps.python.org/pep-0702/)
- [ ] Trigger deprecation warning (with test), this might
Upvote & Fund
We're using Polar.sh so you can upvote and help fund this issue.
We receive the funding once the issue is completed & confirmed by you.
Thank you in advance for helping prioritize & fund our backlog.
@Birdi7 Please feel free to go ahead with this 😊 If you need assistance or a review, feel free to ping me
|
gharchive/issue
| 2024-01-24T10:15:45 |
2025-04-01T06:40:30.043879
|
{
"authors": [
"erikwrede",
"patrick91"
],
"repo": "strawberry-graphql/strawberry",
"url": "https://github.com/strawberry-graphql/strawberry/issues/3359",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2319373394
|
default_factory doesn't work
Hi!
I use default_factory to initialize my variable, but variable always returns the same result. It seems like default_factory doesn't work and it returns always the same result of function.
Here is example to reproduce:
https://play.strawberry.rocks/?gist=a7a5e62ffe4e68696b44456398d11104
Upvote & Fund
We're using Polar.sh so you can upvote and help fund this issue.
We receive the funding once the issue is completed & confirmed by you.
Thank you in advance for helping prioritize & fund our backlog.
I can reproduce this, I think it's because we store the default value when creating the object, see:
I don't think this can be supported (and still be compliant with the spec). GraphQL defaults are statically in the schema. strawberry export-schema ...:
type Mutation {
update1(fields: MyInputType!): String!
update2(fields: MyInputType!): String!
}
input MyInputType {
field1: String = "318fbf6e-73b6-40eb-932f-0b66ba935b75"
}
type Query {
hello: String!
}
Another issue is that the client sending an explicit null is valid and semantically different. So MyInputType would need logic like
if field1 in (None, UNSET):
field1 = uuid_pkg.uuid4()
That may seem like a workaround, but is actually the only correct implementation.
@patrick91 and also can I ask in what situations then we need to use default_factory if this field is static, when we can use only default?
@patrick91 and also can I ask in what situations then we need to use default_factory if this field is static, when we can use only default?
I'm not sure to be honest, I'll need to think about this a bit
I do think it might be a flaw, or at least something surprising, so maybe we need to reconsider it
I do understand the schema's default value issue, but I do agree with this comment. I actually had to do workaround a similar issue on strawberry-resources when exporting form data for the field as a dynamic default value would not actually make sense there.
So my vote would be to actually change the behavior to fix this issue, specially since we still are 0.x =P, and mention as a "possible breaking change" in the changelog, mentioning the use of default as the correct way of relying on the older behavior
just throwing out some ideas to make the change less painful
we could a static_factory or schema_factory which will have the current behaviour, default_factory will mimic dataclasses' and pydantic's behaviour
have a configuration option to disable static defaults
just change the behaviour
@coady sorry to ping you again, but do you use default_factory for defaults in the schema level? 😊 what's your use case exactly?
@coady sorry to ping you again, but do you use default_factory for defaults in the schema level? 😊 what's your use case exactly?
The only use I'm aware of (and use) is for mutables, as dataclasses requires. Any valid value is a valid default value, including [] and {}.
I'd be happy with a cleaner alternative for mutables. This is forbidden (but is valid GraphQL):
q: list[float] = [0.5]
So instead I have to used default_factory or this:
q: list[float] = (0.5,) # type: ignore
which mypy complains about.
just throwing out some ideas to make the change less painful
we could a static_factory or schema_factory which will have the current behaviour, default_factory will mimic dataclasses' and pydantic's behaviour
have a configuration option to disable static defaults
just change the behaviour
option 1 could also have a codemod to make the update easier 😊
I think we could do option 2 for the time being
|
gharchive/issue
| 2024-05-27T15:13:17 |
2025-04-01T06:40:30.056813
|
{
"authors": [
"ShtykovaAA",
"bellini666",
"coady",
"patrick91"
],
"repo": "strawberry-graphql/strawberry",
"url": "https://github.com/strawberry-graphql/strawberry/issues/3517",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1957218713
|
Can't build
In raising this issue I confirm that
[X] I have fully completed the issue template
[X] I have searched open and closed issues for duplicates
[X] I have read the Contribution Guidelines
[X] I have read the Code of Conduct
[X] I have read the Documentation
Describe the bug
Trying to build but failing
Steps to reproduce the behavior
First command of the guide
git clone <https://github.com/streamdeck-linux-gui/streamdeck-linux-gui.git>
Gives file or directory not existing error
If i try
git clone https://github.com/streamdeck-linux-gui/streamdeck-linux-gui.git
and then trying to build i get
/usr/bin/python: No module named build
ALSO
the fedora script results in this error
` The headers or library files could not be found for zlib,
a required dependency when compiling Pillow from source.
Please see the install instructions at:
https://pillow.readthedocs.io/en/latest/installation.html
Traceback (most recent call last):
File "<string>", line 852, in <module>
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/__init__.py", line 103, in setup
return distutils.core.setup(**attrs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/_distutils/core.py", line 185, in setup
return run_commands(dist)
^^^^^^^^^^^^^^^^^^
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/_distutils/core.py", line 201, in run_commands
dist.run_commands()
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/_distutils/dist.py", line 969, in run_commands
self.run_command(cmd)
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/dist.py", line 989, in run_command
super().run_command(command)
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/wheel/bdist_wheel.py", line 364, in run
self.run_command("build")
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/dist.py", line 989, in run_command
super().run_command(command)
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/_distutils/command/build.py", line 131, in run
self.run_command(cmd_name)
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/_distutils/cmd.py", line 318, in run_command
self.distribution.run_command(command)
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/dist.py", line 989, in run_command
super().run_command(command)
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/_distutils/dist.py", line 988, in run_command
cmd_obj.run()
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/_distutils/command/build_ext.py", line 345, in run
self.build_extensions()
File "<string>", line 687, in build_extensions
RequiredDependencyException: zlib
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/alessandro/.local/lib/python3.12/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 353, in <module>
main()
File "/home/alessandro/.local/lib/python3.12/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 335, in main
json_out['return_val'] = hook(**hook_input['kwargs'])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/alessandro/.local/lib/python3.12/site-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py", line 251, in build_wheel
return _build_backend().build_wheel(wheel_directory, config_settings,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/build_meta.py", line 434, in build_wheel
return self._build_with_temp_dir(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/build_meta.py", line 419, in _build_with_temp_dir
self.run_setup()
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/build_meta.py", line 507, in run_setup
super(_BuildMetaLegacyBackend, self).run_setup(setup_script=setup_script)
File "/tmp/pip-build-env-s3ikdncm/overlay/lib/python3.12/site-packages/setuptools/build_meta.py", line 341, in run_setup
exec(code, locals())
File "<string>", line 903, in <module>
RequiredDependencyException:
The headers or library files could not be found for zlib,
a required dependency when compiling Pillow from source.
Please see the install instructions at:
https://pillow.readthedocs.io/en/latest/installation.html
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for pillow
Failed to build hidapi pillow
ERROR: Could not build wheels for hidapi, pillow, which is required to install pyproject.toml-based projects
`
Expected behavior
building from source
Screenshots
No response
System Information
Fedora 39
Stream Deck Version
No response
Could you try
python -m pip install ./
In the streamdeck-linux-gui directory?
python -m pip install ./
ERROR: Package 'streamdeck-linux-gui' requires a different Python: 3.12.0 not in '<3.12,>=3.11'
Also, i had streamdeck installed already, but now gives this error
Traceback (most recent call last): File "/home/alessandro/.local/bin/streamdeck", line 5, in <module> from streamdeck_ui.gui import start ModuleNotFoundError: No module named 'streamdeck_ui'
Try installing python 3.11 or change pyproject.toml on line 15 to include
your current version (that's easier in my opinion)
Also remove the old streamdeck_ui it's way out of date.
Razer @.***> schrieb am Mo., 23. Okt. 2023, 17:33:
python -m pip install ./
ERROR: Package 'streamdeck-linux-gui' requires a different Python: 3.12.0
not in '<3.12,>=3.11'
Also, i had streamdeck installed already, but now gives this error
Traceback (most recent call last): File
"/home/alessandro/.local/bin/streamdeck", line 5, in from
streamdeck_ui.gui import start ModuleNotFoundError: No module named
'streamdeck_ui'
—
Reply to this email directly, view it on GitHub
https://github.com/streamdeck-linux-gui/streamdeck-linux-gui/issues/107#issuecomment-1775471106,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/ABJEWWX7X7O62ND5552ABSLYA2E3RAVCNFSM6AAAAAA6MDPRCKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONZVGQ3TCMJQGY
.
You are receiving this because you commented.Message ID:
@.***
com>
Try installing python 3.11 or change pyproject.toml on line 15 to include your current version (that's easier in my opinion) Also remove the old streamdeck_ui it's way out of date. Razer @.> schrieb am Mo., 23. Okt. 2023, 17:33:
…
python -m pip install ./ ERROR: Package 'streamdeck-linux-gui' requires a different Python: 3.12.0 not in '<3.12,>=3.11' Also, i had streamdeck installed already, but now gives this error Traceback (most recent call last): File "/home/alessandro/.local/bin/streamdeck", line 5, in from streamdeck_ui.gui import start ModuleNotFoundError: No module named 'streamdeck_ui' — Reply to this email directly, view it on GitHub <#107 (comment)>, or unsubscribe https://github.com/notifications/unsubscribe-auth/ABJEWWX7X7O62ND5552ABSLYA2E3RAVCNFSM6AAAAAA6MDPRCKVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONZVGQ3TCMJQGY . You are receiving this because you commented.Message ID: @. com>
Tried uninstalling
pip3 uninstall streamdeck-ui
WARNING: Skipping streamdeck-ui as it is not installed.
Tried installing with the command, but same error as before
note: This error originates from a subprocess, and is likely not a problem with pip. ERROR: Failed building wheel for pillow Successfully built streamdeck-linux-gui Failed to build pillow ERROR: Could not build wheels for pillow, which is required to install pyproject.toml-based projects
Fixed installing libjpeg-turbo-devel and zlib-devel
|
gharchive/issue
| 2023-10-23T13:48:03 |
2025-04-01T06:40:30.096631
|
{
"authors": [
"Quintar",
"Razer0123"
],
"repo": "streamdeck-linux-gui/streamdeck-linux-gui",
"url": "https://github.com/streamdeck-linux-gui/streamdeck-linux-gui/issues/107",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
278073697
|
expose Readable & Writable
fixes #1
There is a way to add streams to unionfs, but it is not trivial.
Your implementation is obviously invalid, all it does is it creates some "placeholders".
What is the reason for doing so?
graceful-fs tries to get those streams (fs.prototype.ReadStream, fs.prototype.WriteStream), if they are not there it fails hard.
yes it seems wrong as it still doesnt work correctly here (just a few steps later)
I am OK including this hack. However this introduces a dependency:
import {Readable, Writable} from "stream";
Which may break browser users. It has to be done somehow as to not break browser builds.
Maybe something like this
const isBrowser = typeof __filename === 'undefined';
?
yeah made it conditional...
another question here:
i think im using it alltogether somehow wrong:
const {ufs} = require('unionfs');
const {Volume} = require('memfs');
const fs = require('fs');
ufs
.use(fs)
.use(Volume.fromJSON({"foo.js": ""}, "/tmp"))
console.log(ufs.existsSync(__filename)); // false
console.log(fs.existsSync(__filename)); // true
console.log(ufs.existsSync("/tmp/foo.js")); // true
why does it fail when trying to stat an existing file? the in memory volume is mounted somewhere else __filename shouldnt be affected?!
@streamich any idea on this one?
I will take a look at it this evening.
@streamich thanks for merging and fixing! did you have some time to look into my example above?
@digitalkaoz Sorry, I keep forgetting about this, will create an issue.
@digitalkaoz Should be fixed here: https://github.com/streamich/unionfs/issues/10#issuecomment-350269633
Thanks for the find.
|
gharchive/pull-request
| 2017-11-30T10:42:09 |
2025-04-01T06:40:30.102680
|
{
"authors": [
"digitalkaoz",
"streamich"
],
"repo": "streamich/unionfs",
"url": "https://github.com/streamich/unionfs/pull/8",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
}
|
2008315073
|
🛑 WSDL SAM SCS Demo is down
In 77c4f7a, WSDL SAM SCS Demo (https://wsdl-demo.streamlined-scs.cl/api/Scssalidabodega/getcontadorsalidabodega) was down:
HTTP code: 502
Response time: 452 ms
Resolved: WSDL SAM SCS Demo is back up in cdfd912 after 1 hour, 54 minutes.
|
gharchive/issue
| 2023-11-23T14:14:15 |
2025-04-01T06:40:30.105572
|
{
"authors": [
"streamlined-scs"
],
"repo": "streamlined-scs/upptime",
"url": "https://github.com/streamlined-scs/upptime/issues/819",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1666397435
|
Is it possible to know how why have 20 seconds delay between issuing the stressapptest command and system running
We run the command to enable the StressAppTest with testing time, take our 100 seconds test for example, after issuing the command it is 04/13 21:02:32, but it start from 21:02:51 so it takes around 20 seconds before running, is it possible to know why we have the 20 seconds before running?
2023/04/13-21:02:32(CST) Log: Prefer plain malloc memory allocation.
2023/04/13-21:02:32(CST) Log: Using mmap() allocation at 0x7f0e4b600000.
2023/04/13-21:02:32(CST) Stats: Starting SAT, 243912M, 100 seconds
2023/04/13-21:02:50(CST) Log: region number 8 exceeds region count 8
2023/04/13-21:02:51(CST) Log: Region mask: 0xff
2023/04/13-21:03:01(CST) Log: Seconds remaining: 90
2023/04/13-21:03:11(CST) Log: Seconds remaining: 80
On Thu, Apr 13, 2023 at 6:12 AM AlecHsiao @.***> wrote:
is it possible to know why we have the 20 seconds before running?
2023/04/13-21:02:32(CST) Stats: Starting SAT, 243912M, 100 seconds
Stressapptest fills memory with patterns before starting the test. You have
a large amount of memory
and it's likely that initializing the memory takes 20 seconds. Do you have
a full log with memory bandwidth indicated?
Message ID: @.***>
Yes, we change the log level to 20 to capture more data, I think you're correct, that is under the process to fill data in that period. Thanks for your help for that.
2023/02/22-08:00:46(CST) Starting Fill Threads 0: 30511 pages
2023/02/22-08:00:46(CST) Starting Fill Threads 1: 30511 pages
2023/02/22-08:00:46(CST) Starting Fill Threads 2: 30511 pages
2023/02/22-08:00:46(CST) Starting Fill Threads 3: 30511 pages
2023/02/22-08:00:46(CST) Starting Fill Threads 4: 30511 pages
2023/02/22-08:00:46(CST) Starting Fill Threads 5: 30511 pages
2023/02/22-08:00:46(CST) Starting Fill Threads 6: 30511 pages
2023/02/22-08:00:46(CST) Starting Fill Threads 7: 30518 pages
2023/02/22-08:00:46(CST) Log: Thread 0 running on core ID 81 mask FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF (FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF).
2023/02/22-08:00:46(CST) Log: Starting fill thread 0
2023/02/22-08:00:46(CST) Log: Thread 2 running on core ID 103 mask FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF (FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF).
2023/02/22-08:00:46(CST) Log: Starting fill thread 2
2023/02/22-08:00:46(CST) Log: Thread 1 running on core ID 0 mask FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF (FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF).
2023/02/22-08:00:46(CST) Log: Starting fill thread 1
2023/02/22-08:00:46(CST) Log: Thread 3 running on core ID 19 mask FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF (FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF).
2023/02/22-08:00:46(CST) Log: Starting fill thread 3
2023/02/22-08:00:46(CST) Log: Thread 6 running on core ID 44 mask FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF (FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF).
2023/02/22-08:00:46(CST) Log: Starting fill thread 6
2023/02/22-08:00:46(CST) Log: Thread 5 running on core ID 121 mask FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF (FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF).
2023/02/22-08:00:46(CST) Log: Starting fill thread 5
2023/02/22-08:00:46(CST) Log: Thread 4 running on core ID 116 mask FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF (FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF).
2023/02/22-08:00:46(CST) Log: Starting fill thread 4
2023/02/22-08:00:46(CST) Log: Thread 7 running on core ID 52 mask FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF (FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF).
2023/02/22-08:00:46(CST) Log: Starting fill thread 7
2023/02/22-08:01:05(CST) Log: Co
1.0.10.log
mpleted 0: Fill thread. Status 1, 30511 pages filled
2023/02/22-08:01:05(CST) Log: Completed 7: Fill thread. Status 1, 30518 pages filled
2023/02/22-08:01:05(CST) Log: Completed 5: Fill thread. Status 1, 30511 pages filled
2023/02/22-08:01:05(CST) Log: Completed 6: Fill thread. Status 1, 30511 pages filled
2023/02/22-08:01:05(CST) Log: Completed 3: Fill thread. Status 1, 30511 pages filled
2023/02/22-08:01:05(CST) Log: Completed 1: Fill thread. Status 1, 30511 pages filled
2023/02/22-08:01:05(CST) Log: Completed 4: Fill thread. Status 1, 30511 pages filled
2023/02/22-08:01:05(CST) Log: Completed 2: Fill thread. Status 1, 30511 pages filled
1.0.10.log
For whatever reason, initialization uses only 8 threads, which
probably isn't appropriate on large systems such as yours.
I guess a better approach would be to scale the number of initialization
and teardown threads to the number of cores
available, similar to the default for copy threads. I'll keep it in mind as
a feature request, or you can send a PR.
If you just want it to be faster and can compile yourself, you can change
the hardcoded thread counts here:
https://github.com/stressapptest/stressapptest/blob/fd4ae17eaad7fde69e1308abbe5af3181ec6ce15/src/sat.cc#L719
On Sun, Apr 16, 2023 at 9:01 PM AlecHsiao @.***> wrote:
1.0.10.log
https://github.com/stressapptest/stressapptest/files/11246074/1.0.10.log
—
Reply to this email directly, view it on GitHub
https://github.com/stressapptest/stressapptest/issues/107#issuecomment-1510660957,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/ADMRIIJTDAMGBYSYGA5RHRDXBS6ANANCNFSM6AAAAAAW5CELRA
.
You are receiving this because you commented.Message ID:
@.***>
Thanks, I think that works, now I can reduce the delay time , so I think that would be enough for me, appreciated for your help !!!
|
gharchive/issue
| 2023-04-13T13:12:24 |
2025-04-01T06:40:30.191824
|
{
"authors": [
"AlecHsiao",
"nickjsanders"
],
"repo": "stressapptest/stressapptest",
"url": "https://github.com/stressapptest/stressapptest/issues/107",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
51143142
|
Update Facebook to use v2 of their API as the current v1 will expire in a few months
facebookTokenURL string = "https://graph.facebook.com/v2.2/oauth/access_token"
facebookEndpointProfile string = "https://graph.facebook.com/v2.2/me?fields=email,first_name,last_name,link,about,id,name,picture,location"
@tmsoft Was this resolved
Doesn't look like it. Is anyone maintaining the codebase?
I'm not sure anyone is assigned to maintain the package. Perhaps we should find some new people who are interested?
On 24 Jul 2015, at 05:40, Todd notifications@github.com wrote:
Doesn't look like it. Is anyone maintaining the codebase?
—
Reply to this email directly or view it on GitHub.
@matryer Feel free to add me as one and I'll take a stab at it.
@matryer: I would like to volunteer to be a maintainer if possible.
|
gharchive/issue
| 2014-12-05T20:36:35 |
2025-04-01T06:40:30.195254
|
{
"authors": [
"bscott",
"maggit",
"matryer",
"tmsoft"
],
"repo": "stretchr/gomniauth",
"url": "https://github.com/stretchr/gomniauth/issues/26",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1582342770
|
[WIP] Roslyn based serialization source generator
PR Details
Opening draft PR in order to be able to reference it in documentation, once I make some progress on this I'll update the description.
Description
TODO
Related Issue
TODO
Motivation and Context
TODO
Types of changes
[ ] Docs change / refactoring / dependency upgrade
[ ] Bug fix (non-breaking change which fixes an issue)
[x] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
Checklist
[ ] My change requires a change to the documentation.
[ ] I have added tests to cover my changes.
[ ] All new and existing tests passed.
what is the current state?
do you need help with it?
i have some expierence with source generators but i struggle with strides structure still.
But it would be a huge step forward.
what is the current state?
do you need help with it?
i have some expierence with source generators but i struggle with strides structure still.
But it would be a huge step forward.
I'm currently busy with another project and haven't made much progress on this. You can see the TODOs in my code. The main hassle is getting parity of output on edge cases. The way I was comparing output: compiling Stride main branch and viewing generated code with DotPeak, then running the source generator and comparing.
Once the GlobalDataSerializer attributes are correctly emitted, the next part is generating the method in the class they're on with the object IDs later used by the runtime serializer.
Since that may be difficult to handle - you could give it a try to refactor my code a bit so that it feels nicer to read (lack of readability was main issue of the previous implementation making it hard to change anything).
|
gharchive/pull-request
| 2023-02-13T13:18:56 |
2025-04-01T06:40:30.208407
|
{
"authors": [
"IXLLEGACYIXL",
"manio143"
],
"repo": "stride3d/stride",
"url": "https://github.com/stride3d/stride/pull/1609",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1928481187
|
[Native] - Implement some existing C++ methods in C#
PR Details
Description
This PR focuses on migrating most of the C++ code in Stride.Native to C#.
My change also applies to Stride.Graphics, so with a few gimmicks it should be possible to build this library outside of Windows.
Related Issue
#1394
Types of changes
[x] Docs change / refactoring / dependency upgrade
TODO
[ ] My change requires testing, so It could break existing code.
Thanks for the PR.
Before we merge this in, I would like to make sure we don't loose much performance. I remember it was in C++ for real perf reason (but at the time mono mobile was very slow so it might not be necessary anymore).
If I remember correctly, the 'FastTextRenderer' is only used for the debug text. Is this correct?
A long time since I've seen this code, but I think the regular spritebatch inherit from BatchBase, and is all C#.
I test the NativeInvoke.xnGraphicsFastTextRendererGenerateVertices with native and cs metod based on .net8.0
Hey @ly3027929699 thanks a bunch for testing this out for us, can you share the benchmark source as well ?
here is the source
TestAOTUnitTests.zip
this is a .rar file.
change .zip to .rar
here is the source TestAOTUnitTests.zip this is a .rar file. change .zip to .rar
the folder is empty, could you make a repo instead?
Nah, it works fine, make sure to change it to rar
is your result similar to me? @Eideren
https://github.com/ly3027929699/TestNativeVSCharp
here is repo of the source
@IXLLEGACYIXL
Fixed the benchmark usage of unsafe, results are still significantly better for c# though
TestAOTUnitTests.zip
Method
num
Mean
NativeMethod
100
30.00 us
CSharpMethod
100
21.51 us
NativeMethod
500
160.88 us
CSharpMethod
500
107.07 us
NativeMethod
1000
299.19 us
CSharpMethod
1000
213.93 us
Here's the result of another benchmark following the suggestions @froce made
Method
num
Mean
NativeMethod
100
32.42 us
Span
100
19.18 us
SpansFor
100
24.21 us
Optimized
100
10.54 us
Ptr
100
18.92 us
NativeMethod
500
163.00 us
Span
500
98.99 us
SpansFor
500
124.67 us
Optimized
500
54.89 us
Ptr
500
96.45 us
NativeMethod
1000
331.56 us
Span
1000
196.77 us
SpansFor
1000
250.98 us
Optimized
1000
107.84 us
Ptr
1000
195.63 us
Where
Ptr is with the fixed VertexPositionNormalTexture* vertexBuffer signature.
Span is using a span instead of a pointer.
SpansFor is a for loop instead of the manually unrolled loop in source.
I wrote an Optimized version where loop-constants are pre-computed, making it almost two times faster than pointer:
public static unsafe void Optimized(RectangleF constantInfos, RectangleF renderInfos, string textPointer, ref int textLength, Span<VertexPositionNormalTexture> vertexBuffer)
{
float fX = renderInfos.X / renderInfos.Width;
float fY = renderInfos.Y / renderInfos.Height;
float fW = constantInfos.X / renderInfos.Width;
float fH = constantInfos.Y / renderInfos.Height;
RectangleF destination = new(fX, fY, fW, fH);
RectangleF source = new(0.0f, 0.0f, constantInfos.X, constantInfos.Y);
// Copy the array length (since it may change during an iteration)
int textCharCount = textLength;
float scaledDestinationX;
float scaledDestinationY = -(destination.Y * 2f - 1f);
float invertedWidth = 1f / constantInfos.Width;
float invertedHeight = 1f / constantInfos.Height;
Span<(Vector2 Position, Vector2 TextureCoordinate)> baseData = stackalloc (Vector2, Vector2)[4]
{
( new(-destination.Width, +destination.Height), new(0 * source.Width * invertedWidth, 0 * source.Height * invertedHeight) ),
( new(+destination.Width, +destination.Height), new(1 * source.Width * invertedWidth, 0 * source.Height * invertedHeight) ),
( new(-destination.Width, -destination.Height), new(0 * source.Width * invertedWidth, 1 * source.Height * invertedHeight) ),
( new(+destination.Width, -destination.Height), new(1 * source.Width * invertedWidth, 1 * source.Height * invertedHeight) ),
};
int j = 0;
for (int i = 0; i < textCharCount; i++)
{
char currentChar = textPointer[i];
if (currentChar == '\v')
{
// Tabulation
destination.X += 8 * fX;
--textLength;
continue;
}
else if (currentChar >= 10 && currentChar <= 13) // '\n' '\v' '\f' '\r'
{
destination.X = fX;
destination.Y += fH;
scaledDestinationY = -(destination.Y * 2f - 1f);
--textLength;
continue;
}
else if (currentChar < 32 || currentChar > 126)
{
currentChar = ' ';
}
source.X = (currentChar % 32 * constantInfos.X) * invertedWidth;
source.Y = (currentChar / 32 % 4 * constantInfos.Y) * invertedHeight;
scaledDestinationX = destination.X * 2f - 1f;
// 0
vertexBuffer[j].Position.X = scaledDestinationX + baseData[0].Position.X;
vertexBuffer[j].Position.Y = scaledDestinationY + baseData[0].Position.Y;
vertexBuffer[j].TextureCoordinate.X = source.X + baseData[0].TextureCoordinate.X;
vertexBuffer[j].TextureCoordinate.Y = source.Y + baseData[0].TextureCoordinate.Y;
j++;
// 1
vertexBuffer[j].Position.X = scaledDestinationX + baseData[1].Position.X;
vertexBuffer[j].Position.Y = scaledDestinationY + baseData[1].Position.Y;
vertexBuffer[j].TextureCoordinate.X = source.X + baseData[1].TextureCoordinate.X;
vertexBuffer[j].TextureCoordinate.Y = source.Y + baseData[1].TextureCoordinate.Y;
j++;
// 2
vertexBuffer[j].Position.X = scaledDestinationX + baseData[2].Position.X;
vertexBuffer[j].Position.Y = scaledDestinationY + baseData[2].Position.Y;
vertexBuffer[j].TextureCoordinate.X = source.X + baseData[2].TextureCoordinate.X;
vertexBuffer[j].TextureCoordinate.Y = source.Y + baseData[2].TextureCoordinate.Y;
j++;
// 3
vertexBuffer[j].Position.X = scaledDestinationX + baseData[3].Position.X;
vertexBuffer[j].Position.Y = scaledDestinationY + baseData[3].Position.Y;
vertexBuffer[j].TextureCoordinate.X = source.X + baseData[3].TextureCoordinate.X;
vertexBuffer[j].TextureCoordinate.Y = source.Y + baseData[3].TextureCoordinate.Y;
j++;
destination.X += destination.Width;
}
}
Thanks !
|
gharchive/pull-request
| 2023-10-05T14:47:04 |
2025-04-01T06:40:30.226864
|
{
"authors": [
"Eideren",
"Ethereal77",
"IXLLEGACYIXL",
"Jklawreszuk",
"ly3027929699",
"xen2"
],
"repo": "stride3d/stride",
"url": "https://github.com/stride3d/stride/pull/1896",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
527451532
|
Go server updates
Used FormValue
Renamed method
Handled Stripe error
Ran gofmt
Thanks!
@andybons
Thanks again @andybons ! 👏
Yay thanks, @adreyfus-stripe! Hope my lack of context in the suggestions or comments came off poorly. Sometimes I forget to say why I'm suggesting changes.
Not at all @andybons! I appreciate attention to detail especially since I'm brand new to Go and still learning best practices.
|
gharchive/pull-request
| 2019-11-22T22:50:23 |
2025-04-01T06:40:30.241165
|
{
"authors": [
"adreyfus-stripe",
"andybons"
],
"repo": "stripe-samples/checkout-subscription-and-add-on",
"url": "https://github.com/stripe-samples/checkout-subscription-and-add-on/pull/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
406451413
|
Demo for how to Save Card with PaymentIntents
I can't seem to work out a way to save a card when making payment using the new PaymentIntents.
The old way of making a token on the client-side and passing this to the backend meant I could use the token as a source to make a customer and the card would be saved.
With PaymentIntents however, if I try to create a customer from the source returned from the Stripe.js handleCardPayment API call it throws an error: "The source you provided cannot be attached to the customer. It must be chargeable or pending.".
This has left me at an impasse. The only documentation I could find that covers saving a source with PaymentIntents assumes you already have a customer and source.
Edit: Creating a customer when you create a PaymentIntent is the only way I can see to have cards saveable with the new PaymentIntents flow. Although minor it's a little annoying because a customer is made whenever a user hits the checkout rather than when they've made at least one payment.
@vinnyvimto thanks for raising this, we're still actively working on the docs.
When using handleCardPayment[0] it does a couple of things under the hood:
It creates a card source (tokenisation of card details)
It confirms the PaymentIntent with the source at which point the radar and SCA rules[1] are evaluated
Based on the outcome of the rules:
do Strong Customer Authentication with 3D Secure if required & create the charge
if no authentication is required create the charge
For your scenario, do you always want to create a customer object if there was a successful payment, or do you have a checkbox in your checkout form that decides whether a customer object is created or not?
[0] https://stripe.com/docs/stripe-js/reference#stripe-handle-card-payment
[1] https://stripe.com/guides/strong-customer-authentication
@thorsten-stripe thanks for coming back to me.
For this scenario, there is a checkbox for saving a card that like you say used to decide whether a customer is created after submitting a form to the backend.
For the most seamless user experience with PaymentIntents, in the end, I just decided it was easier to make a call to my backend before calling handleCardPayment. This 'pre-call' does a number of things such as create a "pending" order on our end, update the PaymentIntent with shipping information, attach other metadata and create a customer if required to save a card.
If a customer leaves that page between this call and handleCardPayment they're then given an option to try and pay again. Moving an order from pending to completed happens via webhook and then updates the frontend via WebSocket.
Because this ended up being the route I took, it actually didn't matter too much what got sent through with the handleCardPayment call. At the time of opening this issue, however, I was really looking for a way to add shipping details and other metadata without having to hit my backend first. Shipping data being something that can only be added with new card sources.
I'd be interested to know what other flows people have generally come up with.
@vinnyvimto that sounds like a great solution for your use-case, thanks for sharing! We're thinking about allowing the creation of a customer object from a successful PaymentIntent which would allow you to shift the customer creation request outside of the checkout path. Would this be of interest to you?
YES, it would be of great interest !
+1 it would be great to be able to create a customer object and attach the source used in the paymentIntent
I think migrating to intents when there's application code already written would be much easier this way. By easier I mean less code and workflow would need to be changed.
+1 👍 I'm in the same position that @vinnyvimto initially described when this thread was opened.
We're using Stripe for charity donations; on the donation page, we're taking both one-off and recurring gifts using Stripe Elements. When the user-facing donation form is submitted, the payment details are checked and, if successful, a client-side Stripe token is return to the donation form and then the whole form (token, donor details and gift details) is passed to the server. The server then creates a customer and, depending on if the gift is a one-off gift or recurring gift, either a charge is applied (for a one-off gift) or a new subscription plan is created and added to the customer (for a recurring gift). If everything has happened correctly, the user is then sent to a 'Thank you' page and sent a receipt e-mail confirming their gift.
With PaymentIntents, I can't see that a corresponding workflow is possible without a serious code re-write. Any thoughts or advice would be very welcome.
@fthuk thanks for outlining your integration path. You can achieve this via the manual confirmation flow: https://stripe.com/docs/payments/payment-intents/quickstart#manual-confirmation-flow
Just note that you will need to go back to the client for performing authentication if required. Alternatively you can use the new Checkout, which also supports subscriptions now: https://stripe.com/docs/payments/checkout/server#create-subscriptions
@thorsten-stripe So I need to create a payment method and not let Elements make an intent out of it straight away?
Then add the method to a customer, subscribe to a plan and then somehow create the intent with or without 3DS?
If so, why can't we just attach intents to a customer and then subscribe them?
Same problem for me with Checkout. I can integrate it in a few minutes, but without proper handling of EU VAT, it useless :(
So right now I'm trying to integrate flow with a subscription manually.
@kernio after hours of messing around I finally figured it out manually. Maybe I can save you some time.
Use Elements, to create a payment method (not intent), send that id to PHP and add it to a customer, also add it as their default payment method otherwise (weirdly) it wont work.
"invoice_settings" => ["default_payment_method" => $paymentMethod]
then create the subscription.
Quick follow-up that post-payment attachment of a payment method to a customer object is now available via the SCA off-session payment APIs: https://stripe.com/docs/payments/cards/saving-cards#saving-card-after-payment
+1 it would be great to be able to create a customer object and attach the source used in the paymentIntent
Hi, is there an update on this one?
@it-creed yes, see the details in my last comment: https://github.com/stripe/stripe-payments-demo/issues/43#issuecomment-508920663
https://stripe.com/docs/payments/cards/saving-cards#save-payment-method
It is crazy how the first (hence the only) added PaymentMethod is not the default and you have to set it explicitly.
@PranayShah thanks for the feedback. While we understand that this adds complexity, we want our users to be more aware and explicit of what method they are charging. Payments regulation is evolving globally and requires merchants in certain regions to explicitly set up new cards via a non-payment authentication during which they have to present the terms of service. You can find more information on this here: https://stripe.com/en-US/guides/sca-payment-flows
We've published a video that specifically looks at customer management / card-on-file with regard to SCA which you might find helpful if you landed here: https://youtu.be/52oinv6BZ34
I am saving PaymentMethod using: https://stripe.com/docs/payments/cards/saving-cards#saving-card-after-payment. However, when I query customer data, I don't see cards under sources. How can I show my users which cards are stored with our platform.
@thandaanda our recommendation is to store a list of your customer's payment methods (including the fingerprint, expiration date, billing address etc) in your own database. That way you don't have to ping the API for that data. Should you need to list the customer's payment methods, you can do so via the payment_methods endpoint: https://stripe.com/docs/api/payment_methods/list
Hi, I might have some issues with saving cards after payment using the PaymentIntents integration. For customers who don't yet have a saved payment method, the flow looks like this:
Customer hits a "Pay" button
UI calls stripe.createPaymentMethod('card', element)
UI calls backend, passing the ID of the payment method obtained in the previous step
Backend calls stripe.paymentIntents.create(), passing in the payment method ID, as well as setup_future_usage: 'off_session', confirmation_method: 'manual', confirm: true
If this fails because the card needs SCA, backend passes the client secret back to frontend, otherwise end
Frontend calls stripe.handleCardAction(clientSecret) and passes the ID of the returned payment intent to backend
Backend calls const intent = stripe.paymentIntents.confirm(intentId)
If payment method isn't saved yet, backend calls stripe.customers.create({ payment_method: intent.payment_method }) and saves customer ID along with payment method ID, otherwise end
Payment using these saved credentials looks like this:
Backend calls stripe.paymentIntents.create(), passing in stored customer ID and payment method ID
Continue from 5) above
The issue is that when charging a saved card, SCA is always triggered - even for the 4000 0025 0000 3155 test card, which should only require SCA the first time around. I think it might be because the first paymentIntents.create() backend call, where I pass in setup_future_usage, fails because it requires SCA and the payment intent is actually really created when I call handleCardAction() - but I can't pass the setup_future_usage to that, so the intent is created without that option.. or is it something else?
@jahudka the setup steps sound correct. For step 1) of "Payment using these saved credentials" are you passing the off_session:true[0] flag to the payment intent creation?
[0] https://stripe.com/docs/payments/cards/charging-saved-cards#create-payment-intent-off-session
@thorsten-stripe omg I wasn't, can't believe I missed that.. It works now! Although the off_session property of Stripe.paymentIntents.IPaymentIntentCreationOptions is missing in @types/stripe version 6.31.23 (which is the latest), so I had to convince TypeScript a little, but that's no biggie.. Thanks a lot!
Okay, now there's another issue - if I test with a card which always requires SCA, during off-session attempts when the card has already been saved, things start to get wonky..
The initial stripe.paymentIntents.create() fails as expected (well, except for the fact that now instead of returning an IPaymentIntent object with the appropriate state, it throws a StripeCardError - but the client secret can be extracted from that too). But then I pass the client secret to the client side and call stripe.handleCardAction() and it fails; my console says that "[t]he PaymentIntent supplied is not in the requires_action state". Inspecting the intent object found on the StripeCardError exception thrown by stripe.paymentIntents.create() indeed shows that the intent is in the requires_source state, but the code of the error is authentication_required and the message says Your card was declined. This transaction requires authentication.. This is in accord with what the docs say: charging a saved card is supposed to fail with an authentication_required code and a requires_payment_method (resp. requires_source, in older API versions) state.
The docs say I should now "follow the on-session payment instructions from step 2". So I pass the client secret extracted from the error object back to my frontend, my frontend calls stripe.handleCardAction().. and fails, telling me in the console that "[t]he PaymentIntent supplied is not in the requires_action state". This, too, is in accord with the docs - namely the fact that handleCardAction() only works with an intent in the requires_action state - but that effectively means that I have to switch to automatic confirmation now, doesn't it, because I'll have to call stripe.handleCardPayment() instead... what am I missing? SCA in off-session payments can't only support automatic confirmation, or does it?
@jahudka off_session is a confirm-time parameter and is not stored on the state of the PaymentIntent. You can extract the PaymentIntent ID from the error object and call confirm on it again with the saved payment method and omitting off_session:true which will default to false (on-session payment) which will then move the PaymentIntent into requires_action state. I'd recommend that you watch our Dev Chat on this topic, especially starting from 26:55[0] where we talk about the recovery flows.
[0] https://youtu.be/52oinv6BZ34?t=1616
Hi everybody !
I have been in the comfortable situation of implementing a checkout flow from scratch today. I wanted to achieve something similar to what @jahudka wanted to migrate: users can either do a one-off payment or go for a subscription.
I ended up with this flow, which looks pretty more lightweight than manual PaymentIntent handling:
Client requests a PaymentIntent to the server as soon as we know the kind of payment we want to go for. For that, I'm sending a payload like this:
{
userData: {
name: "Some name",
email: "some@email.com",
},
amount: 123123,
currency: "eur",
isSubscription: true,
}
The userData field contains information that I want to attach to the customer I may create later on if the payment succeeds. Note that if you need to charge existing customers you could add to this payload another key to send the customerId to use in the creation of the PaymentIntent.
Server generates the PaymentIntent. The userData is stored temporally in the metadata field and if isSubscription === true we need to set setup_future_usage = "off_session". Also, it's required to send payment_method_types = ["card"]. Something like this:
const payload = {
metadata: {
userData: JSON.stringify(userData),
},
amount,
currency,
payment_method_types: ['card'],
};
if (isSubscription) {
payload.setup_future_usage = 'off_session';
}
stripe.paymentIntents.create(payload);
Then it sends the client_secret field of the created PaymentIntent to the client.
The client receives the client_secret (setting it in the state or any other medium depending on your technology of choice), captures the values of the inputs (billing address and card data mainly) and uses stripe.handleCardPayment() passing the payment_method_data when the user presses the "pay" button. In my case looks something like this:
const { paymentIntent, error } = await stripe.handleCardPayment(
clientSecret,
{
payment_method_data: {
billing_details: {
... some details
},
},
}
);
Note that handleCardPayment handles the authentication part of SCA for you, so you don't need to do anything else on the client except providing feedback to the user based on the error key returned.
Finally, in a async way, the server receives via a webhook the payment_intent.succeeded event and takes care of creating a Customer using the data saved in metadata and based on the setup_future_usage value of the PaymentIntent creates a subscription or not (which we could use another field in metadata to differentiate). Note that when creating the customer you need to check if it's a subscription or not in order to add to the payload the payment_method:
const customer = await stripe.customers.create({
payment_method: intent.payment_method, // only if subscription, otherwise it throws an error
});
For the webhooks handling of the events, it was really useful to use the Stripe CLI and this code snippet. The endpointSecret secret is a value that I was actually no able to find anywhere on the dashboard, but the Stripe CLI displays it when executing stripe listen (it has the structure "whsec_....`).
Hope it helps people coming in this SCA storm under deadlines 😄
Hello all, im working with Stripe and cloning cards, everything works for the first step, but when i want to select an stored card im getting "The payment method you provided has already been attached to a customer"
This is what im doing:
$getPaymentMethod = \Stripe\PaymentMethod::retrieve(
$paymentMethodId
);
$payment_method = \Stripe\PaymentMethod::retrieve(
$getPaymentMethod->id
);
$getPaymentMethod->attach([
'customer' => $this->customer->getStripeId(),
]);
@lucasgsati please reach out to support using the form at https://support.stripe.com/ (preferred) or via email to support@stripe.com.
Closing this out as we now have dedicated demos for this at https://github.com/stripe-samples?q=sav
|
gharchive/issue
| 2019-02-04T18:06:02 |
2025-04-01T06:40:30.307302
|
{
"authors": [
"Fossil01",
"PranayShah",
"SpaceyRezum",
"dsampaolo",
"enzoferey",
"fthuk",
"it-creed",
"jahudka",
"kernio",
"loctrice",
"lucasgsati",
"thandaanda",
"thorsten-stripe",
"vinnyvimto"
],
"repo": "stripe/stripe-payments-demo",
"url": "https://github.com/stripe/stripe-payments-demo/issues/43",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
503263570
|
Issue 1500: Use the built-in formatting in logging methods
Pull Request Description
This pull request closes #1500
Acceptance Test
[X] Building the code with mvn clean install -Dintegration.tests still works.
[X] Running mvn spring-boot:run in the strongbox-web-core still starts up the application correctly.
[X] Building the code and running the strongbox-distribution from a zip or tar.gz still works.
[X] The tests in the strongbox-web-integration-tests still run properly.
Questions
Does this pull request break backward compatibility?
[ ] Yes
[X] No
Does this pull request require other pull requests to be merged first?
[ ] Yes, please see #...
[X] No
Does this require an update of the documentation?
[ ] Yes, please see strongbox/strongbox-docs#{PR_NUMBER}
[X] No
Hi @raksit31667 !
Thank you for your contribution!
Would you mind signing the ICLA, as described in the Contributing page?
Also, please, feel free to join our chat channel, if you'd like to learn more about the project and/or like to find out what else you could help with.
Kind regards,
Martin
@raksit31667 ,
Thanks for signing the ICLA!
@ptirador ,
Would you like to review this? :)
|
gharchive/pull-request
| 2019-10-07T06:50:21 |
2025-04-01T06:40:30.329753
|
{
"authors": [
"carlspring",
"raksit31667"
],
"repo": "strongbox/strongbox",
"url": "https://github.com/strongbox/strongbox/pull/1506",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
945775296
|
Command Groups
Hey, I've tested your Help command and I had one problem, when I added a command group, the bot just put the typing status and after that nothing happens, after I removed the command group all was working again. I checked the Cog, if there was any mistake in it, but no. So let me know, if it is my fault or the one from the help command.
Oh, i found the error was a mistake from my site.
|
gharchive/issue
| 2021-07-15T21:46:09 |
2025-04-01T06:40:30.397192
|
{
"authors": [
"Sluggyy"
],
"repo": "stroupbslayen/discord-pretty-help",
"url": "https://github.com/stroupbslayen/discord-pretty-help/issues/49",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
350115908
|
Migrate global datepicker_options into Datepicker component, limit jQuery UI imports
Who is this PR for?
developers
What problem does this PR fix?
It's a step towards https://github.com/studentinsights/studentinsights/issues/1758. This came up as I was looking to see what it would take to upgrade to jQuery 3.x, since Firefox raises CSP violations on load for jQuery 1.12.x.
What does this PR do?
Removes two global bits that <Datepicker /> relied on - window.datepicker_options and an asset path sent down in application.html.erb. Also scopes down the sprockets imports for jQuery UI.
selfie
|
gharchive/pull-request
| 2018-08-13T17:09:44 |
2025-04-01T06:40:30.412651
|
{
"authors": [
"kevinrobinson"
],
"repo": "studentinsights/studentinsights",
"url": "https://github.com/studentinsights/studentinsights/pull/1971",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
377507320
|
Profile: Allow showing MCAS in summary if no STAR, remove SGP graphs if no data
Who is this PR for?
Bedford educators
What problem does this PR fix?
In Bedford, they don't use STAR assessments, but the profile still assumes these are meaningful.
Separately, there's no SGP data for students in younger grades (or other students taking MCAS for the first time). Yet these charts still appear.
What does this PR do?
Adds a PerDistrict.js function that enables showing MCAS in place of STAR for the summary tabs on the student profile.
Updates the details sections to hide the SGP charts if there is no data.
Screenshot (if adding a client-side feature)
Checklists
Which features or pages does this PR touch?
[x] Student Profile
Does this PR use tests to help verify we can deploy these changes quickly and confidently?
[x] Included specs for changes
[x] Manual testing made more sense here
selfie
|
gharchive/pull-request
| 2018-11-05T17:37:28 |
2025-04-01T06:40:30.416123
|
{
"authors": [
"kevinrobinson"
],
"repo": "studentinsights/studentinsights",
"url": "https://github.com/studentinsights/studentinsights/pull/2234",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
202257717
|
Remove eager includes from schools_controller
These aren't needed on the precomputed path, so are probably pulling in more data than is needed.
Tests pass locally and kevin says 🚢
|
gharchive/pull-request
| 2017-01-20T22:24:18 |
2025-04-01T06:40:30.417084
|
{
"authors": [
"alexsoble",
"kevinrobinson"
],
"repo": "studentinsights/studentinsights",
"url": "https://github.com/studentinsights/studentinsights/pull/827",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
589549634
|
20200205 Go: Image Rendering in 2D Video Games with Ebiten
标题:Go: Image Rendering in 2D Video Games with Ebiten
原文链接:https://medium.com/a-journey-with-go/go-image-rendering-in-2d-video-games-with-ebiten-912cc2360c4f
译者自己根据原文翻译,译文保存为 markdown 格式,并附上尾部的签名。
---
via: https://medium.com/a-journey-with-go/go-image-rendering-in-2d-video-games-with-ebiten-912cc2360c4f
作者:[Vincent Blanchon](https://medium.com/@blanchon.vincent)
译者:[译者ID](https://github.com/译者ID)
校对:[校对者ID](https://github.com/校对者ID)
本文由 [GCTT](https://github.com/studygolang/GCTT) 原创编译,[Go 中文网](https://studygolang.com/) 荣誉推出
翻译 by alandtsang
|
gharchive/issue
| 2020-03-28T10:25:25 |
2025-04-01T06:40:30.428319
|
{
"authors": [
"alandtsang",
"polaris1119"
],
"repo": "studygolang/GCTT",
"url": "https://github.com/studygolang/GCTT/issues/1650",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
645411460
|
Nuget包上面的目前是.Net Core和Framework,能不能改成.Net Standard?
Nuget包上面的目前是.Net Core和Framework,能不能改成.Net Standard?
.net standard 请使用以前的版本,.net core 3.0+ 元包机制的更改,无法继续使用 .net standard
|
gharchive/issue
| 2020-06-25T09:40:07 |
2025-04-01T06:40:30.433126
|
{
"authors": [
"stulzq",
"xxdqr"
],
"repo": "stulzq/IPTools",
"url": "https://github.com/stulzq/IPTools/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
38108812
|
manifest.json
This gem definitely, solved my problem in precompiling my assets with same fingerprinting to be deployed to multiple instances of app server code. I am however seeing a side effect when trying to redeploy by seeing the following exception:
`parse': (<unknown>): mapping values are not allowed in this context at line 1 column 13 (Psych::SyntaxError)
from /Users/hootan/.rvm/rubies/ruby-2.0.0-p451/lib/ruby/2.0.0/psych.rb:205:in `parse_stream'
from /Users/hootan/.rvm/rubies/ruby-2.0.0-p451/lib/ruby/2.0.0/psych.rb:153:in `parse'
from /Users/hootan/.rvm/rubies/ruby-2.0.0-p451/lib/ruby/2.0.0/psych.rb:129:in `load'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/recipes/deploy/assets.rb:26:in `parse_manifest'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/namespaces.rb:191:in `method_missing'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/namespaces.rb:191:in `method_missing'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/recipes/deploy/assets.rb:93:in `block (3 levels) in load'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:138:in `instance_eval'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:138:in `invoke_task_directly'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:25:in `invoke_task_directly_with_callbacks'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:89:in `execute_task'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:101:in `find_and_execute_task'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/callback.rb:38:in `call'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:141:in `block in trigger'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:141:in `each'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:141:in `trigger'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:23:in `invoke_task_directly_with_callbacks'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:89:in `execute_task'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:101:in `find_and_execute_task'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/callback.rb:38:in `call'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:141:in `block in trigger'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:141:in `each'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:141:in `trigger'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:27:in `invoke_task_directly_with_callbacks'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:89:in `execute_task'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/namespaces.rb:191:in `method_missing'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/namespaces.rb:110:in `block in define_task'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/recipes/deploy.rb:234:in `block (3 levels) in load'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:56:in `transaction'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/namespaces.rb:191:in `method_missing'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/recipes/deploy.rb:233:in `block (2 levels) in load'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:138:in `instance_eval'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:138:in `invoke_task_directly'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:25:in `invoke_task_directly_with_callbacks'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:89:in `execute_task'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/namespaces.rb:191:in `method_missing'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/namespaces.rb:110:in `block in define_task'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/recipes/deploy.rb:201:in `block (2 levels) in load'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:138:in `instance_eval'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:138:in `invoke_task_directly'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/callbacks.rb:25:in `invoke_task_directly_with_callbacks'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:89:in `execute_task'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/configuration/execution.rb:101:in `find_and_execute_task'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/cli/execute.rb:46:in `block in execute_requested_actions'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/cli/execute.rb:45:in `each'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/cli/execute.rb:45:in `execute_requested_actions'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/cli/help.rb:19:in `execute_requested_actions_with_help'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/cli/execute.rb:34:in `execute!'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/lib/capistrano/cli/execute.rb:14:in `execute'
from /Users/hootan/.rvm/gems/ruby-2.0.0-p451@global/gems/capistrano-2.15.5/bin/cap:4:in `<top (required)>'
from /Users/hootan/.rvm/rubies/ruby-2.0.0-p451/bin/cap:23:in `load'
from /Users/hootan/.rvm/rubies/ruby-2.0.0-p451/bin/cap:23:in `<main>'
I removed manifest-.json from app/shared/assets folder and the redeploy works fine, but this is not practical to login to every single instance and remove. Is this a known problem ? please help.
This error occurs in :update_asset_mtimes if there is more than 1 manifest*.json in the shared_path/shared_assets_prefix. The above fix should remove the manifest from this location instead.
I think we should change to:
desc "remove manifest file from remote server"
task :remove_manifest, roles: :web do
run "rm -f #{shared_path}/#{shared_assets_prefix}/manifest*.json"
end
Thanks @steakchaser, would you mind submitting a pull request?
+1
@stve Have you had a chance to take a look at the PR from @steakchaser ?
Nevermind. It looks like @steakchaser did submit that PR after all. The gem just hasn't been updated on rubygems.
|
gharchive/issue
| 2014-07-17T18:00:07 |
2025-04-01T06:40:30.450636
|
{
"authors": [
"hootan-nikbakht",
"nozpheratu",
"steakchaser",
"stve"
],
"repo": "stve/capistrano-local-precompile",
"url": "https://github.com/stve/capistrano-local-precompile/issues/8",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2080247753
|
🛑 Panel is down
In 80e5d17, Panel (https://portal.xoro-hosting.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Status is back up in 9b1b003 after 43 minutes.
|
gharchive/issue
| 2024-01-13T11:19:35 |
2025-04-01T06:40:30.455721
|
{
"authors": [
"stxrlite"
],
"repo": "stxrlite/xoro-updates",
"url": "https://github.com/stxrlite/xoro-updates/issues/1265",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2731179994
|
[EXCLUSIVE CLIP] Five Girls Five Rocket Viral Video
In the fast-paced world of social media, it doesn’t take long for a unique or intriguing clip to capture the imagination of millions. The latest phenomenon? The “Five Girls Five Rocket Viral Video”, a dazzling piece of content that has become an instant sensation across platforms. This video, which showcases a group of five girls performing a synchronized and visually stunning stunt involving five rockets, has sparked a frenzy of shares, discussions, and imitations worldwide.
🔴 ➤► WATCH THE VIDEO HERE✅👉 https://shorturl.at/H0aYQ (VideoLink)
🔴 ➤► WATCH THE VIDEO HERE✅👉 https://shorturl.at/H0aYQ (VideoLink)
What is the “Rocket Viral Video”?
The Rocket Viral Video features five young women showcasing a perfectly choreographed sequence involving rockets. While the exact origins of the video remain unclear, its captivating visuals and flawless execution have made it impossible for viewers to look away. Each girl holds a rocket, launching them in precise coordination, creating a jaw-dropping spectacle that combines science, artistry, and entertainment.
The synergy of the performers, their confidence, and the dramatic backdrop have made the 5 girls 5 rocket breakout video an unforgettable watch. From Instagram reels to TikTok duets, the video has become a cultural touchstone for creativity and teamwork.
The Rise of the “5 Rocket 5 Girl Viral Video Breakout”
The viral journey of the 5 rocket 5 girl video breakout is a textbook case of how powerful social media can be. The video initially surfaced on platforms like YouTube and TikTok, where its unique concept and mesmerizing execution drew significant attention. Within hours, it began trending under hashtags like #RocketGirls, #5Rockets5Girls, and #ViralVideoBreakout.
As more users began sharing the video, it transcended language and geographical barriers. Memes, reaction videos, and recreations quickly followed, further amplifying its reach. Celebrities and influencers also jumped on the bandwagon, sharing their admiration and even attempting to replicate the stunt.
Why Did the “5 Girl 5 Rocket Video Breakout” Go Viral?
Several factors contributed to the five girls five rocket viral video becoming such a sensation:
Unconventional Concept: Combining five rockets and five performers in a synchronized act was something fresh and unexpected.
Visual Appeal: The video’s stunning visuals, paired with an engaging soundtrack, made it ideal for social sharing.
Relatability and Aspiration: The performers’ teamwork and determination resonated with audiences, inspiring them to recreate the stunt.
Global Accessibility: Short, impactful, and easy to share, the video was perfectly suited for platforms like TikTok, Instagram, and Twitter.
The Impact of the “Five Girls Five Rocket Viral Video”
Beyond its entertainment value, the video has sparked broader discussions about the power of collaboration and innovation in social media content. Schools, dance troupes, and content creators have used it as a template to inspire their own creations.
🔴 ➤► WATCH THE VIDEO HERE✅👉 https://shorturl.at/H0aYQ (VideoLink)
🔴 ➤► WATCH THE VIDEO HERE✅👉 https://shorturl.at/H0aYQ (VideoLink)
Moreover, the clip has demonstrated the potential for short-form content to captivate global audiences in an era of dwindling attention spans. The 5 girl 5 rocket video breakout serves as a reminder that creativity and originality remain key drivers of online virality.
Conclusion
The five girls five rocket viral video is more than just a trending clip; it’s a celebration of creativity, teamwork, and the boundless potential of digital platforms. As it continues to dominate timelines and inspire countless recreations, it cements its place as one of the most talked-about viral phenomena of the year.
If you haven’t seen it yet, prepare to be amazed by this unique and unforgettable display of talent and ingenuity!
🔴 ➤► WATCH THE VIDEO HERE✅👉 https://shorturl.at/H0aYQ (VideoLink)
🔴 ➤► WATCH THE VIDEO HERE✅👉 https://shorturl.at/H0aYQ (VideoLink)
NEW NEW
.
.
.
.
.
.
.
.
.
.
.
.
.
.
NEW NEW
.
.
.
.
.
.
.
.
.
.
.
.
.
.
NEW NEW
.
.
.
.
.
.
.
.
.
.
.
.
.
.
NEW NEW
.
.
.
.
.
.
.
.
.
.
.
.
.
.
cvcvc
|
gharchive/issue
| 2024-12-10T20:59:10 |
2025-04-01T06:40:30.482199
|
{
"authors": [
"Hussainuse",
"biliman001"
],
"repo": "styled-components/styled-components",
"url": "https://github.com/styled-components/styled-components/issues/4465",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
317850704
|
Perf with responsive styling and pure components
I'm a big fan of styled-system so first, thanks!
I was curious about the use of the array literal in responsive styling: <RoundedBox borderRadius={[1,2]} /> I created a little sandbox to show this: a responsive-styled pure component will re-render even though incoming props from the parent component do not change: https://codesandbox.io/s/m7p59mv3xx
The example is kind of contrived - I want the RoundedBox component to be pure so it's pretty obvious this will fail when you pass a prop with an array literal.
Perf should be measured before optimizing - however - given that object/array literals in renders is an infamous way of accidentally re-rendering in React, do you think there should be a note in the docs to mention that using responsive styling through array literals will break pure components? A simple suggestion is, if you're using a pure component, break out the array literal into a const...
Sorry for the delay, I haven't looking into the performance implications at this level, but I'd guess that it's negligible for most use-cases. Feel free to experiment on a branch with the benchmarks and let me know if you find anything interesting. I think a potential future solution for some of this might be in using babel plugins, but going to close this issue out for now
For what it's worth, we found a pretty significant performance impact with inline array literals, in that it effectively ruins pure. With a relatively large app, it becomes important to use pure to keep performance reasonable, and any prop literal is inherently a reference mismatch -- having a note in there might be valuable to others with larger-scale projects.
Also interested in which Babel plugins you're referring to, as we'd love to utilize them to improve our performance.
@AaronBuxbaum How did you identify this performance issue? Am trying to figure out a sporadic bug at the moment that I think might be related to performance but don't know how to identify.. any leads much appreciated
@AaronBuxbaum How did you identify this performance issue? Am trying to figure out a sporadic bug at the moment that I think might be related to performance but don't know how to identify.. any leads much appreciated
If you're seeing a lot of renders in your react dev tools, something is updating too much, and this is one possibility. I think your first step should be to figure out if your renders are expensive, or if you're re-rendering too much. Once you're there, you can dig into the problem -- multiple renders implies problems with things like this (which means that you can resolve a lot by hoisting object/function literals and using pure), but expensive few renders implies very different problems.
|
gharchive/issue
| 2018-04-26T01:59:14 |
2025-04-01T06:40:30.487882
|
{
"authors": [
"AaronBuxbaum",
"adiun",
"georgefeast",
"jxnblk"
],
"repo": "styled-system/styled-system",
"url": "https://github.com/styled-system/styled-system/issues/172",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1913005683
|
🛑 Cubari (Dynasty Scans) is down
In ce2a1ea, Cubari (Dynasty Scans) (https://cubari.moe/read/dynasty/the_engagement_of_the_disgraced_witch_and_the_cross_dressing_princess/?cache_buster=$GITHUB_RUN_NUMBER) was down:
HTTP code: 500
Response time: 430 ms
Resolved: Cubari (Dynasty Scans) is back up in 312dc9b after 12 minutes.
|
gharchive/issue
| 2023-09-26T08:43:10 |
2025-04-01T06:40:30.540070
|
{
"authors": [
"funkyhippo"
],
"repo": "subject-f/cubari-status-page",
"url": "https://github.com/subject-f/cubari-status-page/issues/118",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2559783391
|
🛑 Cubari (MangaSee) is down
In 92eb439, Cubari (MangaSee) (https://cubari.moe/read/mangasee/One-Piece/?cache_buster=$GITHUB_RUN_NUMBER) was down:
HTTP code: 500
Response time: 306 ms
Resolved: Cubari (MangaSee) is back up in 5e879b2 after 1 minute.
|
gharchive/issue
| 2024-10-01T17:10:16 |
2025-04-01T06:40:30.542661
|
{
"authors": [
"funkyhippo"
],
"repo": "subject-f/cubari-status-page",
"url": "https://github.com/subject-f/cubari-status-page/issues/3274",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2074728501
|
🛑 Cubari (Gist) is down
In 34f5103, Cubari (Gist) (https://cubari.moe/read/gist/OPM/?cache_buster=$GITHUB_RUN_NUMBER) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Cubari (Gist) is back up in 258f0ad after 23 minutes.
|
gharchive/issue
| 2024-01-10T16:25:30 |
2025-04-01T06:40:30.545168
|
{
"authors": [
"funkyhippo"
],
"repo": "subject-f/cubari-status-page",
"url": "https://github.com/subject-f/cubari-status-page/issues/769",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2561227063
|
⚠️ Guya.moe Proxied has degraded performance
In 10e7b04, Guya.moe Proxied (https://ice.guya.moe/) experienced degraded performance:
HTTP code: 200
Response time: 12904 ms
Resolved: Guya.moe Proxied performance has improved in 173a2e9 after 2 hours, 27 minutes.
|
gharchive/issue
| 2024-10-02T10:35:28 |
2025-04-01T06:40:30.547597
|
{
"authors": [
"funkyhippo"
],
"repo": "subject-f/guya-status-page",
"url": "https://github.com/subject-f/guya-status-page/issues/181",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1466851033
|
No color decorators for syntax from SublimeText/Tailwind CSS
When I set the syntax of my stylesheet files in Sublime Text to one of the default ones (CSS, CSS3) the color decorators work just fine. For working with Tailwind CSS I use https://github.com/SublimeText/TailwindCSS/ because directives like @apply and nesting are not supported by the other two. However, color decorators don't work with that syntax.
I tried to change my LSP-tailwindcss settings to include the source.css.tailwindcss scope but it still doesn't work:
{
"selector": "source.jsx | source.js.react | source.js | source.tsx | source.ts | source.css.tailwind | source.css | source.scss | source.less | text.html.vue | text.html.svelte | text.html.basic | text.html.twig | text.blade | embedding.php | text.html.rails | text.html.erb | text.haml | text.jinja | text.html.elixir",
"languages": [
{
"languageId": "css",
"scopes": ["source.css", "source.css.tailwind"],
"syntaxes": [
"Packages/CSS/CSS.sublime-syntax",
"Packages/CSS3/CSS3.sublime-syntax",
"Packages/Tailwind CSS/Tailwind CSS.sublime-syntax"
]
}
]
}
Do I have the wrong setup? How do I get the "Tailwind CSS" syntax to show color decorators as well?
Ouptut from "Troubleshooting: LSP-tailwindcss"
Troubleshooting: LSP-tailwindcss
Version
LSP: 1.20.0
Sublime Text: 4141
Server Test Run
exit code: 0
output
Server Configuration
command
[
"/usr/bin/node",
"/home/jnns/.cache/sublime-text/Package Storage/LSP-tailwindcss/18.7.0/language-server/node_modules/@tailwindcss/language-server/bin/tailwindcss-language-server",
"--stdio"
]
shell command
/usr/bin/node "/home/jnns/.cache/sublime-text/Package Storage/LSP-tailwindcss/18.7.0/language-server/node_modules/@tailwindcss/language-server/bin/tailwindcss-language-server" --stdio
selector
source.jsx | source.js.react | source.js | source.tsx | source.ts | source.css | source.scss | source.less | text.html.vue | text.html.svelte | text.html.basic | text.html.twig | text.blade | embedding.php | text.html.rails | text.html.erb | text.haml | text.jinja | text.html.elixir
priority_selector
(source.css)|(source.css.tailwind)
init_options
{}
settings
{
"tailwindCSS": {
"classAttributes": [
"class",
"className",
"ngClass"
],
"colorDecorators": true,
"emmetCompletions": false,
"experimental": {
"classRegex": []
},
"files": {
"exclude": [
"**/.git/**",
"**/node_modules/**",
"**/.hg/**"
]
},
"includeLanguages": {},
"lint": {
"cssConflict": "warning",
"invalidApply": "error",
"invalidConfigPath": "error",
"invalidScreen": "error",
"invalidTailwindDirective": "error",
"invalidVariant": "error",
"recommendedVariantOrder": "warning"
},
"rootFontSize": 16,
"showPixelEquivalents": true,
"validate": true
}
}
env
{
"PATH": "/usr/bin:/usr/bin:"
}
Active view
File name
/home/jnns/<...>/forms.css
Settings
{
"auto_complete_selector": "meta.tag, source - comment - string.quoted.double.block - string.quoted.single.block - string.unquoted.heredoc",
"lsp_active": true,
"syntax": "Packages/Tailwind CSS/Tailwind CSS.sublime-syntax"
}
base scope
source.css.tailwind
## Project / Workspace
- folders
```json
[
"/home/jnns/<...>"
]
is project: True
project data:
{}
LSP configuration
{
"clients": {
"LSP-pyright": {
"enabled": false
}
}
}
System PATH
/home/jnns/.local/bin
/usr/local/bin
/usr/local/sbin
/usr/bin
/usr/sbin
Add this to LSP-tailwindcss settings.
"initializationOptions": {
"userLanguages": {
"tailwind": "css",
}
},
Note that this is something that should be fixed in this package still.
|
gharchive/issue
| 2022-11-28T18:02:12 |
2025-04-01T06:40:30.600085
|
{
"authors": [
"jnns",
"rchl"
],
"repo": "sublimelsp/LSP-tailwindcss",
"url": "https://github.com/sublimelsp/LSP-tailwindcss/issues/51",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1279699910
|
Fix image tagging
Images should be cross-tagged without v prefix, make sure it's so.
Signed-off-by: Mike Kolesnik mkolesni@redhat.com
Backported as #409
|
gharchive/pull-request
| 2022-06-22T07:35:24 |
2025-04-01T06:40:30.619654
|
{
"authors": [
"mkolesnik"
],
"repo": "submariner-io/releases",
"url": "https://github.com/submariner-io/releases/pull/407",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
716503753
|
Document everything OVN
The OVN support will bring some changes that need to be documented from the architecture point of view.
See: https://github.com/submariner-io/enhancements/pull/11
Related-Issue: https://github.com/submariner-io/submariner/issues/778
I was discussing this with @sridhargaddam, and wanted to summarize our thinking here:
At least from a user perspective, OVN should "just work". We do need to:
Add OVN to the list of CNI support matrix: https://submariner.io/getting_started/#support-matrix
Mention that for OVN-based clusters there is no need to open 4800 UDP ports (which normally being used for vxlan-submariner)
Document known issues/limitations (for e.g, OVN is not supported with Globalnet): https://submariner.io/operations/known_issues/
We also need to update some of the architecture pages (and diagrams?), in particular the Route Agent: https://submariner.io/getting_started/architecture/route-agent/
Document known issues/limitations (for e.g, OVN is not supported with Globalnet): https://submariner.io/operations/known_issues/
PR raised:
https://github.com/submariner-io/submariner-website/pull/402
* Add OVN to the CNI support matrix: https://submariner.io/getting_started/#support-matrix
PR: https://github.com/submariner-io/submariner-website/pull/403
|
gharchive/issue
| 2020-10-07T12:53:56 |
2025-04-01T06:40:30.625557
|
{
"authors": [
"mangelajo",
"nyechiel",
"sridhargaddam"
],
"repo": "submariner-io/submariner-website",
"url": "https://github.com/submariner-io/submariner-website/issues/304",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1305998326
|
Decode extrinsic error at ingestion
Hi guys,
For the moment an extrinsic module error is coming in the form
{
"__kind": "Module",
"value": {
"error": "0x06000000",
"index": 9
}
}
Unfortunately, this is not very helpful when we try to show it in the UI of an explorer. It would be nice to have it in the decoded form like what @polkadotjs/api api.registry.findMetaError returns:
{
args: [],
docs: [ 'Contract trapped during execution.' ],
fields: Type(0) [ registry: TypeRegistry {}, initialU8aLength: 1 ],
index: 11,
method: 'ContractTrapped',
name: 'ContractTrapped',
section: 'contracts'
}
Is this something you guys plan on supporting?
Hey! It's likely not feasible to support on the archive explorer side as it simply presents the raw data saved by the ingester. I recommend setting up a separate squid for these purposes and then decode the error using
(ctx._chain as any).scaleCodec.decodeBinary(...)
Take a look at this squid to get an idea of how to handle events and extrinsic at the squid side
https://github.com/subsquid/talisman-squidtest/tree/firesquid-migration
Hey, thanks for the explanation!
There's no problem subscribing directly to succeeded and failed extrinsics in the squid but the error returned is still not decoded. Is it possible to expose a method like ctx._chain.decodeError() similar to ctx._chain.decodeCall() that could be used in the squid processor to get the decoded error? If I understood well from this article: https://wiki.polkadot.network/docs/maintain-errors#polkascan-and-subscan, the index 9 in my example above indicates the index of the pallet as described in the chain metadata and the error 0x06000000 indicates the index of the error in the error field of the pallet description. Since the squid archive explorer already has the chain metadata, it should be relatively easy to resolve the error?
We'll look into supporting this on our end so I'll close the issue.
|
gharchive/issue
| 2022-07-15T12:48:24 |
2025-04-01T06:40:30.636565
|
{
"authors": [
"XY-Wang",
"dzhelezov"
],
"repo": "subsquid/squid-sdk",
"url": "https://github.com/subsquid/squid-sdk/issues/76",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
66588261
|
Update Travis CI tests to use newer Node.js versions
As the ^1.4.1 syntax used for currently including the escodegen module was first introduced in npm 1.3, the npm 1.2 verison included with node 0.8.x fails at installing the dependency.
The escodegen module itself also in version 1.3.3 changed from requiring node >=0.4.0 to requiring node >=0.10.0, so it explicitly doesn't support 0.8.x anymore either.
The pull request to update the required acorn dependency ran into this issue and since that very same PR preferably should use the caret operator for the dependency it would be better to stop testing in node 0.8.x and start testing in newer node versions instead.
Related PR: https://github.com/substack/node-detective/pull/41
Thanks! Fixed by https://github.com/substack/node-detective/pull/46
|
gharchive/pull-request
| 2015-04-06T11:43:10 |
2025-04-01T06:40:30.640007
|
{
"authors": [
"voxpelli",
"zertosh"
],
"repo": "substack/node-detective",
"url": "https://github.com/substack/node-detective/pull/43",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
362276149
|
Find and Replace: Select All in one of the inputs should not select all in the document.
At the moment, keyboard events bubble up to the editor which handles certain combos, such as CtrlOrCommand-a, having the effect that Select All is applied to the area where the internal selection is.
We should either prevent all events from bubbling up, or consume certain combos.
Yeah I saw that too. I wonder if it would make sense to use Substance TextInput instead of the native input fields. Maybe then we have things better under control? (also enabling commands cmd+g etc.)
Command-G is more relevant as a short-cut for next, when the cursor is inside the doc.
Have you noticed, that you can travers with ENTER? Shift-Enter is not handled yet, though.
Don't think that Substance TextInput is solving this problem for us.
In this case it is really important to bind a keydown handler for CtrlOrCommand+a on the input. The SelectAllCommand is working only for Surfaces.
Maybe I didn't understand your suggestion... It would make sense to have a native input component, which is having all the common handlers. That's true.
If Substance TextInput is doing this already, I'd suggest to move it into the kit.
But don't use it for a model! :)
Surfaces are bound to models. So the inputs in the FaRD(ialog) ;) are not candidates for Surfaces.
The would need a detached (temporary model) of course. Anyways.. as said it was just a thought, since now this is the only place where we have <input> elements. :)
That is an overkill. We are perfectly able to create a native input with enough handlers attached to be compliant with the rest of the Substance actors.
Still it makes sense to have such a thing, so we do not need to reinvent it each time.
now this is the only place where we have elements.
The key difference is that FaRD is app chrome, not content.
Fair enough. Let's stick with what we have then.
Yeah. Still I am in favor of creating a Component around Input which behaves nicely. :)
Fixed via https://github.com/substance/texture/commit/64860304b46f5df232771efb81ddc059f934d298
|
gharchive/issue
| 2018-09-20T16:54:59 |
2025-04-01T06:40:30.646806
|
{
"authors": [
"michael",
"oliver----"
],
"repo": "substance/texture",
"url": "https://github.com/substance/texture/issues/912",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
831086983
|
add rubocop to travis
Seems like a shame to have defined rubocop config (and implemented rubocop changes) and not enforce it.
I've pinned the version of rubocop to the latest version to prevent arbitrary correctness changes.
Thank you so much for contributing.
We run Rubocop via Code Climate, though, (see https://codeclimate.com/github/substancelab/route_downcaser/pull/47 for an example). I wonder if it is worth running it on Travis as well?
Ah ok, I didn't see that. But it would be helpful to make it a development dependency pinned to the version being used. (just to make it easier to develop on). Shall I adjust the PR to remove travis and keep the dev dependency? Or if you're not convinced, you can just close it :)
Have made this change in case helpful
I am toying with the idea of switching to https://github.com/testdouble/standard instead of raw Rubocop. It would require us to not rely on Code Climate, but that's not a dealbreaker for me.
What do you think of this instead?
I don't see a problem with testdouble/standard -- the standards seem more sensible than rubocop defaults.
And in terms of moving the rubocop from codeclimate to travis -- I would generally prefer this as I feel like a codeclimate fail is advisory (and needs informed judgement) whereas rubocop fail (like test suite fail) tend to be more absolute.
We've switched to standard by now.
|
gharchive/pull-request
| 2021-03-14T07:21:25 |
2025-04-01T06:40:30.650771
|
{
"authors": [
"koppen",
"timdiggins"
],
"repo": "substancelab/route_downcaser",
"url": "https://github.com/substancelab/route_downcaser/pull/47",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
222682008
|
Use programmaitc SS API instead of REST API from Hub Manager module
Currently there are some classes in hub manager module that address SS API via REST. However we have programmatic code API that could be used. REST API is for external clients
Done
|
gharchive/issue
| 2017-04-19T09:51:47 |
2025-04-01T06:40:30.652909
|
{
"authors": [
"Dilshat"
],
"repo": "subutai-io/base",
"url": "https://github.com/subutai-io/base/issues/1817",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
139073456
|
[UI] [Monitoring] Too many values impacts usability.
For replay this improvement you need:
Test case
Go to SS and login under any account.
Go to Monitoring page.
Improvement: On the Monitoring graphs (i.e. disk) usability is impacted by excessive values. The popup displayed is out of control. Let's condense some of the values into rows where it makes sense. For example here:
On this popup let's condense the "avialable", "used" and "total" metrics in a nice way onto a single row instead of have each one of these on its own row.
Tested.
|
gharchive/issue
| 2016-03-07T19:40:47 |
2025-04-01T06:40:30.655493
|
{
"authors": [
"NikolaiB",
"akarasulu"
],
"repo": "subutai-io/base",
"url": "https://github.com/subutai-io/base/issues/189",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1500555444
|
Add DellyBoard theme
DellyBoard
https://github.com/Dellyare/DellyThemes
Theme for keyboard in Apple style
Checklist
Failure to complete this checklist or deleting boxes from the checklist will result in the closing of your pull request unless this is a theme update. Please write any comments regarding this checklist at the bottom of your pull request.
Check every box.
[x] I am the original author of this theme or have permission from the original author to make this pull request.
[x] All copyright of this theme's contents belong to the listed author or is cited in the repository linked above.
[x] This theme's target has been marked appropriately and only styles said target.
[x] This theme works properly on the latest versions of SteamOS for Steam Deck, decky-loader and SDH-CssLoader.
[x] This theme only uses * or !important if absolutely necessary.
[x] This theme is under 4MB in size and uses the least disk space possible.
[x] This theme's preview image does not include text unless it is necessary to describe changes that can be made.
[x] This theme is safe for work and does not contain any sexual, drug-related, or profane content.
[x] This theme prefixes any CSS variables with a unique identifier.
Check one box.
[x] I am not bundling a part of another theme with this theme to encourage mixing and matching themes.
[ ] Themes included with this theme are toggleable using a patch.
Check one box.
[x] This is a keyboard theme applied to the default keyboard.
[ ] This is a system-wide theme applied to the default keyboard. The keyboard is toggleable using a patch.
[ ] This theme does not target the keyboard.
Hi, would you mind joining this discord? We are planning to migrate all themes to a custom built site soon, and want to discuss how that should happen.
For the PR, I'll review it tomorrow
|
gharchive/pull-request
| 2022-12-16T16:59:56 |
2025-04-01T06:40:30.662646
|
{
"authors": [
"Dellyare",
"suchmememanyskill"
],
"repo": "suchmememanyskill/CssLoader-ThemeDb",
"url": "https://github.com/suchmememanyskill/CssLoader-ThemeDb/pull/140",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
483095576
|
"Community Guidelines" footer link clicked on mobile loads that forum page scrolled down rather than at its top
Same issue that once existed for the main forum link, turbo-links whitelist or something?
Closed by https://github.com/sudara/alonetone/commit/b2a0979f7075e6b56652e80eb7606e5bdf0b1999
|
gharchive/issue
| 2019-08-20T21:37:42 |
2025-04-01T06:40:30.668289
|
{
"authors": [
"ofsound",
"sudara"
],
"repo": "sudara/alonetone",
"url": "https://github.com/sudara/alonetone/issues/684",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2318143646
|
chromadb问题
大神 好, 因为docker镜像太大 ,所以我目前采用的是 "Use with Nuxt 3 Development Server" 安装方式 。 我目前只想把本地RAG跑通 ,因此用了 qwen:0.5b + nomic-embed-text 的组合 ,我用了两台虚拟机 ,一个安装:ollama , 一个安装了:chat-ollama + chromadb + redis . 我的.env大概如下:
问题 : 是因为这个域名导致我不能保存向量数据吗 , 我被卡住了,麻烦大神指导下 ,谢谢。
|
gharchive/issue
| 2024-05-27T03:20:58 |
2025-04-01T06:40:30.690016
|
{
"authors": [
"Raul1044"
],
"repo": "sugarforever/chat-ollama",
"url": "https://github.com/sugarforever/chat-ollama/issues/487",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
378283075
|
WOS client should log non-200 responses
It currently does not check the response code or log anything. As a result, the user has no way of knowing if there was an error.
|
gharchive/issue
| 2018-11-07T13:16:38 |
2025-04-01T06:40:30.758261
|
{
"authors": [
"justinlittman"
],
"repo": "sul-dlss/rialto-etl",
"url": "https://github.com/sul-dlss/rialto-etl/issues/285",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1179537375
|
🛑 SISTEMA ROYAL-HOLIDAY is down
In 155cbf9, SISTEMA ROYAL-HOLIDAY (https://www.royalholiday.com.ar/reservas/admin/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: SISTEMA ROYAL-HOLIDAY is back up in a0c6d0f.
|
gharchive/issue
| 2022-03-24T13:42:21 |
2025-04-01T06:40:30.800794
|
{
"authors": [
"sumito74"
],
"repo": "sumito74/upptime",
"url": "https://github.com/sumito74/upptime/issues/416",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1213280047
|
🛑 SISTEMA ROYAL-HOLIDAY is down
In ece920f, SISTEMA ROYAL-HOLIDAY (https://www.royalholiday.com.ar/reservas/admin/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: SISTEMA ROYAL-HOLIDAY is back up in 4064cbb.
|
gharchive/issue
| 2022-04-23T11:45:37 |
2025-04-01T06:40:30.803266
|
{
"authors": [
"sumito74"
],
"repo": "sumito74/upptime",
"url": "https://github.com/sumito74/upptime/issues/820",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1289681304
|
一些3.4.0新增功能的语法检查异常
单独写lua没能复现
另外有没有强制修改类型的配置方式,比如
`---@class A
local a = {}
---@class B:A
---@type string|number
local c = 1
---comment
---@return A
---@return string|number
local function func1()
return a, c
end
---comment
---@return B
---@return number
local function func2()
local ta, tb = func1()
return ta, tb
end`
在这种情况下我怎么去配置使ta、ta符合返回值类型?
感觉上可以加个配置还忽略可控类型验证,不然现有代码出现太多错误了
可以使用 ---@cast var type 和 --[[@as type]] 转换类型,我晚点写个文档。
所有分支都return了
你再看看?
呃,最后一个不是else,我的错
---@type UnityEngine.Vector3
local x = Vector3.zero
x = x + x 已显式定义变量的类型为 `UnityEngine.Vector3` ,不能再将其类型转换为 `number`
算术运算都变成number了
另外number和integer不兼容,这个检查太头疼了
另外number和integer不兼容,这个检查太头疼了
设置里有让 number 和 integer 兼容的选项
算术运算都变成number了
3.5.0 会支持重载运算符,可以解决这个问题
单独写lua没能复现
你在原来的代码上删除业务代码,只保留一个结构看看
我家里没能复现,周一上班公司的提示也没了
诊断速度变慢,写代码时十几秒才能诊断完
诊断速度变慢,写代码时十几秒才能诊断完
请开一个新的issue并提供日志与可复现此问题的演示工程
500个应该是没问题的,我的工作项目有1500+个文件也无压力,应该是有什么隐藏的bug导致的,这个只有用示例调试才好查
500个应该是没问题的,我的工作项目有1500+个文件也无压力,应该是有什么隐藏的bug导致的,这个只有用示例调试才好查
这个版本我们项目有 800多个诊断问题 会不会和这个有关系,我们有1500多个类型注解
有没有方法实现可选返回值?
有没有方法实现可选返回值?
---@return number?
有没有方法实现可选返回值?
---@return number?
这种怎么处理,手动写个 return nil?
---@return BattleRecordOne?
function BattleEnemy:GetBattleRecord()
if self.battleRecord then
return self.battleRecord.record
end--这里提示要返回值
end
---@return BattleRecordOne?
function BattleEnemy:GetBattleRecord()
if self.battleRecord then
return self.battleRecord.record
end--这里提示要返回值
end
#1257
local tm = os.date("*t", os.time())
local t = os.time(tm) -- 不能将 `string|osdate` 赋给参数 `osdate?`。
local tm = os.date("*t", os.time())
local t = os.time(tm) -- 不能将 `string|osdate` 赋给参数 `osdate?`。
开个单独的issue,以后版本改为分析参数来决定返回值
Is there any other issue?
|
gharchive/issue
| 2022-06-30T06:52:47 |
2025-04-01T06:40:30.817715
|
{
"authors": [
"Whimsyduke",
"septbr",
"sniper00",
"sumneko"
],
"repo": "sumneko/lua-language-server",
"url": "https://github.com/sumneko/lua-language-server/issues/1260",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
840569449
|
"Lua.workspace.ignoreDir" doesn't ignore my file
I use custom format to store my settings, similar to json but in lua notation. I called it "luat" and when I added "*.luat" pattern to "Lua.workspace.ignoreDir" it is still highlighted red:
The language service currently has a setting. If a lua file is ignored, but you actively open the file, the language service will temporarily assume that the file has not been ignored (in order to provide functions such as automatic completion and floating prompts). But this does cause inconvenience in some cases, and I will consider how to improve it.
|
gharchive/issue
| 2021-03-25T05:23:03 |
2025-04-01T06:40:30.819789
|
{
"authors": [
"sumneko",
"winterwolf"
],
"repo": "sumneko/lua-language-server",
"url": "https://github.com/sumneko/lua-language-server/issues/465",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
637449351
|
Which is PHPUnit coverage type, C0 or C1 or C2?
Coverage type:
Type
Name
Description
C0
Statement coverage
Tỉ lệ bao phủ câu lệnh
C1
Branch coverage
Tỉ lệ bao phủ nhánh - Mỗi nhánh của if phải được test ít nhất 1 lần
C2
Condition Coverage
Tỉ lệ bao phủ điều kiện - Mỗi điều kiện trong lên if đều được test 2 case: 1 là true 2 là false
[ ] 1. Dựa vào kết quả coverage của Issue #1, kết luận PHPUnit coverage thuộc loại nào?
[ ] 2. Câu lệnh ở đây là lệnh trong function hay tính cả class property?
[ ] 3. Test case tests/C0/DumpExampleTest.php đã cover 100% C0 coverage, nhưng đã đủ test case cho class DumpExample chưa?
Theo kết quả test thì hiện tại coverage đang thuộc C0 do khi test class DumpExample
Test đã chạy qua hết các lệnh của function examine.
Chưa đạt được C1 do funtion examine không có nhánh else.
Chưa đạt C2 do chưa test đk false cho funtion examine
Câu lệnh chỉ tính lệnh trong function, không tính đến các class property.
PHP Unit coverage thuộc loại C0, Vì:
Test đã chạy qua hết các câu lệnh của function examine
Chưa đạt được C1 do chưa test case else của function examine
Chưa đạt được C2 do điều kiện if mới chỉ được test case true, chưa được test case false
Câu lệnh ở đây là lệnh trong function, không tính đến class property
@taipt-0504
Chưa đạt được C1 do funtion examine không có nhánh else
Em thấy chưa chuẩn lắm anh nhể, vì if thì luôn ngầm định có nhánh else.
if ($condition) {
// ...
}
=> 2 test cases: chạy vào trong block if và không chạy vào trong block if
if ($condition) {
// ...
} else {
// ...
}
=> 2 test cases: chạy vào trong block if và chạy vào trong block else
@taipt-0504 @huongvnq-1722
A/e update thêm câu 3 nhé
C0 do đã phủ toàn bộ xử lý nhưng chưa chú ý đến điều kiện.
Em không hiểu lắm class property là gì :p. Đang hiểu nó là kiểu getAttribute, ... có đúng không ạ?
Đủ test case theo điều kiện C0 nhưng C1, C2 chưa đủ
@nganltb-1258
PHPUnit nó tính coverage là C0 (Statement coverage) ấy, chưa support tính C1, C2 coverage
Ý anh là class property hay class field, VD như public $fillable = []
Trường hợp này 100% code coverage nhưng vẫn bị lack case đúng không, gây ra bug nếu ko test đủ
|
gharchive/issue
| 2020-06-12T03:01:52 |
2025-04-01T06:40:30.842987
|
{
"authors": [
"huongvnq-1722",
"nganltb-1258",
"taipt-0504",
"tuanpt-0634"
],
"repo": "sun7pro/phpunit-training-coverage",
"url": "https://github.com/sun7pro/phpunit-training-coverage/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1687124330
|
棋譜形式一括変換機能
せっかく CSA や KIF, SFEN, USI に対応したので、棋譜一括変換機能を作っても良いと思う。
ただ、 Electron 将棋に搭載すべき機能なのか微妙なのと、自分が頻繁に使うわけではないのでモチベーションもそこまで高くはない。
https://github.com/yaneurao/YaneuraOu/wiki/定跡の作成
やねうら王は拡張子 .sfen で SFEN (というか USI) を行ごとに記載しているらしい。
|
gharchive/issue
| 2023-04-27T15:50:19 |
2025-04-01T06:40:30.847489
|
{
"authors": [
"sunfish-shogi"
],
"repo": "sunfish-shogi/electron-shogi",
"url": "https://github.com/sunfish-shogi/electron-shogi/issues/482",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
155826301
|
About me and Contact sections are not clearly delineated on Home page
On home navigation, about me and contacts do not have distinctive sections.
Created headers to separate these items and display them with very clear sections.
|
gharchive/issue
| 2016-05-19T20:22:32 |
2025-04-01T06:40:30.854423
|
{
"authors": [
"sungkim3"
],
"repo": "sungkim3/portfolio",
"url": "https://github.com/sungkim3/portfolio/issues/18",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
610730522
|
Hard to find docs
Do docs exist for this project? If so some sort of link on the github page (maybe in the Readme) would be helpful, as I can't obviously find one.
In case you haven't found them already they are here: https://docs.sunpy.org/projects/radiospectra/en/latest/ :rofl:
|
gharchive/issue
| 2020-05-01T12:36:40 |
2025-04-01T06:40:30.916668
|
{
"authors": [
"Cadair",
"dstansby"
],
"repo": "sunpy/radiospectra",
"url": "https://github.com/sunpy/radiospectra/issues/36",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
479723389
|
Update to official latest mergeable
Our mergeable is full of bugs. We should update our config and move to the official hosted version.
Well, I thought mergeable had implemented my feature, but it's actually still sitting in the issue queue. Fuck.
https://github.com/mergeability/mergeable/issues/145
https://github.com/mergeability/mergeable/issues/62
Tagging as upstream. I tried clobbering something up for the proposed "unless" operator, but it's a bit non-trivial.
See #434 .
Blocked on https://github.com/mergeability/mergeable/pull/227
|
gharchive/issue
| 2019-08-12T15:54:29 |
2025-04-01T06:40:30.932566
|
{
"authors": [
"roblabla"
],
"repo": "sunriseos/SunriseOS",
"url": "https://github.com/sunriseos/SunriseOS/issues/414",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
547299335
|
v2.50 release
What features and issues are more urgent for version 2.5.0? And what is your plan for 2.5.0?
@dota17 @jjenkov @ChristianWilkie
I like you guys feedback on this. In my mind it is
get jenkins up and running with code coverage
its been a while since last release, so get a release out soon
maybe we should upgrade to java 8 ,just like @jjenkov says java 6 or java 7 is too old #169
I would say, that the highest priority should be to just get a release out. As far as I can see, v. 2.4.0 is 3 years old. I would go for a minimum of functionality in 2.5.0, just to bring SuperCSV "back to life" .
If upgraded to J8, would SuperCSV work correctly under J9+? My application uses SCSV, and a problem that I run into is its incompatibility with J9+ just generally. It would be a big upgrade for me if that issue was resolved. (as always, thank you for maintaining this!)
@DraqueT Is the approach to make a J9 specific release? or will the https://github.com/super-csv/super-csv/pull/152 suffice?
Hi, @kbilsted
I don`t think super-csv version 2.5.0 make a J9 specific release. I think it is enough that ensuring super-csv 2.5.0 can be imported and used in java9 module application(#138 ).
Because java9 modules is important and influential. When we upgrade Java 9, we need to design reasonable modules for super-csv,
And
Hi, @kbilsted
Are you plan to upgrade to java8 in 2.5?
If so, we can incorporate super CSV java8 into super CSV and address #152. that will ensuring super-csv can be imported and used in java9 module application.
Adding support for escaping quotes using \" instead of leaving this as unhandled seems reasonable, very hard to recover from errors based on this otherwise.
@dota17 I'd vote for caution. Can you make things work in a side branch for JDK 9 for now and get experince with it with some users?
isn't java8-specifc jar a part of the existing mvn build?
I can probably help out on getting this to compile on Java 8. This is used on one of the Apache projects I'm working on and there are some bug fixes I could really stand to have. Glad to see active development again.
The master branch does build just fine on Java 8 though. Not sure if what some of the other issues actually were.
@shawnweeks thanks. I'll install jdk8 and retry
@shawnweeks thanks. I'll install jdk8 and retry
I just ran "mvn clean install" using adoptopenjdk on macos. All the test cases ran and everything. Jars appear to be there in the build not sure what else needs to be tested but it should at least build until we move the compile versions forward.
Hi, @kbilsted
I am sorry. Can you change the name of the branch from java9 to java8 :).
sorry yes. I have been away for some days. Had some business I had to do.
Any chance for a 2.5.0 release soon?
What issues/things are still open to do so?
Any chance for a 2.5.0 release soon?
What issues/things are still open to do so?
Same question here, do you have a 2.5.0 release soon or is anything open here ?
|
gharchive/issue
| 2020-01-09T07:26:15 |
2025-04-01T06:40:31.406733
|
{
"authors": [
"DraqueT",
"apatrida",
"dota17",
"edu-de",
"jjenkov",
"kbilsted",
"patric-r",
"shawnweeks"
],
"repo": "super-csv/super-csv",
"url": "https://github.com/super-csv/super-csv/issues/176",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
122486429
|
Delete File
Just want to know whether we can delete a file using supercsv.
First created the ICsvBeanWriter
Write the data using ICsvBeanWriter
Close the ICsvBeanWriter
even if the data is not available in the step 2 we are creating the file and their is not way to delete the file.
any idea how can we achieve this?
Thanks,
Nagaraju
HI i'm not fully sure i understand the problem. Try submitting some Java code for how you think you could use the framework. Why not simply delete the file yourself? why start wrtiting to a file before you check if you have data to write?
ICsvBeanWriter is just an abstraction for writing any bean as CSV row. CsvBeanWriter class pushes CSV String to Writer object which is given in constructor. You probably uses FileWriter class which creates file at the beginning.
You have two solutions:
You can use StringWriter which writes data to memory. At the end, after writing, you can check the length of the String and save it to file if needed. Consumes time and memory but doesn't create file if it is not needed.
You can close the stream and remove file if it is empty.
Summary: does not look like an issue in super-csv library. I vote to close this.
+1 @ZioberMichal
|
gharchive/issue
| 2015-12-16T11:45:31 |
2025-04-01T06:40:31.411279
|
{
"authors": [
"ZioberMichal",
"kbilsted",
"nagraj321"
],
"repo": "super-csv/super-csv",
"url": "https://github.com/super-csv/super-csv/issues/87",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
94581798
|
Add java.time parsers and formatters
Fixes #53 . I have tried to match the structure of the java.time package, which diverges from joda.time. I also tried to design locale-independent tests that also use the current locale.
Hi. It looks really good! I've noticed that you are using different imports order than it is in project.
Project's order:
static
java.*
org.*
In your commit:
org.*
java.*
static
I think that we should use the same order.
I’ll see what I can do about the imports. Importing the eclipse configuration does not seem to solve all format issues :-)
This is a reason why I've crated this issue #61 :smirk:
|
gharchive/pull-request
| 2015-07-12T17:05:25 |
2025-04-01T06:40:31.414352
|
{
"authors": [
"ZioberMichal",
"ludovicofischer"
],
"repo": "super-csv/super-csv",
"url": "https://github.com/super-csv/super-csv/pull/60",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1987373388
|
🛑 Cabinet Place Online is down
In 7c6f6cd, Cabinet Place Online (https://www.cabinetplaceonline.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Cabinet Place Online is back up in 3721f57 after 6 minutes.
|
gharchive/issue
| 2023-11-10T10:42:14 |
2025-04-01T06:40:31.503331
|
{
"authors": [
"jflores1"
],
"repo": "superwebpros/uptime",
"url": "https://github.com/superwebpros/uptime/issues/566",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1274325740
|
远程唤醒
基于RT-Thread+RA6M4的远程开机助手概要
https://club.rt-thread.org/ask/article/e086a4b93cd3b143.html
WOL 远程唤醒,需要网卡带电支持。
进去BIOS打开这个功能
参考
https://mp.weixin.qq.com/s/9p_LTT-7w14MqGiDDNRVmA
设置一下电源按钮,然后记录一下mac地址
然后关闭计算机
然后进入小程序
输入计算机的mac地址即可
#小程序://网络唤醒WakeOnLan/mPDLCY0OakAkNif
https://blog.csdn.net/m0_37624610/article/details/108196206
|
gharchive/issue
| 2022-06-17T01:15:40 |
2025-04-01T06:40:31.507665
|
{
"authors": [
"supperthomas"
],
"repo": "supperthomas/bluetoothlover_doc",
"url": "https://github.com/supperthomas/bluetoothlover_doc/issues/342",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1396704368
|
blackjackgame
this will help to learn the javascript concept by making this game
Thanks for PR. Merged.
|
gharchive/pull-request
| 2022-10-04T18:50:07 |
2025-04-01T06:40:31.510532
|
{
"authors": [
"mdfaisalahmed057",
"surajondev"
],
"repo": "surajondev/JavaScript",
"url": "https://github.com/surajondev/JavaScript/pull/60",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
235856143
|
Only health at container level
Right now pod spec allows you to do both livenessProbe and readinessProbe and since both the fields represent the same data strcuture behind the scene which is called probes.
So a container will look like the following:
containers:
- image: wordpress:4
health:
httpGet:
path: /
port: 80
initialDelaySeconds: 120
timeoutSeconds: 5
the pod that is generated out of this will look like this:
containers:
- image: wordpress:4
livenessProbe:
httpGet:
path: /
port: 80
initialDelaySeconds: 120
timeoutSeconds: 5
readinessProbe:
httpGet:
path: /
port: 80
initialDelaySeconds: 120
timeoutSeconds: 5
So essentially you just specify health field which is probe type and then it is populated into readinessProbe and livenessProbe in the output. But if user provides the fields readinessProbe or livenessProbe explicitly then health field is ignored! Those fields are used directly.
agreed, with one exception. lets fail if there is health field with readinessProbe or livenessProbe defined
|
gharchive/issue
| 2017-06-14T12:02:15 |
2025-04-01T06:40:31.513283
|
{
"authors": [
"kadel",
"surajssd"
],
"repo": "surajssd/opencomposition",
"url": "https://github.com/surajssd/opencomposition/issues/23",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2237591398
|
feat: Added a service for displaying snacks
Checklist
A service has been written that will help you use flutter_easy_dialogs to display snacks "out of the box".
https://github.com/surfstudio/surf-flutter-app-template/assets/39735343/142d53c3-d05f-4d2c-b4de-421ab6411766
[x] Have you added an explanation of what your changes do and why you'd like us to include them?
[x] Is there an existing issue for this PR?
link issue here (use keywords like fix, close, resolve etc. if necessary)
[x] Have the files been linted and formatted?
[ ] Have the docs been updated to match the changes in the PR?
[ ] Have the tests been updated to match the changes in the PR?
[x] Attached videos/screenshots demonstrating the fix/feature.
[ ] Have you run the tests locally to confirm they pass?
New Features
What new capabilities does this PR add?
What docs changes are needed to explain this?
Codecov Report
Attention: Patch coverage is 0.77519% with 128 lines in your changes are missing coverage. Please review.
Project coverage is 23.25%. Comparing base (339c48c) to head (59786b9).
Report is 17 commits behind head on main.
Files
Patch %
Lines
...tures/snack_queue/presentation/snack_queue_wm.dart
0.00%
61 Missing :warning:
...k_queue/presentation/default_snack_controller.dart
4.16%
23 Missing :warning:
...tures/debug/presentation/ui_kit/ui_kit_screen.dart
0.00%
18 Missing :warning:
.../features/debug/presentation/ui_kit/ui_kit_wm.dart
0.00%
8 Missing :warning:
...s/snack_queue/presentation/snack_queue_widget.dart
0.00%
5 Missing :warning:
...snack_queue/presentation/snack_queue_provider.dart
0.00%
4 Missing :warning:
...atures/snack_queue/presentation/top_snack_bar.dart
0.00%
4 Missing :warning:
lib/l10n/app_localizations_en.g.dart
0.00%
2 Missing :warning:
lib/l10n/app_localizations_ru.g.dart
0.00%
2 Missing :warning:
...es/snack_queue/presentation/snack_queue_model.dart
0.00%
1 Missing :warning:
Additional details and impacted files
@@ Coverage Diff @@
## main #114 +/- ##
==========================================
- Coverage 26.39% 23.25% -3.14%
==========================================
Files 55 61 +6
Lines 917 1045 +128
==========================================
+ Hits 242 243 +1
- Misses 675 802 +127
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
|
gharchive/pull-request
| 2024-04-11T12:16:41 |
2025-04-01T06:40:31.563391
|
{
"authors": [
"codecov-commenter",
"internetova"
],
"repo": "surfstudio/surf-flutter-app-template",
"url": "https://github.com/surfstudio/surf-flutter-app-template/pull/114",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2126803933
|
feat: Changed readme file
Changes
updated readme file
Checklist for self-check
[x] Commits and PRs have been filed according to the rules on the project.
[x] The author is marked as an assigne and assigned mandatory reviewers.
[x] Required labels marked
[ ] Specified related tasks and/or related PRs.
[x] Specified Changes.
[ ] Attached videos/screenshots demonstrating the fix/feature.
[x] All unspecified fields in the PR description deleted.
[ ] New code covered by tests.
Checklist for reviewers
[ ] CI passed successfully (with a green check mark).
[ ] PR is atomic, by volume no more than 400 (+-) corrected lines (not including codogen).
Design:
[ ] System design corresponds to the agreements on structure and architecture on the project.
[ ] The code is decomposed into necessary and sufficient components.
Functionality:
[ ] The code solves the problem.
[ ] Any changes to the user interface are reasonable and look good.
Complexity:
[ ] The code is clear, easy to read, functions are small, no more than 50 lines.
[ ] The logic is not overcomplicated, there is no overengineering (no code sections that may be needed in the future, but no one knows about it).
Tests:
[ ] Updated or added tests for mandatory components.
[ ] The tests are correct, helpful, and well designed/developed.
Naming:
[ ] The naming of variables, methods, classes and other components is understandable.
Comments:
[ ] The comments are understandable and helpful.
Documentation:
[ ] All labels are correct
[ ] Technical documentation updated (after approval, updates last reviewer).
@Evgenia-bit I do not have permission for creating a PR or commiting to a branch. Please, contact SA and lease me with access.
|
gharchive/pull-request
| 2024-02-09T09:53:53 |
2025-04-01T06:40:31.571924
|
{
"authors": [
"Evgenia-bit",
"samir-a-ts"
],
"repo": "surfstudio/surf-flutter-faded-text",
"url": "https://github.com/surfstudio/surf-flutter-faded-text/pull/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
280158499
|
widget request "lokesh pretty checkbox"
https://lokesh-coder.github.io/pretty-checkbox/#basic-checkbox
from the https://github.com/surveyjs/editor/issues/232
I've added pretty-checkbox custom widget basic implementation. It works for checkboxes and radiobuttons with limited set of options. The widget inner markup highly depends on the certain pretty-checkbox options. This makes common implementation of this widget too complex IMO.
@tsv2013, I can't make your version of the widget to work for my case. I am not sure what I miss.
I attached a simplified version of my code, "suyveyjs-form.html" and "surveyjs-form.js" here: https://github.com/surveyjs/editor/issues/232, to show how it works for my case.
Please advise.
@tjtaojiang Sorry for delay. I've added example how to integrate pretty-checkbox custom widget. It resides on our site - https://surveyjs.io/Examples/Library/?id=custom-widget-pretty-checkbox&platform=Angular2
As for your implementation of the custom widget - I cann't find binding between question value and the widget control.
I suggest you try to integrate our custom widgets pack into your application.
We'll provide you with the necessary support. Please feel free to contact us in case of any difficulties.
@tsv2013, binding by: row.value and row.text. In my simplified code, you can run it and see the result.
@tsv2013, I noticed one thing, your alter the look and feel for the pretty checkbox. Check https://lokesh-coder.github.io/pretty-checkbox/, and look for "Svg" and "Recurring" checkbox, it supposed to be curved.
@tsv2013, I copy your sample code from your link above and click "Edit in Plunker", for some reason, I still can't make your code to work.
I attached a simplified version from my implementation here for you to see how it work for me:
surveyjs_widget_for_pretty_checkbox.zip.
To answer your previous concern, check row.value and row.text in the js, which I used for data binding.
I think my way has full control in term of custom widget. I just do not know if it will break some surveyjs process , like required filed property, etc. I will keep testing.
Any advise is much appreciated.
@tsv2013, I noticed a very big bug in your recent changes.
All other types are blocked from addCustomWidget, except "radiogroup" and "checkbox". In other words, in "isFit:", you can only see "radiogroup" and "checkbox" type, other types like "text" will be blocked.
I think you need to fix this issue.
@tjtaojiang On this page (https://surveyjs.io/Examples/Library/?id=custom-widget-pretty-checkbox&platform=Angular2) you can find the "Edit in Plunker" button. In this case our site creates a new "ready to run" plunk containing the current sample code.
E.g. after I clicked this button I've got the following plunk with the pretty-checkbox example for angular platform:
https://plnkr.co/edit/chHf8x?p=preview
It works for me ok.
Does this plunk work for you?
@tjtaojiang In the sampe plunk - https://plnkr.co/edit/chHf8x?p=preview
the following lines from the index.html files are responsible for adding the "pretty-checkbox" custom widget functionality:
<link rel="stylesheet" href="https://cdn.jsdelivr.net/npm/pretty-checkbox@3.0/dist/pretty-checkbox.min.css"/>
<script src="https://unpkg.com/surveyjs-widgets"></script>
because of the code of our "pretty-checkbox" custom widget implementation is resides in the "https://unpkg.com/surveyjs-widgets" script.
@tjtaojiang I've also created a plunk with the code you sent me in the "surveyjs_widget_for_pretty_checkbox.zip" archive.
This is the link to the plunk with your code - https://plnkr.co/edit/e3uqCK?p=preview
We can use this plunk as a starting point for the further discussion.
Note that I've commented reference to our custom widgets package (in the index.html file) because you are using your own implementation of the "pretty-checkbox" custom widget:
<!--<script src="https://unpkg.com/surveyjs-widgets@0.98.4/surveyjs-widgets.js" type="text/javascript"></script>-->
Could you check this plunk also?
My thoughts about the pretty-checkbox cutom widget implementation
"pretty-checkbox" is the CSS library and requires different markup for different features.
This makes us to drop our default-generated markup of the "checkbox", "radiogroup" (and that much more worse the "matrix") questions and replace it with the markup required to make "pretty-checkbox" work properly.
If we remove our default markup, we become responsible for
the two-way binding of the custom widget: subscribe on the question value changes and reflect them in the UI, and subscribe the UI control input and pass entered values to the question;
the whole SurveyJS question functionality as read-only behaviour, "other item", "column count" and other options
Our basic implementation of the "pretty-checkbox" custom widget at this moment contains basic options and binding functionality and supports "checkbox" and "radiogroup" quesions only.
The full functionality can be added to the certain custom widget implementation by the united efforts of the SurveyJS team and the community. I hope on your understanding and support.
@tsv2013, for some reason, this Plunker version does not work when I copy them into my local: https://plnkr.co/edit/?p=preview. I still see the old radio and checkbox.
@tsv2013, for some reason, this Plunker version does not work when I copy them into my local: https://plnkr.co/edit/chHf8x?p=preview. I still see the old radio and checkbox.
@tjtaojiang Could you tell me whether the plunker https://plnkr.co/edit/chHf8x?p=preview works for you or not?
@tjtaojiang hello! Could you please describe your use case? Does the example in our site looks good? Do your problem with integrate widget in your project or with widget functionality?
@dmitrykurmanov, as I mentioned above, I copy three files in https://plnkr.co/edit/chHf8x?p=preview into my local, because some websites are blocked by my company, I need to use https://go.weblife.io to download all blocked css and js into my local. I attach zip file contain all three I modified and used in my local. You can see duplicated due to the websites are block and I have downloaded. After I put all three in one directory and click html, I still only see the default (old fashion) radio and checkbox.
@dmitrykurmanov, as I mentioned above, I copy three files in https://plnkr.co/edit/chHf8x?p=preview into my local, because some websites are blocked by my company, I need to use https://go.weblife.io to download all blocked css and js into my local. I attach zip file contain all three I modified and used in my local. You can see duplicated due to the websites are block and I have downloaded. After I put all three in one directory and click html, I still only see the default (old fashion) radio and checkbox.
test-surveyjs-form.zip
Hi, @dmitrykurmanov, @tsv2013,
After getting the sample code from https://plnkr.co/edit/chHf8x?p=preview, I do see the default checkbox and radiogroup are changed to pretty checkbox.
But at least I noticed those issues:
(1). You have modified the default checkbox to a square shaped pretty checkbox. If I need other shapes like "Svg" -> "Recurring", a curve shape checkbox, on https://lokesh-coder.github.io/pretty-checkbox/, or if I need to use the old square checkbox, I still need to create a custom widget, which is actually what users will need: how to create custom widget to cope their requests.
(2). Your new implementation deactivate colCount. Check your sample case.
(3). https://unpkg.com/surveyjs-widgets, or actually https://unpkg.com/surveyjs-widgets@0.98.4/surveyjs-widgets.js, has disabled any custom widget implementation. Uncomment addCustomWidgetForDatePicker() in test-surveyjs-forms.js in my attachment:
test surveyjs widget.zip
, you will see isFit: never gets called.
(4). For lengthy label radio or checkbox, the current pretty checkbox cuts the total description. Check my attachment above.
@tjtaojiang Answering your questions:
(1) - I've implemented "addOn" feature for the "pretty-checkbox" custom widget. It allows to insert any additional markup before the "label" tag
For "SVG" mode you will need to add the following code:
var prettyCheckbox = Survey.CustomWidgetCollection.Instance.getCustomWidgetByName(
"pretty-checkbox"
);
var checkboxSettings = prettyCheckbox.widgetJson.settings.checkbox;
checkboxSettings.rootClass = "pretty p-svg p-curve";
checkboxSettings.addOn =
"<!-- svg path -->" +
'<svg class="svg svg-icon" viewBox="0 0 20 20">' +
'<path d="M7.629,14.566c0.125,0.125,0.291,0.188,0.456,0.188c0.164,0,0.329-0.062,0.456-0.188l8.219-8.221c0.252-0.252,0.252-0.659,0-0.911c-0.252-0.252-0.659-0.252-0.911,0l-7.764,7.763L4.152,9.267c-0.252-0.251-0.66-0.251-0.911,0c-0.252,0.252-0.252,0.66,0,0.911L7.629,14.566z" style="stroke: white;fill:white;">' +
"</path>" +
"</svg>";
(2) I've implemented the "colCount" feature
(3) I've checked different custom widgets (including the "pretty-checkbox") on the same page of a survey - they all works fine for me.
I've checked the last code you sent us and found that you commented your custom widget initialization function:
function onAngularComponentInit() {
// addCustomWidgetForDatePicker();
Survey.SurveyNG.render("surveyElement", {model: survey});
}
After I've uncommented this call, everything works ok for me.
(4) I've checked your attachment, and see that long item descriptions are not trimmed. At this point I'd like to clarify - what behaviour for the long item descriptions is suitable for you? Could you share a screenshot or drawing with desired view.
@tjtaojiang The changes mentioned above will be available since 0.98.6. For more details you can check the 0d7f9b9195ee6a07cc578981e4ec54cb48b26341 commit.
Hi @tsv2013,
(1). Due to the (2) - (3), I can't test this: "If I need other shapes like "Svg" -> "Recurring", a curve shape checkbox, on https://lokesh-coder.github.io/pretty-checkbox/, or if I need to use the old square checkbox..."
(2). colCount is still not working. I attach my code here for reference
test surveyjs widget.zip. You can check the attached image file, test-surveyjs-form.jpg .
(3). Check my attachment above, https://unpkg.com/surveyjs-widgets, or actually now: https://unpkg.com/surveyjs-widgets@0.98.5/surveyjs-widgets.js, still disable radiogroup and checkbox. Much worse, in my real code (not attached), it disables any custom widget implementation.
(4). Check the attached image file, lengthy label still stretch out and trimmed.
@tjtaojiang
As I wrote in this comment (https://github.com/surveyjs/widgets/issues/12#issuecomment-353289703) - "The changes mentioned above will be available since 0.98.6". After 0.98.6 will be released, you will be able to try "addOn" feature and "colCount" option. I'll also add corresponding demo on our site.
As for (3) - I'm not able to reproduce this issue on my side. Can you prepare an example in plunker or any other sandbox or project or github repo? So I will be able to reproduce the issue and fix it.
(4) - I do understand that things are not so good as they can be. But could you give your opinion - how it should be fixed?
@tjtaojiang I've checked your code and screenshot you sent. And I can say that datePickerWidget is applied on your page. You see an empty space instead of date picker control because you disable default rendering via setting the isDefaultRender: false option and nothing created in the afterRender: function (question, el) { } function as I can see from your custom datePickerWidget code from your file:
function addCustomWidgetForDatePicker() {
var datePickerWidget = {
isFit: function (question) {
alert("question = |" + question + "|\n"
+ "question.getType() = |" + question.getType() + "|\n"
+ "question.inputType = |" + question.inputType + "|\n"
+ "question.class = |" + question.class + "|\n"
+ "question.name = |" + question.name + "|\n"
+ "question.title = |" + question.title + "|\n"
);
return (question.getType() === 'text') && (question.inputType === 'date');
},
isDefaultRender: false,
afterRender: function (question, el) {
}
};
Survey.CustomWidgetCollection.Instance.addCustomWidget(datePickerWidget);
}
@tjtaojiang In order to use custom widgets, their code sould be loaded after the core SurveyJS library:
<script src="https://surveyjs.azureedge.net/0.98.5/survey.angular.js" type="text/javascript"></script>
<script src="https://unpkg.com/surveyjs-widgets@0.98.5/surveyjs-widgets.js" type="text/javascript"></script>
@tsv2013,
Refer to your: "As for (3) - I'm not able to reproduce this issue on my side. Can you prepare an example in plunker or any other sandbox or project or github repo? So I will be able to reproduce the issue and fix it."
Please use my previous zip file, test surveyjs widget.zip, the 4th post above this one, click test-surveyjs-form.html after you unzip. You can see radio and checkbox element never got passed into isFit: in addCustomWidgetForDatePicker() .
I will post another post to illustrate how do I cope my issues (10 - (4) next.
Hi @tsv2013 and @dmitrykurmanov,
Here is how I solve my issues from (1) - (4). I need to comment out , otherwise it will disable all custom widget I created, as I mentioned before.
Here is the attachement for it:
test surveyjs widget-3.zip
click test-surveyjs-form.html after you unzip it.
I am not sure my way will introduce any other issues, which may break some calculation and dependency, like required property etc.
Please offer advise.
Hi @tsv2013 and @dmitrykurmanov,
Here is how I solve my issues from (1) - (4). You need to comment out https://unpkg.com/surveyjs-widgets@0.98.5/surveyjs-widgets.js as in test-surveyjs-form.html, otherwise it will disable all custom widget I created, as I mentioned before.
Here is the attachement for it:
test surveyjs widget-3.zip
, click test-surveyjs-form.html after you unzip it.
I am not sure if my code will introduce any other issues, which may break some surveyjs calculation and dependency, like required property etc.
Please offer advise.
@tjtaojiang The "radio and checkbox element never got passed into isFit in addCustomWidgetForDatePicker()" because they are already fit for another custom widget. It is hard to apply two custom widgets on the same question. This can lead to unexpected side effects.
But until now our pretty-checkbox custom widget was activated always by question type if pretty-checkbox css styles were loaded on the page. I've changed this behaviour to activation by property: renderAs that should be set to "prettycheckbox" value.
As for your implementation of the pretty-checkbox custom widgets. Your certain implementation can better be suitable for solving your certain tasks. I don't see any problems in your implementation for now except of absence of two-way bindings: UI will not updated in case of model has been changed and choosen in UI values are not passed in the model. The last problem is crucial because if you choose some items and press Complete button you will see validation complains "Please answer the question." because the values you have choosen haven't been stored in the question.
@tsv2013, I will look into the two-way bindings issue you mentioned.
As far as "radio and checkbox" are already fit for another custom widget, then user will not be able to create a widget for their requirement? This may not be a good idea. Can we remove the limit?
@tsv2013, I will look into the two-way bindings issue you mentioned. Do you have any recommendation how should I modify my implementation to cope with the issue?
As far as "radio and checkbox" are already fit for another custom widget, then user will not be able to create a widget for their requirement? This may not be a good idea. Can we remove the limit?
Hi @tsv2013 and @dmitrykurmanov,
Here is my current version which will fix two-way bindings issue. It also contains a custom widget for a date picker (pikaday object):
test surveyjs widget-4.zip. If you wish, I can move the date picker (pikaday object) custom widget to a separate thread.
But there is a minor issue which I have not figured out. For some reason, the checkbox does not show the white check sign inside the checkbox. Do you know why?
I will be on vacation. I won't reply to you until I come back. Thanks.
Hi @tsv2013 and @dmitrykurmanov,
Here is my current version which will fix two-way bindings issue. It also contains a custom widget for a date picker (pikaday object):
test surveyjs widget-4.zip
. If you prefer, I can move the date picker (pikaday object) custom widget to a separate thread.
But there is a minor issue which I have not figured out. For some reason, the checkbox does not show the white check sign inside the checkbox. Do you know why?
I will be on vacation. I won't be able to reply to you until I come back. Thanks.
|
gharchive/issue
| 2017-12-07T14:53:32 |
2025-04-01T06:40:31.704565
|
{
"authors": [
"dmitrykurmanov",
"tjtaojiang",
"tsv2013"
],
"repo": "surveyjs/widgets",
"url": "https://github.com/surveyjs/widgets/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
506141399
|
Shadab M S
I have added my code here please accept the pull request.
Hey @shadabms ,
Could you please change the font size?
|
gharchive/pull-request
| 2019-10-12T06:31:57 |
2025-04-01T06:40:31.725696
|
{
"authors": [
"Inframind",
"shadabms"
],
"repo": "sushant-j/my-first-pull-request",
"url": "https://github.com/sushant-j/my-first-pull-request/pull/152",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
683069915
|
New version
New version should be released in order to start using SPM. I would suggest doing a minor bump to 2.3.0.
Yes indeed, that also occurred to me while I was drinking my beer this evening 😀 Will do tomorrow!
Thanks a lot and enjoy that beer 💪
Done! Since nothing else changed, I labeled the version 2.2.2. Thanks again!
|
gharchive/issue
| 2020-08-20T20:18:32 |
2025-04-01T06:40:31.789170
|
{
"authors": [
"LucianoPolit",
"svdo"
],
"repo": "svdo/ReRxSwift",
"url": "https://github.com/svdo/ReRxSwift/issues/14",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1879005804
|
Wallet Standard support, bug fixes and upgrades
Wallet Standard support
Dependencies upgraded to recent versions
Accessibility warnings fixed
If you can't wait for the merge of this PR, consider using aztemi/svelte-on-solana-wallet-adapter repo instead. It is a fork and already has these commits with additional features integrated.
Thanks!
Will try this out now
That'll great @cogoo. Thanks.
@cogoo
have you also tried
https://github.com/portalpayments/svelte-on-solana-wallet-adapter
do u know if these support solana mobile adapters?
which one do you recommend?
|
gharchive/pull-request
| 2023-09-03T09:16:27 |
2025-04-01T06:40:31.792560
|
{
"authors": [
"JayeshVP24",
"cogoo",
"t4top"
],
"repo": "svelte-on-solana/wallet-adapter",
"url": "https://github.com/svelte-on-solana/wallet-adapter/pull/57",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2642176520
|
feat: simple testing library
Alternative to #32
This takes more of a lightweight approach where we provide a bare minimum api for testing rather than controlling the whole testing pipeline. This should allow for better flexibility as we're no longer tied to any particular testing framework.
Notes:
Had to rename the directory of the vitest addon as it was somehow interfering with vitest or tinypool during runtime, causing shit to randomly break
Storybook has been pinned to v8.3.x as v8.4.x flat out doesn't work. I haven't had a moment to fully investigate the issue yet.
TODO:
Snapshots?
look into using vitest workspaces
Should we continue to use degit to clone the vite templates? If so, should we pin it to a specific commit hash (see here)? Or should this be something we manage fully (perhaps as part of create)?
Ohhh and there is another thing I just noticed: How are the directory names inside .test-output/_tests generated? Here are my directories inside drizzle
-1930824653_0
-1930824653_1
-1930824653_2
-1930824653_3
-1930824653_4
-1930824653_5
-1930824653_6
-1930824653_7
They are semi-useful at least. In #32 I generated them based on the options passed for each test, so that if a test failed you could identify it by directory name and run it by hand.
How are the directory names inside .test-output/_tests generated? Here are my directories inside drizzle
they're based on the task id provided by vitest. Each test gets their own meta.json (similar to how we had it in svelte-add) that documents the options that were passed in.
They are semi-useful at least. In #32 I generated them based on the options passed for each test, so that if a test failed you could identify it by directory name and run it by hand.
not sure if that's a scalable solution. imagine having multiple test cases that share the same addon options but they evaluate different things
|
gharchive/pull-request
| 2024-11-07T20:55:14 |
2025-04-01T06:40:31.800143
|
{
"authors": [
"AdrianGonz97",
"manuel3108"
],
"repo": "sveltejs/cli",
"url": "https://github.com/sveltejs/cli/pull/275",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1202238620
|
Support for subdomains in SvelteKit
Describe the problem
I do want to start this off with the acknowledgment that subdomains with SvelteKit is possible. However, there is no clear documentation nor have I found any clear and concise examples on how to do so. The only reason I even assume it is possible is because of kit.svelte.dev.
If I am wrong, and there is a clear and documented way of adding support for subdomains to SvelteKit websites then please do let me know and I will close this issue.
Describe the proposed solution
The proposed solution is that of the title.
Alternatives considered
Subdirectories are fine, just not ideal.
Importance
nice to have
Additional Information
No response
I don't know what this is asking for. You have access to the hostname via the event that's sent to endpoints and the $page store that's available to pages.
kit.svelte.dev is its own app/site that is completely separate from the app/site at svelte.dev.
Is the request to be able to have one app serve multiple hostnames, and have the hostname be part of what distinguishes routes, instead of just the path?
I understand I have access to the site's hostname. Even then there is still no clear documentation.
I am using cloudflare pages and see no clear way of using multiple subdomains with one or even multiple SvelteKit projects.
subdomains are something that you configure with DNS (not with your frontend framework), so there's nothing that SvelteKit would do here
If you'd like help or are confused about something, please ask on Discord
Hi!
I think that the question being asked is if someone wants to use dynamic subdomain routing can they do that in Svelte Kit as part of their routing scheme:
I.e.
http://mysubdomain.example.com => '/subdomain/mysubdomain/'
http://myexcellentsubdom.example.com/homepage => '/subdomain/myexcellentsubdom/homepage'
http://first.second.example.com => '/subdomain/first/second/'
http://first.second.example.com/another/page => '/subdomain/first/second/another/page'
In express this is possible using express-subdomain-handler
In other words: Just like you can use src/routes/about/+page.svelte and src/routes/blog/[slug]/+page.svelte to make /blog/foo and about pages on mysite.com, so too you should be able to create a route for a subdomain such as /src/routes/[subdomains]/login/+page.svelte or /routes/search.site/+page.svelte to make a login.mysite.com or search.mysite.com subdomain.
This seems like a reasonable and understandable request, so I think that this issue should maybe be reopened.
|
gharchive/issue
| 2022-04-12T18:28:50 |
2025-04-01T06:40:31.807026
|
{
"authors": [
"Conduitry",
"benmccann",
"cowboycodr",
"rocketinventor"
],
"repo": "sveltejs/kit",
"url": "https://github.com/sveltejs/kit/issues/4598",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1363961052
|
1.0.0-next.455 causing broken Netlify response header behavior
Describe the bug
With a simple upgrade to @sveltejs/kit from 1.0.0-next.454 to 1.0.0-next.455 with no code changes except the ones needed for the upgrade, the response headers of the pages as served from Netlify are different. I suspect that something changed with the build to where now the netlify.toml is not being read (even though it seems to be detected when looking at the Netlify build logs for the branch).
The response headers (truncated) for an initial html page are changed as follows:
- cache-control: public, max-age=0, must-revalidate
- content-type: text/html; charset=UTF-8
- cross-origin-embedder-policy: require-corp
- cross-origin-opener-policy: same-origin
- strict-transport-security: max-age=31536000; includeSubDomains; preload
+ cache-control: no-cache
+ content-type: text/html
+ link: <./_app/immutable/assets/_layout-7e2cdc5f.css>; rel="preload";as="style"; nopush, <./_app/immutable/assets/modalProvider-2c7dffb9.css>; rel="preload";as="style"; nopush, <./_app/immutable/assets/outputLoadingMessage-6ab61537.css>; rel="preload";as="style"; nopush, <./_app/immutable/assets/tooltip-0a1c217e.css>; rel="preload";as="style"; nopush, <./_app/immutable/start-9e2d9bc6.js>; rel="modulepreload"; nopush, <./_app/immutable/chunks/index-f9b80742.js>; rel="modulepreload"; nopush, ...
The only custom ones set in the netlify.toml are the cross-origin headers you see there (COOP and COEP).
Reproduction
Upgrade from 1.0.0-next.454 to 1.0.0-next.455 and follow the simple changes described here
Logs
No response
System Info
The build image used is the default one in Netlify Ubuntu Focal 20.04.
Severity
blocking an upgrade
Additional Information
No response
Please provide a repro as described in the issue template
https://github.com/papiro/svelte-kit-netlify-issue
netlify: https://master--aquamarine-palmier-db1503.netlify.app/
I haven't been able to get the old behaviour working, despite countless permutations of kit and adapter-netlify versions, and the provided repro won't build on Netlify at all, so unfortunately I have no idea what changed. The https://master--aquamarine-palmier-db1503.netlify.app/ deployment doesn't appear to be active.
I can't spend any more time on this, so if it's not possible to create a reliable repro then I'm afraid we'll have to close this. For all we can tell, it could be a change to Netlify itself. It seems quite unlikely that a change to Kit would result in this behaviour — far more likely to be a change to adapter-netlify.
Hmmmm, i left a clean git commit history so you could see what changes i made. Barely any. I apologize i took the site down. Got overzealous cleaning stuff up. I can redeploy it, but it sounds like you're confident this isn't a Svelte-Kit issue anyways. Also, I moved past the issue by creating a +page.ts so it's not an issue which is bothering me any longer so we can close it if you'd like. I'd be okay with that.
|
gharchive/issue
| 2022-09-07T01:05:58 |
2025-04-01T06:40:31.814495
|
{
"authors": [
"Rich-Harris",
"papiro"
],
"repo": "sveltejs/kit",
"url": "https://github.com/sveltejs/kit/issues/6626",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1535632866
|
Adapter Vercel Image Optimizations
Describe the problem
Would be really awesome if adapter-vercel package accepted images option for using Vercel's native image optimization. It only requires simple addition of images object to build config file .vercel/output/config.json. I would submit a pull request but I am sure the options, naming conventions, etc would not be as the team would desire. Given it is such a straightforward task it probably wouldn't prove helpful.
Describe the proposed solution
Here are the options
type ImageFormat = 'image/avif' | 'image/webp';
type RemotePattern = {
protocol?: 'http' | 'https';
hostname: string;
port?: string;
pathname?: string;
};
type ImagesConfig = {
sizes: number[];
domains: string[];
remotePatterns?: RemotePattern[];
minimumCacheTTL?: number; // seconds
formats?: ImageFormat[];
dangerouslyAllowSVG?: boolean;
contentSecurityPolicy?: string;
};
https://vercel.com/docs/concepts/image-optimization
https://vercel.com/docs/build-output-api/v3#build-output-configuration/supported-properties/images
Alternatives considered
No response
Importance
nice to have
Additional Information
No response
I'm not sure just adding the options is enough, there's also work involved to create a specific image component I guess? Related to #241
I'm not sure just adding the options is enough, there's also work involved to create a specific image component I guess? Related to #241
you can leave that to the community, because thats more subjective and that would be separate from an adapter package anyway, but without the ability to put these options in the build config then you can't use it, so this shouldn't be the breakpoint of using features
Also all the solutions I have seen are for local images, not remote images.
I'm not sure just adding the options is enough, there's also work involved to create a specific image component I guess? Related to #241
Next and Nuxt have supported image components, but from my understanding once enabled you could also just update your image srcs manually from eg /public/images/logo.png to /_vercel/image?url=/public/image/logo.png?w=320&q=80. A fancy image component could come later if the community wants it, but just enabling this would be a great enhancement IMO
Next and Nuxt have supported image components, but from my understanding once enabled you could also just update your image srcs manually from eg /public/images/logo.png to /_vercel/image?url=/public/image/logo.png?w=320&q=80. A fancy image component could come later if the community wants it, but just enabling this would be a great enhancement IMO
Exactly
@lettucebowler has made a PR for this #8667.
Someone shared an example on Discord (https://discord.com/channels/457912077277855764/1068354495240741005/1070047413898850424)
// Image.svelte
<script lang="ts">
import { dev } from '$app/environment';
let className = ""
export { className as class };
export let image:WpImage
export let showTitle = false
export let lazy = true
export let height = 1280
export let quality = 70
export let width = 720
export let unoptimized = false
const sizeArr = [640, 768, 1024, 1280, 1536, 2048]
const vercelImg = (sourceUrl:string,size:number) =>
(`/_vercel/image?url=${encodeURIComponent(sourceUrl)}&w=${size}&q=${quality} ${size}w`)
$:({ srcSet, sourceUrl, title, altText} = image)
$: vercelSrcSet = sizeArr.map(size=>vercelImg(image?.sourceUrl,size)).join(',')
// dyanmic srcset only for optimized (w/vercel)
$: imageSrcSet = (unoptimized || dev) ? srcSet : vercelSrcSet
let element
</script>
<img
{width}
{height}
title={showTitle ? title : undefined}
srcset={imageSrcSet}
src={sourceUrl}
alt={altText ?? title}
class={className ?? undefined}
loading={lazy ? 'lazy' : 'eager'}
{...$$restProps}
bind:this={element}
/>
// scripts/vercel-images.js
import fs from 'node:fs';
import dotenv from 'dotenv'
dotenv.config()
// sizes should be shared variable with Image component
import {sizes} from './theme.cjs'
const file = '.vercel/output/config.json';
const config = {
...JSON.parse(fs.readFileSync(file, 'utf-8')),
images: {
"sizes": sizes,
"domains": [],
"minimumCacheTTL": 60,
"formats": ["image/avif", "image/webp"],
"remotePatterns": [
{
"hostname": "yourcdn\\.cloudfront\\.net$"
},
]
}
};
fs.writeFileSync(file, JSON.stringify(config, null, 2));
// package.json
"build": "vite build && node scripts/vercel-images",
@benmccann do you have working example hoisted anywhere (e.g. GitHub, Code sandbox etc. ). I wanted to see the configuration details. Thanks.
There's a draft PR for this: https://github.com/sveltejs/kit/pull/10323
Hi @hartwm, thanks so much for providing these examples! I do have one question I wonder if anyone knows the answer to.
Vite lets you import images like this:
<script>
import logo from '$images/example.png';
</script>
That means that a hash gets put in their name so that they can be cached forever. Is it possible to use such an image with the Vercel image solution?
When I go to https://vercel-images-sveltekit.vercel.app/ I notice that the local image can't be cached as strongly as the Vite processed images we have today. If you hit the page a second time, it returns a 304 response as opposed to avoiding the request entirely. I wonder if it's possible to use an imported image so that Vite processes the request before Vercel takes over possibly allowing us to skip the request entirely still. If it's possible, it'd be a great example to add to https://github.com/hartwm/vercel-images-sveltekit
@benmccann You can set Cache-Control headers with Vercel...
https://github.com/leoj3n/svelte-vercel-optimized-images?tab=readme-ov-file#avoiding-304-network-requests
It looks like you merged https://github.com/sveltejs/kit/pull/8667#pullrequestreview-1730990777 which relates to the OP... Considering the ability to set Cache-Control with Vercel, your remaining question may be answered, and so you might now be able to close this open issue.
@benmccann Regarding "Vite processes the request before Vercel takes over"...
It kind of defeats the purpose of using Vercel to send the correct image type (avif/webp) based on headers sent from the browser, as well as the (pre-defined) dynamic image sizes. If you circumvent Vercel you would have to generate all those permutations ahead of time yourself which could be prohibitive for user uploaded images, as mentioned.
I do believe Next.js implements their own code that emulates what the Vercel image optimization endpoint does, so it may be possible to add such functionality to SvelteKit itself, but you probably will have a very similar approach.
|
gharchive/issue
| 2023-01-17T00:38:50 |
2025-04-01T06:40:31.828517
|
{
"authors": [
"aakash14goplani",
"benmccann",
"dummdidumm",
"hartwm",
"izznatsir",
"leoj3n",
"lettucebowler"
],
"repo": "sveltejs/kit",
"url": "https://github.com/sveltejs/kit/issues/8561",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
851621816
|
Fetching works on initial page load, not on page reload
Describe the bug
I'm fetching data from an API in my route's load function, but I can't find a way to make it work on subsequential loads.
Logs
using ctx.fetch: TypeError: Failed to execute 'fetch' on 'Window': Illegal invocation at load
using fetch: ReferenceError: fetch is not defined at load
To Reproduce
install sveltekit
create a component that loads data in the <script> tag, as follows:
export async function load(ctx) { // Call our fetch let postsData; postsData = await fetch( 'https://jsonplaceholder.typicode.com/posts' ).then((x) => x.json()); console.log(postsData); }
Expected behavior
I expect the response to be logged to the console on each pageload.
Using fetch, it works on initial load, not reloads.
Using ctx.fetch, it works on reloads, but not inital load.
Stacktraces
Unfamiliar with stacktraces. This is my first time publishing an issue to a repo, so any information about this/how I can better format my issues would be much appreciated. I have more bugs/issues I've found that I'm too scared to post. Really hate bugging you guys when you're working so hard, but I'm lost on a few things.
Information about your SvelteKit Installation:
Diagnostics
The output of npx envinfo --system --npmPackages svelte,@sveltejs/kit,vite --binaries --browsers
System:
OS: Windows 10 10.0.19042
CPU: (8) x64 Intel(R) Core(TM) i7-6700K CPU @ 4.00GHz
Memory: 8.92 GB / 15.89 GB
Binaries:
Node: 15.11.0 - C:\Program Files\nodejs\node.EXE
Yarn: 1.22.10 - ~\AppData\Roaming\npm\yarn.CMD
npm: 7.7.5 - C:\Program Files\nodejs\npm.CMD
Browsers:
Edge: Spartan (44.19041.423.0), Chromium (89.0.774.68)
Internet Explorer: 11.0.19041.1
npmPackages:
@sveltejs/kit: next => 1.0.0-next.69
svelte: ^3.29.0 => 3.37.0
vite: ^2.1.0 => 2.1.5
Your browser
(Using NPM & Brave Web Browser)
Your adapter (e.g. Node, static, Vercel, Begin, etc...)
Node adapter
Severity
How severe an issue is this bug to you? Is this annoying, blocking some users, blocking an upgrade or blocking your usage of SvelteKit entirely?
I guess I could store content statically, but this is a big problem for me as I'm trying to use Sveltekit to refactor some of my projects that use APIs everywhere. Axios might work as a workaround, but I'm trying to use as many Svelte/Sveltekit features as possible because I'm a big believer in what you guys are doing. If I could figure this out & some other problems I'm having, I'm ready to start refactoring some of my projects to use Svelte/Sveltekit instead of react & related helped frameworks
Additional context
I love what you guys are doing & would love to help if I can. Svelte is a breath of fresh air after getting annoyed & leaving the development community years ago. Thank you so much for everything youre doing, even if you can't help me out
One thing to remember with the load function is that it can run on both the server OR the client.
So when you say you see it in the console, do you mean the terminal window where you are running npm run dev or the browser console?
With SSR enabled (the default) the load function is called on the server and the response embedded into the page so that it doesn't need to re-run on the client (but in dev, it does re-run). One the app has been loaded then navigating between routes will call the appropriate load functions for them but only on the client at that point. So, be sure to check both.
Minor tip: as you already have the function marked as async it's easier to use await than ye-olde promise chaining, so you could write:
export async function load(ctx) {
const resp = await fetch( 'https://jsonplaceholder.typicode.com/posts' )
const post = await resp.json()
return {
props: post
}
}
(assuming the route component has a property named post)
@evan-thurston You're forgetting to use the fetch parameter passed to your load function. Because load can run on the server or the client, but the built-in window.fetch function is only available on the client, Svelte-Kit includes a fetch function that will work in either environment, and passes it to your load function as a parameter. You're supposed to use that function rather than the built-in fetch. So change your load function to use ctx.fetch instead of fetch and it should work. Or rewrite it the way the documentation recommends (and the way CaptainCodeman showed in his example) using destructuring to get the fetch parameter, i.e. export async function load({ fetch }) and then you can just write await fetch('...') and you'll be using Svelte-Kit's supplied function.
|
gharchive/issue
| 2021-04-06T17:01:49 |
2025-04-01T06:40:31.841533
|
{
"authors": [
"CaptainCodeman",
"evan-thurston",
"rmunn"
],
"repo": "sveltejs/kit",
"url": "https://github.com/sveltejs/kit/issues/906",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1716926676
|
Component styles flicker/blink with SSR mode on
Describe the bug
When applying stylings to components or pages, on SSR mode, first server stylings get rendered, and then client ones.
In some cases, when components / pages use more 'dynamic' data like slots, $$restProps with class names, lists, etc., what oddly happens is that there is a 'flicker', where server stylings do not seem to match client ones.
Below is a video with two examples of when this happens:
https://github.com/sveltejs/kit/assets/61620873/e5688342-1c23-4994-a4ba-807fc1247e1b
Reproduction
About twenty lines: https://www.sveltelab.dev/vg1t4x2cb8n84f4
Logs
No response
System Info
I've spoken to a number of people so I don't think that it is a device-specific thing. But anyway:
System:
OS: macOS 13.3.1
CPU: (8) arm64 Apple M1
Memory: 83.23 MB / 16.00 GB
Shell: 5.9 - /bin/zsh
Binaries:
Node: 19.6.0 - /opt/homebrew/bin/node
Yarn: 1.22.11 - /opt/homebrew/bin/yarn
npm: 9.4.0 - /opt/homebrew/bin/npm
Browsers:
Brave Browser: 113.1.51.114
Safari: 16.4
npmPackages:
@sveltejs/adapter-auto: ^1.0.0-next.91 => 1.0.0-next.91
@sveltejs/kit: next => 1.0.0-next.589
svelte: ^3.55.1 => 3.55.1
vite: ^4.0.0 => 4.1.1
Severity
annoyance
Additional Information
No response
The same issue happens even when setting export const csr = false. Interestingly, the only workaround seems to be wrapping for loops in a flexbox.
<div style="display:flex;">
{#each item as item}
<!-- represent the item details in a tailwind card -->
<div class="w-64 inline-block bg-pink">
<div>
<a href="/media/{item}">Page</a>
</div>
</div>
{/each}
</div>
But I would suggest this approach be avoided. When setting e.g. 5 divs (without the for loop), there are no spaces. Only through the above do they have spaces on the server-rendered HTML, and then revert to no spaces / gaps as expected (see repro). So this seems to be more of a workaround than expected behavior.
This issue is not caused by when styles are applied, but differences in the whitespace present in the SSR markup and the client-side rendered elements.
The SSR markup inserts a text node with "\n " between each child div:
But when client-side rendered, those nodes are removed:
There are multiple whitespace issues open in the Svelte repo, so not sure if this is a duplicate or not (or what workarounds would be). Transferring over there for now.
|
gharchive/issue
| 2023-05-19T09:22:37 |
2025-04-01T06:40:31.848147
|
{
"authors": [
"Destaq",
"geoffrich"
],
"repo": "sveltejs/kit",
"url": "https://github.com/sveltejs/kit/issues/9984",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1484994739
|
Inspector - open parent component
Describe the problem
If I have a button and i open it with inspector, it opens the button component, but usually I dont care about the button component, but about component where the button is used.
Describe the proposed solution
If i click on component with inspector, i could get list of parents for this component and i could click on one of them to open them instead. It would open parent component on line, where selected component is used.
Inspector could have current behavior on left click and this more advanced one could be on right click.
Alternatives considered
Importance
would make my life easier
you can use the up arrow key to select the parent component
@ToP29 does arrow selection help with your usecase?
Yes, it is ok. Thank you
|
gharchive/issue
| 2022-12-08T16:33:33 |
2025-04-01T06:40:31.935679
|
{
"authors": [
"ToP29",
"dominikg"
],
"repo": "sveltejs/vite-plugin-svelte",
"url": "https://github.com/sveltejs/vite-plugin-svelte/issues/531",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1991097478
|
Svelte 5 - css nesting broke
Describe the bug
Not sure if this is the right spot to report, but since the error window mentions this plugin specifically, figured I'd start here.
Anyways, not sure if I'm doing something wrong but in the current stable version of Svelte, the following component compiles and functions as expected
<h1>Welcome to SvelteKit</h1>
<p>Visit <a href="https://kit.svelte.dev">kit.svelte.dev</a> to read the documentation</p>
<style>
h1 {
&:hover {
color:red;
}
}
</style>
However in the current Svelte 5 beta, an error is triggered saying Expected a valid CSS identifier. This is tested from a fresh install of SvelteKit running
npm create svelte@latest
afterwards running
npm install svelte@next
to install Svelte 5
Not sure how support was implemented in Svelte 4 but it'd be great to have this working in the final release without having to download another package.
Reproduction URL
https://github.com/sortofsleepy/svelte5-vite-css-error
Reproduction
Svelte 5 should already be setup in package.json so error should pop up immediately
Rerun npm i svelte@latest to compare and contrast.
Logs
[plugin:vite-plugin-svelte] C:/Users/Joseph/Documents/projects/tests/svelte-test/src/routes/+page.svelte:9:1 Expected a valid CSS identifier
C:/Users/Joseph/Documents/projects/tests/svelte-test/src/routes/+page.svelte:9:1
7 | color:red;
8 | }
9 | }
| ^
10 | </style>
CompileError: Expected a valid CSS identifier
at error (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/errors.js:568:8)
at read_identifier (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:491:3)
at read_selector (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:289:11)
at read_selector_list (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:157:17)
at read_rule (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:139:12)
at read_body (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:64:18)
at read_style (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/read/style.js:25:19)
at tag (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/state/element.js:264:20)
at new Parser (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/phases/1-parse/index.js:65:12)
at parse (file:///C:/Users/Joseph/Documents/projects/tests/svelte-test/node_modules/svelte/src/compiler/ph
System Info
System:
OS: Windows 10 10.0.19045
CPU: (24) x64 AMD Ryzen 9 5900X 12-Core Processor
Memory: 16.00 GB / 31.90 GB
Binaries:
Node: 20.8.0 - C:\Program Files\nodejs\node.EXE
npm: 10.1.0 - C:\Program Files\nodejs\npm.CMD
pnpm: 6.32.3 - ~\AppData\Roaming\npm\pnpm.CMD
Browsers:
Edge: Spartan (44.19041.3570.0), Chromium (119.0.2151.58)
Internet Explorer: 11.0.19041.3570
native css nesting isn't supported in svelte4 by itself, some parts of it work but thats because svelte4 uses css-tree. svelte5 does not and hasn't implemented native css nesting yet.
To work around this, use a preprocessor that converts nested syntax to unnested. This can be done with postcss or lightningcss (experimental option on vite)
see https://github.com/sveltejs/svelte/issues/8587 and https://github.com/sveltejs/svelte/pull/9343
closing here as it is not caused by vite-plugin-svelte. Please follow the issues on svelte repo for updates.
|
gharchive/issue
| 2023-11-13T17:14:42 |
2025-04-01T06:40:31.942119
|
{
"authors": [
"dominikg",
"sortofsleepy"
],
"repo": "sveltejs/vite-plugin-svelte",
"url": "https://github.com/sveltejs/vite-plugin-svelte/issues/792",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2428430294
|
Cannot find svelte inspector when running dev server
Describe the bug
Recently had to reinstall Windows and am trying to pick my svelte-kit project back up again. I get the following issue when trying to run the dev server:
error when starting dev server:
file:///C:/Users/llewe/Documents/web/projects/reg-message/reg-message-frontend/node_modules/.pnpm/@sveltejs+vite-plugin-svelte@3.1.1_svelte@4.2.18_vite@5.3.4/node_modules/@sveltejs/vite-plugin-svelte/src/index.js:3
import { svelteInspector } from '@sveltejs/vite-plugin-svelte-inspector';
^^^^^^^^^^^^^^^
SyntaxError: Named export 'svelteInspector' not found. The requested module '@sveltejs/vite-plugin-svelte-inspector' is a CommonJS module, which may not support all module.exports as named exports.
CommonJS modules can always be imported via the default export, for example using:
import pkg from '@sveltejs/vite-plugin-svelte-inspector';
const { svelteInspector } = pkg;
at ModuleJob._instantiate (node:internal/modules/esm/module_job:171:21)
at async ModuleJob.run (node:internal/modules/esm/module_job:254:5)
at async ModuleLoader.import (node:internal/modules/esm/loader:474:24)
at async load_config (file:///C:/Users/llewe/Documents/web/projects/reg-message/reg-message-frontend/node_modules/.pnpm/@sveltejs+kit@2.5.18_@sveltejs+vite-plugin-svelte@3.1.1_svelte@4.2.18_vite@5.3.4__svelte@4.2.18_vite@5.3.4/node_modules/@sveltejs/kit/src/core/config/index.js:70:17)
at async sveltekit (file:///C:/Users/llewe/Documents/web/projects/reg-message/reg-message-frontend/node_modules/.pnpm/@sveltejs+kit@2.5.18_@sveltejs+vite-plugin-svelte@3.1.1_svelte@4.2.18_vite@5.3.4__svelte@4.2.18_vite@5.3.4/node_modules/@sveltejs/kit/src/exports/vite/index.js:130:24)
at async Promise.all (index 0)
at async asyncFlatten (file:///C:/Users/llewe/Documents/web/projects/reg-message/reg-message-frontend/node_modules/.pnpm/vite@5.3.4/node_modules/vite/dist/node/chunks/dep-D8YhmIY-.js:17369:12)
at async resolveConfig (file:///C:/Users/llewe/Documents/web/projects/reg-message/reg-message-frontend/node_modules/.pnpm/vite@5.3.4/node_modules/vite/dist/node/chunks/dep-D8YhmIY-.js:66463:27)
at async _createServer (file:///C:/Users/llewe/Documents/web/projects/reg-message/reg-message-frontend/node_modules/.pnpm/vite@5.3.4/node_modules/vite/dist/node/chunks/dep-D8YhmIY-.js:63050:18)
at async CAC.<anonymous> (file:///C:/Users/llewe/Documents/web/projects/reg-message/reg-message-frontend/node_modules/.pnpm/vite@5.3.4/node_modules/vite/dist/node/cli.js:735:20)
When creating a new sveltekit project using the same package versions, the issue is not present.
Reproduction
Install git bash for windows (v2.45.2) here
Install pnpm v9.6.0 with npm install -g pnpm@9.6.0
Clone the repo found here: f-llewellyn/reg-message-frontend
Navigate to the cloned repo
Run pnpm i to install the packages
Run pnpm dev to run the dev server
Logs
No response
System Info
System:
OS: Windows 11 10.0.22631
CPU: (12) x64 AMD Ryzen 5 5600 6-Core Processor
Memory: 18.20 GB / 31.91 GB
Binaries:
Node: 22.3.0 - C:\Program Files\nodejs\node.EXE
npm: 10.8.1 - C:\Program Files\nodejs\npm.CMD
pnpm: 9.6.0 - C:\Program Files\nodejs\pnpm.CMD
Browsers:
Edge: Chromium (126.0.2592.113)
Internet Explorer: 11.0.22621.3527
npmPackages:
@sveltejs/adapter-auto: ^3.2.2 => 3.2.2
@sveltejs/kit: ^2.5.18 => 2.5.18
@sveltejs/vite-plugin-svelte: ^3.1.1 => 3.1.1
svelte: ^4.2.18 => 4.2.18
vite: ^5.3.4 => 5.3.4
Severity
blocking all usage of SvelteKit
Additional Information
No response
duplicate of https://github.com/sveltejs/vite-plugin-svelte/issues/922
this seems to be caused by pnpm on windows , try reinstalling it and clearing its cache
yep, pnpm issue.
|
gharchive/issue
| 2024-07-24T09:48:25 |
2025-04-01T06:40:31.948172
|
{
"authors": [
"dominikg",
"f-llewellyn",
"rrrrigel"
],
"repo": "sveltejs/vite-plugin-svelte",
"url": "https://github.com/sveltejs/vite-plugin-svelte/issues/948",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
116690645
|
update peer dependencies
This is so that npm doesnt complain when using react 0.14
Thanks :)
|
gharchive/pull-request
| 2015-11-13T02:58:52 |
2025-04-01T06:40:31.949214
|
{
"authors": [
"svenanders",
"tikotzky"
],
"repo": "svenanders/react-iframe",
"url": "https://github.com/svenanders/react-iframe/pull/2",
"license": "isc",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2465850589
|
Non-zero exit status 3
Using pdf-sign file.pdf, I receive the error below. I haven't run into this before the new version (maybe when pdftk was used?). Seems to be for only some files.
WARNING: file.pdf, object 52 0 at offset 395659: this widget annotation is not reachable from /AcroForm in the document catalog
WARNING: file.pdf, object 55 0 at offset 387012: this widget annotation is not reachable from /AcroForm in the document catalog
WARNING: file.pdf, object 58 0 at offset 378362: this widget annotation is not reachable from /AcroForm in the document catalog
WARNING: file.pdf (object 10 0): object has offset 0
WARNING: file.pdf (object 15 0): object has offset 0
WARNING: file.pdf (object 414 0): object has offset 0
WARNING: file.pdf, object 52 0 at offset 395659: this widget annotation is not reachable from /AcroForm in the document catalog
WARNING: file.pdf, object 55 0 at offset 387012: this widget annotation is not reachable from /AcroForm in the document catalog
WARNING: file.pdf, object 58 0 at offset 378362: this widget annotation is not reachable from /AcroForm in the document catalog
qpdf: operation succeeded with warnings; resulting file may have some problems
Traceback (most recent call last):
File "/nix/store/ic3q5cr8jx5ghwhw2k8sccvn6dhd2jzj-pdf-sign-git/bin/.pdf-sign-wrapped", line 755, in <module>
main(parser.parse_args(sys.argv[1:] or ['-h']))
File "/nix/store/ic3q5cr8jx5ghwhw2k8sccvn6dhd2jzj-pdf-sign-git/bin/.pdf-sign-wrapped", line 38, in main
qpdfOrPdftk([
File "/nix/store/ic3q5cr8jx5ghwhw2k8sccvn6dhd2jzj-pdf-sign-git/bin/.pdf-sign-wrapped", line 487, in qpdfOrPdftk
subprocess.run(cmd, check=True)
File "/nix/store/7hnr99nxrd2aw6lghybqdmkckq60j6l9-python3-3.11.9/lib/python3.11/subprocess.py", line 571, in run
raise CalledProcessError(retcode, process.args,
subprocess.CalledProcessError: Command '['qpdf', '--flatten-annotations=all', '--generate-appearances', 'file.pdf', '/tmp/tmpmu_7ot87/input.pdf']' returned non-zero exit status 3.
@GregorySchwartz Would be great if you could provide 1) An example problematic file without sensitive content and 2) The nix flake producing ic3q5cr8jx5ghwhw2k8sccvn6dhd2jzj or versions for pdf-sign and qpdf.
@GregorySchwartz Bump. An example file might be enough.
Unfortunately the only files I could reproduce this with have sensitive information...
@GregorySchwartz Could you please test branch dev?
Executing as before, you should see the same warnings, but no traceback. pdf-sign should print qpdf exited with code 3, indicating warnings. Use --continue-on-warnings to ignore. and exit with code 1.
Adding --continue-on-warnings to the invocation should cause pdf-sign to function normally, except the warnings should still be printed.
|
gharchive/issue
| 2024-08-14T13:32:01 |
2025-04-01T06:40:31.963607
|
{
"authors": [
"GregorySchwartz",
"svenssonaxel"
],
"repo": "svenssonaxel/pdf-sign",
"url": "https://github.com/svenssonaxel/pdf-sign/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1909862225
|
RoadRunner integration
[x] #29
[x] #30
[ ] #31
[x] #36
Ryan is working on this
|
gharchive/issue
| 2023-09-23T12:50:30 |
2025-04-01T06:40:31.980010
|
{
"authors": [
"JJTech0130"
],
"repo": "svhsrobotics/FtcRobotController",
"url": "https://github.com/svhsrobotics/FtcRobotController/issues/28",
"license": "BSD-3-Clause-Clear",
"license_type": "permissive",
"license_source": "github-api"
}
|
125299736
|
MethodError: |(::Pattern, ::SubstitutionString{ASCIIString})
_
_ _ _(_)_ | A fresh approach to technical computing
(_) | (_) (_) | Documentation: http://docs.julialang.org
_ _ _| |_ __ _ | Type "?help" for help.
| | | | | | |/ _` | |
| | |_| | | | (_| | | Version 0.4.2 (2015-12-06 21:47 UTC)
_/ |\__'_|_|_|\__'_| | Official http://julialang.org/ release
|__/ | x86_64-w64-mingw32
julia> using Lisp
ERROR: LoadError: LoadError: MethodError: `|` has no method matching |(::ParserCombinator.Pattern, ::Base.SubstitutionString{ASCIIString})
Closest candidates are:
|(::Any, ::Any, ::Any, ::Any...)
|(::ParserCombinator.Matcher, ::ParserCombinator.Alt)
|(::ParserCombinator.Matcher, ::ParserCombinator.Matcher)
in include at boot.jl:261
in include_from_node1 at loading.jl:304
in include at boot.jl:261
in include_from_node1 at loading.jl:304
in require at loading.jl:243
while loading C:\Users\Peter\.julia\v0.4\Lisp\src\parser.jl, in expression starting on line 9
while loading C:\Users\Peter\.julia\v0.4\Lisp\src\Lisp.jl, in expression starting on line 2
julia>
This is the line: https://github.com/swadey/Lisp.jl/blob/master/src/parser.jl#L9
This was due to a change in ParserCombinator's API. Should be fixed now. I also updated everything to support 0.4 and higher.
Thank you very much!
|
gharchive/issue
| 2016-01-07T01:07:16 |
2025-04-01T06:40:31.999245
|
{
"authors": [
"Ismael-VC",
"swadey"
],
"repo": "swadey/Lisp.jl",
"url": "https://github.com/swadey/Lisp.jl/issues/2",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
281277106
|
"oneOf" property in @Schema annotation ignored in resource response
Q
A
Bug or feature request?
Unknown
Which Swagger-Core version?
v2.0.0-rc2
Which Java version?
1.8.0_131
Which JAX-RS framework & version?
2.0.1
I try to annotate my jaxrs class in order to generate an open api 3.0.0 spec that supports inheritance / polymorphism. For some reason I'm not able to generate a responses object that reflects the semantic of "oneOf" in order to show the user that different "subtypes" of the response type may be returned.
In detail, I want to achieve the following result
{
"openapi": "3.0.0",
"paths": {
"/test/bean": {
"get": {
"summary": "Test inheritance / polymorphism",
"operationId": "getBean",
"parameters": [
{
"name": "number",
"in": "query",
"description": "Test inheritance / polymorphism",
"required": true,
"schema": {
"type": "integer",
"format": "int32"
},
"example": 1
}
],
"responses": {
"200": {
"description": "bean answer",
"content": {
"application/json": {
"schema": {
"oneOf": [
{
"$ref": "#/components/schemas/MultipleSub1Bean"
},
{
"$ref": "#/components/schemas/MultipleSub2Bean"
}
]
}
}
}
}
}
}
}
},
"components": {
"schemas": {
"MultipleSub2Bean": {
"type": "object",
"properties": {
"d": {
"type": "integer",
"format": "int32"
}
},
"description": "MultipleSub2Bean",
"allOf": [
{
"$ref": "#/components/schemas/MultipleBaseBean"
}
]
},
"MultipleBaseBean": {
"type": "object",
"properties": {
"beanType": {
"type": "string"
},
"a": {
"type": "integer",
"format": "int32"
},
"b": {
"type": "string"
}
},
"description": "MultipleBaseBean",
"example": {
"a": 0,
"b": "test"
}
},
"MultipleSub1Bean": {
"type": "object",
"properties": {
"c": {
"type": "integer",
"format": "int32"
}
},
"description": "MultipleSub1Bean",
"allOf": [
{
"$ref": "#/components/schemas/MultipleBaseBean"
}
]
}
}
}
}
At present I use the following java code:
@Path("/test")
public class TestSwaggerApi {
@GET
@Path("/bean")
@Operation(summary = "Test inheritance / polymorphism",
responses = {
@ApiResponse(description = "bean answer",
responseCode = "200",
content = @Content(
mediaType = MediaType.APPLICATION_JSON,
schema = @Schema(
oneOf = { MultipleSub1Bean.class, MultipleSub2Bean.class }
)
)
)
})
@Produces({ MediaType.APPLICATION_JSON })
public MultipleBaseBean getBean(
@Context HttpServletRequest req,
@Parameter(description = "Test inheritance / polymorphism",
required = true,
example = "1")
@QueryParam("number") final int beanNumber) {
// do something...
...
}
}
@io.swagger.v3.oas.annotations.media.Schema(
description = "MultipleBaseBean",
subTypes = { MultipleSub1Bean.class, MultipleSub2Bean.class }
)
public class MultipleBaseBean {
public String beanType;
public int a;
public String b;
}
@io.swagger.v3.oas.annotations.media.Schema(
description = "MultipleSub1Bean"
)
public class MultipleSub1Bean extends MultipleBaseBean {
public int c;
}
@io.swagger.v3.oas.annotations.media.Schema(
description = "MultipleSub2Bean"
)
public class MultipleSub2Bean extends MultipleBaseBean {
public int d;
}
But the generated api holds not the expected responses object:
{
"openapi": "3.0.0",
"paths": {
"/test/bean": {
"get": {
"summary": "Test inheritance / polymorphism",
"operationId": "getBean",
"parameters": [
{
"name": "number",
"in": "query",
"description": "Test inheritance / polymorphism",
"required": true,
"schema": {
"type": "integer",
"format": "int32"
},
"example": 1
}
],
"responses": {
"200": {
"description": "bean answer",
"content": {
"application/json": {
"schema": {
"type": "string"
}
}
}
}
}
}
}
},
"components": {
"schemas": {
"MultipleSub2Bean": {
"type": "object",
"properties": {
"d": {
"type": "integer",
"format": "int32"
}
},
"description": "MultipleSub2Bean",
"allOf": [
{
"$ref": "#/components/schemas/MultipleBaseBean"
}
]
},
"MultipleBaseBean": {
"type": "object",
"properties": {
"beanType": {
"type": "string"
},
"a": {
"type": "integer",
"format": "int32"
},
"b": {
"type": "string"
}
},
"description": "MultipleBaseBean",
"example": {
"a": 0,
"b": "test"
}
},
"MultipleSub1Bean": {
"type": "object",
"properties": {
"c": {
"type": "integer",
"format": "int32"
}
},
"description": "MultipleSub1Bean",
"allOf": [
{
"$ref": "#/components/schemas/MultipleBaseBean"
}
]
}
}
}
}
I played around with the swagger test classes code, but I was not able to produce the requested result. I also tried to use the "implementation" property, but by doing so my "oneOf" property definition gets completely ignored.
Am I doing something wrong, or is this a missing feature?
It was indeed a missing feature / bug; correct support has been added in #2585, and is part of latest 2.0.0-SNAPSHOT version available on sonatype, and will be part of next release; closing ticket, please reopen if your issue isn't fixed.
Appreciate your work! Thank you very much!!!
|
gharchive/issue
| 2017-12-12T06:56:38 |
2025-04-01T06:40:32.039279
|
{
"authors": [
"duerrm",
"frantuma"
],
"repo": "swagger-api/swagger-core",
"url": "https://github.com/swagger-api/swagger-core/issues/2579",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2603066083
|
Improved footer
Issue resolved: #630
Pull Request Format
PR Title
Issue #0 : ISSUE NAME solved
Type of PR
Add X in the box to specify the improvement type.
[X] Bug fix
[ ] Feature enhancement
[ ] Documentation update
[ ] Other (specify): ___________
Description
This pull request aims to [briefly describe what the pull request does, e.g., "add a new feature," "fix a bug," "improve documentation," etc.].
Screenshots / Videos (if applicable)
Before:
[Describe the state before the changes, e.g., "The section headings in the footer were misaligned, leading to an inconsistent and cluttered appearance."]
After:
[Describe the state after the changes, e.g., "The section headings in the footer are now properly aligned, providing a clean and professional look."]
Checklist
Add X in the box to specify.
[X] I have performed a self-review of my code.
[ ] I have tested the changes thoroughly before submitting this pull request.
[ ] I have provided relevant issue numbers, screenshots, and videos after making the changes.
[ ] I have commented my code, particularly in hard-to-understand areas.
Additional Context
[Provide any additional context about the changes, such as specific challenges you faced or decisions you made. This can help reviewers understand the rationale behind your changes.]
Thank you for reviewing my pull request!
@Kajalmehta29 Too poor footer. If you can make it better. Not need separate page for feedback.
|
gharchive/pull-request
| 2024-10-21T16:20:58 |
2025-04-01T06:40:32.113701
|
{
"authors": [
"Kajalmehta29",
"swaraj-das"
],
"repo": "swaraj-das/Collect-your-GamingTools",
"url": "https://github.com/swaraj-das/Collect-your-GamingTools/pull/633",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2595715716
|
exportToExcel getting Call to undefined method EDL::isGroup() on v2.1.27
Before we get into the details, a quick question: does v2.1.27 require a newer PHP version?
I'm still running PHP v8.3.11.
Also, my installed older version was showing as 2.1.25 but I think it was a bugfix dev release that rolled into v2.1.26 if I remember correctly (could be wrong).
Describe the bug
I ran a command that worked fine the last time I used it (then-current-release or possibly dev release v2.1.25 downloaded 2 Sep 2024), but after upgrading to v2.1.27 latest release today I get the error in the title.
Expected behavior
Normal output per previous versions.
Current behavior
> php -r "require_once 'C:/tools/pan/pan-os-php/utils/pan-os-php.php';" type=rule in=api://panorama.domain.com location=any ruletype=security 'actions=exportToExcel:DevQA-Rules.xls' 'filter=(src has.recursive network1) or (src has.recursive host1) or (src has.recursive host2) or (src has.recursive host3) or (src has.recursive host4) or (dst has.recursive network1) or (dst has.recursive host1) or (dst has.recursive host2) or (dst has.recursive host3) or (dst has.recursive host4)'
***********************************************
*********** pan-os-php.php type=rule UTILITY **************
- PAN-OS-PHP version: 2.1.27 [WIN] [8.3.11]
- Downloading config from API...
- Detected platform type is 'panorama'
- No 'template' provided so using default ='any'
- filter after sanitization : (src has.recursive network1) or (src has.recursive host1) or (src has.recursive host2) or (src has.recursive host3) or (src has.recursive host4) or (dst has.recursive network1) or (dst has.recursive host1) or (dst has.recursive host2) or (dst has.recursive host3) or (dst has.recursive host4)
- Loading configuration through PAN-OS-PHP library...
(0.97 seconds, 52.43 mb memory)
- PAN-OS-PHP APP-ID version: 8902-9003
- PAN-OS APP-ID version: 8904-9011
- PAN-OS AV version: 4974-5492
- PAN-OS WF version: 515890-518890
- PAN-OS THREAT version: 8904-9011
- PAN-OS version: 110
- PAN-OS Device timezone: US/Pacific is used. actual time: 2024/10/17 13:09:44
- action 'exportToExcel' has tasks to process before start.
- action 'exportToExcel' has tasks to process before start.
- action 'exportToExcel' has tasks to process before start.
* processing ruleset 'PanoramaConf: / RuleStore:Security' that holds 454 rules
PHP Fatal error: Uncaught Error: Call to undefined method EDL::isGroup() in C:\tools\pan\pan-os-php\lib\container-classes\AddressRuleContainer.php:555
Stack trace:
#0 C:\tools\pan\pan-os-php\lib\misc-classes\RQuery.php(199) : eval()'d code(1): AddressRuleContainer->hasObjectRecursive()
#1 C:\tools\pan\pan-os-php\lib\misc-classes\RQuery.php(199): eval()
#2 C:\tools\pan\pan-os-php\lib\misc-classes\RQuery.php(331): RQuery->matchSingleObject()
#3 C:\tools\pan\pan-os-php\utils\lib\RULEUTIL.php(371): RQuery->matchSingleObject()
#4 C:\tools\pan\pan-os-php\utils\lib\RULEUTIL.php(37): RULEUTIL->time_to_process_objects()
#5 C:\tools\pan\pan-os-php\utils\lib\UTIL.php(215): RULEUTIL->utilStart()
#6 C:\tools\pan\pan-os-php\lib\misc-classes\PH.php(1060): UTIL->__construct()
#7 C:\tools\pan\pan-os-php\utils\pan-os-php.php(118): PH::callPANOSPHP()
#8 Command line code(1): require_once('...')
#9 {main}
thrown in C:\tools\pan\pan-os-php\lib\container-classes\AddressRuleContainer.php on line 555
Possible solution
On v2.1.25 (again, not sure if this was the stable or dev release from my last issue report), I get this:
> php -r "require_once 'C:/tools/pan/pan-os-php/utils/pan-os-php.php';" type=rule in=api://panorama.domain.com location=any ruletype=security 'actions=exportToExcel:DevQA-Rules.xls' 'filter=(src has.recursive network1) or (src has.recursive host1) or (src has.recursive host2) or (src has.recursive host3) or (src has.recursive host4) or (dst has.recursive network1) or (dst has.recursive host1) or (dst has.recursive host2) or (dst has.recursive host3) or (dst has.recursive host4)'
***********************************************
*********** pan-os-php.php type=rule UTILITY **************
- PAN-OS-PHP version: 2.1.25 [WIN] [8.3.11]
- Downloading config from API...
- Detected platform type is 'panorama'
- No 'template' provided so using default ='any'
- filter after sanitization : (src has.recursive network1) or (src has.recursive host1) or (src has.recursive host2) or (src has.recursive host3) or (src has.recursive host4) or (dst has.recursive network1) or (dst has.recursive host1) or (dst has.recursive host2) or (dst has.recursive host3) or (dst has.recursive host4)
- Loading configuration through PAN-OS-PHP library...
(1.20 seconds, 52.43 mb memory)
- PAN-OS-PHP APP-ID version: 8867-8824
- PAN-OS APP-ID version: 8904-9011
- PAN-OS AV version: 4974-5492
- PAN-OS WF version: 515890-518890
- PAN-OS THREAT version: 8904-9011
- PAN-OS version: 110
- PAN-OS Device timezone: US/Pacific is used. actual time: 2024/10/17 13:14:09
- action 'exportToExcel' has tasks to process before start.
- action 'exportToExcel' has tasks to process before start.
- action 'exportToExcel' has tasks to process before start.
* processing ruleset 'PanoramaConf: / RuleStore:Security' that holds 454 rules
- object 'rule001' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'File-Sharing-01' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'File-Sharing-02' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule002' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule003' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule004' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule005' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule006' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule007' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule008' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule009' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
* objects processed in DG/Vsys '' : 11 filtered over 454 available
* processing ruleset 'PanoramaConf: / DeviceGroup:FW1 / RuleStore:Security' that holds 950 rules
- object 'rule010' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule011' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule012' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule013' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014b' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule015' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule016' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule017' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule018' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule018a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule019' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule019a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule020' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule021' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule021a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule022' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule023' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule024' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule025' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule025a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule026' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule027' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule028' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule029' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule030' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
* objects processed in DG/Vsys 'FW1' : 28 filtered over 950 available
* processing ruleset 'PanoramaConf: / DeviceGroup:FW2 / RuleStore:Security' that holds 1469 rules
- object 'rule011' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule010' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule015' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule016' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule012' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule029' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule030' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule018' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule018a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule013' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule017' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule019' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule014b' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule020' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule021' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule021a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule023' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule024' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule025' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule025a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule026' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule031' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule028' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule019a' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule013-1' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule012b' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'rule013-2' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
- object 'File-Sharing-01-T2' passing through Action='exportToExcel' Args: filename=DevQA-Rules.xls, additionalFields=,
* objects processed in DG/Vsys 'FW2' : 30 filtered over 1469 available
- action 'exportToExcel' has tasks to process before shutdown.
**** PROCESSED 69 objects over 2873 available ****
************* END OF SCRIPT pan-os-php.php type=rule ************
Your Environment
Environment name and version: PowerShell v5.1
Operating System and version (desktop or mobile): Windows 10
Thanks for sharing this.
Sorry I did not saw your ticket.
This bug is now fixed in latest develop container, and it will be published in the next few week to latest container.
in your case please use:
https://github.com/swaschkut/pan-os-php/tree/develop_2-1-X
No worries - I just assumed you were busy or on holiday. Thanks for getting to it.
I'll check out the dev build in a few hours.
Sorry, I was doing some punishing PANOS upgrades that did not go well (nice 16 hour day). I'll check it out later today (Wednesday).
My apologies for the delay - I finally was able to test the dev build and it worked fine.
Thanks!
version 2.1.28 is now published and include this fix
#30
|
gharchive/issue
| 2024-10-17T20:45:37 |
2025-04-01T06:40:32.129330
|
{
"authors": [
"bethatasitmay",
"swaschkut"
],
"repo": "swaschkut/pan-os-php",
"url": "https://github.com/swaschkut/pan-os-php/issues/29",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
}
|
2327689154
|
doc: add @swc/core dependency to usage example
On a fresh repo, running the previous example would result in a Error: Cannot find module '@swc/core' otherwise.
I think it'd also be good to add @swc/core as a peer dependency to @swc-node/register, so that @swc/core will be installed automatically when missing.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
@swc/core in peerDependencies works as expected
I see it: https://github.com/swc-project/swc-node/blob/a1098be19a37077c419e42a60cf81889c5c94704/packages/register/package.json#L51
I guess my npm is not set up to automatically install peer dependencies.
Found the reason why npm did not install @swc/core: I had legacy-peer-deps=true in my .npmrc. Removing it made it install the peer dep.
|
gharchive/pull-request
| 2024-05-31T12:04:01 |
2025-04-01T06:40:32.140162
|
{
"authors": [
"CLAassistant",
"silverwind",
"yeliex"
],
"repo": "swc-project/swc-node",
"url": "https://github.com/swc-project/swc-node/pull/782",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1160189902
|
Unicode character U+309B not accepted in identifiers
Describe the bug
Some unicode character not accepted in identifiers.
If using unicode character, it will get the error “”Invalid character in identifier"
If using encoded character, it will get the error: Unexpected character '゛'
Input code
var ゛; // U+309B
var \u309B;
var \u309C;
var a\u309B;
var a\u309C;
Config
{
"jsc": {
"parser": {
"syntax": "ecmascript"
},
"target": "es5"
},
"module": {
"type": "commonjs"
},
"isModule": false
}
Playground link
https://play.swc.rs/?version=1.2.148&code=H4sIAAAAAAAAAytLLFJ43DTbWkFfXyFU29jA0gkAY3O%2FZhIAAAA%3D&config=H4sIAAAAAAAAAy2LQQqAMAwE%2F5KzVy%2F9g48INUpL05YkgkX6dw142t0Z9oGsEcIDHUVJvOmohjcEoMioUVI3mAsYyknmWFff3ParkB9s9C8hNuZWs7pMuv36wKI0X3tLUmhnAAAA
https://play.swc.rs/?version=1.2.148&code=H4sIAAAAAAAAAytLLFKIKTU2sHSy5iqDsZ0h7ERkiUSoDAASLisXMQAAAA%3D%3D&config=H4sIAAAAAAAAAy2LQQqAMAwE%2F5KzVy%2F9g48INUpL05YkgkX6dw142t0Z9oGsEcIDHUVJvOmohjcEoMioUVI3mAsYyknmWFff3ParkB9s9C8hNuZWs7pMuv36wKI0X3tLUmhnAAAA
Expected behavior
SWC should properly parse these unicode as these code are valid identifiers.
Actual behavior
error: Unexpected character '゛'
--> input.js:1:5
|
1 | var ゛; // U+309B
| ^
Caused by:
0: failed to process js file
1: failed to process js file
2: Syntax Error
error: Invalid character in identifier
--> input.js:1:5
|
1 | var \u309B;
| ^^^^^^
Version
1.2.148
Additional context
I run swc use cmd as follows
npx swc --config-file=.swcrc poc.js -o poc.js
The environment is ubuntu 20.04 amd64.
Isn't this same as https://github.com/swc-project/swc/issues/3341 except charpoint?
@kwonoj Sounds like yes, maybe we can union them
It's an identical issue iff U+309B unicode id and not unicode xid.
I didn't close this as a duplicate as I need to check it.
I found https://github.com/Boshen/unicode-id
I'll patch it.
|
gharchive/issue
| 2022-03-05T02:12:17 |
2025-04-01T06:40:32.146447
|
{
"authors": [
"alexander-akait",
"clhiker",
"kdy1",
"kwonoj"
],
"repo": "swc-project/swc",
"url": "https://github.com/swc-project/swc/issues/3862",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1976290497
|
JSX automatic contains classic output when jsx spread comes before other props
Describe the bug
It seems the emit output contains the classic output when a jsx spread comes before other props.
Input code
const TabButton = (tab) => {
const props = { className: tab };
return (
<button {...props} key={tab}>
{tab}
</button>
)
}
const TabButton2 = (tab) => {
const props = { className: tab };
return (
<button key={tab} {...props}>
{tab}
</button>
)
}
Config
{
"jsc": {
"parser": {
"syntax": "ecmascript",
"jsx": true
},
"target": "es2022",
"loose": false,
"minify": {
"compress": false,
"mangle": false
},
"transform": {
"react": {
"importSource": "npm:react",
"runtime": "automatic"
}
}
},
"module": {
"type": "es6"
},
"minify": false,
"isModule": true
}
Playground link (or link to the minimal reproduction)
https://play.swc.rs/?version=1.3.96-nightly-20231103.1&code=H4sIAAAAAAAAA0vOzysuUQhJTHIqLSnJz1OwVdAoSUzSVLC1U6jmUlBIBksXFOUXFAOlqhWScxKLi%2F0Sc1OtFIDKFGqtuYCKilJLSovyFDSATAUFmySIQdV6enpgfbUK2amVttVA5bV2YBUKCmAORLU%2BRDlIRpOrlisZ1TlG1HMP3BFILiPoHgB6b3fKHgEAAA%3D%3D&config=H4sIAAAAAAAAA1WQPQ7CMAyFd05ReWZAHRi4AxMnsIJbBdVJZDsSFerdSdMfyhIl33vPjv05NQ281MGt%2BZRreSQUJdnfhegYDN%2BFADlGdeKTwXlTXzpLJpkqmRYBDKUnqyFtL227BmCIUangDgellbEPvhuPLV3kJKT6b5ytGPphz%2F93FAzaReFjISF0dgAFeU5R7BGzuLkQhMS3xXb%2BmSQH81x1zBYZzTtY5em0nbU1cHzm%2Bqd1gzYmWga%2Fws%2B0zbjPA17vW7Kub%2FoCEngkRo0BAAA%3D
SWC Info output
No response
Expected behavior
Should use jsx automatic for everything.
Actual behavior
import { jsx as _jsx } from "npm:react/jsx-runtime";
import { createElement as _createElement } from "npm:react";
const TabButton = (tab)=>{
const props = {
className: tab
};
return /*#__PURE__*/ _createElement("button", {
...props,
key: tab
}, tab);
};
const TabButton2 = (tab)=>{
const props = {
className: tab
};
return /*#__PURE__*/ _jsx("button", {
...props,
children: tab
}, tab);
};
Version
1.3.96-nightly-20261103.1
Additional context
No response
This is intentional and the source code has comments for this
Huh, interesting. Thanks! Seems like both TypeScript and Babel also do this:
https://www.typescriptlang.org/play?jsx=4#code/MYewdgzgLgBAKgQwEYCECuUrhgXhgCimQEpcA+GAbwFgAoGGUSWABwCcQWJcrGAbBBAgA5BAFsApgC4YRJDAC+Abjp0GbCVDRswBNQxgAeJBiy7KAOivtOEBTADWEgJ45KchWX0GqH70YB6E0xwL3oYYjoFOiZoeGR0ELAAJh5CEnIqfVjWDi4eSn5BEXFpWWRFFVp9DS0dPXCGY1NsJ1d3ZHtLazy7MJ8GDqRoxsDgs36IqKA
https://babeljs.io/repl#?browsers=&build=&builtIns=false&corejs=3.6&spec=false&loose=false&code_lz=MYewdgzgLgBAKgQwEYCECuUrhgXhgCimQEpcA-GAbwCgYZRJYAHAJxCYlyvoBsEIIAOQQBbAKYAuGESQwAvgG5qtGCzFQ0LMARV0APEgxZtlAHTnW7CHJgBrMQE8clGXLK66VVx70B6Q5jg7nTE1HLUDNDwyOiBYABMXIQk5FQqkcxsHFyUvPxCopLSyPJKKmoaWjqeMAZG2PZOLsg2ZhZZ1sE1Xi0-_vVgXaFyQA&debug=false&forceAllTransforms=false&modules=false&shippedProposals=false&circleciRepo=&evaluate=false&fileSize=false&timeTravel=false&sourceType=module&lineWrap=true&presets=es2015%2Creact%2Cstage-2&prettier=false&targets=&version=7.23.2&externalPlugins=&assumptions={}
|
gharchive/issue
| 2023-11-03T14:10:01 |
2025-04-01T06:40:32.153170
|
{
"authors": [
"dsherret",
"kdy1"
],
"repo": "swc-project/swc",
"url": "https://github.com/swc-project/swc/issues/8215",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1044536758
|
fix(node-swc/types): add missing es6 config type(#2633)
Closes #2633
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
Should I rebase it?
No, I rebased it.
|
gharchive/pull-request
| 2021-11-04T09:53:28 |
2025-04-01T06:40:32.155908
|
{
"authors": [
"CLAassistant",
"kdy1",
"magic-akari"
],
"repo": "swc-project/swc",
"url": "https://github.com/swc-project/swc/pull/2648",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1889360731
|
feat(es/compat): impl pure_getters for optional_chaining
Description:
BREAKING CHANGE:
Related issue (if exists):
Closes #7921
I don't believe it would be easy. Both are referenced many times.
|
gharchive/pull-request
| 2023-09-10T22:24:09 |
2025-04-01T06:40:32.157830
|
{
"authors": [
"Austaras"
],
"repo": "swc-project/swc",
"url": "https://github.com/swc-project/swc/pull/7933",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1537277079
|
fixes branch ref
currently refs
ref: 912/merge
which is incorrect branch name.
Frontend and Release Workflow Started here
Frontend and Release Workflow Started here
|
gharchive/pull-request
| 2023-01-18T01:52:48 |
2025-04-01T06:40:32.197859
|
{
"authors": [
"adrian-kong",
"swiftnav-travis"
],
"repo": "swift-nav/swift-toolbox",
"url": "https://github.com/swift-nav/swift-toolbox/pull/914",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
273076454
|
Use Dockerfile.template as root source of truth.
This commit takes the lead from
docker-node, separating
supported versions into their own folders, using Dockerfiles generated
from template. This commit closes #92 by backporting Clang 3.8 into 3.1.
I have also removed two soon to expire PGP keys from the file and added the new automation key.
@swizzlr i wanted to do this forever!!!! redis, ruby, and a bunch of other docker images are doing the same thing! I think i mentioned this in some discussion earlier Thank you!
I've smoke tested both 3 and 4 with two separate Swift 3 and 4 codebases respectively.
|
gharchive/pull-request
| 2017-11-10T22:16:56 |
2025-04-01T06:40:32.200017
|
{
"authors": [
"hamin",
"swizzlr"
],
"repo": "swiftdocker/docker-swift",
"url": "https://github.com/swiftdocker/docker-swift/pull/95",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2598404536
|
Tests can’t jump to generated interface of stdlib
I needed to disable a few tests in SwiftInterfaceTests because they can’t jump to generated interfaces of the stdlib. Investigate why.
Synced to Apple’s issue tracker as rdar://138210215
|
gharchive/issue
| 2024-10-18T21:19:58 |
2025-04-01T06:40:32.205433
|
{
"authors": [
"ahoppen"
],
"repo": "swiftlang/sourcekit-lsp",
"url": "https://github.com/swiftlang/sourcekit-lsp/issues/1771",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2669873402
|
Long cache retrieval when resolving dependencies
Description
SPM dependency resolution takes a very long time to extract. Each packet takes on average ~1 second to extract.
As I understand it, the problem is that the cache is stored in a compressed form and takes time to extract. But it seems to me that this is of less practical use than a faster resolution.
I would like the user cache to store the original artifacts (the same as in the local .build directory)
I assume you mean "package" not "packet". But yes, it's actually doing a git clone and checkout from the cache which is a very expensive operation.
I think we're probably at the point where we're doing the best we can. SwiftPM does a complete git clone into the user cache under .swiftpm/cache/repositories. We need that so we have all the versions available during dependency resolution and allows you to have multiple projects using different versions. We then do a shallow clone of the selected revision over to the package scratch directory. One second to do is about as fast as git can go.
The solution we're hoping for is the previous work with package registries where the registry would have the metadata we need and a source archive for each available version. Then we can cache those which would be much faster. But I'm not sure when a community wide solution for that will happen.
|
gharchive/issue
| 2024-11-18T20:46:42 |
2025-04-01T06:40:32.207829
|
{
"authors": [
"Ernest0-Production",
"dschaefer2"
],
"repo": "swiftlang/swift-package-manager",
"url": "https://github.com/swiftlang/swift-package-manager/issues/8123",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2545186033
|
[DO NOT MERGE] Check the codepage used on Windows in Jenkins
[One line description of your change]
Motivation:
[Explain here the context, and why you're making that change. What is the problem you're trying to solve.]
Modifications:
[Describe the modifications you've done.]
Result:
[After your change, what will change.]
Checklist:
[ ] Code and documentation should follow the style of the Style Guide.
[ ] If public symbols are renamed or modified, DocC references should be updated.
@swift-ci test Windows
@swift-ci test
*** CODEPAGE IN CI: 1252 -- CONSOLE: 65001
|
gharchive/pull-request
| 2024-09-24T11:54:30 |
2025-04-01T06:40:32.211447
|
{
"authors": [
"grynspan"
],
"repo": "swiftlang/swift-testing",
"url": "https://github.com/swiftlang/swift-testing/pull/727",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2420916565
|
[Macros] Remove SwiftOperators dependency
Looks like this got missed by #74502
@swift-ci please test
|
gharchive/pull-request
| 2024-07-20T14:15:56 |
2025-04-01T06:40:32.212323
|
{
"authors": [
"hamishknight"
],
"repo": "swiftlang/swift",
"url": "https://github.com/swiftlang/swift/pull/75384",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2485777492
|
[Concurrency] Add missing Sendable conformances to nested value types in Async{Throwing}Stream.
Termination, BufferingPolicy, and YieldResult all have trivial conformances to Sendable. The conformance on YieldResult is conditional on Element: Sendable because one of the enum cases stores an element in its associated value.
@swift-ci please smoke test
|
gharchive/pull-request
| 2024-08-26T03:24:46 |
2025-04-01T06:40:32.213791
|
{
"authors": [
"hborla"
],
"repo": "swiftlang/swift",
"url": "https://github.com/swiftlang/swift/pull/76079",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2508296069
|
scan-deps-deterministic-check-windows
Test commit to see the test failure on windows.
@swift-ci please smoke test windows platform
Closing as I have an idea what is going on.
|
gharchive/pull-request
| 2024-09-05T16:52:35 |
2025-04-01T06:40:32.214996
|
{
"authors": [
"cachemeifyoucan"
],
"repo": "swiftlang/swift",
"url": "https://github.com/swiftlang/swift/pull/76284",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2515264450
|
[Diagnostics] Add -print-diagnostic-groups flag
This change adds the -print-diagnostic-groups flag as described by SE-0443.
@DougGregor please take a look
I've also added the flag to many relevant tests. This way we test that the messages we're getting are expected and belong to the expected group. IMO it's nice to have some extra coverage so we can be sure we don't change the group of a warning accidentally. But let me know if you think otherwise.
Also, there's a use of DiagnosticOptions and DiagnosticEngine in LLDB so I've opened a PR there as well https://github.com/swiftlang/llvm-project/pull/9240
Only a small comment and a question, this looks good!
@swift-ci please smoke test
@DougGregor Since there are changes in the public members of DiagnosticOptions and DiagnosticEngine LLDB won't build without this patch https://github.com/swiftlang/llvm-project/pull/9240
I don't know how to deal with it, both of these PRs depend on each other and should be tested and merged together.
Can you rerun the tests here with that PR?
https://github.com/swiftlang/llvm-project/pull/9240
@swift-ci please smoke test
I don't know how to deal with it, both of these PRs depend on each other and should be tested and merged together.
Can you rerun the tests here with that PR?
Yes, of course. I've kicked off the multi-PR test, sorry I missed that earlier.
@DougGregor The windows tests failed because of the missing quote marks in ucrt.swift. https://github.com/swiftlang/swift/pull/76363/files#diff-b68568b1ae0d461890b6f117d611184202d92a42af3a9caad91e666d6e542415
https://github.com/swiftlang/llvm-project/pull/9240
@swift-ci please smoke test
|
gharchive/pull-request
| 2024-09-10T02:29:50 |
2025-04-01T06:40:32.221440
|
{
"authors": [
"DmT021",
"DougGregor"
],
"repo": "swiftlang/swift",
"url": "https://github.com/swiftlang/swift/pull/76363",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2348852925
|
Always register the swift-lldb debug adapter provider
Issue: #848
We can always register this and that way the user does not need to reload the window when use change to using lldb-dap. Now they can use the updated debug configurations right away
@adam-fowler the debug session does nothing if the path does not exist or cannot find in toolchain, so added a check to verify that the the debug adapter does exist so the user gets an error message instead of silently failing. I see that we are already listening to changes to swift.debugger.useDebugAdapterFromToolchain and swift.debugger.path settings, else where so seemed a little more complex than necessary to listen to this again to register the debug provider
|
gharchive/pull-request
| 2024-06-12T13:58:33 |
2025-04-01T06:40:32.223341
|
{
"authors": [
"award999"
],
"repo": "swiftlang/vscode-swift",
"url": "https://github.com/swiftlang/vscode-swift/pull/886",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
284299670
|
Brackets in SetFrom are missing
Q
A
Bug report?
yes
Feature request?
no
RFC?
no
How used?
in Yii 2
Swiftmailer version
5.4.3
PHP version
7.1.1
Observed behaviour
$mailer->setFrom(["name@mail.com" => "Name (Something in Brackets) xxx"])
generates this header information:
From: Name (Something in Brackets) xxx <name@mail.com>
But the received mail has this header information:
From: Name
xxx <name@mail.com>
The brackets are replaced with a line break.
Expected behaviour
The received message should have this header information:
From: Name (Something in Brackets) xxx <name@mail.com>
or
From: "Name (Something in Brackets) xxx" <name@mail.com>
Example
Yii::$app->mailer->setTextBody("This is a test mail.")
->setFrom(["name@mail.com" => "Name (Something in Brackets) xxx"])
->setTo("receiver@mail.com")
->setSubject('test mail')
->setHtmlBody('<b>This is a test mail.</b>')
->send();
@cech12 Please add which transport you used. (Maybe the brackets filtered for security reasons.)
@Rotzbua Sorry, but I don't know, what you mean with "transport". :/
I tested the normal php mail function on 2 different web servers
mail("receiver@mail.com", "test mail", "This is a test mail.", "From: Name (Something in Brackets) xxx <name@mail.com>");
and the result is the same.
The received message had this header information:
From: Name
xxx <name@mail.com>
Maybe there are some security mechanisms. How can I test it?
I asked for the sending method: mail(), sendmail or smtp. Swiftm. calls them Swift_SmtpTransport, Swift_SendmailTransport, Swift_MailTransport.
I suggest your swiftmailer use phps mail(), because it shows the same behaviour. mail() is just a proxy function in php. It is processed very different on different computer systems. That's the reason why Swift_MailTransport is removed in v6.
Try smtp?
By the way: I tested your code on my webspace (php7.0) and the brackets are not removed. Seems that my provider has a different setup than your provider.
Ah okay! Yes I use the mail transport Swift_MailTransport.
If this function is removed in the new swiftmailer version, in the future nobody will have this problem. I reached that I wanted to reach. :D
I will try the SMTP mail transport and if there is a Problem I will open a new Issue.
So this Issue can be closed. :)
Thanks for your help! :)
|
gharchive/issue
| 2017-12-23T10:36:44 |
2025-04-01T06:40:32.231787
|
{
"authors": [
"Rotzbua",
"cech12"
],
"repo": "swiftmailer/swiftmailer",
"url": "https://github.com/swiftmailer/swiftmailer/issues/1033",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.