id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
1077086690
|
Let the user decide which certificate is the one which is currently used
Current Implementation
The app decides which is the currently used certificate. It does so depending on some values, like how old a test is, whiter it's a RAT or a PCR test, etc.
Suggested Enhancement
Let the user decide which certificate is the one which is currently used.
Expected Benefits
The user can decide which certificate is the most important to them and select the currently used certificate accordingly.
Internal Tracking ID: EXPOSUREAPP-11160
Thanks for your suggesiton @Ein-Tim.
Internal Tracking ID: EXPOSUREAPP-11160
Corona-Warn-App Open Source Team
I had a few situations where I accidentally showed the wrong certificate by opening the "main" certificate (the first that will be shown in the app).
IMO, it would be best to just stick to the last vaccine certificate. The test result can still be opened with "certificates".
As the CWA project went into ramp-down mode, I don't expect this feature to be implemented. I'm therefore closing this issue.
|
gharchive/issue
| 2021-12-10T17:44:39 |
2025-04-01T04:33:53.067368
|
{
"authors": [
"CodingAir",
"Ein-Tim",
"larswmh"
],
"repo": "corona-warn-app/cwa-wishlist",
"url": "https://github.com/corona-warn-app/cwa-wishlist/issues/718",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2456366072
|
Updated Twitter icon to X logo
Changes Made
Updated the InfoCard.tsx component to replace the Twitter SVG logo with the new X logo.
Ensured compatibility across different browsers and devices.
Screenshots
Before:
To:
Fixed #484
Closing as duplicate
|
gharchive/pull-request
| 2024-08-08T18:13:04 |
2025-04-01T04:33:53.076213
|
{
"authors": [
"its-harshitgoel",
"rithviknishad"
],
"repo": "coronasafe/leaderboard",
"url": "https://github.com/coronasafe/leaderboard/pull/485",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
550107864
|
Add the ability to change the styles of a label
Is your feature request related to a problem? Please describe.
@tmdoit (a community member on latest.cortezaproject.org) wants to highlight mandatory fields.
Describe the solution you'd like
It would be nice to be able to set a style to the label, like what we have with automation buttons (primary, secondary, warning,...).
We added a required field indicator and we improved error messages on the form if the required field is empty for better user experience. In version 2020.3.
|
gharchive/issue
| 2020-01-15T10:45:29 |
2025-04-01T04:33:53.165566
|
{
"authors": [
"lennyhorstink",
"zmija"
],
"repo": "cortezaproject/corteza-webapp-compose",
"url": "https://github.com/cortezaproject/corteza-webapp-compose/issues/171",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1754954033
|
Fix namespace description and subtitle translations not highlighted when translation modal is opened
The following changes are implemented
TODO: Summary
Changes in the user interface:
TODO: Add screenshots, recordings or remove this section
Checklist when submitting a final (!draft) PR
[ ] Commits are tidied up, squashed if needed and follow guidelines in CONTRIBUTING.md
[ ] Code builds
[ ] All existing tests pass
[ ] All new critical code is covered by tests
[ ] PR is linked to the relevant issue(s)
[ ] Rebased with the target branch
LGTM
|
gharchive/pull-request
| 2023-06-13T13:54:31 |
2025-04-01T04:33:53.168299
|
{
"authors": [
"katrinDY",
"kelanik8"
],
"repo": "cortezaproject/corteza",
"url": "https://github.com/cortezaproject/corteza/pull/1255",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
188851951
|
Future-proofing for Webpack 2
Node version: 7.0.0+
npm version: 3.10.8
Operating system: Windows Server 2003 [[ 0_o ]]
Command line used: npm run build
Steps to reproduce: I was having a lot of issues getting the production build to work after upgrading to Webpack 2.0 beta (so I could use tree-shaking). Development was working fine, but I kept getting "l is not a function" errors when attempting to load the production bundle.
What looks like is happening is that your dynamic inclusion for configureStore() is not playing well with Webpack 2.0 with CommonJS replacement turned off.
I'm honestly not sure exactly what is going on -- I tried replacing the require() calls with System.import() calls but got the same problem. What did work was to combine the dev and prod versions of configureStore() into a single file and just use static inclusion.
If it isn't a huge efficiency to have these split up with dynamic inclusion, it would probably be more compatible with Webpack in the long run to just use a single file.
I'd be happy to make up a pull request if you'd like to review, or if you have a better idea I'm all ears. Love Slingshot, by the way! Thanks so much for making this available to the community!
Hi Matt - Thanks for the issue. This is good stuff. Yes, the dynamic inclusion is a problem because it's not statically analyzable. I'd welcome a PR with your change since I plan to move us to Webpack 2 as soon as it comes out of Beta.
I'd also love to see the changes you made to support Webpack 2 if you're willing to share your fork.
Thanks again! 👍
Absolutely! I'll put it together either tonight or tomorrow and get it to you!
This was merged with #313. Closing.
|
gharchive/issue
| 2016-11-11T21:11:41 |
2025-04-01T04:33:53.174289
|
{
"authors": [
"coryhouse",
"kwelch",
"mlwigdahl"
],
"repo": "coryhouse/react-slingshot",
"url": "https://github.com/coryhouse/react-slingshot/issues/309",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
241810605
|
Upgrade to webpack 3, and upgrade all other dependencies
Haven't done a ton of testing on this branch yet, but so far it looks ok.
This will upgrade to Webpack 3 and upgrades a ton of other packages too. If you'd like, I can try to separate those dependency upgrades out, but figured it'd be useful to get everything up to the latest versions.
Resolves #442
Coverage decreased (-1.0%) to 94.286% when pulling 27169b3f21780f25ac686aeed4f182cff6de88e2 on egdelwonk:webpack3 into 46b7d26ff8fa8de8fbb682d93446f0f1cc230cf7 on coryhouse:master.
Odd that coverage would go down for an non-code change. I will pull this down and test it tonight.
Thanks for the PR, Will!
Related, #442.
I made a couple minor tweaks, but looked great. Thanks @egdelwonk 👍
|
gharchive/pull-request
| 2017-07-10T18:53:00 |
2025-04-01T04:33:53.177859
|
{
"authors": [
"coryhouse",
"coveralls",
"egdelwonk",
"kwelch",
"nickytonline"
],
"repo": "coryhouse/react-slingshot",
"url": "https://github.com/coryhouse/react-slingshot/pull/444",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2286481083
|
Add dotplot tutorial slideshow to stage 1
This PR adds the dotplot tutorial slideshow to stage 1. This adds the dotplot viewer, but I haven't yet wired up the hoverover line in the viewer and the various line showing/hiding that takes place - I'll take care of that in a future PR.
This depends on https://github.com/cosmicds/cosmicds/pull/277, as it pulls in the ViewerLayout component that I've added there. It also depends on the newest version of glue-plotly for the dotplot layer artist, and so I've added that to the dependencies in setup.cfg.
Looks great, thanks @Carifio24!
|
gharchive/pull-request
| 2024-05-08T21:10:10 |
2025-04-01T04:33:53.186465
|
{
"authors": [
"Carifio24",
"nmearl"
],
"repo": "cosmicds/hubbleds",
"url": "https://github.com/cosmicds/hubbleds/pull/365",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2303610494
|
validation urgently needed
definition and schema files can be committed today without any validation, which causes many issues downstream.
We need to put the work on the devs submitting PRs, not on the maintainers to correct basic JSON Schema or JSON errors.
I have tooling that I can add that will fail and throw errors under these conditions, which we do want (IMO). I will propose a PR soon to take care of this. Documenting the issue here as it's a very general widespread issue. It's not necessarily specific to any particular issue (e.g. assetlist validation vs. versions schema definition validation).
My suggestion:
I'll submit a PR with workflow that can run under workflow dispatch only — we manually clean up issues — then agree to turn on the automation with will cause any new PRs to fail with incorrect data.
First Pass
[ ] schema definition validations (non-strict)
[ ] schema validations (non-strict)
Second Pass
This will require a bit more manual updates, but ideally we have strict validations
[ ] schema definition validations (strict)
[ ] schema validations (strict)
|
gharchive/issue
| 2024-05-17T21:39:53 |
2025-04-01T04:33:53.195971
|
{
"authors": [
"pyramation"
],
"repo": "cosmos/chain-registry",
"url": "https://github.com/cosmos/chain-registry/issues/4661",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1679949809
|
Add GetNearest method to the DB interface
Summary
There is an issue in cosmos-sdk: Make KVStore interface have methods to getNearest entry.
To address this issue, we can add a new method, GetNearest to the DB interface and implement it for all supported storage engines. This method should have a similar cost to the Get operation and provide a more efficient alternative to iterators for finding the nearest key.
// GetNearest retrieves the nearest key to the given key.
// If 'ascending' is true, the method returns the smallest key greater than the given key.
// If 'ascending' is false, the method returns the largest key smaller than the given key.
// In case there is no key in the desired direction, the method returns nil.
// CONTRACT: key, value read-only []byte
GetNearest(key []byte, ascending bool) ([]byte, error)
Once implemented, the method can be used in the cosmos-sdk repo to add a GetNearest function to the KVStore interface, as described in the original issue.
@babadro are you actively working on this?
@elias-orijtech
Not yet.
I was uncertain whether I correctly understood the concept and if the problem is still relevant considering the age of the original issue.
Hence, I decided to verify it by opening this discussion first.
I still believe that this would be valuable
|
gharchive/issue
| 2023-04-23T09:41:05 |
2025-04-01T04:33:53.199371
|
{
"authors": [
"ValarDragon",
"babadro",
"elias-orijtech"
],
"repo": "cosmos/cosmos-db",
"url": "https://github.com/cosmos/cosmos-db/issues/79",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1304984020
|
Simplify SendPacket API
Description
closes: #1395
Before we can merge this PR, please make sure that all the following items have been
checked off. If any of the checklist items are not applicable, please leave them but
write a little note why.
[x] Targeted PR against correct branch (see CONTRIBUTING.md)
[x] Linked to Github issue with discussion and accepted design OR link to spec that describes this work.
[ ] Code follows the module structure standards.
[ ] Wrote unit and integration tests
[ ] Updated relevant documentation (docs/) or specification (x/<module>/spec/)
[ ] Added relevant godoc comments.
[ ] Added a relevant changelog entry to the Unreleased section in CHANGELOG.md
[ ] Re-reviewed Files changed in the Github PR explorer
[ ] Review Codecov Report in the comment section below once CI passes
Marking as draft just so it is easier to know when suggestions and merge conflicts have been addressed and the pr is ready for a full review again
Codecov Report
Merging #1703 (7a94343) into main (7dcfc2c) will increase coverage by 0.02%.
The diff coverage is 82.00%.
Additional details and impacted files
@@ Coverage Diff @@
## main #1703 +/- ##
==========================================
+ Coverage 78.63% 78.65% +0.02%
==========================================
Files 178 178
Lines 12354 12296 -58
==========================================
- Hits 9714 9671 -43
+ Misses 2208 2197 -11
+ Partials 432 428 -4
Impacted Files
Coverage Δ
...7-interchain-accounts/controller/ibc_middleware.go
70.96% <0.00%> (ø)
...27-interchain-accounts/controller/keeper/keeper.go
92.53% <ø> (ø)
.../apps/27-interchain-accounts/host/keeper/keeper.go
83.80% <ø> (ø)
modules/apps/29-fee/ibc_middleware.go
91.83% <0.00%> (ø)
modules/apps/29-fee/keeper/keeper.go
92.48% <ø> (ø)
modules/apps/29-fee/keeper/relay.go
88.23% <0.00%> (ø)
modules/apps/transfer/keeper/keeper.go
91.17% <ø> (ø)
modules/apps/transfer/keeper/relay.go
87.45% <50.00%> (-1.22%)
:arrow_down:
.../27-interchain-accounts/controller/keeper/relay.go
75.00% <66.66%> (+0.53%)
:arrow_up:
modules/core/04-channel/keeper/packet.go
98.20% <92.59%> (+1.61%)
:arrow_up:
... and 2 more
@crodriguezvega Are there any changes left you think we should make? With regards to your current suggestions, I think it probably makes sense to continue using the packet variables so as to ensure we are referencing the correct values
Otherwise I think this pr is ready to be merged
@crodriguezvega Are there any changes left you think we should make?
Yes, I just updated the documentation. I would appreciate a review to check it all looks ok. After that, I agree we can merge it.
|
gharchive/pull-request
| 2022-07-14T15:58:11 |
2025-04-01T04:33:53.270721
|
{
"authors": [
"AdityaSripal",
"codecov-commenter",
"colin-axner",
"crodriguezvega"
],
"repo": "cosmos/ibc-go",
"url": "https://github.com/cosmos/ibc-go/pull/1703",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2079257423
|
let any layout able to be homepage
Type of change
[ ] Bug fix (non-breaking change which fixes an issue)
[x] New feature (non-breaking change which adds functionality)
[x] Improvement (refactoring and improving code)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
[ ] Documentation update
Description
separate posts from the homepage:
other suffixes of index is able to use
now any layout can be homepage by change page.layout of index.html
add posts layout same as the old home layout
retained the home layout for compatibility
Additional context
Known issues:
non index posts can not be paginated
neither of jekyll-paginate or jekyll-paginate-v2(production) can not work it out
Todo:
remove home layout
a breaking change
this is my very first pr.
sorry about commitlint and everything.
closing it and open a new one
|
gharchive/pull-request
| 2024-01-12T16:41:52 |
2025-04-01T04:33:53.281090
|
{
"authors": [
"bonjour-py"
],
"repo": "cotes2020/jekyll-theme-chirpy",
"url": "https://github.com/cotes2020/jekyll-theme-chirpy/pull/1478",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1588873594
|
🛑 Mainnet Smartpy (mainnet.smartpy.io) is down
In 2126c6b, Mainnet Smartpy (mainnet.smartpy.io) (https://mainnet.smartpy.io/chains/main/blocks/head/header) was down:
HTTP code: 503
Response time: 15411 ms
Resolved: Mainnet Smartpy (mainnet.smartpy.io) is back up in 5cb4f27.
|
gharchive/issue
| 2023-02-17T07:24:17 |
2025-04-01T04:33:53.284456
|
{
"authors": [
"copolycube"
],
"repo": "cotezos/teznodes",
"url": "https://github.com/cotezos/teznodes/issues/1509",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
997559411
|
Use correct service type name in query context
The N1QL service query was using the nameof operator
which doesn't match in the orphan reporter that is
looking for the value of the property and not the name
This has been merged. Thanks for contributing!
|
gharchive/pull-request
| 2021-09-15T22:07:02 |
2025-04-01T04:33:53.293918
|
{
"authors": [
"jeffrymorris",
"smitcham"
],
"repo": "couchbase/couchbase-net-client",
"url": "https://github.com/couchbase/couchbase-net-client/pull/116",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1887403
|
the id method in embeddable returns nil if there's no base doc
In the embeddable module, the id method returns nil if there is no base doc.
def id
return nil if base_doc.nil?
base_doc.id
end
I think this means that for the model in which you're embedding stuff, id is always nil. Which is bad.
I may be missing the point but should it be something like this:
def id
unless base_doc.nil?
base_doc.id
else
self['_id']
end
end
Sorry - I had got the wrong end of the stick. I was using it wrongly. Please ignore. :)
Haha, yeah, embeddable docs are not meant to have an ID as they are not accessible from the database directly. As a type of hack in some situations (such as forms, IIRC) we provide the id of the parent document. You should be able to overwrite the id with a property if you wanted.
Cheers, sam
Thanks for getting back to me. I'd just included the module on the wrong 'end' (i.e. the embedder not the embedee). Apologies for rashly creating an issue! :)
|
gharchive/issue
| 2011-10-12T13:45:02 |
2025-04-01T04:33:53.332282
|
{
"authors": [
"RicSwirrl",
"samlown"
],
"repo": "couchrest/couchrest_model",
"url": "https://github.com/couchrest/couchrest_model/issues/120",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
599896463
|
cs install with url channel fails
Hi, I tried to use cs install to install from an application definition on gh, thus:
cs install --channel https://raw.githubusercontent.com/gemini-hlsw/itac/cli/cs/itac.json itac
And it failed, claiming a decoding error :
https://raw.githubusercontent.com/gemini-hlsw/itac/cli/cs/itac.json
No modified time in response
Exception in thread "main" java.lang.Exception: Error decoding /Users/rnorris/.coursier/cache/v1/https/raw.githubusercontent.com/gemini-hlsw/itac/cli/cs/itac.json (https://raw.githubusercontent.com/gemini-hlsw/itac/cli/cs/itac.json): Expected object: CursorHistory(List(El(CursorOpDownField(repositories),true)))
at coursier.install.Channels.$anonfun$find$20(Channels.scala:141)
at scala.util.Either$LeftProjection.map(Either.scala:573)
at coursier.install.Channels.$anonfun$find$19(Channels.scala:141)
at coursier.install.Channels.$anonfun$find$19$adapted(Channels.scala:135)
at coursier.util.Task$.$anonfun$flatMap$2(Task.scala:14)
at scala.concurrent.Future.$anonfun$flatMap$1(Future.scala:307)
at scala.concurrent.impl.Promise.$anonfun$transformWith$1(Promise.scala:41)
at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:64)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
at com.oracle.svm.core.thread.JavaThreads.threadStartRoutine(JavaThreads.java:497)
at com.oracle.svm.core.posix.thread.PosixJavaThreads.pthreadStartRoutine(PosixJavaThreads.java:193)
However if I then try to use the downloaded file referenced in the error message it works!
~$ cs install --channel /Users/rnorris/.coursier/cache/v1/https/raw.githubusercontent.com/gemini-hlsw/itac/cli/cs itac
Wrote itac
Any ideas?
@tpolecat That's because https://raw.githubusercontent.com/gemini-hlsw/itac/cli/cs/itac.json designates an app descriptor, not a channel per se. That would be the same with /Users/rnorris/.coursier/cache/v1/https/raw.githubusercontent.com/gemini-hlsw/itac/cli/cs/itac.json for example (JSON file, not directory).
To make it a channel, you can transform it like
{
"itac": {
"repositories": [
"central",
"sonatype:public",
"https://github.com/gemini-hlsw/maven-repo/raw/master/releases",
"ivy2Local"
],
"dependencies": [
"edu.gemini:itc-cli-main_2.12:latest.release"
]
}
}
That also allows to add more apps to it.
But maybe support for using a single app descriptor as a channel should be added… I've been bitten by that too in the past.
Aha ok, I misread the documentation. Thanks!:+1:
|
gharchive/issue
| 2020-04-14T22:27:16 |
2025-04-01T04:33:53.360480
|
{
"authors": [
"alexarchambault",
"tpolecat"
],
"repo": "coursier/coursier",
"url": "https://github.com/coursier/coursier/issues/1663",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1072367928
|
ci: migrated pull requests CI to GitHub Actions
https://coveord.atlassian.net/browse/KIT-1268
Follow-up to https://github.com/coveo/ui-kit/pull/1562.
If you wish to compare with the previous PR, here's the reviewable diff: https://github.com/coveo/ui-kit/pull/1567/files/a432ecfef6d52147e7fefe096297d9e6eaa09608..68cdc8d790105ee14afa8be60ff18aa52f488d56
Checks passed, but one old check was still showing up despite no longer being required (as proven by the master branch PR). I tried to close and reopen the PR, but that didn't fix it, so I force merged.
|
gharchive/pull-request
| 2021-12-06T16:48:37 |
2025-04-01T04:33:53.365816
|
{
"authors": [
"btaillon"
],
"repo": "coveo/ui-kit",
"url": "https://github.com/coveo/ui-kit/pull/1567",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1395205901
|
feat(atomic): add atomic-result-localized-text component
Component that allows to output localized string more easily.
Can be used that way:
searchInterface.i18n.addResourceBundle('en', 'translation', {
foo: 'Hello world {{replace_me_with_field_value}} !',
foo_plural: 'Hello worlds {{replace_me_with_field_value}}'
});
<atomic-result-localized-text key="foo" field-count="the_field_which_will_be_used_to_determine_if_plural_should_be_used" field-value-fieldname="replace_me_with_field_value" ></atomic-result-localized-text>
https://coveord.atlassian.net/browse/KIT-2064
I don't think it's a big deal but the downside here is it's forcing implementer to make a choice:
use atomic-result-text: gain highlighting but lose the ability to pass a count value
use atomic-result-localized-text: gain count value, multiple field values, but lose highlighting
Could we add a bit more options to the new component so it can highlight, and perhaps deprecate atomic-result-text & remove it in the next version? Or having those 2 separate component better?
👍 If we have an example in our default result template (index.html)
|
gharchive/pull-request
| 2022-10-03T19:18:55 |
2025-04-01T04:33:53.368702
|
{
"authors": [
"ThibodeauJF",
"olamothe"
],
"repo": "coveo/ui-kit",
"url": "https://github.com/coveo/ui-kit/pull/2483",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1394136715
|
Try to unify levelSet and categories for COMMUNITY_RISK_LEVEL metric.
See https://github.com/covid-projections/hackathon-september-2022/pull/3/files#diff-ce3a93efd046e7a95d61484997047a9987f83e8947ca7abfdd6f3669cfed455eR97
Basically to set up the COMMUNITY_RISK_LEVEL metric in the hackathon, I had to specify a levelSet so that we could render a thermometer for it, and also categories (so that the 0,1,2 values could be rendered as Low,Medium,High and colored correctly).
This feels kinda' redundant / hacky. I'm wondering if we could make the thermometer fall back to rendering from the categories or something. Or maybe categories should be tied to levels somehow (you specify a value and a level instead of value,label,color)? Not sure.
I have been thinking about this recently, I think we could consolidate the notion of level for metrics with categories and thresholds
Levels have a color, a name and a description or extended name
There is a notion of order in the levels. For example: high risk to low risk, or high, medium high, medium low, and low income for categorical, and levels sorted by thresholds in the case of threshold metrics
In both cases we have a "default" or "missing data" level that will likely be treated differently in the UI (it shouldn't be part of the thermometer, for example)
Both types of metrics should have a way to resolve a level given metric data
we could update MetricCategory to be defined in terms of MetricLevel objects, and separate the default level from the ordered sequence of levels for the metric.
👍 to all of that. I think it's worth trying and we can see how it looks!
|
gharchive/issue
| 2022-10-03T04:57:29 |
2025-04-01T04:33:53.374077
|
{
"authors": [
"mikelehen",
"pnavarrc"
],
"repo": "covid-projections/act-now-packages",
"url": "https://github.com/covid-projections/act-now-packages/issues/259",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1202611834
|
Weekly new cases chart status text
From copy doc:
Looks good with minor suggestion. Also, can we remove the // TODO(8.2) : Update with real metric + content: comment now or is there more to fix?
removing the todos! that should be everything
|
gharchive/pull-request
| 2022-04-13T00:40:00 |
2025-04-01T04:33:53.375965
|
{
"authors": [
"chasulin"
],
"repo": "covid-projections/covid-projections",
"url": "https://github.com/covid-projections/covid-projections/pull/5550",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
597839383
|
Page Number can be set negative in essentials page
Describe the bug
A clear and concise description of what the bug is.
To Reproduce
Steps to reproduce the behavior:
Go to '...'
Click on '....'
Scroll down to '....'
See error
Expected behavior
A clear and concise description of what you expected to happen.
Screenshots
If applicable, add screenshots to help explain your problem.
Desktop (please complete the following information):
OS: [e.g. iOS]
Browser [e.g. chrome, safari]
Version [e.g. 22]
Smartphone (please complete the following information):
Device: [e.g. iPhone6]
OS: [e.g. iOS8.1]
Browser [e.g. stock browser, safari]
Version [e.g. 22]
Additional context
Add any other context about the problem here.
The issue was reproducible. I am taking this up and working on this. I will submit a PR for this soon.
Hi. Since we are done away with the essentials page pagination. We can close this issue. We now have infinite scroll implemented.
|
gharchive/issue
| 2020-04-10T11:21:20 |
2025-04-01T04:33:53.385219
|
{
"authors": [
"nikita-11",
"robintom",
"the-sajith"
],
"repo": "covid19india/covid19india-react",
"url": "https://github.com/covid19india/covid19india-react/issues/901",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
588406395
|
Added setup commands for windows
Done for issue: https://github.com/covid19india/covid19india-react/issues/171
True.
Temporarily, what I do is:
Remove PORT=3000 from here "react-scripts start css-watch"
npm start
Restore PORT=3000 in "react-scripts start css-watch"
@manigandan-dorairaj Please check the new update.
@sudevschiz Inspired from your PR https://github.com/covid19india/covid19india-react/pull/197. This works.
Yes, it works. Thanks!
Hi Rajan - can you check PR 197 again ? It has updates for setting port too without needing temporary port.
Closing since #197. Feel free to reopen an issue if you're still having trouble with installing and starting the application :)
|
gharchive/pull-request
| 2020-03-26T13:05:35 |
2025-04-01T04:33:53.388284
|
{
"authors": [
"Rajanpandey",
"aar-1",
"jeremyphilemon",
"manigandan-dorairaj"
],
"repo": "covid19india/covid19india-react",
"url": "https://github.com/covid19india/covid19india-react/pull/177",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
874001357
|
Doctors section should be improved
Now it's quite messy need to fix its margin, padding, etc. You can find this section on homepage.
Hi! I can fix this issue. Just assign it to me.
Sure @polgarj, go ahead.
It's done, check the pull request :)
|
gharchive/issue
| 2021-05-02T17:44:13 |
2025-04-01T04:33:53.389875
|
{
"authors": [
"imanishbarnwal",
"polgarj"
],
"repo": "covidoff/covidoff",
"url": "https://github.com/covidoff/covidoff/issues/4",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2549003234
|
Fix: Unable to elevate system while "cl7h" repository is active
By submitting pull requests to this repo, I agree to the Contributor License Agreement which can be found at: https://github.com/cpanel/elevate/blob/main/docs/cPanel-CLA.pdf
What is the source of the cl7h repo? We'd need to review it before we green light it.
I'm replacing this with #518
|
gharchive/pull-request
| 2024-09-25T21:08:07 |
2025-04-01T04:33:53.452509
|
{
"authors": [
"lkabuci",
"toddr"
],
"repo": "cpanel/elevate",
"url": "https://github.com/cpanel/elevate/pull/516",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
110298904
|
Circle example improvements
The circle example is pretty cool, and could benefit from some text addition and small refactoring! What I changed
The equation of the circle was changed to
(x - a).pow(2) / 4 + (y - b).pow(2) The addition of / 4 takes into account that rows and columns are not evenly spaced in terminals. The distance between rows is much greater than the distance between each column for the terminal cells, this gives it a more circular shape.
Decreasing the Radius no longer works beyond 1 by use of saturating_sub
Text at the bottom of the screen lists controls, which are rendered through a simple write_out function
Small refactoring
Picture of new look
Fantastic, thank you!
|
gharchive/pull-request
| 2015-10-07T19:25:56 |
2025-04-01T04:33:53.469684
|
{
"authors": [
"Syntaf",
"cpjreynolds"
],
"repo": "cpjreynolds/rustty",
"url": "https://github.com/cpjreynolds/rustty/pull/18",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1417421546
|
Document bootstrap
Provide a high level overview of how facade is initialized, the major decisions involved, and the most important implications.
https://github.com/cpp-gamedev/facade/wiki/How-It-Works
|
gharchive/issue
| 2022-10-20T22:48:01 |
2025-04-01T04:33:53.476185
|
{
"authors": [
"karnkaul"
],
"repo": "cpp-gamedev/facade",
"url": "https://github.com/cpp-gamedev/facade/issues/21",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1340932640
|
build: cmake dependency approach refactor
The current git submodule approach for project dependencies tightly
integrates subprojects as part of the main project. CMake's FetchContent
is used as a configurable alternative to explicit submodule inclusion.
Users can freely modify the dependency origin, and also facilitates
further refactors and separation of depedency build configuration
from the cpp-sc2 library CMake files.
Not looking for a merge right now - would like some feedback of the approach and next steps. This refactor can effectively eliminate keeping submodules in project, as they are fetched as part of the build configuration. There are some additional smaller improvements I can make, but the work done so far is one step short of working out packaging cpp-sc2 through CMake so it can be imported by projects like BlankBot with a single fetch and configuration.
The cleanup also handled some minor carryover from switching off the SDL mirror (introduced some additional builds which are not necessary), replacing all file globbing with explicit sources (some unnecessary headers were being globbed in for compilation...), and a variety of other CMake related improvements.
Naturally there is a small tradeoff, as a clean configuration build (running cmake -B build for the first time) will introduce some download overhead to fetch the external subprojects, but I think the increased configuration time is negligible. Updating pins, for both development and maintenance is also simplified as it is handled in the CMake configuration instead of updating the submodule itself.
Regarding the proposal:
I am ok with removal of files globbing.
I am mostly ok with the fetch approach but have some questions:
Is it really more familiar to other people? Honestly I haven't seen it widely used.
How does it help in packaging?
What if I change the hash after the full build? Should I recreate the build folder and fetch everything from scratch?
Here is CMake's documentation outlining the high level approaches available for managing dependencies, but it can be summarized to:
find_package for pre-built or installed packages (so especially useful for containerized images where you can reliably download and configure the dependencies, and then build the project).
FetchContent, which is an abstraction over ExternalProject - but will download the dependencies at configuration time instead of build time.
Both can be used together with the FIND_PACKAGE_ARGS setting for FetchContent to run find_package for NAMES before downloading the dependency.
So to answer your questions in order:
Is it really more familiar to other people? Honestly I haven't seen it widely used.
FetchContent was added in version 3.11, which released a little under 4 years ago. The only alternative at the time for ("superbuilds")[https://www.kitware.com/cmake-superbuilds-git-submodules/] strategy would be to use ExternalProject, which has the drawback of pulling in dependencies at build time.
In a perfect world every project using CMake would configure export and install targets, and find_package would be the standard. civetweb, SDL and protobuf all actually define these targets, so cpp-sc2 could easily enough have the user install these projects separately - but as we are currently pinning to not only git hashes but also a mixture of forked and original repositories, I would hesitate to go all in on the find_package route. If we were pinning to specific packaged releases of a dependency it would be more appropriate, but this leads into the next question.
how does it help in packaging
As described in the CMake dependency guide, there is more than one way to include dependencies in a project. Currently, dependencies are maintained at git submodules - so outside of the context of the build system. CMake has no explicit knowledge of the dependencies outside of the manually set add_subdirectory paths we maintain. The side-effects are already present in the build configuration, such as manual checks in sc2protocol:
if ("${proto_files}" STREQUAL "")
message(FATAL_ERROR "Missing root protocol submodule.\nRemove your protocol directory and run:\ngit submodule update --recursive --init")
endif ()
Relying on git submodules to manage dependencies for the build requires the inclusion of git submodule logic, whereas with FetchContent we can isolate dependency management to the build itself, as well as provide a more configurable interface.
What if I change e.g. SDL2 hash after the full build? Should I recreate the build folder and fetch everything from scratch?
CMake looks for changes between the build tree files and the configuration. Using the example of file globbing, adding a file to a source directory would not be detected by CMake globbing with CONFIGURE_DEPENDS on MSVC as it isn't supported, and would necessitate a full rebuild.
FetchContent_Declare and FetchContent_MakeAvailable have well defined procedures and expected behaviors. Changing the git hash for a dependency will download and update the build tree sources only for the changed dependency. All FetchContent commands will always run on configuration builds, but FetchContent_MakeAvailable checks to see if the dependency requirements are already met (if it is already available and up to date, and in the correct directories, do nothing).
This is where I think the tradeoffs for FetchContent are most powerful, as the overhead is minimal and now all dependency logic is restricted to CMake. Users can freely change dependency repos (e.g. between forks) and hashes without needing to step into submodules and the edge-case headaches associated with it. Returning the the "perfect world" example, projects would publish/maintain a minimal repository branch / package for download which only contains the sources, headers, and CMake configuration required to include as a subproject using FetchContent. Just as with gitsubmodules, FetchContent unfortunately will still drag in all documentation, miscellaneous scripts, etc. from the subproject, but we at least still have the configuration capability to turn off all unnecessary target builds (tests, shared libraries).
I hope this answers your questions and sufficiently makes the case for using FetchContent over git submodules. There is still some target level work I have left to do (after reading more documentation, we should be able to use properly namespaced targets from dependencies such as civetweb::civetweb-c-library, but I have not tested this yet).
Thank you! Now the subject is much more clearer.
civetweb, SDL and protobuf all actually define these targets, so cpp-sc2 could easily enough have the user install these projects separately
Oh well, I had lots of issues with people who have no idea how to install e.g. boost, especially on Windows. The submodules are more or less common now so usually don't rise questions. So I would prefer to keep the current approach (i.e. with fetching) in any case. Partly because it doesn't force people to install dependencies manually (and come to me with questions like where one can find protobuf and how to compile it :) ), partly because I want to have convenient access to the sources especially if I want to patch something or experiment with different version without altering the packages in the system.
Another use case for this fetching mechanic is patching of the dependencies. Currently we use a patched fork of civetweb with several patches from Blizzard. While I have no idea do we really need to keep them, I would prefer to explicitly include these patches in cpp-sc2, apply during checkout and drop that fork.
Thus, green light from me :)
Sounds good. I too prefer FetchContent over find_package in this case. If we were publishing docker images with everything cleanly packaged, then it would make more sense (and would really tighten up the includes as we wouldn't need to target include subproject header files, as they would all be exported to the namespaced target like civetweb::civetweb-c-library). But CMake is a realm of trade-offs, so just trying to make the best of what works for this project.
I'll just be applying some additional adjustments + documentation updates, and the submodules can then be removed once everything is verified.
As an aside on FetchContent, I can't find the original mailing list thread but it is my understanding the author of the module - Craig Scott - implemented FetchContent precisely because he didn't like working with git submodules and ExternalProject on very large projects (he described managing 40+ dependencies and needing to work within the dependencies in parallel to the main project while testing), and the module seems to be the logical resolution to the problem.
Yeah, the management overhead is smaller with FetchContent. Totally agree on this.
So I did a final pass on my changes this weekend alongside working on cross compilation support for Windows under WSL2 using the x86_64-w64-mingw32 toolchain. Naturally that work will be presented in a different PR, but some additional cleanup to the sc2protocol configuration has been included to facilitate the upcoming changes.
I have also normalized the usage of quotes for CMake variable expansion and setting according to the consolidated tips found in this thread. The big takeaway is quotes are most important for expanding variables which may have spaces or colons - so any paths with variable expansion should be quoted. cpp-sc2 generally does not concern itself with colons for project configuration.
I'm content with the PR as it is (most recent rebase has dealt with the outstanding issues), so I have nothing else to add.
|
gharchive/pull-request
| 2022-08-16T22:10:11 |
2025-04-01T04:33:53.493796
|
{
"authors": [
"alkurbatov",
"jlknauer"
],
"repo": "cpp-sc2/cpp-sc2",
"url": "https://github.com/cpp-sc2/cpp-sc2/pull/118",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
345562294
|
Something screwy in CompareGroups?
I've been trying to track down an issue where I'm able to find a function definition once - but not a second time.
I think I've pinned it on CompareGroups(), which:
sorts the vectors current_data and previous_data
then walks them in tandem, looking for differences
The problem, I believe, is that the sorting is by id, whereas the checking-for-differences is by usr.
The consequence is that if the lists are not identical, but there is some thing that really is in both lists, and you get unlucky - then that thing will be first treated as added, then as removed, and removed will win.
So if I'm right then I reckon that the sorting wants to be by usr.
Or am I way off base?
In the hope that I might have got this right, I've submitted #749
id should be derived from usr, so it should be safe to sort on it. Ideally usr should be almost entirely unused in querydb, everything should be relative to id.
But either one should be fine to base off, as both are unique.
I'll take a look at the existing querydb code and analyze usage of id vs usr before making a decision w.r.t. merging.
Thanks for the investigation :)
in the debugger, usr is some random-looking number and id seems to be just the same as the array index. So I'm not seeing how id can have been derived from usr.
I could believe that it would work to sort on id and compare ids, if you say so. But the current mix-and-match looks wrong...
Take a look at the IdMap constructor to see how id is derived from usr.
https://github.com/cquery-project/cquery/blob/master/src/query.cc#L390-L409
Specifically, if we haven't seen the usr before, we allocate a new id, otherwise we find and reuse the old id.
I could believe that it would work to sort on id and compare ids, if you say so. But the current mix-and-match looks wrong...
I'm a little surprised it is causing issues, but it seems reasonable as well that this is causing issues.
Spelling out an example of what I think is happening, suppose:
prev_list has entries [{id=1, usr=200}, {id=2, usr=100}]
current_list has entries [{id=2, usr=100]
then we'll go:
200 > 100, so element id=1 gets on_added and curr_it is incremented
curr_it is now exhausted
both elements are still in prev_it so both get on_removed()
We've ended up removing both elements, which we didn't want to do.
|
gharchive/issue
| 2018-07-29T20:59:47 |
2025-04-01T04:33:53.508345
|
{
"authors": [
"dimbleby",
"jacobdufault"
],
"repo": "cquery-project/cquery",
"url": "https://github.com/cquery-project/cquery/issues/748",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1052566971
|
Update scalafmt-core to 3.1.1
Updates org.scalameta:scalafmt-core from 3.0.8 to 3.1.1.
GitHub Release Notes - Version Diff
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.scalameta", artifactId = "scalafmt-core" } ]
labels: library-update, early-semver-minor, semver-spec-minor
Superseded by #123.
|
gharchive/pull-request
| 2021-11-13T07:19:38 |
2025-04-01T04:33:53.517702
|
{
"authors": [
"scala-steward"
],
"repo": "cquiroz/scalajs-react-resizable",
"url": "https://github.com/cquiroz/scalajs-react-resizable/pull/121",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
962977260
|
Convert to webpack 5 & Docker
This is conversion of the Craft CMS marketing demo site Europa Museum to a Docker-ized setup that uses a modularized webpack 5 setup for frontend tooling.
💯
|
gharchive/pull-request
| 2021-08-06T19:03:41 |
2025-04-01T04:33:53.543851
|
{
"authors": [
"brandonkelly",
"khalwat"
],
"repo": "craftcms/europa-museum",
"url": "https://github.com/craftcms/europa-museum/pull/3",
"license": "0BSD",
"license_type": "permissive",
"license_source": "github-api"
}
|
201106371
|
Leaf Dependencies
Copied from PR #28
[...] I think we need some sort of a plan-apply command and/or a way to build dependency tree with the branch runs.
plan-apply: automatically (maybe a confirmation input and/or a flag to accept) run apply after you have planned (assuming no error)
dependency tree: If possible, do not run plan/apply on dependent leafs until required changes are applied
Here is the scenario where 1 or both of these things would be handy:
Assume you have a project where you have a leaf called security_groups where you define your security groups for your environment. Then you have another leaf that builds your EC2 instances, called webservers. You will need output from security_groups in order to correctly launch your webservers. You create a build_order.txt to make this a branch, where you have both leafs listed, so that it runs security_groups and then webservers.
Now, if you run a colonize plan -e dev to build out your dev environment, it will go through and run the plan on each leaf. The issue is, webservers cannot plan correctly, because the output from security_groups does not exist yet. So, you have to plan, apply, plan, and apply, in order to correctly build your webservers, which is the same as just running at the leaf level each time.
There is a workaround for this to work...albeit a dangerous one.
If you do not "plan" first and just apply, then it will execute correctly, since they are applied before the next implicit plan
Just adding some color to what @jyore said, and throwing out a crazy idea.
plan fails due to missing state files of dependency resources.
Background
We have a project broken up into a number of resources. e.g.:
|- ebs
| |-- main.tf
|- ec2
| |-- main.tf
|- security_groups
| |-- main.tf
|- kms
| |-- main.tf
where each has its own state file. However, there are state file dependencies such as (non-exhaustive list):
ec2 depends on the state file for security_groups being present so it can read in SG ids.
ebs depends on the state file for kms being present so it can read in KMS ids for encrypted volumes.
ec2 depends on the state file for ebs being present so it can read in ebs volume ids.
Crazy Idea
Of course since there are benefits to breaking out the state files, it would be great to retain this structure while still being able to run a (Colonize or Terraform) plan prior to building anything.
My crazy idea just to continue the convo: what if Colonize recognized some of these dependencies and just created some 'fake' or empty state files without actually creating (apply) anything?
Not sure if this will work at all, i.e. if Terraform will still balk at either fake or empty values.
My Current Workaround
A wrapper script that checks if there's a remote state file first. If there's one, great, then it runs colonize plan. If not, it runs colonize apply.
Of course, this assumes that you always want to create the TF described in the project, and makes plan an action that is only for updating/modifying an existing project.
|
gharchive/issue
| 2017-01-16T19:56:33 |
2025-04-01T04:33:53.565612
|
{
"authors": [
"jyore",
"kjhosein"
],
"repo": "craigmonson/colonize",
"url": "https://github.com/craigmonson/colonize/issues/33",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
183079590
|
Add RTL Styles
https://github.com/MohammadYounes/gulp-rtlcss
Some themes need right-to-left styles, and this package can help generate them.
The gotcha here, is being able to generate the right sourcemap when building SCSS into CSS, AND minifying, AND producing RTL versions.. Separating the styles task up into subtasks that save an intermediary file might be needed.
Like I did with my separate zip tasks, we could have two separate tasks (build:cssleft, build:cssright, and either a master task (build:css) that does both, or otherwise checks for a config flag that will produce just one or both.
Now we've got #41 (Allow multiple CSS files), we may have more flexibility to identify those which should have RTL files generated as well.
The default behaviour is to save the output to rtl.css, which WordPress will autoload if it exists.
rtl.css is meant to contain the additional changes, such that style.css followed by rtl.css gives the styles needed for an RTL site.
However, this not only means loading two stylesheets, but the rtl.css generated is not just the additions, but the original style.css with a few changes made, such is rtlcss()'s behaviour.
We might consider outputting to style-rtl.css, but we'd need to direct theme authors to add in something like the following, so that style-rtl.css is loaded instead of style.css when in an RTL context:
\add_action( 'wp_enqueue_scripts', __NAMESPACE__ . '\\enqueue_assets' );
/**
* Allow style-rtl.css to be referenced instead of style.css.
*
* @since 1.0.0
*
* @link http://www.poststat.us/automating-i18n-wordpress-themes/
*/
function enqueue_assets(): void {
$handle = defined( 'CHILD_THEME_NAME' ) && CHILD_THEME_NAME ? \sanitize_title_with_dashes( CHILD_THEME_NAME ) : 'child-theme';
\wp_style_add_data( $handle, 'rtl', 'replace' );
}
|
gharchive/issue
| 2016-10-14T15:27:06 |
2025-04-01T04:33:53.570865
|
{
"authors": [
"GaryJones",
"craigsimps"
],
"repo": "craigsimps/gulp-wp-toolkit",
"url": "https://github.com/craigsimps/gulp-wp-toolkit/issues/3",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
130786108
|
kernel-devel different than kernel-devel-VERSION for Centos needs adding to module.. Hiera perhaps?
Hi Craig!
I run into a problem when the module tries to install kernel-devel for a kernel-lt version.
I want to avoid touchihng your module so i was wondering if the params could be filled with hiera.
I tried adding vmwaretools::redhat_devel_package to hiera for that node only but no luck. Keep seeing
Dependency Package[kernel-devel-3.10.95-1.el6.elrepo.x86_64] Where there should be, kernel-lt-devel-3.10.95-1.el6.elrepo.x86_64
Thanks for the help!
Hey,
I can add the lt kernel to the params.pp if statement - can you post the output of facter -p kernelrelease?
Thanks,
Craig
Sure!
facter -p kernelrelease
3.10.95-1.el6.elrepo.x86_64
Let me put more context. This kernel is from a special repo (elrepo). it has this kernel packages
kernel-lt.x86_64 : The Linux kernel. (The core of any Linux-based operating system.)
kernel-lt-devel.x86_64 : Development package for building kernel modules to match the kernel.
and
kernel-ml.x86_64 : The Linux kernel. (The core of any Linux-based operating system.)
kernel-ml-devel.x86_64 : Development package for building kernel modules to match the kernel.
for the ml, it is
Version : 4.4.1
Release : 1.el6.elrepo
Just so you are aware of. Not sure hohw to make both.. I only need the kernel-lt, which i already did, but perhhaps its useful for others.
Just curious, how difficult it is to make that available through hiera? Just starting.
Thanks Craig!
Hi,
Apologies for the delay in looking into this properly!
I've had a look through the code and made a lot of improvements to the way that the module handles installing packages, so can you try with the latest version uploaded to the Puppet Forge - 2.3.0?
If you're running a RedHat-family OS, you shouldn't need kernel development packages installed - the module doesn't manage these unless you explicitly ask it to by setting the install_devel parameter to true - did you set this in your code?
Thanks,
Craig
Hi,
Can you please confirm if this is still an issue for you, and if the issue still exists when running version 2.3.0 of the module?
Thanks,
Craig
Hi,
As nobody has confirmed that this problem still exists, I'm closing the issue. Please feel free to re-open if the problem still exists.
Thanks,
Craig
|
gharchive/issue
| 2016-02-02T19:16:10 |
2025-04-01T04:33:53.579211
|
{
"authors": [
"brahama",
"craigwatson"
],
"repo": "craigwatson/puppet-vmwaretools",
"url": "https://github.com/craigwatson/puppet-vmwaretools/issues/76",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
803578562
|
Create es.json
Hopefully everything works alright.
Looks good, thanks for the help! I'll get this out with the next patch, probably sometime this weekend.
Is there any particular way you'd like to be credited (discord tag, real name, github username, etc)?
No problem. MS-PBS is fine, but crediting me is not necessary. I do this for fun :)
|
gharchive/pull-request
| 2021-02-08T13:43:29 |
2025-04-01T04:33:53.588645
|
{
"authors": [
"MS-PBS",
"crash1115"
],
"repo": "crash1115/5e-training",
"url": "https://github.com/crash1115/5e-training/pull/47",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1967964076
|
Add pycaret automl timeseries forecasting mlflow-cratedb example (currently only python 3.10)
Provides an example for how to use pycaret to automatically benchmark, select and register a timeseries forecasting model.
Uses sales forecasting as example data.
Based on the canonical example track_merlion.py from @amotl .
Requires MLFLOW_TRACKING_URI to be set, as the example attempts to register the final model with MLflow. This does not work, if only file-based tracking is available.
Hey @amotl the branch itself is fine. Github just does not rebase the commit history automatically in pull requests. Will close this one and create a new one, there the file diff will show the correct history.
|
gharchive/pull-request
| 2023-10-30T10:02:54 |
2025-04-01T04:33:53.592829
|
{
"authors": [
"andnig"
],
"repo": "crate-workbench/mlflow-cratedb",
"url": "https://github.com/crate-workbench/mlflow-cratedb/pull/45",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
478506866
|
Add CODEOWNERS file
this will auto-request a review from the @crate/docs team (but a review from the team is not mandatory)
The Jenkins tests failed. It is unrelated to this change. It seems that the issue has something to do with the change in the geo point representation.
See https://github.com/crate/crate/commit/3fb0dd2170f4533cab018b5084717b16b84c3521
Probably, we would have to fix this first before merging the current PR. I'll create an impediment for it.
|
gharchive/pull-request
| 2019-08-08T14:28:11 |
2025-04-01T04:33:53.594409
|
{
"authors": [
"kovrus",
"mechanomi"
],
"repo": "crate/crate-npgsql",
"url": "https://github.com/crate/crate-npgsql/pull/32",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
126213157
|
updated documentation for default schema support
in database connection string
ack
|
gharchive/pull-request
| 2016-01-12T16:09:09 |
2025-04-01T04:33:53.595415
|
{
"authors": [
"chaudum",
"kovrus"
],
"repo": "crate/crate-pdo",
"url": "https://github.com/crate/crate-pdo/pull/31",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1318426792
|
Error: failed to turn off translog retention on startup
CrateDB version
5.0.0
CrateDB setup information
No response
Steps to Reproduce
After upgrading a cluster from 4.7 to 5.0.0 we see lots of the following error messages on CrateDB startup
2022-07-26 17:18:09 | [2022-07-26T15:18:09,879][WARN ][o.e.i.s.IndexShard ] [data-hot-0] [schema.table][0]failed to turn off translog retention
-- | --
| | 2022-07-26 17:18:08 | at java.lang.Thread.run(Thread.java:833) [?:?]
| | 2022-07-26 17:18:08 | at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) [?:?]
| | 2022-07-26 17:18:08 | at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) [?:?]
| | 2022-07-26 17:18:08 | at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) [crate-server.jar:?]
| | 2022-07-26 17:18:08 | at org.elasticsearch.index.shard.IndexShard$3.doRun(IndexShard.java:1861) [crate-server.jar:?]
| | 2022-07-26 17:18:08 | at org.elasticsearch.index.shard.IndexShard.trimTranslog(IndexShard.java:1050) ~[crate-server.jar:?]
| | 2022-07-26 17:18:08 | at org.elasticsearch.index.shard.IndexShard.getEngine(IndexShard.java:2453) ~[crate-server.jar:?]
| | 2022-07-26 17:18:08 | org.apache.lucene.store.AlreadyClosedException: engine is closed
| | 2022-07-26 17:18:08 | [2022-07-26T15:18:08,977][WARN ][o.e.i.s.IndexShard ] [data-hot-0] [.partitioned.schema.table2][0]failed to turn off translog retention
| | 2022-07-26 17:18:08 | at java.lang.Thread.run(Thread.java:833) [?:?]
| | 2022-07-26 17:18:08 | at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) [?:?]
| | 2022-07-26 17:18:08 | at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) [?:?]
| | 2022-07-26 17:18:08 | at org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37) [crate-server.jar:?]
| | 2022-07-26 17:18:08 | at org.elasticsearch.index.shard.IndexShard$3.doRun(IndexShard.java:1861) [crate-server.jar:?]
| | 2022-07-26 17:18:08 | at org.elasticsearch.index.shard.IndexShard.trimTranslog(IndexShard.java:1050) ~[crate-server.jar:?]
| | 2022-07-26 17:18:08 | at org.elasticsearch.index.shard.IndexShard.getEngine(IndexShard.java:2453) ~[crate-server.jar:?]
| | 2022-07-26 17:18:08 | org.apache.lucene.store.AlreadyClosedException: engine is closed
Expected Result
no errors / warnings on startup after migrating from 4.x to 5.0
Actual Result
9 | [2022-07-26T15:18:09,879][WARN ][o.e.i.s.IndexShard ] [data-hot-0] [schema.table][0]failed to turn off translog retention
-- | --
Those are WARN messages, not errors.
Did the ugprade run smoothly nevertherless?
If the update runs smoothly, then the issue is only the WARN logs which I wouldn't consider changing anything regarding those logs, as it could be helpful to troubleshoot real issues during upgrade.
If it causes much trouble to users, maybe we could define another logger for those, or mention them in an upgrade guide/how-to
I think it would make sense to maybe mention them.
@proddata
The warning log message should not be written in some cases, fixed by #12982, but if so it can be ignored.
Thank you for reporting.
|
gharchive/issue
| 2022-07-26T15:42:42 |
2025-04-01T04:33:53.599852
|
{
"authors": [
"matriv",
"proddata",
"seut"
],
"repo": "crate/crate",
"url": "https://github.com/crate/crate/issues/12828",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2271497183
|
Mt/session job
Previously, we were using mostRecentJobID in listeners and lamdas
to mark a job as failed/ended, but there is a chance that the client
issues a next statement, before this marking has taken place, and
the mostRecentJobID gets overwritten with a brand new UUID, so
the listener/lambda would mark a wrong job, and the first job would
never be marked as failed/ended and stay forever in sys.jobs table.
Should I try to somehow unit test this? maybe a dummy ResultReceiver with a sleep() and issue a new stmt execution?
Should I try to somehow unit test this? maybe a dummy ResultReceiver with a sleep() and issue a new stmt execution?
I think you can adapt test_bulk_operations_result_in_jobslog_entries slightly by adapting the future trigger of the executeBulk
Should I try to somehow unit test this? maybe a dummy ResultReceiver with a sleep() and issue a new stmt execution?
I think you can adapt test_bulk_operations_result_in_jobslog_entries slightly by adapting the future trigger of the executeBulk
Thx a lot, please check.
(tested and without the fix, overwriting the mostRecentJobID, makes the bulk job stay in activeJobs
|
gharchive/pull-request
| 2024-04-30T13:34:54 |
2025-04-01T04:33:53.604099
|
{
"authors": [
"matriv",
"mfussenegger"
],
"repo": "crate/crate",
"url": "https://github.com/crate/crate/pull/15936",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2542161462
|
Introduce capacity parameter for topk
Summary of the changes / Why this improves CrateDB
This changes the topk max limit to 5000 with capacity 8192 so the limit is less than 0.75 % of the capacity to have a better default behaviour for performance reasons. It also make the capacity configurable for the user.
Background:
I tried out the nyc taxi datset with topk and found out that there is potential to improve performance significant by increasing the capacity of number of tracked items. Initializing an item is expensive while increasing a counter is cheap on the sketch datastructure. Small capacity means a small memory footprint, but If the capacity is too small items get dropped out and recreated which is slow because it requires multiple hashings. We should increase the default capacity size and make the parameter configurable for users like Spark does. On the nyc-taxi dataset this variant is much faster, and the other benchmarks are not slower.
V1: 5.9.0-9c2b7637a90c1271858560e0ce4288c569e7ade5
V2: 5.9.0-00be8919e533067034a1733d1c799101e57ce993
Q: select topk(pulocationid) from nyc_taxi
C: 1
| Version | Mean ± Stdev | Min | Median | Q3 | Max |
| V1 | 84.861 ± 21.829 | 81.305 | 82.471 | 82.880 | 300.588 |
| V2 | 42.820 ± 20.812 | 26.345 | 40.736 | 41.025 | 247.978 |
├---------┴-------------------------┴------------┴------------┴------------┴------------┘
| - 65.85% - 67.75%
There is a 100.00% probability that the observed difference is not random, and the best estimate of that difference is 65.85%
The test has statistical significance
System/JVM Metrics (durations in ms, byte-values in MB)
| YOUNG GC | OLD GC | HEAP | ALLOC
| cnt avg max | cnt avg max | initial used | rate total
V1 | 2 3.04 3.14 | 0 0.00 0.00 | 268 2001 | 3049.95 23119
V2 | 0 0.00 0.00 | 0 0.00 0.00 | 268 0 | 5657.67 17064
Top frames (by count)
V1
ReversePurgeLongHashMap.purge(int) total=4601564458, count=2218
ReversePurgeLongHashMap.adjustOrPutValue(long, long) total=944, count=944
DocValuesAggregates.getRow(...) total=922, count=922
ReversePurgeLongHashMap.keepOnlyPositiveCounts() total=262, count=262
QuickSelect.partition(...) total=158, count=158
AbstractMemorySegmentImpl.checkBounds(long, long) total=110, count=110
ReversePurgeLongHashMap.hashDelete(int) total=90, count=90
QuickSelect.select(...) total=49, count=49
CompositeBatchIterator$AsyncCompositeBI.lambda$loadNextBatch$0(int, int) total=1906284128, count=18
SingletonSortedNumericDocValues.nextValue() total=17, count=17
V2
DocValuesAggregates.getRow(...) total=546, count=546
SingletonSortedNumericDocValues.nextValue() total=299, count=299
ReversePurgeLongHashMap.adjustOrPutValue(long, long) total=237, count=237
AbstractMemorySegmentImpl.checkBounds(long, long) total=129, count=129
ParserATNSimulator.getEpsilonTarget(...) total=26216552, count=9
ReversePurgeLongHashMap.purge(int) total=332351600, count=8
ObjectName.construct(String) total=19064, count=8
LinkedList.listIterator(int) total=14652848, count=6
StreamSupport.stream(...) total=12304, count=6
ObjectName.setCanonicalName(...) total=10344, count=5
For the implementation, there are a few options:
Let the user use any value for capacity and convert internally to a power of number (@proddata favourite)
Make it explicit, so capacity must be a value in the power of two (current approach)
Use a load factor like clickhouse, this may be difficult to fine-tune for large datasets where you want a small topk.
Checklist
[x] Added an entry in the latest docs/appendices/release-notes/<x.y.0>.rst for user facing changes
[x] Updated documentation & sql_features table for user facing changes
[x] Touched code is covered by tests
[x] This does not contain breaking changes, or if it does:
It is released within a major release
It is recorded in the latest docs/appendices/release-notes/<x.y.0>.rst
It was marked as deprecated in an earlier release if possible
You've thought about the consequences and other components are adapted
(E.g. AdminUI)
|
gharchive/pull-request
| 2024-09-23T10:08:23 |
2025-04-01T04:33:53.610347
|
{
"authors": [
"mkleen"
],
"repo": "crate/crate",
"url": "https://github.com/crate/crate/pull/16651",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2623938194
|
Figure out a CircuitPython approach / story.
CircuitPython is another very popular runtime for Python on microcontrollers, especially Adafruit products. Figure out how to get the driver working in that environment and provide install instructions and example code for it.
I've started playing with this on CircuitPython 9.2.0 on a Raspberry Pi Pico W and will add notes here.
I copied crate.py into /lib on the device.
I tried out importing the other libraries that crate.py needs: requests and base64 and found that neither is installed by default.
Having looked at this, I think we'll need a separate repo for circuitpython-cratedb as the networking and requests are sufficiently different. For example we have to pass networking information into the requests calls, see here.
Thanks for your investigations and for sharing them. First of all, bummer, but on the other hand, c'est la vie, and somehow expected, still unfortunate. However, the library's use only differs on the initialization phase, right?
I think it'll be a bit messier and be poorly named to try and get 2 for one here. I'll have a go at adapting the current code for CircuitPython and see how it looks. I can also ask someone I know at Adafruit for guidance.
I spent a little time looking at other existing CircuitPython libraries that use networked resources. These generally pass in a connection pool and an SSL context, so do work differently than MicroPython and with CircuitPython specific libraries. Here's an MQTT example.
pool = adafruit_connection_manager.get_radio_socketpool(esp)
ssl_context = adafruit_connection_manager.get_radio_ssl_context(esp)
# Set up a MiniMQTT Client
mqtt_client = MQTT.MQTT(
broker="io.adafruit.com",
username=aio_username,
password=aio_key,
socket_pool=pool,
ssl_context=ssl_context,
)
The best approach looks like the creation of a separate library circuitpython-cratedb.
The best approach looks like the creation of a separate library circuitpython-cratedb.
Hi. I don't get it yet, because apparently it would just be about a different interface, right? We could still host it within the same package / repository?
Background: I am taking a strong position here, biased towards monorepo's where applicable, to keep the number of repositories-to-be-maintained down. If you think my concerns are not applicable, please ignore my admonitions.
With a breaking change (now's as good a time as any to do this if we're going to)... I got this to work on CircuitPython.
First, quick change to the imports in cratedb.py... remove everything to do with requests, and add this to account for platform differences in where the base64 encoder is:
try:
# MicroPython
from base64 import b64encode
except ImportError:
# CircuitPython
from circuitpython_base64 import b64encode
Then, change the constructor in cratedb.py:
class CrateDB:
def __init__(
self, requests, host, port=4200, user=None, password=None, schema="doc", use_ssl=True
):
self.requests = requests
self.user = user
self.password = password
self.schema = schema
self.host = host
self.port = port
...
and change our one usage of requests to use the version passed in:
response = self.requests.post(request_url, headers=headers, json=payload)
Then, usage looks like this on CircuitPython and works fine:
import adafruit_connection_manager
import adafruit_requests
import os
import sys
import time
import wifi
import cratedb
socket_pool = adafruit_connection_manager.get_radio_socketpool(wifi.radio)
ssl_context = adafruit_connection_manager.get_radio_ssl_context(wifi.radio)
requests = adafruit_requests.Session(socket_pool, ssl_context)
crate = cratedb.CrateDB(
requests=requests, # This is different from the current approach.
host="<redacted>", # We can advise these go in the TOML file used to store env vars in CircuitPython.
user="<redacted>",
password="<redacted>"
)
# Set up the wifi network and connect.
wifi.radio.connect(os.getenv('CIRCUITPY_WIFI_SSID'), os.getenv('CIRCUITPY_WIFI_PASSWORD'))
ip_addr = wifi.radio.ipv4_address
print(f"Connected to wifi as {ip_addr}")
# Create a table in CrateDB if necessary.
try:
crate.execute(
"""
CREATE TABLE IF NOT EXISTS picow_test (
id TEXT,
ts TIMESTAMP WITH TIME ZONE GENERATED ALWAYS AS current_timestamp,
temp DOUBLE PRECISION
)
""",
return_response=False,
)
except Exception as e:
print("Error creating table:")
print(e)
sys.exit(1)
while True:
# TODO calculate this properly...
temperature = 22.1
response = crate.execute(
"INSERT INTO picow_test (id, temp) VALUES (?, ?)",
[
f"{ip_addr}",
temperature
]
)
if response["rowcount"] == 1:
print("Inserted record into CrateDB.")
time.sleep(10)
For installation with dependencies on CircuitPython we'll want to look at circup as well as listing dependencies in the README -- see Adafruit's requests for example..
Dependencies we'll need:
circuitpython-base64
adafruit-circuitpython-requests
Excellent Simon, thanks for the updates. Adafruit even provides a thoroughly encapsulated GitHub Actions Workflow workflows-circuitpython-libs/build. It would be cool if your package, even in its hybrid form of supporting both MicroPython and CircuitPython, could adhere to this specification, so builds will succeed when slapping that workflow recipe onto this repository.
|
gharchive/issue
| 2024-10-30T12:25:59 |
2025-04-01T04:33:53.621875
|
{
"authors": [
"amotl",
"simonprickett"
],
"repo": "crate/micropython-cratedb",
"url": "https://github.com/crate/micropython-cratedb/issues/24",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
3122418
|
Move events from FormFlow-class into real events
Fixes #11.
I will open up a new PR for this.
You could also just update the branch. Do whatever you want. ;)
Yeah but I somehow managed to completely screw the PR. Trying to solve this issue before open up a new one.
|
gharchive/issue
| 2012-02-07T11:14:27 |
2025-04-01T04:33:53.623630
|
{
"authors": [
"craue",
"daFish"
],
"repo": "craue/CraueFormFlowBundle",
"url": "https://github.com/craue/CraueFormFlowBundle/issues/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
850626946
|
Using a subfolder over cleanup_repository
Is possible to use a subfolder?. I'm using artifactory-cleanup without any issue to clean a repository called x which is Package Type: | RPM
If I try to use cleanup_repository= x/y or y I'm getting: ERR Cannot retrieve repo configuration for sgd error="400 Bad Request" policy=cleanup-policy
I duplicated the project which is working well and for the new one I only changed the parameters cleanup_repository as I described before, plus I set cleanup_include_names=* and cleanup_dryrun=true for my testing.
I tried just creating a virtual repository for the folder and I got ERR x-virtual is not a local repository
|
gharchive/issue
| 2021-04-05T20:05:29 |
2025-04-01T04:33:53.635451
|
{
"authors": [
"Klaudioz"
],
"repo": "crazy-max/artifactory-cleanup",
"url": "https://github.com/crazy-max/artifactory-cleanup/issues/64",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
240033999
|
Django 1.11.3 ?
Hi after upgrading to Django 1.11.3 i get this error :
Exception Value: context must be a dict rather than Context.
when using in my template :
{% social_widget_render "facebook/share_button.html" href=meta.url %}
Getting the same issue here on 1.11
the same issue on 1.11.3
the same issue on 1.11.4
I am still getting the same error in 1.11.5
|
gharchive/issue
| 2017-07-02T20:58:13 |
2025-04-01T04:33:53.662894
|
{
"authors": [
"chxiaoxin",
"luispaulot",
"moccand",
"naveenreddyin",
"wedwardbeck"
],
"repo": "creafz/django-social-widgets",
"url": "https://github.com/creafz/django-social-widgets/issues/9",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
168615315
|
$NODE_PATH is not being set, so it can not find global modules
For whatever reasons, nvm is not setting correct NODE_PATH, and hence my builds can not find the globablly installed modules.
my app does this require('ibrik') but it fails, even if the ibrik is installed globally
I think this is same problem as http://stackoverflow.com/questions/27876557/node-js-configuring-node-path-with-nvm
Am i supposed to set it manually ?
MacBook-Pro:coffee sudhir$ nvm ls
-> v6.3.1
default -> 6.3.1 (-> v6.3.1)
node -> stable (-> v6.3.1) (default)
stable -> 6.3 (-> v6.3.1) (default)
iojs -> N/A (default)
lts/* -> lts/argon (-> N/A)
lts/argon -> v4.4.7 (-> N/A)
MacBook-Pro:coffee sudhir$ nvm use 6.3.1
Now using node v6.3.1 (npm v3.10.3)
MacBook-Pro:coffee sudhir$ which npm
/Users/sudhir/.nvm/versions/node/v6.3.1/bin/npm
MacBook-Pro:coffee sudhir$ which node
/Users/sudhir/.nvm/versions/node/v6.3.1/bin/node
MacBook-Pro:coffee sudhir$ echo $NODE_PATH
<EMPTY>
MacBook-Pro:coffee sudhir$ npm ls -g -depth 0
/Users/sudhir/.nvm/versions/node/v6.3.1/lib
├── bower@1.7.9
├── coffee-script@1.10.0
├── grunt-cli@1.2.0
├── ibrik@2.0.0
├── karma-cli@1.0.1
└── npm@3.10.3
This is my exact problem - https://github.com/karma-runner/karma-cli/issues/22
Not sure if its problem because of nvm or its a karma bug
This is intentional. NODE_PATH should never be set, and global modules both are not requireable by default, and should never be required. Something you require is a dependency, and thus, must be installed locally.
Doesnt work even when i have ibrik installed locally..
@snimavat when it's locally installed, you should use an npm run-script so that it sets up the PATH for you. Alternatively, ./node_modules/.bin/ibrik should work.
I will tryout and let u knw, this is the only module which doesnt work... other locally installed works just fine..
@snimavat according to https://npmcdn.com/ibrik@2.0.0/package.json there's no "main" - so that entire package is not requireable. It is only a command-line tool.
Crazy.. i am not a node expert, just trying to get coverage for the coffee files, and karma coverage examples used it the same way as i posted above...
https://github.com/karma-runner/karma-coverage/blob/master/examples/coffee/karma.conf.coffee
This is strange, but this setup works on other machine..
Note sure where this issue belongs to.. here, or karma, or karma-coverage..
It looks like v1 did in fact have a "main" - https://npmcdn.com/ibrik@1.1.1/package.json - so I think you're just using the wrong version of ibrik. I'd file the issue on karma-coverage - it's definitely not related to nvm.
Thansk for clarifying, karma-coverage asks to use ibrik 2.. i will try to get help from them.. thanks
question on this - shouldn't / couldn't NODE_PATH get set for repl usage?
otherwise you can't really include global modules easily when using a REPL with nvm unless you do some hackery. sorry to reanimate a dead thread...
@illegalnumbers you shouldn't be including global modules ever under any circumstances. If you want to play with something in a repl, install it locally, and then it'll be available to require.
why though? sometimes it does make sense to install and use global dependencies outside of the scope of a project especially if you're using it in a repl. there are many points of common functionality from parsing files of specific types to outputting in specific ways that could be common. it feels rather arbitrary to say to never use a part of npm.
No, it never does make sense to do that; if it's required, it's a dependency, and it should be installed locally.
local to what though that's the question - it's not a project it's just a repl. it's a global usage area
What use case do you have for playing with require that isn't going to end up being a project?
exploratory testing a new module, using something for a script that isn't part of the project (parsing files with a syntax that is less terrible than doing bash scripting), etc
npm install --no-save new-module && node; as for the script, if it requires anything, it'd need a package.json to be able to be reproducible on another machine, so it's a bad idea to use globals for that too.
TIL --no-save which i think would fulfill what I needed for that particular use case and true, but I guess the prerequisite to what I was thinking would be that you aren't putting as part of the project and just running it as a one off
@maxatwork if it’s requireable in a project, it is a dependency of a project, full stop. Separately, global things are never requireable, only runnable as command line tools. If an editor wants to do something absurd like require something that isn’t a dependency of the editor, it’d be the editor’s responsibility to employ hacks to make that work, not nvm’s.
I found this thread because I have a requirement to use the appdynamics node module. The npm install for that module pulls a version specific to the os. This breaks my shrinkwrap file because it hashes the module and it's always different between windows, mac, and linux. I can either install it globally, or I have to set the user-agent when I build so that it pulls the version for the server I'm deploying to. It's a mess. I was exploring using require with a global module for this specific package.
@dustinaleksiuk typically using a shrinkwrap file also means every consumer has to be on the same OS (when you have compiled deps).
Appreciate the concise schooling on global vs. local modules. Especially late at night and doubly so when trying to use global packages when nvm was designed specifically to only locally define the node environment.
|
gharchive/issue
| 2016-08-01T10:30:58 |
2025-04-01T04:33:53.686415
|
{
"authors": [
"dustinaleksiuk",
"illegalnumbers",
"ljharb",
"seantcanavan",
"snimavat"
],
"repo": "creationix/nvm",
"url": "https://github.com/creationix/nvm/issues/1186",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
175876859
|
When using .nvmrc, $2 is undefined/empty
Fix for issue #1201
Outdated PR for NVM. Will rebase and rework.
I'm not sure what this is fixing - could you file an issue, with repro steps and nvm debug output, before submitting a PR? Thanks!
|
gharchive/pull-request
| 2016-09-08T22:05:39 |
2025-04-01T04:33:53.688092
|
{
"authors": [
"ljharb",
"robbwagoner"
],
"repo": "creationix/nvm",
"url": "https://github.com/creationix/nvm/pull/1222",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
299028420
|
Remove duplicate code in nvm_npm_global_modules()
No matter what's the result of the condition here, the path is always the same.
hmmm ... did I just miss something?
- if [ "_$VERSION" = "_system" ]; then
- NPMLIST=$(nvm use system > /dev/null && npm list -g --depth=0 2> /dev/null | command sed 1,1d)
- else
- NPMLIST=$(nvm use "$VERSION" > /dev/null && npm list -g --depth=0 2> /dev/null | command sed 1,1d)
- fi
+ NPMLIST=$(nvm use system > /dev/null && npm list -g --depth=0 2> /dev/null | command sed 1,1d)
they looks pretty the same to me
|
gharchive/pull-request
| 2018-02-21T16:04:16 |
2025-04-01T04:33:53.689429
|
{
"authors": [
"PeterDaveHello"
],
"repo": "creationix/nvm",
"url": "https://github.com/creationix/nvm/pull/1742",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
560178947
|
Only send analytics data in production mode
Signed-off-by: Olga Bulat obulat@gmail.com
Description
Currently, google analytics data is sent every time the app is run, even in testing and development mode. This change adds a check for production mode before setting up google analytics, as well as checks before sending each event, to minimize false events monitoring.
Checklist:
[x] My pull request has a descriptive title (not a vague title like Update index.md).
[x] My pull request targets the master branch of the repository.
[x] My commit messages follow best practices.
[ ] I added tests for the changes I made (if applicable).
[ ] I added or updated documentation (if applicable).
[x] I tried running the project locally and verified that there are no
visible errors.
Developer Certificate of Origin
Version 1.1
Copyright (C) 2004, 2006 The Linux Foundation and its contributors.
1 Letterman Drive
Suite D4700
San Francisco, CA, 94129
Everyone is permitted to copy and distribute verbatim copies of this
license document, but changing it is not allowed.
Developer's Certificate of Origin 1.1
By making a contribution to this project, I certify that:
(a) The contribution was created in whole or in part by me and I
have the right to submit it under the open source license
indicated in the file; or
(b) The contribution is based upon previous work that, to the best
of my knowledge, is covered under an appropriate open source
license and I have the right under that license to submit that
work with modifications, whether created in whole or in part
by me, under the same open source license (unless I am
permitted to submit under a different license), as indicated
in the file; or
(c) The contribution was provided directly to me by some other
person who certified (a), (b) or (c) and I have not modified
it.
(d) I understand and agree that this project and the contribution
are public and that a record of the contribution (including all
personal information I submit with it, including my sign-off) is
maintained indefinitely and may be redistributed consistent with
this project or the open source license(s) involved.
Looks great! Assuming you've already verified that events are sent in prod, and are not sent in dev, merge whenever!
I've tested it by opening analytics realtime screen and localhost version at first (dev), and analytics and serve -s dist with the new changes after a while. It did only show events that happened on dist (prod) version, but I will still monitor analytics just in case after I merge.
|
gharchive/pull-request
| 2020-02-05T07:00:11 |
2025-04-01T04:33:53.697401
|
{
"authors": [
"obulat"
],
"repo": "creativecommons/cc-chooser",
"url": "https://github.com/creativecommons/cc-chooser/pull/90",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
294119339
|
Deadlink in Japanese page
https://creativecommons.org/licenses/by-sa/3.0/deed.ja
This Japanese page contains deadlink that is in “これは人が読んでわかりやすいようにしたライセンスの要約です。”
Source
<span class="summary">
これは人が読んでわかりやすいようにした<a href="legalcode.ja" class="fulltext">ライセンス</a>の要約です。(ライセンスの代わりになるものではありません。)
</span>
Probably ‘href="legalcode"’ is correct.
@foomin10 It looks like there is no existing Japanese translation for the 3.0 licenses. I recommend using the 4.0 version located here: https://creativecommons.org/licenses/by-sa/4.0/deed.ja . The link to the license works there.
Can you tell me what led you to the 3.0 deed? How did you find a link to this page?
The definitive list of ported (we don't say translated for pre-4.0 ported licenses) is here: https://wiki.creativecommons.org/wiki/CC_Ports_by_Jurisdiction
The definitive list of translated legal code for the 4.0 licenses is here: https://wiki.creativecommons.org/wiki/Legal_Tools_Translation#Translation_status_of_the_4.0_licenses_and_of_CC0
Japan doesn't have a ported 3.0 license suite. The deed, however, is a different matter. Even without a port of the 3.0 legal code, there is a Japanese translation of the deed that is the same as the 4.0 deed but is rendered as 3.0 when a work is licensed under a 3.0 license whether generic or if ported to another jurisdiction. The version number is updated, but the text of all deeds for the pre-4.0 ports (1.0, 2.0 and 2.1) is identical to the deed that is rendered relative to the 4.0 legal code in Japanese.
Has been fixed. There is nothing I want to claim.
|
gharchive/issue
| 2018-02-03T15:28:28 |
2025-04-01T04:33:53.702265
|
{
"authors": [
"aldenstpage",
"foomin10",
"peterspdx"
],
"repo": "creativecommons/creativecommons.org",
"url": "https://github.com/creativecommons/creativecommons.org/issues/746",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
392966363
|
Weaviate Panics on incorrect user input on schema import
Weaviate panics on incorrect user input on schema imports
For example
when adding a cross-ref that is not a valid className (i.e. doesn't pass the Regex validation rules)
... (possibly more, haven't investigated yet)
Todos:
[ ] Find all (most common?) occurrences
[ ] Respond with a correct (error) HTTP response, but don't panic while serving the http request.
Closing as I wasn't reproduce this with either POST localhost:8080/weaviate/v1/schema/things or POST localhost:8080/weaviate/v1/schema/things/<className>/properties and various incorrect values. I assume this issue has been fixed in the mean time. Tracking similar issues in #826.
|
gharchive/issue
| 2018-12-20T09:49:56 |
2025-04-01T04:33:53.724130
|
{
"authors": [
"etiennedi"
],
"repo": "creativesoftwarefdn/weaviate",
"url": "https://github.com/creativesoftwarefdn/weaviate/issues/654",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
1338461376
|
🛑 Settletech [DEV] is down
In 9610266, Settletech [DEV] (https://dev.settletech.io/login) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Settletech [DEV] is back up in f5fc4fc.
|
gharchive/issue
| 2022-08-15T02:34:56 |
2025-04-01T04:33:53.772909
|
{
"authors": [
"wefundtech"
],
"repo": "credtechau/credtech-monitor",
"url": "https://github.com/credtechau/credtech-monitor/issues/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
335710631
|
Integrate Mailcatcher
Integrate Mailcatcher to provide an easy test environment for sending mails.
done.
|
gharchive/issue
| 2018-06-26T08:26:24 |
2025-04-01T04:33:53.773655
|
{
"authors": [
"benjaminknauer"
],
"repo": "creios/creiwork",
"url": "https://github.com/creios/creiwork/issues/5",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
404466598
|
Give some love to the docs
References:
Alabaster customization
The docs look way better now.
|
gharchive/issue
| 2019-01-29T20:11:59 |
2025-04-01T04:33:53.774979
|
{
"authors": [
"MaxHalford"
],
"repo": "creme-ml/creme",
"url": "https://github.com/creme-ml/creme/issues/9",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
707049772
|
Data stream generators
This PR is to port the existing synthetic stream generators:
[x] Agrawal
[x] AnomalySine
[x] ConceptDriftStream
[x] Friedman
[x] Hyperplane
[x] LED
[x] LEDDrift
[x] Logical
[x] Mixed
[x] RandomRBF
[x] RandomRBFDrift
[x] RandomTree
[x] SEA
[x] Sine
[x] STAGGER
[x] Waveform
Not ported generators are:
DataStream and FileStream. Functionality is already available via iter_csv and iter_array.
MultilabelGenerator and RegressionGenerator. They rely on scikit-learn data generators. The same functionality can be achieved using the scikit-learn generator and iter_array.
Notes: Some methods have (minor) changes that result in (expected) changes in usage/data values.
LEDDrift. Setting the seed value was (unnecessarily) set twice. The new implementation fixes this.
RandomTree. The new implementation returns the nominal features in single columns. Previously, these features were one-hot encoded.
SEA. This is a different implementation that uses Python's random number generator.
I think the PR is ready. Let me know if there is something else to do @MaxHalford :-)
|
gharchive/pull-request
| 2020-09-23T05:02:58 |
2025-04-01T04:33:53.781576
|
{
"authors": [
"jacobmontiel"
],
"repo": "creme-ml/creme",
"url": "https://github.com/creme-ml/creme/pull/356",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
262699812
|
Horizontal Scroll on Pages
Ideally located on the bottom of pages for scrolling to the left and right on pages that ARE wider than the active window size.
I don't understand this bug. What websites are not showing horizontal scroll that do on other browsers when the browsers are the same size? Can you give an example? I have confirmed that Doogie does properly show horizontal scrollbars on the browser for wide pages.
Yes, when text is used within a div that is longer than the page horizontally.
I try to provide an example soon, if still necessary.
@tfont - definitely necessary. Screenshot too would help.
|
gharchive/issue
| 2017-10-04T08:06:29 |
2025-04-01T04:33:53.790755
|
{
"authors": [
"cretz",
"tfont"
],
"repo": "cretz/doogie",
"url": "https://github.com/cretz/doogie/issues/60",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
694905194
|
Use anyhow crate for error handling
To simplify error handling we can utilize the anyhow crate:
https://github.com/dtolnay/anyhow
This means we can add an error context and reduce the type overhead in
function signatures. It also formats the errors in a more sane way, for
example if build.rs would fail:
Error: compile CRI protocol buffers
Caused by:
No such file or directory (os error 2)
/release-note-none
|
gharchive/pull-request
| 2020-09-07T09:23:31 |
2025-04-01T04:33:53.801363
|
{
"authors": [
"saschagrunert"
],
"repo": "cri-o/cri",
"url": "https://github.com/cri-o/cri/pull/10",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1266378882
|
TimeTaken ordering is being done alphabetically.
TimeTaken is not being converted into a numeric field, and as such the ordering is being done in an alphabetical way.
Change the type of the TimeTaken column to be a numeric one.
Closed as fixed by https://github.com/cristian-clamsen/FrebSbS/releases/tag/V1.3
|
gharchive/issue
| 2022-06-09T16:20:24 |
2025-04-01T04:33:53.816152
|
{
"authors": [
"HugoDurana",
"cristian-clamsen"
],
"repo": "cristian-clamsen/FrebSbS",
"url": "https://github.com/cristian-clamsen/FrebSbS/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
238542527
|
Blowfish implemented
Blowfish implemented and generator.py updated (argument for blowfish -f BLOWFISH) and added
This PR is replaced with #8
|
gharchive/pull-request
| 2017-06-26T13:26:54 |
2025-04-01T04:33:53.843604
|
{
"authors": [
"mitko501",
"rozsa117"
],
"repo": "crocs-muni/eacirc-streams",
"url": "https://github.com/crocs-muni/eacirc-streams/pull/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1410412401
|
Navigation Menu overlapping on Web Mobile
Issue: the navigation Menu Discord overlaps the edit button on mobile.
OS: Android.
Thanks @highb33kay for reporting. We need to only show 3 icons in the home screen and not show these icons in other screens. Please feel free to fix it if anyone is interested. This needs to be fixed in the web branch.
Assign me then. Thank you @kamalkishor1991
Can you point me in the right direction as to location the code source?
I think I see it Navigation
Yes it is defined here
https://github.com/croma-app/croma/blob/51f1793c20597d1463a4fb4314f25fb8bcc69d1d/navigation/NavigatorHelper.js#L161
Hi again @kamalkishor1991 . I don't know enough react to fix this alone. So any suggestions?
So here is what I have so far:
I think the navigationHelper.js file is okay and the issue lies in the different screen the function is being called. A sure fix would be to remove any screen that isnt the home screen that calls this function.
If 1 is not the case, another theory is that the navigation menus somehow got tangled together and are being bundled as one, which is causing the issue. a fix would be to untangle them (Sorry if i am not using the right words, I dont really understand how react works)
That is all I have as far as theories are concerned.
I tried using chrome inspect tools and I found out that doesn't work for react lol.
@highb33kay
I think way to implement it is to remove it from navigation and implement a custom header. You can refer to custom header here.
https://github.com/croma-app/croma/blob/51f1793c20597d1463a4fb4314f25fb8bcc69d1d/screens/ColorListScreen.js#L53
That is a big help, thank you.
is this issue fixed?
@nisilachandunu no.
Hi @kamalkishor1991, I'd like to take a stab at this. Is that alright? :)
@willdelorm sure. Thanks
Let me know if you need any help
Hi @kamalkishor1991, I'd like to try to fix this issue, will that be alright?
|
gharchive/issue
| 2022-10-16T08:14:27 |
2025-04-01T04:33:53.856926
|
{
"authors": [
"highb33kay",
"kamalkishor1991",
"mtalhabaig3",
"nisilachandunu",
"willdelorm"
],
"repo": "croma-app/croma",
"url": "https://github.com/croma-app/croma/issues/160",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
111919916
|
tuple index out of range if raising ApplicationError without args
Simplest example:
from autobahn.twisted.wamp import ApplicationSession
from autobahn.wamp import ApplicationError
from twisted.internet.defer import inlineCallbacks
class App(ApplicationSession):
def test(self):
raise ApplicationError(u'error')
@inlineCallbacks
def onJoin(self, details):
yield self.register(self.test, u'test')
yield self.call(u'test')
tuple index out of range appears in line errmsg = u"{0}".format(err.value.args[0]) (file autobahn/wamp/protocol.py)
Duplicate of #484
|
gharchive/issue
| 2015-10-16T22:11:03 |
2025-04-01T04:33:53.870542
|
{
"authors": [
"jvdm",
"schoonc"
],
"repo": "crossbario/autobahn-python",
"url": "https://github.com/crossbario/autobahn-python/issues/528",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
347646597
|
Add getting-started examples code to the dockers, it will be used in the Getting Started page of new documentation
Contains Pubsub and RPC pair example using twisted.
Codecov Report
:exclamation: No coverage uploaded for pull request base (master@9b6fb57). Click here to learn what that means.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #1032 +/- ##
=========================================
Coverage ? 61.27%
=========================================
Files ? 74
Lines ? 13141
Branches ? 2091
=========================================
Hits ? 8052
Misses ? 4386
Partials ? 703
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 9b6fb57...c3e351c. Read the comment docs.
@om26er when we move to multi-arch, we only need 1 dir instead of 3 (as of now .. for each CPU arch), right? if so, the example code of @codelectron wouldn't need 3 copies ..
@meejah yes, agreed: we should have the AB docs (which now are promoting the Component API https://autobahn.readthedocs.io/en/latest/wamp/programming.html#application-components) and the new getting started examples (the docker based stuff @codelectron is working on) "in sync" - that is both use the Component API. Sorry @codelectron for 2nd round .. blame me .. forgot to mention that ABPy has 2 APIs at the WAMP level: the "old" one, which is ApplicationSession (inheritance) based (also me to blame), and with this certain limitations or unwanted user code intrusiveness results. The "new API" is not using inheritance, and I consider it a much better design. The "problem" is: we have a lot of code, examples etc that is based on ApplicationSession - but we certainly want to avoid creating even more stuff based on this ..
|
gharchive/pull-request
| 2018-08-04T20:35:56 |
2025-04-01T04:33:53.876725
|
{
"authors": [
"codecov-io",
"codelectron",
"oberstet"
],
"repo": "crossbario/autobahn-python",
"url": "https://github.com/crossbario/autobahn-python/pull/1032",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
436846022
|
[WIP] Add push/pop lockless
This PR will add non-deadlocking push/pop operations to queues -- they will never deadlock even if a concurrent push/pop stalls for whatever reason indefinitely.
@Restioson I have a high-level question about your use case with interrupts pushing keyboard events into the queue.
So if your interrupt handler is defined like this:
fn handler(event: KeyboardEvent) {
queue.push(event);
}
What happens if an event handler (call it A) is interrupted before push is called by another interrupt handler (call it B)? Does that mean keyboard event from A will end up in the queue after event from B?
That sounds wrong to me so I wonder how keyboard drivers handle situations like these? Are you sure this is the way to go?
My guess would be that event handlers shouldn't be interruptable and maybe the hardware guarantees so, but I've no idea if what I'm saying is true.
Yes, event handlers cannot be interrupted (interrupts are disabled by the os). This is because if nesting is allowed, excessive nesting depth could cause a stackoverflow. However, the local APIC is able to buffer one IRQ to be dispatched if the cpu has disabled interrupts, which will be dispatched when the cpu re-enables them.
I do believe that I still need the function though, for instance in case the push interrupts a pop from the handler. And nevertheless, it could be useful for others.
@Restioson Thanks for clarifying!
And just a couple more questions to see if I understood the problem domain correctly:
Is it true that there is only a single consumer?
Is it true that there is only a single producer?
Is it true that even though you're using the queue in the SPSC fashion, you want a MPSC or MPMC interface only to get around Rust's ownership/sharing rules?
yes
yes
yes pretty much :) it has to be static unfortunately
In that case, the way to go would really be https://github.com/crossbeam-rs/crossbeam/pull/338, which is ~5 times faster than ArrayQueue anyway :)
It's unfortunate that the SPSC queue cannot be used in your use case without unsafe {} blocks, but that is really the nature of the problem of interrupts. So my suggestion would be to use static mut variable then...
Aren't those slated to be removed? Perhaps a RefCell. I will consider it though. Do you still think that this has a use case?
Aren't those slated to be removed?
Possibly -- I'm not sure. The alternative to static mut would be a static UnsafeCell<T> variable.
Do you still think that this has a use case?
I don't think so. The SPSC queue is really designed to be wait-free and applicable in embedded systems, device drivers, and similar use cases. Others queues are a bit more high-level and can't provide strong progress guarantees without jumping through hoops or having tricky caveats.
Sorry, I haven't had much time to work on this because of exams... I will get back to it when they are finished though.
|
gharchive/pull-request
| 2019-04-24T18:32:30 |
2025-04-01T04:33:53.900958
|
{
"authors": [
"Restioson",
"stjepang"
],
"repo": "crossbeam-rs/crossbeam",
"url": "https://github.com/crossbeam-rs/crossbeam/pull/363",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1943646180
|
Implement entries cli subcommand
Add a subcommand to access the entries feedbin endpoint:
https://github.com/feedbin/feedbin-api/blob/master/content/entries.md
Being worked on in this feature branch
The feedbin entries endpoint can also take ids to retrieve. Not sure how that's supposed to interact with flags such as read and starred. Integrating with this subcommand will be a little weird.
Also need to handle the sinceparameter. Should use one of the following modules for human timetamp input:
arrow
pendulum
|
gharchive/issue
| 2023-10-15T02:17:04 |
2025-04-01T04:33:53.904206
|
{
"authors": [
"crossjam"
],
"repo": "crossjam/feedbin-tools",
"url": "https://github.com/crossjam/feedbin-tools/issues/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1644361955
|
fix(rds/docdb)!: ParameterGroups remove ApplyMethod isUpToDate check
Description of your changes
RDS: DBParameterGroup, DBClusterParametergroup and DocDB DBClusterParametergroup
Breaking: switch to custom struct CustomParameters with only fields applyMethod, parameterName and parameterValue (terraform does it the same way)
Breaking: make parameter fields required in CRD (for RDS these 3 fields where already required by the controller before; for RDS and DocDB those fields were also already required indirectly by AWS )
reduce Parameter isUpToDate check to only consider parameterValue
Fixes #1708
Breaking changes ToDos/notes for users:
for resources RDS: DBParameterGroup, DBClusterParametergroup and DocDB DBClusterParametergroup claims need to be checked:
each entry in parameters no longer uses the previously read-only fields allowedValues, applyType, dataType, description, isModifiable, minimumEngineVersion, source and supportedEngineModes
each entry in parameters requires the only use of all 3 the fields applyMethod, parameterName and parameterValue
I have:
[x] Read and followed Crossplane's [contribution process].
[x] Run make reviewable test to ensure this PR is ready for review.
How has this code been tested
Manually
@haarchri Considering the breaking changes, I added a potential ToDo note, that can be added to the release notes for users.
Sounds good to me - thanks
|
gharchive/pull-request
| 2023-03-28T17:12:57 |
2025-04-01T04:33:53.910502
|
{
"authors": [
"haarchri",
"wotolom"
],
"repo": "crossplane-contrib/provider-aws",
"url": "https://github.com/crossplane-contrib/provider-aws/pull/1709",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
428455010
|
image build broken on jenkins
Multiple pull requests are seeing the image build fail on Jenkins, e.g.,
https://jenkinsci.upbound.io/blue/organizations/jenkins/crossplane%2Fbuild/detail/PR-363/1/pipeline
https://jenkinsci.upbound.io/blue/organizations/jenkins/crossplane%2Fbuild/detail/PR-362/2/pipeline
[2019-04-02T16:45:46.752Z] Step 2/8 : RUN apk --no-cache add ca-certificates bash
[2019-04-02T16:45:46.752Z] ---> Running in c627ed56a895
[2019-04-02T16:45:46.752Z] standard_init_linux.go:190: exec user process caused "no such file or directory"
[2019-04-02T16:45:46.752Z] The command '/bin/sh -c apk --no-cache add ca-certificates bash' returned a non-zero code: 1
[2019-04-02T16:45:46.752Z] 16:45:46 [FAIL]
[2019-04-02T16:45:46.752Z] Makefile:16: recipe for target 'img.build' failed
[2019-04-02T16:45:46.752Z] make[6]: *** [img.build] Error 1
[2019-04-02T16:45:46.752Z] make[5]: *** [do.build.artifacts.linux_arm64] Error 2
Google searches indicate it's a line endings issue, but I don't think the Dockerfile has changed at all recently. Has the Jenkins instance/slave or docker daemon been updated recently? Do they need to be restarted?
This is blocking the v0.2 release that we want to get out this week.
This issue occurred when the qemu container stopped running on the slave Jenkins node. As the system is persistent, it is susceptible to gradual changes (OS/Application updates) which seem to have affected the Docker daemon. After rebooting the slave node, builds started running again.
Long Term Fixes:
[ ] Cycle slave node regularly (every other week)
[ ] Use configuration management (Chef) to keep slave configs in a declared state
[ ] Implement automate Packer process to generate updated templates used by slave nodes
Thanks for looking into this issue @Defilan, it looks like the builds are running again OK. Are you tracking those "long term fixes" somewhere already so we don't forget them? Thanks!
Closing this issue for v0.2 since the builds are working again.
|
gharchive/issue
| 2019-04-02T21:33:28 |
2025-04-01T04:33:53.923925
|
{
"authors": [
"Defilan",
"jbw976"
],
"repo": "crossplaneio/crossplane",
"url": "https://github.com/crossplaneio/crossplane/issues/364",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
56216410
|
[atip][XWALK-3539] Update tizen.py to get user id dynamically
Tizen user can be given by param of lite dynamically.
The default tizen user is app.
LGTM, @cicili please help double check. :)
@Honry @cicili can this pull request be landed? It affects the multi-user tasks.
@chenxix @Honry please add the UT result
Verified on latest testkit-lite(v3.1.13) build, the userID can be achieved sucessfully.
thanks, and LGTM
|
gharchive/pull-request
| 2015-02-02T10:02:26 |
2025-04-01T04:33:53.927589
|
{
"authors": [
"Honry",
"chenxix",
"cicili",
"zqzhang"
],
"repo": "crosswalk-project/crosswalk-test-suite",
"url": "https://github.com/crosswalk-project/crosswalk-test-suite/pull/1778",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
92824073
|
Fix Coverity failure caused uninitialized class member.
BUG=XWALK-4543,XWALK-4542
@pozdnyakov PTAL, thanks.
I think @pozdnyakov is on holidays, @rakuco ?
lgtm.
BTW, the xwalk_runtime_features.cc was fixed a long time ago in ef046ff37efdf0d97f12f947aba1fbfa72236c3e, but 69df77eede452f47fc5dcf736d6e3d5212ddd4b1 reintroduced the very same bug :(
|
gharchive/pull-request
| 2015-07-03T08:39:00 |
2025-04-01T04:33:53.929538
|
{
"authors": [
"baleboy",
"rakuco",
"seanlong"
],
"repo": "crosswalk-project/crosswalk",
"url": "https://github.com/crosswalk-project/crosswalk/pull/3120",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
116004748
|
[Android] Support theme color at task switcher on Lollipop+(5.0+)
For Android version older than 5.0, this doesn't work.
BUG=XWALK-4305
(cherry picked from commit c6a8be02852b53b67538316ed2ef0c90890e749f)
We do not release anymore version 13 : see https://crosswalk-project.org/documentation/downloads.html.
... and https://lists.crosswalk-project.org/pipermail/crosswalk-dev/2015-October/003136.html
|
gharchive/pull-request
| 2015-11-10T01:19:08 |
2025-04-01T04:33:53.931516
|
{
"authors": [
"axinging",
"darktears",
"rakuco"
],
"repo": "crosswalk-project/crosswalk",
"url": "https://github.com/crosswalk-project/crosswalk/pull/3374",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
104158632
|
User override
This allows a user logged in as an admin to spontaneously create more users (while specifying worker or requester or both). To test you can make yourself an admin by UPDATE auth_user SET is_staff='t' WHERE id={your_id};
there is a validator crowdsourcing/validators/utils.py RegistrationAllowed
which will invalidate this action, you might have to change that
thanks ill look into it later tonight.
Let's complete registration close as well.
lgtm
lgtm.
thank you!
On Tue, Sep 1, 2015 at 4:33 PM, neilthemathguy notifications@github.com
wrote:
Merged #465
https://github.com/crowdresearch/crowdsource-platform/pull/465.
—
Reply to this email directly or view it on GitHub
https://github.com/crowdresearch/crowdsource-platform/pull/465#event-398390127
.
|
gharchive/pull-request
| 2015-09-01T00:23:28 |
2025-04-01T04:33:53.949911
|
{
"authors": [
"aginzberg",
"dmorina",
"mbernst",
"neilthemathguy"
],
"repo": "crowdresearch/crowdsource-platform",
"url": "https://github.com/crowdresearch/crowdsource-platform/pull/465",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
773215541
|
Jreyesr teleop panel (copy of #561)
Copy of #561
For anyone watching this issue - we recently added a teleop panel to Foxglove Studio (webviz fork) in #1429.
|
gharchive/pull-request
| 2020-12-22T20:09:08 |
2025-04-01T04:33:53.953095
|
{
"authors": [
"amacneil",
"vidaaudrey"
],
"repo": "cruise-automation/webviz",
"url": "https://github.com/cruise-automation/webviz/pull/567",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
354325920
|
Moving away from new Buffer(...)
new Buffer(...) is deprecated and unsafe. Would it be possible to replace its usage with Buffer.from(...)?
yeah we should switch over to safe-buffer
|
gharchive/issue
| 2018-08-27T13:46:37 |
2025-04-01T04:33:53.972514
|
{
"authors": [
"calvinmetcalf",
"gabrielschulhof"
],
"repo": "crypto-browserify/publicEncrypt",
"url": "https://github.com/crypto-browserify/publicEncrypt/issues/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
445410967
|
Correct way to derive a key from a passphrase
Is it possible / reasonable to derive a key for SimpleBox from a password / passphrase? Or do I just need to use a generated random password?
The password hashing algorithms in this library can all be used for cryptographic key derivation. However, generating a strong password is still probably a good idea to ensure minimum entropy (a cryptographic key derived from a password is still only as strong as the original password):
https://github.com/crypto-rb/rbnacl/wiki/Password-Hashing
I've looked at that page and had a go at using it for key derivation, however I understood it to be used for hashing a password so that it can be later verified. If I run through that example for Argon2 :
password = 'password'
salt = RbNaCl::Random.random_bytes(RbNaCl::PasswordHash::Argon2::SALTBYTES)
opslimit = 5
memlimit = 7_256_678
digest_size = 32
digest_1 = RbNaCl::PasswordHash.argon2(
password,
salt,
opslimit,
memlimit,
digest_size
)
salt = RbNaCl::Random.random_bytes(RbNaCl::PasswordHash::Argon2::SALTBYTES)
digest_2 = RbNaCl::PasswordHash.argon2(
password,
salt,
opslimit,
memlimit,
digest_size
)
digest_1 == digest_2
=> false
The calculated digests are different - so they can't be used as the key - right? So how is it supposed to work? Are we supposed to use the same salt each time maybe?
The calculated digests are different - so they can't be used as the key - right? So how is it supposed to work? Are we supposed to use the same salt each time maybe?
Yes, you need to record the salt that was used somewhere, and supply the same one if you intend to derive the same cryptographic key.
I've looked at that page and had a go at using it for key derivation, however I understood it to be used for hashing a password so that it can be later verified
Both scrypt and Argon2 are password-based key derivation functions (KDF), and are explicitly designed for the purpose of generating cryptographic keys from a password. From Wikipedia:
https://en.wikipedia.org/wiki/Key_derivation_function
KDFs happen to have the characteristics desired for a "password hash function", even though they were not originally designed for this purpose.
Awesome - thanks very much. Is it possible to update the wiki page to include that information for key derivation?
|
gharchive/issue
| 2019-05-17T12:09:43 |
2025-04-01T04:33:53.980302
|
{
"authors": [
"dkam",
"tarcieri"
],
"repo": "crypto-rb/rbnacl",
"url": "https://github.com/crypto-rb/rbnacl/issues/193",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
179728440
|
Windows paths bug in Dir.php:260
you should escape strings for preg_replace
$target = preg_replace('~^' . preg_quote($root) . '~', '', $path);
on Windows it does not replace the root path.
see here
Not sure to get what you mean.
Can you give me an example with a real path ? What is the actual result and what you expected ?
Hello Simon,
yes, assume $path is „c:\dir\file.txt”
when concatenated as string to regex pattern some backslashes act as modifiers so “c:\dir” becomes actually “c:[0-9]ir”, which is wrong.
And even if you escape that path string with regex_qoute() this does not solve all problems with windows paths.
There is one more:
For example we want to copy “C:\dir\” to “D:\dir”, the iterator resolves all files and returns an array:
Array(
“C:\dir/subdir/file1”,
“C:\dir/file2”,
);
So replacement fails again, since “/” can’t be matched by “\”.
It results in concatenation of dest path with original source path and becomes something like “D:\dir\C:\dir/subdir/file”, which is simply wrong.
Best regards
Eugen Wesseloh
Von: Simon JAILLET [mailto:notifications@github.com]
Gesendet: Mittwoch, 28. September 2016 12:07
An: crysalead/dir
Cc: Eugen Wesseloh; Author
Betreff: Re: [crysalead/dir] Windows paths bug in Dir.php:260 (#3)
Not sure to get what you mean.
Can you give me an example with a real path ? What is the actual result and what you expected ?
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHubhttps://github.com/crysalead/dir/issues/3#issuecomment-250126094, or mute the threadhttps://github.com/notifications/unsubscribe-auth/AOOoIwsVXQp4b4JwSSvtzgJkS9NZVNvqks5qujxfgaJpZM4KIn-3.
Good catch, so does using:
$target = preg_replace('~^' . preg_quote(rtrim($root, $ds)) . '~', '', $path);
works for both cases ?
Yes. we have a windows webserver, so we had to fix that first.
You may experience problems even on *nix systems without that fix, just try some regex modifiers in filenames (like "fine.+name").
Fixed in https://github.com/crysalead/dir/commit/bb91ed1e8d2db50ac88a62b8ba7b2afdbf94af97
|
gharchive/issue
| 2016-09-28T09:53:12 |
2025-04-01T04:33:54.014149
|
{
"authors": [
"ewesseloh",
"jails"
],
"repo": "crysalead/dir",
"url": "https://github.com/crysalead/dir/issues/3",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1616261718
|
Error: undefined fun 'isatty' for LibC on Windows
Bug Report
Discovered as part of https://github.com/athena-framework/athena/actions/runs/4370118689/jobs/7644691322 within the framework tests:
Creating library C:\Users\runneradmin\AppData\Local\crystal\cache\D-a-_temp-crystal-nightly-true-undefined-src-ecr-process.cr\macro_run.lib and object C:\Users\runneradmin\AppData\Local\crystal\cache\D-a-_temp-crystal-nightly-true-undefined-src-ecr-process.cr\macro_run.exp
Showing last frame. Use --error-trace for full trace.
In D:\a\_temp\crystal-nightly-true-undefined\src\crystal\system\win32\socket.cr:338:10
338 | LibC.isatty(fd) == 1
Error: Error: undefined fun 'isatty' for LibC
I haven't manually confirmed it, but I'd assume this is a resume of doing like STDIN.tty? on windows.
Seems the actual binding for that function was missed in https://github.com/crystal-lang/crystal/pull/10784 and this is first time someone tried using IO#tty? on windows?
Okay no, STDIN.tty? works just fine. But it does fail if you actually require socket:
require "socket"
pp STDIN.tty?
339 | LibC.isatty(fd) == 1
^-----
Error: undefined fun 'isatty' for LibC
Easiest to just copy Crystal::System::FileDescriptor#system_tty?, which uses LibC._isatty instead, to Crystal::System::Socket#system_tty?
I don't think a socket on Windows can ever be a tty. Unlike on POSIX platforms, they're disparate handles. And AFAIK you cannot turn a console handle into a socket handle somehow.
So perhaps the trivial solution is for Socket#system_tty? to return false on windows?
Based on what I read that's probably worth a try. It'll at least get things compiling and can always revisit if someone ever comes across a case where it could be true. Will see what CI things. I pushed up another commit.
|
gharchive/issue
| 2023-03-09T01:54:41 |
2025-04-01T04:33:54.038357
|
{
"authors": [
"Blacksmoke16",
"HertzDevil",
"straight-shoota"
],
"repo": "crystal-lang/crystal",
"url": "https://github.com/crystal-lang/crystal/issues/13165",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
319590806
|
Alternative to BountySource?
First of all, I apologize for creating an issue for a non-technical topic -- to me, it felt like the cleanest way of raising this question/issue with the core team in a manner that makes it publicly visible and findable. If this was a stupid idea, please close this and accept my apologies.
My question is the following: I would like to support Crystal with a monthly donation. Is there an alternative way of doing this besides going through BountySource?
My reason for asking this question is that my experience with BountySource so far has not been great. A short summary:
Signing up via GitHub failed with some error message I unfortunately did not record
Signing up via Twitter asked me for my email address, then gave me an error that my email was already in use (possibly an artifact of my previous attempt at signing up through GitHub)
I eventually gained access to my account through BountySource's password recovery
Setting up a $10/month donation for Crystal gave me a meaningless error message ("Validation failed: Active support levels count Can't modify active rewards"), but only after authorizing with PayPal
Two more attempts at setting up a $10/month subscription had the same result
The next day, I see a $30 charge from BountySource. Lo and behold, they did, in fact, set up three separate $10/month subscriptions.
The UI for cancelling these subscriptions is broken (the submit button doesn't close the modal), but fortunately, the actual cancellation works.
I know these are bugs with BountySource and not Crystal (and I've raised them with BountySource in my slightly angry rant on Twitter), but at this point, I simply don't trust BountySource enough for any kind of transaction.
I'd like to put $25/month into Crystal. I don't need a listing on Crystal's supporters page, but I would receive some sort of receipt (something that BountySource also appears to not be doing?), and I don't want to go through a platform I don't trust. What are my options?
(FWIW, I'm supporting a number of open source projects though OpenCollective, which has been a good experience so far. They even send receipts.)
👍 for setting up OpenCollective account.
I also gave up on BountySource long ago, same frustrations. Website's simply busted.
Some issue here. We used to support via BountySource but it became to painful to use.
We support ImageMagick monthly through PayPal which is a much better setup.
A quick update regarding my own situation with BountySource: they've refunded those $30 (which I feel bad about because I had no problems with giving $30 to Crystal at all) and added $50 of credit to my account. https://twitter.com/Bountysource/status/991960275103334400
I appreciate that they acted on this in some fashion, but it doesn't really fix the issue I'm having with them. I will try to forward those $50 to Crystal, but would still love to set up a regular support payment via a platform that is not BountySource.
🏓 @crystal-lang core team (psss! it's about your money, pay attention ;))
We've had some conversation with OpenCollective in the past far before this issue. There are some unresolved issues regarding how to get the money from OC to Manas, mainly because of regulatory boring things. Maybe we could give another round and see if it is workable.
Regarding where the money goes. It goes to Manas and they are spend only for crystal development hours. As it can be read in a past blog post currently at least twice the hours covered by donations are been allocated.
@wontruefree the link has been fixed
Thanks that explains a lot and is very open of Manas.
Btw. currently visiting https://www.bountysource.com/teams/crystal-lang/fundraisers/702-crystal-language (at least from Germany) results in the page starting to load and then it just hangs and eats up 100% CPU – tested in Safari, Chrome and FF. So I doubt it's a browser bug.
Happens here, too, @felixbuenemann.
@felixbuenemann I think that is the old link that changed on BS side. It was changed to https://salt.bountysource.com/teams/crystal-lang in some places, but if you still find it somewhere let us know where.
It is linked from the bountysource badge at the top of the README.md.
In the past we had struggled with Paypal as they had not find a legal pathway through the tax agency here (Argentina).
I have checked that again and it looks like that has now been solved and we can open a business paypal account to receive payments via that channel.
We like OpenCollective a lot and will start the process of setting it up for Crystal. I'll update here when we have news.
Thanks!
Any update on this? We'd like to donate money
@blitline-dev thanks for your support! 💙
Haven't forgot about this issue, but we're talking Argentinian times here :)
Will let you know here as soon as we have an update.
Thanks for the patience! Here we are, any early feedback on the experience would be appreciated. And thanks a ton for the support 🙏 ! cc @blitline-dev @hmans
https://opencollective.com/crystal-lang
Thanks for setting up an OpenCollective account.
To remind everyone that they should be avoided, BountySource today sent out this email:
Not only are they telling me this a full 16 days after the start date of this fees, they also just deducted this fee from my account balance there right away. (This, fortunately, was "free" money they added to my account after the woes I had with them that triggered the creation of this issue, but if I had known about these inactivity fees in advance, I would have preferred to put it to better use than just letting them claim it back.)
tl;dr Avoid BountySource.
Hi Hendrik,
We had seen a notification of this a while ago and have been doing periodic withdrawals, but thanks for the heads up!
On Dec 17, 2018, at 5:38 AM, Hendrik Mans notifications@github.com wrote:
Thanks for setting up an OpenCollective account.
To remind everyone that they should be avoided, BountySource today sent out this email:
https://user-images.githubusercontent.com/1061/50075369-0f726200-01df-11e9-9952-221519ecbc2b.png
Not only are they telling me this a full 16 days after the start date of this fees, they also just deducted this fee from my account balance there right away. (This, fortunately, was "free" money they added to my account after the woes I had with them that triggered the creation of this issue, but if I had known about these inactivity fees in advance, I would have preferred to put it to better use than just letting them claim it back.)
tl;dr Avoid BountySource.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub https://github.com/crystal-lang/crystal/issues/6051#issuecomment-447763333, or mute the thread https://github.com/notifications/unsubscribe-auth/ABS9VUTSsxTHk-ovaG3qD4V9LiXGl3tyks5u51gBgaJpZM4Tvn66.
Anything left to do here?
|
gharchive/issue
| 2018-05-02T15:01:58 |
2025-04-01T04:33:54.056722
|
{
"authors": [
"Sija",
"arikast",
"bcardiff",
"blitline-dev",
"felixbuenemann",
"hmans",
"jhass",
"matiasgarciaisaia",
"nditada",
"wontruefree"
],
"repo": "crystal-lang/crystal",
"url": "https://github.com/crystal-lang/crystal/issues/6051",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
680197291
|
Document deployment build
Need to document the full deployment flow starting with how to build the new release.
Also, I prefer to separate the staging/production tfweb instance running, as we might have breaking changes while we use a single instance for both,
Already documented at https://github.com/crystaluniverse/publishingtools#build-binaries
|
gharchive/issue
| 2020-08-17T12:28:32 |
2025-04-01T04:33:54.074330
|
{
"authors": [
"abom"
],
"repo": "crystaluniverse/publishingtools",
"url": "https://github.com/crystaluniverse/publishingtools/issues/194",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2439294619
|
IMPORT FAILED CRYSTOOLS on Comfyui SDXL1
Describe the bug
I was using my friend workflow that also contained crystools, but after drag and drop the workflow into my comfyui, i says the error popup, but first i thought it is just a missing node, but after installing the missing node, the crystools node only said import failed in the manager.
To Reproduce
Please attach a workflow file to make it easier for others to reproduce the error!
Expected behavior
just some missing nodes
Screenshots
Error in console:
When loading the graph, the following node types were not found:
Primitive boolean [Crystools] (In group node 'workflow/Switch Detector')
workflow/Switch DetectorRemove from workflow
Switch any [Crystools] (In group node 'workflow/Switch Detector')
Switch conditioning [Crystools]
Nodes that have failed to load will show as red on the graph.
Versions:
Copy the output of the console (4 parts), like this:
** Python version: 3.11.8 (tags/v3.11.8:db85d51, Feb 6 2024, 22:03:32) [MSC v.1937 64 bit (AMD64)]
Total VRAM 8192 MB, total RAM 32555 MB
Set vram state to: NORMAL_VRAM
[Crystools INFO] Crystools version: 1.15.0
[Crystools INFO] CPU: 12th Gen Intel(R) Core(TM) i7-12700K 3.60 GHz - Arch: 64 Bit - OS: Windows 10
[Crystools INFO] GPU/s:
[Crystools INFO] 0) NVIDIA GeForce RTX 2070 8GB
[Crystools INFO] NVIDIA Driver: 555.85
### Loading: ComfyUI-Manager ( V2.46.3)
### ComfyUI Revision: 2418[66d35c] | Released on (2024-07-29)
Additional context
Add any other context about the problem here.
PAK HANG.json
Workflow attached
umm, is strange, please paste all the console output (log of python server)
@VantaHitam
how do i do that, can i take from the cmd instead? im no computer wiz
The same way you copy and paste the report, yeap, from cmd :)
E:\Ai\ComfyUI_windows_portable>.\python_embeded\python.exe -s ComfyUI\main.py --windows-standalone-build
[START] Security scan
[DONE] Security scan
ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-07-31 21:22:58.359464
** Platform: Windows
** Python version: 3.11.8 (tags/v3.11.8:db85d51, Feb 6 2024, 22:03:32) [MSC v.1937 64 bit (AMD64)]
** Python executable: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe
** ComfyUI Path: E:\Ai\ComfyUI_windows_portable\ComfyUI
** Log path: E:\Ai\ComfyUI_windows_portable\comfyui.log
#######################################################################
[ComfyUI-Manager] Starting dependency installation/(de)activation for the extension
Install: pip packages for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Inspire-Pack'
Install: pip packages for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack'
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
[SKIP] Downgrading pip package isn't allowed: scipy (cur=1.14.0)
Install: install script for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack'
[!]
[!] [bold yellow]WARN: The COMFYUI_MODEL_PATH environment variable is not set. Assuming E:\Ai\ComfyUI_windows_portable\ComfyUI\models as the ComfyUI path.[/bold yellow]
ComfyUI-Impact-Pack: Check dependencies
ComfyUI-Impact-Pack: Updating subpack
[Impact Pack] EXECUTE: ['E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe', 'install.py'] in 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\impact_subpack'
req_path: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\impact_subpack\requirements.txt
req_path: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\impact_subpack\requirements.txt
req_path: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\requirements.txt
ComfyUI-Impact-Pack: Check basic models
Install: pip packages for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager'
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
[SKIP] Downgrading pip package isn't allowed: huggingface-hub (cur=0.23.4)
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
Install: pip packages for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui_controlnet_aux'
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
[ERROR] Failed to execute install/(de)activation script: ['E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui_controlnet_aux', '#LAZY-INSTALL-SCRIPT', 'E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe']
/ invalid version number '4.10.0.84'
Install: pip packages for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\was-node-suite-comfyui'
[SKIP] Downgrading pip package isn't allowed: fairscale (cur=0.4.13)
Collecting git+https://github.com/WASasquatch/img2texture.git
Cloning https://github.com/WASasquatch/img2texture.git to c:\users\asyraf-gryft\appdata\local\temp\pip-req-build-12kb5k2r
[!] Running command git clone --filter=blob:none --quiet https://github.com/WASasquatch/img2texture.git 'C:\Users\ASYRAF-GRYFT\AppData\Local\Temp\pip-req-build-12kb5k2r'
Resolved https://github.com/WASasquatch/img2texture.git to commit a546609a915caf1dcb84a15bf44b5b639ad924e6
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
Collecting git+https://github.com/WASasquatch/cstr
Cloning https://github.com/WASasquatch/cstr to c:\users\asyraf-gryft\appdata\local\temp\pip-req-build-gxv72rzs
[!] Running command git clone --filter=blob:none --quiet https://github.com/WASasquatch/cstr 'C:\Users\ASYRAF-GRYFT\AppData\Local\Temp\pip-req-build-gxv72rzs'
Resolved https://github.com/WASasquatch/cstr to commit 0520c29a18a7a869a6e5983861d6f7a4c86f8e9b
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
[!] WARNING: opencv-python-headless 4.7.0.72 does not provide the extra 'ffmpeg'
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
Collecting git+https://github.com/WASasquatch/ffmpy.git
Cloning https://github.com/WASasquatch/ffmpy.git to c:\users\asyraf-gryft\appdata\local\temp\pip-req-build-z7r8fvb9
[!] Running command git clone --filter=blob:none --quiet https://github.com/WASasquatch/ffmpy.git 'C:\Users\ASYRAF-GRYFT\AppData\Local\Temp\pip-req-build-z7r8fvb9'
Resolved https://github.com/WASasquatch/ffmpy.git to commit f000737698b387ffaeab7cd871b0e9185811230d
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
[SKIP] Downgrading pip package isn't allowed: scikit-image (cur=0.24.0)
[SKIP] Downgrading pip package isn't allowed: timm (cur=1.0.7)
Install: pip packages for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-VideoHelperSuite'
Install: pip packages for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\rgthree-comfy'
Install: pip packages for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_essentials'
Collecting transparent-background
Downloading transparent_background-1.3.1-py3-none-any.whl.metadata (13 kB)
Collecting gdown>=4.5.4 (from transparent-background)
Downloading gdown-5.2.0-py3-none-any.whl.metadata (5.8 kB)
Collecting wget>=3.2 (from transparent-background)
Downloading wget-3.2.zip (10 kB)
Preparing metadata (setup.py): started
Preparing metadata (setup.py): finished with status 'done'
Collecting flet>=0.23.1 (from transparent-background)
Downloading flet-0.23.2-py3-none-win_amd64.whl.metadata (3.2 kB)
Collecting opencv-python-headless>=4.9.0.80 (from albumentations>=1.3.1->transparent-background)
Using cached opencv_python_headless-4.10.0.84-cp37-abi3-win_amd64.whl.metadata (20 kB)
Collecting cookiecutter<3.0.0,>=2.6.0 (from flet>=0.23.1->transparent-background)
Downloading cookiecutter-2.6.0-py3-none-any.whl.metadata (7.3 kB)
Collecting fastapi<1,>=0 (from flet>=0.23.1->transparent-background)
Downloading fastapi-0.111.1-py3-none-any.whl.metadata (26 kB)
Collecting flet-runtime==0.23.2 (from flet>=0.23.1->transparent-background)
Downloading flet_runtime-0.23.2-py3-none-any.whl.metadata (1.1 kB)
Collecting packaging<24.0,>=23.1 (from flet>=0.23.1->transparent-background)
Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)
Collecting uvicorn<1,>=0 (from uvicorn[standard]<1,>=0->flet>=0.23.1->transparent-background)
Downloading uvicorn-0.30.3-py3-none-any.whl.metadata (6.5 kB)
Collecting watchdog<5.0.0,>=4.0.0 (from flet>=0.23.1->transparent-background)
Downloading watchdog-4.0.1-py3-none-win_amd64.whl.metadata (37 kB)
Collecting flet-core==0.23.2 (from flet-runtime==0.23.2->flet>=0.23.1->transparent-background)
Downloading flet_core-0.23.2-py3-none-any.whl.metadata (971 bytes)
Collecting oauthlib<4.0.0,>=3.2.2 (from flet-runtime==0.23.2->flet>=0.23.1->transparent-background)
Downloading oauthlib-3.2.2-py3-none-any.whl.metadata (7.5 kB)
Collecting repath<0.10.0,>=0.9.0 (from flet-core==0.23.2->flet-runtime==0.23.2->flet>=0.23.1->transparent-background)
Downloading repath-0.9.0-py3-none-any.whl.metadata (899 bytes)
Collecting beautifulsoup4 (from gdown>=4.5.4->transparent-background)
Downloading beautifulsoup4-4.12.3-py3-none-any.whl.metadata (3.8 kB)
Collecting binaryornot>=0.4.4 (from cookiecutter<3.0.0,>=2.6.0->flet>=0.23.1->transparent-background)
Downloading binaryornot-0.4.4-py2.py3-none-any.whl.metadata (6.0 kB)
Collecting python-slugify>=4.0.0 (from cookiecutter<3.0.0,>=2.6.0->flet>=0.23.1->transparent-background)
Downloading python_slugify-8.0.4-py2.py3-none-any.whl.metadata (8.5 kB)
Collecting arrow (from cookiecutter<3.0.0,>=2.6.0->flet>=0.23.1->transparent-background)
Downloading arrow-1.3.0-py3-none-any.whl.metadata (7.5 kB)
Collecting starlette<0.38.0,>=0.37.2 (from fastapi<1,>=0->flet>=0.23.1->transparent-background)
Downloading starlette-0.37.2-py3-none-any.whl.metadata (5.9 kB)
Collecting fastapi-cli>=0.0.2 (from fastapi<1,>=0->flet>=0.23.1->transparent-background)
Downloading fastapi_cli-0.0.4-py3-none-any.whl.metadata (7.0 kB)
Collecting python-multipart>=0.0.7 (from fastapi<1,>=0->flet>=0.23.1->transparent-background)
Downloading python_multipart-0.0.9-py3-none-any.whl.metadata (2.5 kB)
Collecting email_validator>=2.0.0 (from fastapi<1,>=0->flet>=0.23.1->transparent-background)
Downloading email_validator-2.2.0-py3-none-any.whl.metadata (25 kB)
Collecting httptools>=0.5.0 (from uvicorn[standard]<1,>=0->flet>=0.23.1->transparent-background)
Downloading httptools-0.6.1-cp311-cp311-win_amd64.whl.metadata (3.7 kB)
Collecting python-dotenv>=0.13 (from uvicorn[standard]<1,>=0->flet>=0.23.1->transparent-background)
Downloading python_dotenv-1.0.1-py3-none-any.whl.metadata (23 kB)
Collecting watchfiles>=0.13 (from uvicorn[standard]<1,>=0->flet>=0.23.1->transparent-background)
Downloading watchfiles-0.22.0-cp311-none-win_amd64.whl.metadata (5.0 kB)
Collecting websockets>=10.4 (from uvicorn[standard]<1,>=0->flet>=0.23.1->transparent-background)
Downloading websockets-12.0-cp311-cp311-win_amd64.whl.metadata (6.8 kB)
Collecting soupsieve>1.2 (from beautifulsoup4->gdown>=4.5.4->transparent-background)
Downloading soupsieve-2.5-py3-none-any.whl.metadata (4.7 kB)
Collecting PySocks!=1.5.7,>=1.5.6 (from requests[socks]->gdown>=4.5.4->transparent-background)
Downloading PySocks-1.7.1-py3-none-any.whl.metadata (13 kB)
Collecting dnspython>=2.0.0 (from email_validator>=2.0.0->fastapi<1,>=0->flet>=0.23.1->transparent-background)
Downloading dnspython-2.6.1-py3-none-any.whl.metadata (5.8 kB)
Collecting text-unidecode>=1.3 (from python-slugify>=4.0.0->cookiecutter<3.0.0,>=2.6.0->flet>=0.23.1->transparent-background)
Downloading text_unidecode-1.3-py2.py3-none-any.whl.metadata (2.4 kB)
Collecting types-python-dateutil>=2.8.10 (from arrow->cookiecutter<3.0.0,>=2.6.0->flet>=0.23.1->transparent-background)
Downloading types_python_dateutil-2.9.0.20240316-py3-none-any.whl.metadata (1.8 kB)
Downloading transparent_background-1.3.1-py3-none-any.whl (29 kB)
Downloading flet-0.23.2-py3-none-win_amd64.whl (42.7 MB)
---------------------------------------- 42.7/42.7 MB 18.7 MB/s eta 0:00:00
Downloading flet_runtime-0.23.2-py3-none-any.whl (19 kB)
Downloading flet_core-0.23.2-py3-none-any.whl (399 kB)
--------------------------------------- 399.2/399.2 kB 24.3 MB/s eta 0:00:00
Downloading gdown-5.2.0-py3-none-any.whl (18 kB)
Downloading cookiecutter-2.6.0-py3-none-any.whl (39 kB)
Downloading fastapi-0.111.1-py3-none-any.whl (92 kB)
---------------------------------------- 92.2/92.2 kB ? eta 0:00:00
Using cached opencv_python_headless-4.10.0.84-cp37-abi3-win_amd64.whl (38.8 MB)
Downloading packaging-23.2-py3-none-any.whl (53 kB)
---------------------------------------- 53.0/53.0 kB ? eta 0:00:00
Downloading uvicorn-0.30.3-py3-none-any.whl (62 kB)
---------------------------------------- 62.8/62.8 kB ? eta 0:00:00
Downloading watchdog-4.0.1-py3-none-win_amd64.whl (83 kB)
---------------------------------------- 83.0/83.0 kB 4.5 MB/s eta 0:00:00
Downloading beautifulsoup4-4.12.3-py3-none-any.whl (147 kB)
---------------------------------------- 147.9/147.9 kB 9.2 MB/s eta 0:00:00
Downloading binaryornot-0.4.4-py2.py3-none-any.whl (9.0 kB)
Downloading email_validator-2.2.0-py3-none-any.whl (33 kB)
Downloading fastapi_cli-0.0.4-py3-none-any.whl (9.5 kB)
Downloading httptools-0.6.1-cp311-cp311-win_amd64.whl (55 kB)
---------------------------------------- 55.4/55.4 kB ? eta 0:00:00
Downloading oauthlib-3.2.2-py3-none-any.whl (151 kB)
---------------------------------------- 151.7/151.7 kB 9.4 MB/s eta 0:00:00
Downloading PySocks-1.7.1-py3-none-any.whl (16 kB)
Downloading python_dotenv-1.0.1-py3-none-any.whl (19 kB)
Downloading python_multipart-0.0.9-py3-none-any.whl (22 kB)
Downloading python_slugify-8.0.4-py2.py3-none-any.whl (10 kB)
Downloading soupsieve-2.5-py3-none-any.whl (36 kB)
Downloading starlette-0.37.2-py3-none-any.whl (71 kB)
---------------------------------------- 71.9/71.9 kB ? eta 0:00:00
Downloading watchfiles-0.22.0-cp311-none-win_amd64.whl (281 kB)
---------------------------------------- 282.0/282.0 kB ? eta 0:00:00
Downloading websockets-12.0-cp311-cp311-win_amd64.whl (124 kB)
---------------------------------------- 125.0/125.0 kB ? eta 0:00:00
Downloading arrow-1.3.0-py3-none-any.whl (66 kB)
---------------------------------------- 66.4/66.4 kB ? eta 0:00:00
Downloading dnspython-2.6.1-py3-none-any.whl (307 kB)
--------------------------------------- 307.7/307.7 kB 18.6 MB/s eta 0:00:00
Downloading repath-0.9.0-py3-none-any.whl (4.7 kB)
Downloading text_unidecode-1.3-py2.py3-none-any.whl (78 kB)
---------------------------------------- 78.2/78.2 kB 4.2 MB/s eta 0:00:00
Downloading types_python_dateutil-2.9.0.20240316-py3-none-any.whl (9.7 kB)
Building wheels for collected packages: wget
Building wheel for wget (setup.py): started
Building wheel for wget (setup.py): finished with status 'done'
Created wheel for wget: filename=wget-3.2-py3-none-any.whl size=9678 sha256=ee0f176491b0345d4343853c338aa9d957b95e87ac66aaa450500b6f550c3b68
Stored in directory: c:\users\asyraf-gryft\appdata\local\pip\cache\wheels\40\b3\0f\a40dbd1c6861731779f62cc4babcb234387e11d697df70ee97
Successfully built wget
Installing collected packages: wget, text-unidecode, websockets, watchdog, types-python-dateutil, soupsieve, repath, python-slugify, python-multipart, python-dotenv, PySocks, packaging, opencv-python-headless, oauthlib, httptools, dnspython, binaryornot, watchfiles, uvicorn, starlette, flet-core, email_validator, beautifulsoup4, arrow, gdown, flet-runtime, cookiecutter, fastapi-cli, fastapi, flet, transparent-background
[!] WARNING: The script watchmedo.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
[!] WARNING: The script slugify.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
[!] WARNING: The script dotenv.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
Attempting uninstall: packaging
Found existing installation: packaging 24.1
Uninstalling packaging-24.1:
Successfully uninstalled packaging-24.1
Attempting uninstall: opencv-python-headless
Found existing installation: opencv-python-headless 4.7.0.72
Uninstalling opencv-python-headless-4.7.0.72:
Successfully uninstalled opencv-python-headless-4.7.0.72
[!] WARNING: The script watchfiles.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
[!] WARNING: The script uvicorn.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
[!] WARNING: The script email_validator.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
[!] WARNING: The script gdown.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
[!] WARNING: The script cookiecutter.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
[!] WARNING: The script fastapi.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
[!] WARNING: The script flet.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
[!] WARNING: The scripts transparent-background-gui.exe and transparent-background.exe are installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
Successfully installed PySocks-1.7.1 arrow-1.3.0 beautifulsoup4-4.12.3 binaryornot-0.4.4 cookiecutter-2.6.0 dnspython-2.6.1 email_validator-2.2.0 fastapi-0.111.1 fastapi-cli-0.0.4 flet-0.23.2 flet-core-0.23.2 flet-runtime-0.23.2 gdown-5.2.0 httptools-0.6.1 oauthlib-3.2.2 opencv-python-headless-4.10.0.84 packaging-23.2 python-dotenv-1.0.1 python-multipart-0.0.9 python-slugify-8.0.4 repath-0.9.0 soupsieve-2.5 starlette-0.37.2 text-unidecode-1.3 transparent-background-1.3.1 types-python-dateutil-2.9.0.20240316 uvicorn-0.30.3 watchdog-4.0.1 watchfiles-0.22.0 websockets-12.0 wget-3.2
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
Install: pip packages for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Image-Saver'
[SKIP] Downgrading pip package isn't allowed: transformers (cur=4.43.3)
[SKIP] Downgrading pip package isn't allowed: tokenizers (cur=0.19.1)
[SKIP] Downgrading pip package isn't allowed: safetensors (cur=0.4.3)
ComfyUI-Manager: EXECUTE => ['E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe', '-m', 'pip', 'install', 'pyyaml']
Execute install/(de)activation script for 'E:\Ai\ComfyUI_windows_portable\ComfyUI'
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
ComfyUI-Manager: EXECUTE => ['E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe', '-m', 'pip', 'install', 'Pillow']
Execute install/(de)activation script for 'E:\Ai\ComfyUI_windows_portable\ComfyUI'
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
[SKIP] Downgrading pip package isn't allowed: kornia (cur=0.7.3)
Install: pip packages for 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-KJNodes'
Collecting pillow>=10.3.0
Using cached pillow-10.4.0-cp311-cp311-win_amd64.whl.metadata (9.3 kB)
Using cached pillow-10.4.0-cp311-cp311-win_amd64.whl (2.6 MB)
Installing collected packages: pillow
Attempting uninstall: pillow
Found existing installation: Pillow 9.5.0
Uninstalling Pillow-9.5.0:
Successfully uninstalled Pillow-9.5.0
[!] ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
Successfully installed pillow-10.4.0
[!] img2texture 1.0.6 requires pillow<10,>=9.2, but you have pillow 10.4.0 which is incompatible.
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
Collecting mss
Downloading mss-9.0.1-py3-none-any.whl.metadata (3.9 kB)
Downloading mss-9.0.1-py3-none-any.whl (22 kB)
Installing collected packages: mss
[!] WARNING: The script mss.exe is installed in 'E:\Ai\ComfyUI_windows_portable\python_embeded\Scripts' which is not on PATH.
[!] Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.
Successfully installed mss-9.0.1
[!]
[!] [notice] A new release of pip is available: 24.1.1 -> 24.1.2
[!] [notice] To update, run: E:\Ai\ComfyUI_windows_portable\python_embeded\python.exe -m pip install --upgrade pip
[ComfyUI-Manager] Startup script completed.
#######################################################################
Prestartup times for custom nodes:
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\rgthree-comfy
58.8 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager
Total VRAM 8192 MB, total RAM 32555 MB
pytorch version: 2.4.0+cu121
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 2070 : cudaMallocAsync
Using pytorch cross attention
[Prompt Server] web root: E:\Ai\ComfyUI_windows_portable\ComfyUI\web
E:\Ai\ComfyUI_windows_portable\python_embeded\Lib\site-packages\kornia\feature\lightglue.py:44: FutureWarning: torch.cuda.amp.custom_fwd(args...) is deprecated. Please use torch.amp.custom_fwd(args..., device_type='cuda') instead.
@torch.cuda.amp.custom_fwd(cast_inputs=torch.float32)
Error:
[WinError 1314] A required privilege is not held by the client: 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyLiterals\js' -> 'E:\Ai\ComfyUI_windows_portable\ComfyUI\web\extensions\ComfyLiterals'
Failed to create symlink to E:\Ai\ComfyUI_windows_portable\ComfyUI\web\extensions\ComfyLiterals. Please copy the folder manually.
Source: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyLiterals\js
Target: E:\Ai\ComfyUI_windows_portable\ComfyUI\web\extensions\ComfyLiterals
Adding E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes to sys.path
Efficiency Nodes: Attempting to add Control Net options to the 'HiRes-Fix Script' Node (comfyui_controlnet_aux add-on)...Success!
Loaded Efficiency nodes from E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\efficiency-nodes-comfyui
Loaded ControlNetPreprocessors nodes from E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui_controlnet_aux
Could not find AdvancedControlNet nodes
Could not find AnimateDiff nodes
Loaded IPAdapter nodes from E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_IPAdapter_plus
Loaded VideoHelperSuite from E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-VideoHelperSuite
Loading: ComfyUI-Impact-Pack (V5.18.16)
Loading: ComfyUI-Impact-Pack (Subpack: V0.6)
Loading: ComfyUI-Impact-Pack (V5.18.16)
[Impact Pack] Wildcards loading done.Loaded ImpactPack nodes from
E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack
[Impact Pack] Wildcards loading done.
[Crystools INFO] Crystools version: 1.15.1
Traceback (most recent call last):
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1931, in load_custom_node
module_spec.loader.exec_module(module)
File "", line 940, in exec_module
File "", line 241, in call_with_frames_removed
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools_init.py", line 24, in
from .server import *
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\server_init_.py", line 1, in
from .monitor import *
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\server\monitor.py", line 4, in
from ..general import cmonitor
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\general_init_.py", line 1, in
from .monitor import *
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\general\monitor.py", line 5, in
from .hardware import CHardwareInfo
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\general\hardware.py", line 4, in
from .gpu import CGPUInfo
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\general\gpu.py", line 6, in
from pyrsmi import rocml
ModuleNotFoundError: No module named 'pyrsmi'
Cannot import E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools module for custom nodes: No module named 'pyrsmi'
Loading: ComfyUI-Impact-Pack (V5.18.16)
[Impact Pack] Wildcards loading done.
Loading: ComfyUI-Inspire-Pack (V0.82.6)
Total VRAM 8192 MB, total RAM 32555 MB
pytorch version: 2.4.0+cu121
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 2070 : cudaMallocAsync
Loading: ComfyUI-Manager (V2.48.2)
ComfyUI Revision: 2425 [c24f8973] | Released on '2024-07-31'
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json
Failed to auto update Quality of Life Suit
QualityOfLifeSuit_Omar92_DIR: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-QualityOfLifeSuit_Omar92
A new version of Albumentations is available: 1.4.12 (you have 1.4.11). Upgrade using: pip install --upgrade albumentations
[ReActor] - STATUS - Running v0.5.1-a6 in ComfyUI
Torch version: 2.4.0+cu121
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
(pysssss:WD14Tagger) [DEBUG] Available ORT providers: TensorrtExecutionProvider, CUDAExecutionProvider, CPUExecutionProvider
(pysssss:WD14Tagger) [DEBUG] Using ORT providers: CUDAExecutionProvider, CPUExecutionProvider
Workspace manager - Openning file hash dict
🦄🦄Loading: Workspace Manager (V2.1.0)
3D OpenPose Editor: Loaded
Comfyroll Studio v1.76 : 175 Nodes Loaded
** For changes, please see patch notes at https://github.com/Suzie1/ComfyUI_Comfyroll_CustomNodes/blob/main/Patch_Notes.md
** For help, please see the wiki at https://github.com/Suzie1/ComfyUI_Comfyroll_CustomNodes/wiki
[tinyterraNodes] Loaded
[rgthree] Loaded 40 exciting nodes.
[rgthree] Will use rgthree's optimized recursive execution.
Traceback (most recent call last):
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1931, in load_custom_node
module_spec.loader.exec_module(module)
File "", line 936, in exec_module
File "", line 1073, in get_code
File "", line 1130, in get_data
FileNotFoundError: [Errno 2] No such file or directory: 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\sd-webui-stablesr\init.py'
Cannot import E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\sd-webui-stablesr module for custom nodes: [Errno 2] No such file or directory: 'E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\sd-webui-stablesr\init.py'
WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite Warning: ffmpeg_bin_path is not set in E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\was-node-suite-comfyui\was_suite_config.json config file. Will attempt to use system ffmpeg binaries if available.
WAS Node Suite: Finished. Loaded 217 nodes successfully.
"Opportunities don't happen. You create them." - Chris Grosser
Import times for custom nodes:
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools-save
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\websocket_image_save.py
0.0 seconds (IMPORT FAILED): E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\sd-webui-stablesr
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\sdxl_prompt_styler
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-VideoHelperSuite
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_IPAdapter_plus
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_3dPoseEditor
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Eagle-PNGInfo
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\efficiency-nodes-comfyui
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\stability-ComfyUI-nodes
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-WD14-Tagger
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\wlsh_nodes
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui_controlnet_aux
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyLiterals
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-QualityOfLifeSuit_Omar92
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfy-image-saver
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Custom-Scripts
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_UltimateSDUpscale
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Impact-Pack
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\Derfuu_ComfyUI_ModdedNodes
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Image-Saver
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\rgthree-comfy
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-MimicMotionWrapper
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-workspace-manager
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_Comfyroll_CustomNodes
0.0 seconds (IMPORT FAILED): E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_essentials
0.0 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_tinyterraNodes
0.1 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-KJNodes
0.2 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Inspire-Pack
0.2 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-moondream
0.2 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-KwaiKolorsWrapper
0.2 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Gemini
0.3 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager
0.6 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-reactor-node
0.6 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-SUPIR
0.9 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfyui-art-venture
1.6 seconds: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\was-node-suite-comfyui
Starting server
To see the GUI go to: http://127.0.0.1:8188
FETCH DATA from: E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager\extension-node-map.json [DONE]
HTTP Request: GET https://api.openai.com/v1/models "HTTP/1.1 401 Unauthorized"
Error: OpenAI API key is invalid OpenAI features wont work for you
QualityOfLifeSuit_Omar92::NSP ready
[Crystools INFO] Crystools version: 1.15.1
Traceback (most recent call last):
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\nodes.py", line 1931, in load_custom_node
module_spec.loader.exec_module(module)
File "", line 940, in exec_module
File "", line 241, in call_with_frames_removed
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools_init.py", line 24, in
from .server import *
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\server_init_.py", line 1, in
from .monitor import *
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\server\monitor.py", line 4, in
from ..general import cmonitor
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\general_init_.py", line 1, in
from .monitor import *
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\general\monitor.py", line 5, in
from .hardware import CHardwareInfo
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\general\hardware.py", line 4, in
from .gpu import CGPUInfo
File "E:\Ai\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools\general\gpu.py", line 6, in
from pyrsmi import rocml
ModuleNotFoundError: No module named 'pyrsmi'
Crystools Updated! Please try to updated (should be 1.16.0)
thank you sir much appreciated
can someone help me? https://github.com/crystian/ComfyUI-Crystools/issues/115
|
gharchive/issue
| 2024-07-31T07:02:46 |
2025-04-01T04:33:54.203468
|
{
"authors": [
"VantaHitam",
"crystian",
"rachelcenter"
],
"repo": "crystian/ComfyUI-Crystools",
"url": "https://github.com/crystian/ComfyUI-Crystools/issues/93",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
817032161
|
[tests-only] Bump commit id for API tests
The way the skeleton files happen in core acceptance tests got a refactoring. So bump the commit id here to use that. And it passes, which is a bonus!
Bumped again to include some changes to the core acceptance tests run.sh script.
@ishank011 @labkode ready for review.
CI failed after the last commit. I am not sure how this got merged. I will make another PR to fix it.
|
gharchive/pull-request
| 2021-02-26T04:36:30 |
2025-04-01T04:33:54.236381
|
{
"authors": [
"phil-davis"
],
"repo": "cs3org/reva",
"url": "https://github.com/cs3org/reva/pull/1500",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1255843668
|
[tests-only] [full-ci] Test edge against core branch issue-40093-squashed
Related core PR https://github.com/owncloud/core/pull/40114
https://drone.cernbox.cern.ch/cs3org/reva/7308/24/7
runsh: Total unexpected failed scenarios throughout the test run:
apiShareReshareToShares3/reShareUpdate.feature:152
apiShareReshareToShares3/reShareUpdate.feature:153
https://drone.cernbox.cern.ch/cs3org/reva/7308/15/6
runsh: Total unexpected failed scenarios throughout the test run:
apiShareReshareToShares3/reShareUpdate.feature:152
apiShareReshareToShares3/reShareUpdate.feature:153
That test scenario was introduced in core PR https://github.com/owncloud/core/pull/38652 and demonstrates some existing core behavior. I can look into that separately, because it might be a test that we should not expect to behave the same on reva.
|
gharchive/pull-request
| 2022-06-01T12:15:52 |
2025-04-01T04:33:54.238946
|
{
"authors": [
"phil-davis"
],
"repo": "cs3org/reva",
"url": "https://github.com/cs3org/reva/pull/2912",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1190682200
|
Create Settings Modal
Need modal for setting menu
Settings modal not going to work with check boxes, scrapping the modal just keeping the check boxes
|
gharchive/issue
| 2022-04-02T16:48:57 |
2025-04-01T04:33:54.244855
|
{
"authors": [
"snakeround"
],
"repo": "cs481-ekh/s22-hospsim",
"url": "https://github.com/cs481-ekh/s22-hospsim/issues/95",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
69541600
|
Move to Silex 1.2
Hi Charles,
is there any chance to remove the dependency on Silex 1.1 and move onto Silex "dev-master"?
Tnx!
Hi Cristian,
Sorry I was very busy on subjects other than PHP these days.
I 'm going back to work on this library very soon.
But if you have time, create a PR and i will check it.
Thx.
Composer configuration updated.
Works fine!
Many thanks!
|
gharchive/issue
| 2015-04-20T09:04:58 |
2025-04-01T04:33:54.259006
|
{
"authors": [
"cristianp6",
"csanquer"
],
"repo": "csanquer/PdoServiceProvider",
"url": "https://github.com/csanquer/PdoServiceProvider/issues/4",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2341905654
|
Task: split category and status
Fixes the bug where marking a task as completed in the overview page removes it from the tasks page
Description
Create a new category property in stores/task.js
replace all uses of task.status with task.category in components/task.js
Related Issue
How Has This Been Tested?
I looked at it with my eyes
Checklist
[ ] All automated tests have been passed locally (unit / e2e / linter / coverage)
[ ] New unit and E2E tests have been added to address the changes made
[ ] Manual testing has been performed according to the provided instructions
[ ] Code follows team's coding style, convention and standards
[ ] Documentation has been updated to reflect relevant changes (if applicable)
Screenshots (if applicable):
Addressed in another PR so we won't proceed with this implementation
|
gharchive/pull-request
| 2024-06-08T23:17:56 |
2025-04-01T04:33:54.331223
|
{
"authors": [
"Ant044",
"nadecancode"
],
"repo": "cse110-sp24-group35/journal",
"url": "https://github.com/cse110-sp24-group35/journal/pull/60",
"license": "WTFPL",
"license_type": "permissive",
"license_source": "github-api"
}
|
2606301710
|
bug: blockquote styles not applied to LLM message markdown properly
This is caused by this known issue, open since April of 2021: https://github.com/flutter/flutter/issues/81720
There seems to be an PR to address this issue that wasn't merged: https://github.com/flutter/packages/pull/4396
@domesticmouse would you be willing to take a look at this? you were a reviewer on the PR. it's blocking proper formatting for the LLM messages of the Flutter AI Toolkit, which uses a MarkdownBody to show rich text output. Thanks!
My commitment to this package is to make sure changes have test coverage. I don't actually know the code base. @Gspencergoog may have thoughts on how to fix this.
Thanks to both of you. The bug in question has a suggested fix.
There seems to be an PR to address this issue that wasn't merged: flutter/packages#4396
There is no content in this PR.
The bug in question has a suggested fix.
I've put together https://github.com/flutter/packages/pull/7927 with this fix applied, but the current tests don't pass. Can you debug? And add required tests for your use case?
Thanks, @domesticmouse. I'll take a look ASAP.
more info: https://github.com/flutter/packages/pull/8163/checks?check_run_id=33450035628
|
gharchive/issue
| 2024-10-22T19:29:54 |
2025-04-01T04:33:54.337522
|
{
"authors": [
"csells",
"domesticmouse"
],
"repo": "csells/flutter_ai_toolkit",
"url": "https://github.com/csells/flutter_ai_toolkit/issues/58",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
382102789
|
Duplicate monospace in font-family
https://github.com/csstools/normalize.css/blob/946efd2cde37c939cbdb4bb541773e6527ee8851/normalize.css#L59
Is this some trick or just wrong?
It’s required by most browsers to size fonts correctly. https://stackoverflow.com/questions/38781089/font-family-monospace-monospace
Ah, thanks for the info =)
|
gharchive/issue
| 2018-11-19T08:22:48 |
2025-04-01T04:33:54.397012
|
{
"authors": [
"DanielRuf",
"jonathantneal"
],
"repo": "csstools/normalize.css",
"url": "https://github.com/csstools/normalize.css/issues/5",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2752222935
|
Output did not contain fingerprint if smtp_tls_fingerprint_digest is not set
Starting with postfix 3.6 the default digest setting is changed. Default is now:
postconf -h -p smtp_tls_fingerprint_digest
${{$compatibility_level} <level {3.6} ? {md5} : {sha256}}
This is not working with the script. After setting the smtp_tls_fingerprint_digest = sha256 the main.cf the output is sha256 and working.
Can you please elaborate exactly what the issue is?
|
gharchive/issue
| 2024-12-20T08:35:58 |
2025-04-01T04:33:54.401513
|
{
"authors": [
"csware",
"jnandreae"
],
"repo": "csware/postfix-tls-policy",
"url": "https://github.com/csware/postfix-tls-policy/issues/12",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
419184378
|
Tried flashing (flash3) but didn't work, tried undoing, now device unresponsive - Tonbux switch
I'm kinda freaking out. I successfully flashed the middleware firmware, got this info:
READ FLASH: http://10.42.42.42/get?read=HEX-ADDRESS
ChipID: 258c75
MAC: 3C:71:BF:25:8C:75
SdkVersion: 1.5.3(aec24ac9)
BootVersion: 4
BootMode: 1
FlashMode: 1M QIO @ 40MHz
FlashChipId: 1440c8
FlashChipRealSize: 1048576
FlashChipSize: 1048576
FlashChipMode: 0
system_upgrade_userbin_check: user2 0x81000
Then I did curl http://10.42.42.42/flash3. It said Device should flash http://10.42.42.1/files/thirdparty.bin and restart. But the device didn't restart, it just stayed connected to the vtrust-flash and the "sonoff-xx" AP never came up.
I waited an hour, and then did curl http://10.42.42.42/undo to try again. It said it would undo and restart, and the device disconnected from vtrust-flash but never turned back on. Now I can't get it to turn on at all.
system_upgrade_userbin_check: user2 0x81000
That tells you to flash "flash2" .....this put the flash loader in the correct poistion.
Then flash "flash3".
When it restarts you wont see any lights. You will see sonoffxxx appear in your wifi list to connect to.
Then you will configure tasmota.
Hmm, that wasn’t clear from the documentation. Is there any way I can get the switch to get un-bricked? Is it even bricked?
I ran into the same issue with my Heygo H02, I simply restarted the process from the beginning and the script recognized the plug instantly and I could issue flash2 and then flash3. Similiar worked when the firmware backup failed, restarted the script and could continue.
@AlphaDE How did you get it to go into pairing mode again? I tried to start the process again, but my switch has no lights or clicks or any reaction at all anymore when I hold down the button like before, and it doesn't appear to be sending any probe requests or any sort of packets (judging from Wireshark). No signs of life :(
@melikyuksel
I did not had to get it manually into pairing mode, the device reacted when I just restarted the script and on another (where I assumed that I bricked it, because it did not react in any way anymore, no button, no led) it also connect and resumed after plugging it in.
The stub installed by the script obviously stays up/comes up without manual intervention.
@AlphaDE I followed those steps, but the device isn't responding. Gave up after over an hour. So it looks the undo operation screwed it up in a big way where it doesn't even boot anymore. I don't know what to do to salvage it
The undo removes the stub, so if not told otherwise there is nothing to boot anymore. If you can open it, you could try to flash a firmware via the serial port of that chip. But that's more effort (and needs a USB-serial-convert and more to do it).
I wasn't able to resuscitate the first switch OTA, so I'll try the soldering for it. I have no idea how to do it, but will look at guides. Perhaps I can use a Pi Zero in gadget mode as the USB serial device.
However, I tried the OTA process again on another one of this switch, and I got it to successfully flash! This time, this switch kept timing out while backing up the factory firmware, but the start_flash.sh would just keep going.
Traceback (most recent call last):
File "./backup.py", line 42, in <module>
main()
File "./backup.py", line 27, in main
conn.request("GET", "/get?read=%X" % address )
File "/usr/lib/python3.5/http/client.py", line 1107, in request
self._send_request(method, url, body, headers)
File "/usr/lib/python3.5/http/client.py", line 1152, in _send_request
self.endheaders(body)
File "/usr/lib/python3.5/http/client.py", line 1103, in endheaders
self._send_output(message_body)
File "/usr/lib/python3.5/http/client.py", line 934, in _send_output
self.send(msg)
File "/usr/lib/python3.5/http/client.py", line 877, in send
self.connect()
File "/usr/lib/python3.5/http/client.py", line 849, in connect
(self.host,self.port), self.timeout, self.source_address)
File "/usr/lib/python3.5/socket.py", line 712, in create_connection
raise err
File "/usr/lib/python3.5/socket.py", line 703, in create_connection
sock.connect(sa)
OSError: [Errno 113] No route to host
~/tuya-convert
I tried it over and over like 3 times, and on the 3rd try, it worked to back up the firmware without errors. When I did curl http://10.42.42.42/flash2 as recommended, it said Device is allready booting from userspace 2 (0x81000). So that wasn't the problem with the first one...
I did curl http://10.42.42.42/flash3, and got the Device should flash http://10.42.42.1/files/thirdparty.bin and restart message. I crossed my fingers. And my switch started blinking green! An improvement. Then the sonoff-xxxx access point came up, and I was able to configure the switch successfully!
Going to solder the first one. The second, successfully flashed switch is now blinking weird colors (blinking green/nothing when off, blinking red/yellow when on). How does one go about "fixing" this?
Gonna pass you this vid to watch so it makes clear how to go about configing Tasmota....figuring out the pins.
https://www.youtube.com/watch?v=m_O24tTzv8g
@onebyside @AlphaDE -- Got the first one flashed as well using the serial ports with the Raspberry Pi Zero. Nice! Both have Tasmota now. I realized the first failure never managed to backup the factory firmware. Perhaps some sort of Wi-Fi connection issue with the vtrust-flash AP on the Pi Zero the first time around.
@melikyuksel Do you have any firmware backups of problematic devices? I am tracking down a bug and a copy of firmware where the flash did not go as expected would be very helpful for narrowing this down.
Where did the script store the backup? I run into these problems (timout while saving the original firmware, but resuming was possible) with my HEYGO H02 US plugs and on my Raspberry Pi3B that backup should exist.
The backups are stored in the tuya-convert directory by default, with a timestamp and _readout.bin
I have nothing such there...
Too bad, thank you for checking anyway!
@kueblc I do have firmware from the exact same devices as the one that initially failed, but I’ve flashed 4 of them so far, and only the first one failed. I feel like there was some sort of corruption of the sent firmware files.
The reason I think this is is because I used the Raspberry Pi Zero W’s built-in WiFi to try to flash the first device. There were many disconnects. I then switched to an external USB WiFi adapter I use for penetration testing (monitor mode capable), and the process worked well for the rest of the devices. However, on one of the later devices even with this new adapter, the device was trying to connect to “vtrt-t-flash” instead of the usual “vtrust-flash” (checked with a packet sniffer), so I had to reflash it. So obviously some sort of glitch.
So there appears to be a noticeable potential for data corruption between the script and the ESP device during the OTA flashing.
There was a bug in the intermediate firmware that caused undo to fail if you were already booting from userspace 1. I believed this to be related to the problems you reported, but the bug is now fixed in the current master. I have found identical devices from the same box that boot from different userspaces so it's entirely possible that you just happened to undo on one of the devices that booted from 1.
I have not gotten any other reports of corruption but if you have any data I would be happy to look into it.
|
gharchive/issue
| 2019-03-10T13:13:25 |
2025-04-01T04:33:54.414463
|
{
"authors": [
"AlphaDE",
"kueblc",
"melikyuksel",
"onebyside"
],
"repo": "ct-Open-Source/tuya-convert",
"url": "https://github.com/ct-Open-Source/tuya-convert/issues/127",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2655192379
|
Add nice repr s to the reference metadata objects
Please describe the use case that requires this feature.
The Reference metadata objects currently do not have reprs:
(Pdb) str(inputs[0]["reference_meta"].product)
'<ctapipe.io.metadata.Product object at 0x7f82685366d0>'
Describe the solution you'd like
Add reprs
Some have reprs, but it is not consistent. Could just use the _to_dict() helper function in that directory as a general way.
|
gharchive/issue
| 2024-11-13T11:34:17 |
2025-04-01T04:33:54.416821
|
{
"authors": [
"kosack",
"maxnoe"
],
"repo": "cta-observatory/ctapipe",
"url": "https://github.com/cta-observatory/ctapipe/issues/2649",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
645983902
|
Onnx shower reconstruction
Deep Learning Shower Reconstructor (#1274)
This pull request adds a new reconstructor that allows one or more ONNX models to be used for a shower reconstruction. This reconstructor is an abstract class, that's because different algorithms will use different methods to combine the results from the different models and different inputs.
New dependency
onnxruntime: To load and run inferencing sessions on ONNX models.
New classes
ONNXModel: Loads a model from a given path to make predictions using the predict method, that allows multiple inputs to be used, given ordered or named arguments. The predict method returns a list with the model outputs.
DeepLearningReconstructor(reco.Reconstructor): An abstract Reconstructor class that loads a dictionary of models paths, one model for each kind of camera (FlashCam, LSTCam, ASTRICam, etc.). Then makes a prediction with the corresponding model for each observation made from the event. Once all the predictions have been made, the reconstructor will take all the outputs and merge them into one ReconstructedShowerContainer.
Abstract methods
The abstract methods and properties defined for the DeepLearningReconstructor are:
supported_cameras (property): Supported camera types for this algorithm (FlashCam, LSTCam, ASTRICam, etc.).
_to_input(self, event, tel_id, cam_name): Transforms an event observation (given by the tel_id) into the input that the model will use.
_reconstruct(self, models_outputs): Receives all the predictions from the different telescopes and uses them to create a new ReconstructedShowerContainer with the reconstructed data. This value will be returned by the predict() method.
Tests
For testing purposes I added in the reco tests directory another directory with three models, each one for different camera types (FlashCam, ASTRICam and LSTCam) with two inputs and outputs to test which are used in the DeepLearningReconstructor's unit tests.
@capaths @maxnoe Is this PR still in progress and active? Can I help finalising/testing it? I have onnx models (mono and/or stereo full event reconstruction) at hand that give reasonable performances.
@TjarkMiener we are currently finalizing the Initial version of the classical machine learning module.
I think we should probably do a work meeting discussing requirements and approaches for including deep learning application into ctapipe.
We should probably introduce a plugin system like for the eventsources also for reconstructions at several points in the pipeline for different kinds of models (predicting from waveforms, from dli images plus peak times, models using all telescopes at once, mono predictions that are later combined, ....).
There are a lot of possibilities and we should discuss what to support in a first iteration.
I am mostly for having an initial implementation to discuss but in this case I think we should devote some time to planning how to best integrate and what to support directly in ctapipe and what better lives in plugins.
Shall we close this now that it is possible to have Reconstructor plugins that are external to ctapipe?
Yes, closing this. We discussed in the datapipe meeting that we are doing this for now via the plugin system.
Anyone wanting to contribute and facing issues with the current reconstructor API, please join the discussion in #2291
|
gharchive/pull-request
| 2020-06-26T03:38:44 |
2025-04-01T04:33:54.424874
|
{
"authors": [
"TjarkMiener",
"capaths",
"kosack",
"maxnoe"
],
"repo": "cta-observatory/ctapipe",
"url": "https://github.com/cta-observatory/ctapipe/pull/1371",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
350672124
|
Apollo配置中心页面搜索Chrome浏览器不跳转
在配置中心页面搜索某个字段,微软Edge、IE浏览器可以跳转到关键字的位置,Chrome浏览器不行。
Apollo版本:Apollo V0.6.3 Release,经测试最新release也不跳转。Chrome版本:Version 68.0.3440.106 (Official Build) (64-bit),其他Chrome版本也同样的问题。
最新版本 1.1.1-SNAPSHOT 也是。谷歌浏览器,火狐浏览器,都不能Ctrl+F 跟随,当配置项很多很多的时候,需要手动滚动一点点找,正常情况下是搜索框回车浏览器自动滚动到目标位置。
我们也遇到了这个问题,查了一下是因为用了 jquery.nicescroll 导致 html 标签被加了 overflow:hidden 样式。
搜索了一下似乎各个浏览器现在更改了行为,在ctrl + f 时不再滚动到 overflow:hidden 的元素里面,由于ctrl + f 不会触发浏览器事件,也就是说可能大部分禁止原生滚动条,自定义滚动条的插件都有这个问题。
我不是前端开发,通过我搜到的资料来看,似乎这个问题只能强制修改使用了 nicescroll 滚动标签的元素,把 overflow:hidden 去掉,但是这样可能破坏 nicescroll 的行为,比如功能失效或者出现两个滚动条。还有一种办法就是把 nicescroll 去掉,我看 apollo 使用这个组件似乎也没有特别的目的。
我将NiceScroll换成最新3.7.6 版本,问题依旧。 https://github.com/inuyaksa/jquery.nicescroll/issues/766
正在尝试解决办法。
我将NiceScroll换成最新3.7.6 版本,问题依旧。 inuyaksa/jquery.nicescroll#766
正在尝试解决办法。
不管用,nicescroll 官网都有这问题,浏览器改了默认行为以后这问题应该在 nicescroll 层面无解。你要想设置,可以强行把 html 标签上的 overflow:hidden 删掉,就能翻页了。但是这样会出现两个滚动条,你还要修改样式,隐藏原生的滚动条。但是这样的话可能会干扰 nicescroll 的行为,再用 nicescroll 就没有多少意义了。
想让 ctrl + f 翻页生效很简单,不超过几行代码,但是整个问题很棘手。
多谢几位的反馈,这个问题我们再跟进一下
我们在 portal 里面完全去掉了 nicescroll 这个组件,目前没发现什么使用上的问题,不过不知道会不会影响触屏使用。
有需要我可以提一个 pr
@codepiano
劳烦提个PR吧,这个问题还挺影响使用的,多谢!
1.2.0版本已经发布,去除了nicescroll组件
|
gharchive/issue
| 2018-08-15T03:19:41 |
2025-04-01T04:33:54.493260
|
{
"authors": [
"codepiano",
"maoplus",
"nobodyiam",
"slankka"
],
"repo": "ctripcorp/apollo",
"url": "https://github.com/ctripcorp/apollo/issues/1370",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
253546359
|
怎么添加部门?
转了一圈没发现添加部门的地方
请参考分布式部署指南 - 2.1.3 调整服务端配置,目前需要在表中直接添加。
什么时候可以有页面添加部门?
请参考分布式部署指南 - 2.1.3 调整服务端配置,现在可以通过管理员工具 - 系统参数页面进行配置,修改完一分钟实时生效。
|
gharchive/issue
| 2017-08-29T06:53:32 |
2025-04-01T04:33:54.495836
|
{
"authors": [
"lc273734858",
"nobodyiam",
"zhanghaichang"
],
"repo": "ctripcorp/apollo",
"url": "https://github.com/ctripcorp/apollo/issues/722",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
262223376
|
[typescript] Use ComponentType for more flexibility.
This allows for React component classes or stateless functional components to be used when calling asyncComponent method.
Awesome thanks for this @bdoss!
|
gharchive/pull-request
| 2017-10-02T20:47:34 |
2025-04-01T04:33:54.496806
|
{
"authors": [
"bdoss",
"ctrlplusb"
],
"repo": "ctrlplusb/react-async-component",
"url": "https://github.com/ctrlplusb/react-async-component/pull/59",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2096994578
|
Add support for multiple audiences
The OIDC provider Zitadel returns multiple audiences. Since I use Zitadel I added the option to have more than one audience in the configuration. If none are provided the client_id is used as only valid audience.
Thanks for submitting the PR! That's definitely useful!
I am just wondering if we could simplify this a bit. To my understanding, calling set_other_audience_verifier_fn is a check in addition to checking the client_id already. So we could rename the valid_audiences to additional_required_audiences (or something like that), give it a type of Vec<_> (instead of the additional Option, and then either omit calling set_other_audience_verifier_fn, or return true if the Vec is empty.
What do you think?
Yeah I think that sounds like a good point. I just used that way in a different server side actix code :smile:
I am taking a look at this right now, trying to get it merged.
Just for my understanding: the idea of the set_other_audience_verifier_fn is, to verify each audience present in the token, right? so actually it would not be "additional required audiences", but "additional allowed audiences"?
Yeah that is about it. That is why the client id is always required.
Cool, so the change is released with 0.10.1, maybe you can give it a try and report back if that works for you.
I need to see when I get to it. For the current project I moved the auth into the actix server and its session.
|
gharchive/pull-request
| 2024-01-23T21:15:23 |
2025-04-01T04:33:54.500643
|
{
"authors": [
"DerKnerd",
"ctron"
],
"repo": "ctron/yew-oauth2",
"url": "https://github.com/ctron/yew-oauth2/pull/26",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2249686969
|
[Bug]: The mysql in Dolphinscheduler start failed.
Contact Details
softxcq@sina.com
Is there an existing issue for this?
[X] I have searched all the existing issues
Priority
low (Default)
Current Behavior
branch main , lastest code .
When I use Docker Compose to install DolphinScheduler, MySQL fails to start properly.
Starting MySQL database server mysqld su: warning: cannot change directory to /nonexistent: No such file or directory
Expected Behavior
When installing mysql-server , system will create 'mysql' user with no home dir, it could be the reason.
Steps To Reproduce
1. download code and compile
2. docker compose install
3. install dolphinscheduler (mysql server)
mysql start failed
Compass Log
No response
Anything else? (Additional Context)
No response
@xuchangqun Add build parameters and try again ? docker compose --profile dolphinscheduler up -d --build
/nonexistent
Adding parameters still doesn't work. I solved it by replacing '/nonexistent' in /etc/passwd.
😄
@xuchangqun I also encountered the same problem. What do I need to replace '/nonexistent' with
|
gharchive/issue
| 2024-04-18T04:01:03 |
2025-04-01T04:33:54.522070
|
{
"authors": [
"a278574361",
"nilnon",
"xuchangqun"
],
"repo": "cubefs/compass",
"url": "https://github.com/cubefs/compass/issues/221",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
171643188
|
(Enhancement) inconsistent use of pagination for tables
in ct we have small tables and big tables, regardless i see pagination only at some of them.
Locations, Triggers, Networks, Zones, Splash, Users have none.
(ok you don't need to add them if entrys are limited to a number which fits on page one)
Not all the results accept pagination at the moment, that's why. If you put a list of the ones w/o pagination, we will go through the backend.
@ChristianWinkler
@ariandora is this resolved?
@iamzakr the code is commented out, the commit is more of a setup for the pagination, as the links information was not always present.
@iamzakr -> open a ticket internally to ensure all required
controllers have the _links block.
👍
|
gharchive/issue
| 2016-08-17T12:17:51 |
2025-04-01T04:33:54.534546
|
{
"authors": [
"AndrewMargarit",
"ChristianWinkler",
"ariandora",
"iamzakr",
"simonmorley"
],
"repo": "cucumber-tony/cucumber-frontend",
"url": "https://github.com/cucumber-tony/cucumber-frontend/issues/290",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
158986753
|
Devices : Index Page not Subscribing to API for streaming updates
The devices page should be subscribed to receive streaming updates from CT.
Currently the device#show page is subscribed.
The front is done and subscribes.
Backend needs wiring up.
|
gharchive/issue
| 2016-06-07T18:02:41 |
2025-04-01T04:33:54.535908
|
{
"authors": [
"simonmorley"
],
"repo": "cucumber-tony/cucumber-frontend",
"url": "https://github.com/cucumber-tony/cucumber-frontend/issues/7",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
192777173
|
Drop Rails 3 support
Implements #333.
Also:
Adds an appraisal to specifically test compatibility with the oldest declared versions of dependencies.
Updates the development dependencies to require compatible versions
Updates aruba to ensure cucumber-rails' own scenarios see the correct bundle
Restores compatibility of cucumber-rails' scenarios with cucumber 1.3.x
thank you @mvz for all the hard work :+1:
|
gharchive/pull-request
| 2016-12-01T07:57:19 |
2025-04-01T04:33:54.544275
|
{
"authors": [
"Kosmas",
"mvz"
],
"repo": "cucumber/cucumber-rails",
"url": "https://github.com/cucumber/cucumber-rails/pull/334",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
347715634
|
gherkin: ruby,go: Failing test for strange size limitiation for feature files
Summary
Between 886c1c4...285b244 there seems to have appeared a strange size limitation for feature files when passing them as source (instead of as path) to the gherkin parser.
Details
The failing test says that a feature file with 49 minimal scenarios works as expected (51 messages produced by the parser) but a feature file with 50 minimal scenario does not work (no messages produced by the parser). It does not occur when passing the feature file as path (and it is not specifically about 50 scenarios), since when a use https://github.com/cucumber/cucumber-ruby/blob/master/features/docs/formatters/junit_formatter.feature from the gherkin_spec.rb and send it by path the parser produces 12 messages, but when sending its content as source the parser produces 0 messages.
Motivation and Context
How Has This Been Tested?
Screenshots (if appropriate):
Types of changes
[ ] Bug fix (non-breaking change which fixes an issue).
[ ] New feature (non-breaking change which adds functionality).
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected).
Checklist:
[ ] I've added tests for my code.
[ ] My change requires a change to the documentation.
[ ] I have updated the documentation accordingly.
I ran git bisect and the message was that d4dae16 is the first bad commit.
test.sh of a0d4e9a can be used to check if bin/gherkin can handle source messages for "large" feature files on stdin.
This could be a bug in gogo/protobuf, or it could simply be that it’s not used correctly.
My motivation for switching to gogo was its built-in support for varint-delimited streams. The official protobuf leaves that entirely to the app developer.
Perhaps it’s safer to switch back to the official one.
|
gharchive/pull-request
| 2018-08-05T17:19:37 |
2025-04-01T04:33:54.550465
|
{
"authors": [
"aslakhellesoy",
"brasmusson"
],
"repo": "cucumber/cucumber",
"url": "https://github.com/cucumber/cucumber/pull/449",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
710951954
|
Remove recent actions box
Since the django-admin is public and no actions are taken by users, it makes sense to get the box "Recent Actions" removed.
According to @cuducos's tip here, we can get it done by using display: none: this template uses dashboard.css that has a display: none.
Hej,
I tried to solve this issue in a other way, with Django templates.
Take a look at PR #20 .
|
gharchive/issue
| 2020-09-29T09:55:30 |
2025-04-01T04:33:54.553173
|
{
"authors": [
"anapaulagomes",
"jhrdt"
],
"repo": "cuducos/django-public-admin",
"url": "https://github.com/cuducos/django-public-admin/issues/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1706906222
|
Segfault while importing CUE definitions from Istio API definitions
What version of CUE are you using (cue version)?
$ cue version
cue version v0.5.0
go version go1.20.3
-buildmode exe
-compiler gc
-trimpath true
CGO_ENABLED 1
GOARCH amd64
GOOS darwin
GOAMD64 v1
Does this issue reproduce with the latest stable release?
Yes
What did you do?
cd $(mktemp -d)
(
set -e
cue mod init
go mod init mod.test
go get istio.io/api/networking/v1beta1@v0.0.0-20230511193058-044e6e6ce64a
cue get go istio.io/api/networking/v1beta1
)
What did you expect to see?
No segfault, and definitions that work.
Interestingly, the definitions are still generated in my cue.mod directory but perhaps I'm missing some?
If I try to make use of a definition by importing them, I get errors:
import failed: import failed: cannot find package "google.golang.org/protobuf/types/known/durationpb":
./definitions/routing.cue:7:2
./cue.mod/gen/istio.io/api/networking/v1beta1/virtual_service.pb_go_gen.cue:7:8
What did you see instead?
panic: runtime error: invalid memory address or nil pointer dereference [recovered]
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x20 pc=0x16ff0f9]
goroutine 1 [running]:
cuelang.org/go/cmd/cue/cmd.recoverError(0xc00c4ade90)
cuelang.org/go/cmd/cue/cmd/root.go:261 +0x7b
panic({0x17aeca0, 0x1e09ee0})
runtime/panic.go:884 +0x213
cuelang.org/go/cmd/cue/cmd.(*extractor).extractPkg(0xc00a6234d0, {0xc000040054, 0x3f}, 0x0)
cuelang.org/go/cmd/cue/cmd/get_go.go:438 +0x79
cuelang.org/go/cmd/cue/cmd.(*extractor).extractPkg(0xc00a6234d0, {0xc000040054, 0x3f}, 0xc0000cca80)
cuelang.org/go/cmd/cue/cmd/get_go.go:554 +0xb1a
cuelang.org/go/cmd/cue/cmd.extract(0xc000354ed0, {0xc000392050, 0x1, 0x1})
cuelang.org/go/cmd/cue/cmd/get_go.go:417 +0x3e5
cuelang.org/go/cmd/cue/cmd.mkRunE.func1(0xc000388280?, {0xc000392050, 0x1, 0x1})
cuelang.org/go/cmd/cue/cmd/root.go:72 +0xb3
github.com/spf13/cobra.(*Command).execute(0xc000388280, {0xc000392020, 0x1, 0x1})
github.com/spf13/cobra@v1.4.0/command.go:856 +0x67c
github.com/spf13/cobra.(*Command).ExecuteC(0xc00036ea00)
github.com/spf13/cobra@v1.4.0/command.go:974 +0x3bd
github.com/spf13/cobra.(*Command).Execute(...)
github.com/spf13/cobra@v1.4.0/command.go:902
cuelang.org/go/cmd/cue/cmd.(*Command).Run(0xc000354ed0, {0x3?, 0x3?})
cuelang.org/go/cmd/cue/cmd/root.go:246 +0x6b
cuelang.org/go/cmd/cue/cmd.mainErr({0x1996288, 0xc00019c008}, {0xc00019a050?, 0x1fa9a68?, 0x60?})
cuelang.org/go/cmd/cue/cmd/root.go:183 +0x55
cuelang.org/go/cmd/cue/cmd.Main()
cuelang.org/go/cmd/cue/cmd/root.go:165 +0x7c
main.main()
cuelang.org/go/cmd/cue/main.go:24 +0x19
This also segfaults on the latest alpha, and segfaults on 0.4.3.
Possibly related issue, though it doesn't have any details? https://github.com/istio/api/issues/2620
Apologies, this seems to be a dup of https://github.com/cue-lang/cue/issues/2046
For anyone who wants a workaround, this seems to work:
cue get go google.golang.org/protobuf/types/known/durationpb
You should now be able to import and use the generated CUE definitions.
|
gharchive/issue
| 2023-05-12T04:29:01 |
2025-04-01T04:33:54.557778
|
{
"authors": [
"Niksko"
],
"repo": "cue-lang/cue",
"url": "https://github.com/cue-lang/cue/issues/2402",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2554691008
|
Multiline string rejects line terminator elision
What version of CUE are you using (cue version)?
$ cue version
cue version v0.10.0
go version go1.23.0
-buildmode exe
-compiler gc
-trimpath true
CGO_ENABLED 0
GOARCH amd64
GOOS linux
GOAMD64 v1
cue.lang.version v0.10.0
Does this issue reproduce with the latest stable release?
Yes
What did you do?
Tried to evaluate the string from the spec under String and byte sequence literals:
"""
lily:
out of the water
out of itself
bass
picking \
bugs
off the moon
— Nick Virgilio, Selected Haiku, 1988
"""
What did you expect to see?
According to the spec page it should be equivalent to:
"lily:\nout of the water\nout of itself\n\n" +
"bass\npicking bugs\noff the moon\n" +
" — Nick Virgilio, Selected Haiku, 1988"
What did you see instead?
cannot get syntax for "-": unknown escape sequence:
-:7:14
Notes
Here's a playground with this issue: https://cuelang.org/play/?id=7uimnxuUVGR#w=function&i=cue&f=eval&o=cue
A smaller reproduction case: https://cuelang.org/play/?id=hxhDZo_g4zz#w=function&i=cue&f=eval&o=cue
Ah this seems related to https://github.com/cue-lang/cue/issues/2317
|
gharchive/issue
| 2024-09-29T04:00:10 |
2025-04-01T04:33:54.562283
|
{
"authors": [
"infogulch"
],
"repo": "cue-lang/cue",
"url": "https://github.com/cue-lang/cue/issues/3469",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
897555912
|
fix: bug for last sold mods
What
Fixes bug for last sold mods when Bungie API is down at reset the day before.
:tada: This PR is included in version 1.21.3 :tada:
The release is available on GitHub release
Your semantic-release bot :package::rocket:
|
gharchive/pull-request
| 2021-05-20T23:48:46 |
2025-04-01T04:33:54.570427
|
{
"authors": [
"cujarrett"
],
"repo": "cujarrett/destiny-insights-backend",
"url": "https://github.com/cujarrett/destiny-insights-backend/pull/43",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.