id
stringlengths 4
10
| text
stringlengths 4
2.14M
| source
stringclasses 2
values | created
timestamp[s]date 2001-05-16 21:05:09
2025-01-01 03:38:30
| added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| metadata
dict |
---|---|---|---|---|---|
158685821
|
IE9/IE10 crash in AuthService.js
Hi,
We've encountered an issue in IE<=10 with Aurelia-Authentication (version 1.1.2), Working fine with Edge, and we've included all the necessary polyfills: fetch.js, bluebird.js, etc...
In line 31 of AuthService.js, config is undefined:
this.auth = auth;
this.oAuth1 = oAuth1;
this.oAuth2 = oAuth2;
this.config = config.current; // config is undefined.
this.client = this.config.client;
Which causes an obvious problem. This code is called during the instanciation
of our Authentication class, whose code follows:
import {inject} from 'aurelia-dependency-injection';
import {AuthService} from 'spoonx/aurelia-authentication';
import {Endpoint} from 'spoonx/aurelia-api';
@inject(Endpoint.of('codabox-api'))
export class CodaboxAuthService extends AuthService {
constructor(codaboxApi, ...args){
super(...args); // crashes here.
this.codaboxApi = codaboxApi;
}
...
}
Any idea ?
in that version current is a getter. if you can try to use t eobject directly aka this.config = config._current
Well, the problem is that config is undefined. config.current is trying to access a field from an undefined reference, thus the crash.
hang on a little, i think it'll be fixed in the next version.
you need to inject all again when extending i'm sure
@inject(Endpoint.of('codabox-api'), Authentication, OAuth1, OAuth2, BaseConfig);
This is super urgent. A workaround is:
class Foo extends Bar {
constructor () {
super()
Bar.call(this)
}
}
Note that this calls the parent constructor twice in other browsers, so make sure your parent constructor does nothing too complex (which it shouldn't anyway)
or
I've actually just found something that fixed this problem for me. There's an option called loose for babel-plugin-transform-es2015-classes, that makes it work on IE <= 10. I don't know if on babel 5 this was turned on by default or if the code always worked on older browsers, but it seems like setting this option is what's required now, it all works perfectly as before.
You basically just need to explicitly pass the plugin with the option, instead of just using the es2015 preset, like this:
plugins: [['transform-es2015-classes', {loose: true}]]
Seems like this isn't a bug then, but I still think this should be better documented so people know what to do to make babel 6 compiled code work on all browsers.
Ok, thanks for your investigation, I will try this and confirm that it works on my side.
Hi, I can't get it to work. It builds all-right, but it's still not working in IE.
I'm not sure I am doing things right as I am not familiar with either aurelia nor babel, here's my build/babel-options.js
var path = require('path');
var paths = require('./paths');
exports.base = function() {
return {
filename: '',
filenameRelative: '',
sourceMap: true,
sourceRoot: '',
moduleRoot: path.resolve('src').replace(/\\/g, '/'),
moduleIds: false,
comments: false,
compact: false,
code: true,
presets: [ 'es2015-loose', 'stage-1'],
plugins: [
'syntax-flow',
'transform-decorators-legacy',
'transform-flow-strip-types',
'transform-class-properties',
['transform-es2015-classes', {loose: true}]
]
};
};
exports.commonjs = function() {
var options = exports.base();
options.plugins.push('transform-es2015-modules-commonjs');
return options;
};
exports.amd = function() {
var options = exports.base();
options.plugins.push('transform-es2015-modules-amd');
return options;
};
exports.system = function() {
var options = exports.base();
options.plugins.push('transform-es2015-modules-systemjs');
return options;
};
exports.es2015 = function() {
var options = exports.base();
options.presets = ['stage-1'];
return options;
};
And I installed the following npm packages:
+ "babel-plugin-transform-class-properties": "^6.9.1",
+ "babel-plugin-transform-es2015-classes": "^6.9.0",
without the loose it seems to work for me
presets: [ 'es2015-loose', 'stage-1'],
plugins: [
'syntax-flow',
'transform-decorators-legacy',
'transform-flow-strip-types',
'transform-class-properties',
'transform-es2015-classes'
]
Thanks !
This fixed the issue.
Unfortunately there were lots of other IE issues waiting behind ... but that's none of your concern :)
oho good luck then
|
gharchive/issue
| 2016-06-06T14:04:02 |
2025-04-01T04:55:39.326012
|
{
"authors": [
"cb-fred",
"doktordirk"
],
"repo": "SpoonX/aurelia-authentication",
"url": "https://github.com/SpoonX/aurelia-authentication/issues/185",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
238131618
|
Registering an entity without the @resource() decorator leads to invalid meta data
The documentation about the @resource() decorator says
This decorator is probably the most important one.
But this decorator is never mentioned in the whole Quickstart or Entites guide https://aurelia-orm.spoonx.org/entities.html
Fact is: without the decoratory, the entitymanager doesn't has meta data for this entity. This is a documentation bug.
Resource should have the name of the resource because minification will otherwise mess things up.
But yes, it could use more docs! :)
Minification removes the @resource() decorator completely?
@mreiche No, it renames classes. And without a value, the decorator uses the name of the class. Minification can still work if you disable renaming. Check the bottom of the readme for more info.
Ok, I add that to the documentation.
@mreiche Yeah it can be added with the decorator itself. I agree. Thanks!
|
gharchive/issue
| 2017-06-23T12:43:23 |
2025-04-01T04:55:39.329378
|
{
"authors": [
"RWOverdijk",
"mreiche"
],
"repo": "SpoonX/aurelia-orm",
"url": "https://github.com/SpoonX/aurelia-orm/issues/300",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2831381
|
INT-2400 Fix tool:expected-type Classes in IP XSD
MessageChannel was moved from core to the base package.
InterceptorFactoryChain was incorrect.
More than just IP; closing.
|
gharchive/issue
| 2012-01-13T13:59:36 |
2025-04-01T04:55:39.361164
|
{
"authors": [
"garyrussell"
],
"repo": "SpringSource/spring-integration",
"url": "https://github.com/SpringSource/spring-integration/issues/314",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1729383864
|
Add a New Login Button
Hi @Spyware007, I am a contributor in GSSOC'23.
I want to add a new login button.
can you please assign me this task under label gssoc'23
@Ayush9951 Assigned to you!
|
gharchive/issue
| 2023-05-28T10:38:51 |
2025-04-01T04:55:39.363167
|
{
"authors": [
"Ayush9951",
"Spyware007"
],
"repo": "Spyware007/Animating-Buttons",
"url": "https://github.com/Spyware007/Animating-Buttons/issues/713",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
156177774
|
Remote inventories not seeming to work
I might be derping this up, but I have a wired modem block, connected to an iron chest, a turtle, and vanilla chest. I am attempting to call iron.pushItemRemote("chest_0",1) but it is returning
lua:1: Parameter 1 value not set
Which version of CC, CCTweaks and OpenPeripheral are you running? Also can I have your log? Thanks!
Oh, sorry about that, forgot versions:
ComputerCraft: 1.75
CCTweaks: 0.3.2
OpenPeripheralCore: 1.3
OpenPeripheralIntegration: 0.5
OpenPeripheralAddons: 0.5.1
How much of the log do you need? Just around when I tried the function, or the whole thing?
Oh yea, this is also on a server.
Fixed in 39740a48a2689e8b3aa6b54380a711192094974d. I've pushed it but GitHub is derping.
|
gharchive/issue
| 2016-05-22T22:45:40 |
2025-04-01T04:55:39.367890
|
{
"authors": [
"Selim042",
"SquidDev"
],
"repo": "SquidDev-CC/CCTweaks",
"url": "https://github.com/SquidDev-CC/CCTweaks/issues/98",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1461431973
|
Unable to load DLL 'libSkiaSharp' on iOS
Hello,
When I build for iOS using this package I get the following error:
DllNotFoundException: Unable to load DLL 'libSkiaSharp'. Tried the load the following dynamic libraries: Unable to load dynamic library '/libSkiaSharp' because of 'Failed to open the requested dynamic library (0x06000000) dlerror() = dlopen(/libSkiaSharp, 0x0005): tried: '/libSkiaSharp' (errno=2), '/private/preboot/Cryptexes/OS/libSkiaSharp' (errno=2), '/libSkiaSharp' (errno=2)
Please help.
any luck with this? I get it too
Could you please add __IOS__and __UNIFIED__ to the Scripting Define Symbols and try to build again? Let me know the update.
I will update SkiaSharp's version to the latest one (2.88.3), it's been a long time since we are using an older version (2.80.3).
Could you please add __IOS__and __UNIFIED__ to the Scripting Define Symbols and try to build again? Let me know the update.
I will update SkiaSharp's version to the latest one (2.88.3), it's been a long time since we are using an older version (2.80.3).
I actually tried that already, didn't work.
Alright, thanks for the update.
Hello @LeshM, could you please try this 0.2.0-preview.1 version? Unfortunately, I don't have access to any Mac device right now, so couldn't build the XCode project.
Right now, you can grab it from OpenUPM .
Or, merge these in your manifest.json -
"scopedRegistries": [
{
"name": "package.openupm.com",
"url": "https://package.openupm.com",
"scopes": [
"com.srejonkhan"
]
}
]
"dependencies": {
"com.srejonkhan.asyncimagelibrary": "0.2.0-preview.1"
}
Alternatively, you can try cloning this sandbox project and change the branch of the submodule to develop branch (instructions added) - https://github.com/SrejonKhan/AsyncImageLibrary-Sandbox
The build process would be exactly the same as you've tried earlier with Scripting Define Symbols. In case it fails, you may try tweaking the setting of /AsyncImageLibrary/Dependencies/SkiaSharp/ios/native/libSkiaSharp.framework if any mistakes were made from my side. Any changes/pull request is always appreciated.
Hello @LeshM, could you please try this 0.2.0-preview.1 version? Unfortunately, I don't have access to any Mac device right now, so couldn't build the XCode project.
Right now, you can grab it from OpenUPM .
Or, merge these in your manifest.json -
"scopedRegistries": [
{
"name": "package.openupm.com",
"url": "https://package.openupm.com",
"scopes": [
"com.srejonkhan"
]
}
]
"dependencies": {
"com.srejonkhan.asyncimagelibrary": "0.2.0-preview.1"
}
Alternatively, you can try cloning this sandbox project and change the branch of the submodule to develop branch (instructions added) - https://github.com/SrejonKhan/AsyncImageLibrary-Sandbox
The build process would be exactly the same as you've tried earlier with Scripting Define Symbols. In case it fails, you may try tweaking the setting of /AsyncImageLibrary/Dependencies/SkiaSharp/ios/native/libSkiaSharp.framework if any mistakes were made from my side. Any changes/pull request is always appreciated.
I'm sorry but I'm afraid I can't, I'll need to reintegrate your package and then build for both Android & iOS using UCB (which takes about an hour each) just to find out if this works. I'm on a tight deadline and can't spare the time.
I do appreciate your work though, so thank you! 🙏
BTW I ended up using some Unity coroutine hack as a solution.
@LeshM No problem. Thank you too!
|
gharchive/issue
| 2022-11-23T10:03:39 |
2025-04-01T04:55:39.399701
|
{
"authors": [
"BigGiantHead",
"LeshM",
"SrejonKhan"
],
"repo": "SrejonKhan/AsyncImageLibrary",
"url": "https://github.com/SrejonKhan/AsyncImageLibrary/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1592019731
|
Bottom to top button
Make a bottom to top button for the Home page, which will be appeared after some scrolling.
@Srijita-Mandal I would like to work on it
@sarkartanmay393 Go for it.
I would like to work on it under JWOC
I would Like to work on this issue under JWOC.
#119
Please assign me this issue under jwoc @Srijita-Mandal Srijita-Mandal
@Srijita-Mandal
I would like to work on this issue under JWOC.
Please assign this issue to me.
|
gharchive/issue
| 2023-02-20T14:59:14 |
2025-04-01T04:55:39.404021
|
{
"authors": [
"Awesomeasma",
"LuckySingh9897",
"Srijita-Mandal",
"priyanka1551",
"sarkartanmay393",
"vatan-prakash"
],
"repo": "Srijita-Mandal/fix-your-nums",
"url": "https://github.com/Srijita-Mandal/fix-your-nums/issues/114",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1509335357
|
Questions about Analysis of Disentanglement in the paper
Hi, I read your amazing work: "Adversarial Graph Disentanglement", and then I have a question about the experimental part. In the V. EXPERIMENTAL RESULTS AND ANALYSIS, E. Analysis of Disentanglement, you compute the component confusion matrix C to illustrate the correlations among the distributions in different component spaces.
The correlation score between the i-th and j-th component distributions is given by
.
I don't understand why this formula is used to calculate the correlation of two vector spaces.
Why is cosine similarity calculated between every two elements in these two spaces?
Why not compute the similarity between corresponding elements as a similarity measure for two vector spaces?that is Cij=1/Vsum(hu^ihu^j).
Finally, can you provide your code on computing the cosine similarity of two vector spaces?
I would appreciate it if you could.
Thank you!
Thanks for your attention to our work and the valuable question you have raised!
I don't know if I understand this correctly, I guess the similarity you mentioned Cij=1/Vsum(hu^ihu^j) is the dot product. Actually, the dot product similarity is practically equivalent to the cosine similarity we adopt in this work, since the vector h^i_u and h^j_v are normalized during the process of micro-disentanglement.
As for the associated codes, I will release them after I have reorganized. Plz stay tuned for this project!
|
gharchive/issue
| 2022-12-23T13:12:26 |
2025-04-01T04:55:39.417396
|
{
"authors": [
"SsGood",
"xiaomi4356"
],
"repo": "SsGood/ADGCN",
"url": "https://github.com/SsGood/ADGCN/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1876279094
|
Continue generating the images after an interrumption in electricity
I want to know if we could easily recover from an electricy issue, for example. If the server closes for example... Can I reopen the server again and continue generating images from the same queue? Should we persist the current queue on disk in realtime? Does it make sense?
Thanks a lot for your time
That's... a new one. Have you considered, uh, a UPS? (Uninterruptable Power Supply - hardware device that's essentially a giant battery specifically to protect computers from failure if power fails)
Spent some time considering whether to maybe implement some form of option for drive-caching of the generation queue or something, but I don't think it's going to fit as a valid feature, at least not during the current Alpha stage.
So, for now, yeah best advice is get a UPS lol. Once Swarm is out of alpha and in release phase, can reopen this and reconsider the possibility.
|
gharchive/issue
| 2023-08-31T20:46:18 |
2025-04-01T04:55:39.440421
|
{
"authors": [
"fabianrodrigo",
"mcmonkey4eva"
],
"repo": "Stability-AI/StableSwarmUI",
"url": "https://github.com/Stability-AI/StableSwarmUI/issues/82",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1377680642
|
StabilityInference.generate doesn't supports Sequence[int] as seed
As type hinted here, StabilityInference.generate should supports a Sequence[int] as seed but is overwrited here.
PR in-progress
thanks for the contribution!
|
gharchive/issue
| 2022-09-19T09:40:57 |
2025-04-01T04:55:39.442299
|
{
"authors": [
"Thytu",
"dmarx"
],
"repo": "Stability-AI/stability-sdk",
"url": "https://github.com/Stability-AI/stability-sdk/issues/48",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
228693222
|
blackbox_list_admins
The black_list_admins fails on my machine, even though the blackbox-admins.txt file exists.
$ tree keyrings
keyrings
└── live
├── blackbox-admins.txt
├── blackbox-files.txt
├── pubring.kbx
└── trustdb.gpg
1 directory, 4 files
$ blackbox_list_admins
cat: blackbox-admins.txt: No such file or directory
Wow, that script actually is broken. The quick fix is to edit the line:
Awesome Tom 👌
On Mon, 15 May 2017 at 17:11, Tom Limoncelli notifications@github.com
wrote:
Closed #193 https://github.com/StackExchange/blackbox/issues/193 via
7649c9f
https://github.com/StackExchange/blackbox/commit/7649c9f00b14390e08fb90a4318378058c2d0a3c
.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/StackExchange/blackbox/issues/193#event-1082442530,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AAivm754sLQ_wSyXxAZrRPVoCtCn9dKrks5r6GsFgaJpZM4NbCnc
.
|
gharchive/issue
| 2017-05-15T12:06:47 |
2025-04-01T04:55:39.451565
|
{
"authors": [
"aymericbeaumet",
"tlimoncelli"
],
"repo": "StackExchange/blackbox",
"url": "https://github.com/StackExchange/blackbox/issues/193",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
101783495
|
RFR: Support trace_id in CLI run
(virtualenv)/m/s/s/st2 git:support_trace_id_in_cli_run ❯❯❯ st2 run core.local date --trace-id='55d4c36f32ed3554acf2db4'
ERROR: 400 Client Error: Bad Request
MESSAGE: Unable to find Trace with id="55d4c36f32ed3554acf2db4"
(virtualenv)/m/s/s/st2 git:support_trace_id_in_cli_run ❯❯❯
(virtualenv)/m/s/s/st2 git:master ❯❯❯ st2 run core.local cmd=date --trace-tag="OohLaLaLa" ⏎ ✭ ✱ ◼
.
id: 55d3d5bf32ed3530544a753d
status: succeeded
result:
{
"failed": false,
"stderr": "",
"return_code": 0,
"succeeded": true,
"stdout": "Tue Aug 18 18:02:55 PDT 2015
"
}
(virtualenv)/m/s/s/st2 git:master ❯❯❯ st2 trace list --trace-tag OohLaLaLa ✭ ✱ ◼
+-------------------+-------------------------------------------------------+
| Property | Value |
+-------------------+-------------------------------------------------------+
| id | 55d3d5bf32ed3530544a753e |
| trace_tag | OohLaLaLa |
| action_executions | [ |
| | { |
| | "updated_at": "2015-08-19T01:02:55.018421Z", |
| | "object_id": "55d3d5bf32ed3530544a753d" |
| | } |
| | ] |
| rules | |
| start_timestamp | 2015-08-19T01:02:55.018362Z |
| trigger_instances | |
+-------------------+-------------------------------------------------------+
(virtualenv)/m/s/s/st2 git:master ❯❯❯ st2 run core.local cmd=date --trace-id='55d3d5bf32ed3530544a753e' ✭ ✱ ◼
.
id: 55d3d5ec32ed3530544a7540
status: succeeded
result:
{
"failed": false,
"stderr": "",
"return_code": 0,
"succeeded": true,
"stdout": "Tue Aug 18 18:03:41 PDT 2015
"
}
(virtualenv)/m/s/s/st2 git:master ❯❯❯ st2 trace get 55d3d5bf32ed3530544a753e ✭ ✱ ◼
+-------------------+-------------------------------------------------------+
| Property | Value |
+-------------------+-------------------------------------------------------+
| id | 55d3d5bf32ed3530544a753e |
| trace_tag | OohLaLaLa |
| action_executions | [ |
| | { |
| | "updated_at": "2015-08-19T01:03:40.972020Z", |
| | "object_id": "55d3d5ec32ed3530544a7540" |
| | }, |
| | { |
| | "updated_at": "2015-08-19T01:02:55.018421Z", |
| | "object_id": "55d3d5bf32ed3530544a753d" |
| | } |
| | ] |
| rules | |
| start_timestamp | 2015-08-19T01:02:55.018362Z |
| trigger_instances | [ |
| | { |
| | "updated_at": "2015-08-19T01:03:41.091999Z", |
| | "object_id": "55d3d5ed32ed35305dc72cc5" |
| | } |
| | ] |
+-------------------+-------------------------------------------------------+
(virtualenv)/m/s/s/st2 git:master ❯❯❯
+1
|
gharchive/pull-request
| 2015-08-19T01:07:12 |
2025-04-01T04:55:39.456429
|
{
"authors": [
"lakshmi-kannan",
"manasdk"
],
"repo": "StackStorm/st2",
"url": "https://github.com/StackStorm/st2/pull/1840",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
412129081
|
Switching between Mainnet and Testnet
Hi Stadicus, I have one simple question. Everything is working well on mainnet now for some weeks but if I want to move back to testnet (and later back to mainnet etc.), my understanding of reading all directories and your instructions is that I only need to adjust the following
0) stop lnd and bitcoind
change to testnet in /home/bitcoin/.bitcoin/bitcoin.conf
chainge to testnet in /home/admin/.bitcoin/bitcoin.conf
change to testnet in /home/bitcoin/.lnd/lnd.conf
I think the admin.macaroon is already available in both mainnet and testnet (also for user admin) and also the tls.cert does not have to change?
so after 1), 2) and 3) start bitcoind and than lnd should have it all work on testnet again?
And to go back to mainnet just do the config files back to mainnet
Before I am going to destroy a working node, I just wanted to check, thanks! Walter
Reading through your very thorough list, I think that's exactly right. 👨🚀👍
|
gharchive/issue
| 2019-02-19T21:19:19 |
2025-04-01T04:55:39.461543
|
{
"authors": [
"Stadicus",
"walterkok"
],
"repo": "Stadicus/guides",
"url": "https://github.com/Stadicus/guides/issues/388",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2129892313
|
🛑 Parsoid is down
In ae7dfbe, Parsoid (https://star-citizen.wiki/api.php?action=visualeditor&format=json&paction=parse&page=Star_Citizen_Wiki) was down:
HTTP code: 404
Response time: 632 ms
Resolved: Parsoid is back up in e2d0544 after 9 minutes.
|
gharchive/issue
| 2024-02-12T11:02:21 |
2025-04-01T04:55:39.488327
|
{
"authors": [
"octfx"
],
"repo": "StarCitizenWiki/status",
"url": "https://github.com/StarCitizenWiki/status/issues/334",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1745144702
|
Broker load shall fail immediately with error message on wrong PROPERTIES setting
Enhancement
When there is misconfig in the broker load PROPERTIES(e.g. wrong AWS AK/SK or endpoint), broker load job state becomes QEUEUEING with out and error message to help debug.
Suggest fixing this issue by setting the state to INVALID or FAILED and have details in ErrorMsg fields.
This is essential for the upstream service to check the status and take appropriate measure for failure resilience.
cc @chaoyli
Thanks for reporting this issue. It's highly OK to add a check whether the ak/sk is set.
I'll resolve it.
I've checked the latest version of branch-2.5/branch-3.0 and found that the result has been shown like this. It seems to be what you want.
It is PENDING state in a built from branch-3.0 commit hash https://github.com/StarRocks/starrocks/commits/a5dd8c36
on 5/23/2023
In this test, I issued the load sql with a wrong secret key in the properties.
LOAD LABEL abb_10m.KNA1_4
(
DATA INFILE("s3a://celostar/test_data/abb_10m/fd1c713c-d76f-4113-99d2-a22c38afc7a6/*.parquet")
INTO TABLE KNA1
FORMAT AS "parquet"
******
WITH BROKER
(
"aws.s3.access_key" = "xxx",
"aws.s3.secret_key" = "WRONG SECRET", <===================
"aws.s3.region"="us-west-1"
)
PROPERTIES
(
"timeout" = "36000", -- the timeout time in seconds for the import operation. e.g 3600 means 1 hr time out. Jobs not completed in timeout window will be CANCELLED state.
"max_filter_ratio" = "0.2", -- Data ratio of maximum tolerance filterable (data irregularity, etc.).
"timezone" = "America/Los_Angeles"
);
This is enhanced in https://github.com/StarRocks/starrocks/pull/21982.
Could you check it with a newer version again?
|
gharchive/issue
| 2023-06-07T06:37:05 |
2025-04-01T04:55:39.505960
|
{
"authors": [
"dengliu",
"rickif"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/issues/24794",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1579196287
|
[Doc]add pause routine load
Signed-off-by: hellolilyliuyi hellolilyliuyi123@163.com
What type of PR is this:
[ ] BugFix
[ ] Feature
[ ] Enhancement
[ ] Refactor
[ ] UT
[x] Doc
[ ] Tool
Which issues of this PR fixes :
Fixes #
Problem Summary(Required) :
Checklist:
[ ] I have added test cases for my bug fix or my new feature
[ ] This pr will affect users' behaviors
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto backported to target branch
[x] 2.5
[x] 2.4
[x] 2.3
[x] 2.2
run starrocks_admit_test
@mergify backport branch-2.5
@mergify backport branch-2.4
@mergify backport branch-2.3
@mergify backport branch-2.2
|
gharchive/pull-request
| 2023-02-10T08:03:09 |
2025-04-01T04:55:39.512671
|
{
"authors": [
"hellolilyliuyi"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/17665",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1807936950
|
[BugFix] disable dictionary optimize stream load in branch-23
hard to cp #26463 so we invalid dictionary when streaming load
[FE PR Coverage Check]
:heart_eyes: pass : 1 / 1 (100.00%)
file detail
path
covered_line
new_line
coverage
not_covered_line_detail
:large_blue_circle:
com/starrocks/common/Config.java
1
1
100.00%
[]
|
gharchive/pull-request
| 2023-07-17T14:39:08 |
2025-04-01T04:55:39.515862
|
{
"authors": [
"stdpain",
"wanpengfei-git"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/27395",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2013575236
|
[Enhancement] support mysql key words in where caluse
Why I'm doing:
What I'm doing:
support mysql key words in where caluse
Fixes #issue
What type of PR is this:
[ ] BugFix
[ ] Feature
[x] Enhancement
[ ] Refactor
[ ] UT
[ ] Doc
[ ] Tool
Does this PR entail a change in behavior?
[ ] Yes, this PR will result in a change in behavior.
[x] No, this PR will not result in a change in behavior.
If yes, please specify the type of change:
[ ] Interface/UI changes: syntax, type conversion, expression evaluation, display information
[ ] Parameter changes: default values, similar parameters but with different default values
[ ] Policy changes: use new policy to replace old one, functionality automatically enabled
[ ] Feature removed
[ ] Miscellaneous: upgrade & downgrade compatibility, etc.
Checklist:
[x] I have added test cases for my bug fix or my new feature
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto-backported to the target branch
[x] 3.2
[x] 3.1
[x] 3.0
[x] 2.5
https://github.com/Mergifyio backport branch-2.5
|
gharchive/pull-request
| 2023-11-28T04:05:02 |
2025-04-01T04:55:39.522903
|
{
"authors": [
"zombee0"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/35917",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2158536165
|
adding fpe function
Why I'm doing:
What I'm doing:
Fixes #issue
What type of PR is this:
[ ] BugFix
[x] Feature
[ ] Enhancement
[ ] Refactor
[ ] UT
[ ] Doc
[ ] Tool
Does this PR entail a change in behavior?
[x] Yes, this PR will result in a change in behavior.
[ ] No, this PR will not result in a change in behavior.
If yes, please specify the type of change:
[ ] Interface/UI changes: syntax, type conversion, expression evaluation, display information
[ ] Parameter changes: default values, similar parameters but with different default values
[ ] Policy changes: use new policy to replace old one, functionality automatically enabled
[ ] Feature removed
[ ] Miscellaneous: upgrade & downgrade compatibility, etc.
Checklist:
[ ] I have added test cases for my bug fix or my new feature
[ ] This pr needs user documentation (for new or modified features or behaviors)
[ ] I have added documentation for my new feature or new function
[ ] This is a backport pr
Bugfix cherry-pick branch check:
[x] I have checked the version labels which the pr will be auto-backported to the target branch
[x] 3.2
[ ] 3.1
[ ] 3.0
[ ] 2.5
better to add docs for these functiions, you can refer to: https://github.com/StarRocks/starrocks/tree/main/docs/en/sql-reference/sql-functions
@mergify rebase
|
gharchive/pull-request
| 2024-02-28T09:46:39 |
2025-04-01T04:55:39.530423
|
{
"authors": [
"before-Sunrise",
"liubotao",
"silverbullet233"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/41865",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1293056803
|
jvm gc 机制由cms修改成g1
What type of PR is this:
[ ] bug
[ ] feature
[x] enhancement
[ ] refactor
[ ] others
Which issues of this PR fixes :
Fixes #
Problem Summary(Required) :
#fe.conf jvm配置中gc所使用的cms机制运行效率和性能实测低于g1,cms优化为g1,生产环境实测性能更好,full gc更少。
Thank you for your contributions. Could you give more performance information?
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
Can you make it configurable? We can use cms or g1 according to a config item
|
gharchive/pull-request
| 2022-07-04T11:45:54 |
2025-04-01T04:55:39.534874
|
{
"authors": [
"CLAassistant",
"chaoyli",
"gengjun-git",
"sfwanyi"
],
"repo": "StarRocks/starrocks",
"url": "https://github.com/StarRocks/starrocks/pull/8245",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
2487534832
|
fix(item): Item attachment mods use parent item mod type
This PR fixes mods added on the fly to item attachments having no type defined, and only allowing basic mods instead of weapon/armour/vehicule/all mods.
Changing the item attachment mod type doesn't change the on the fly mods type retroactively, so we could also always set these to "all" instead. If so I'll update the PR.
Closes #1624
Nope, LGTM.
|
gharchive/pull-request
| 2024-08-26T18:49:26 |
2025-04-01T04:55:39.536220
|
{
"authors": [
"obrenckle",
"wrycu"
],
"repo": "StarWarsFoundryVTT/StarWarsFFG",
"url": "https://github.com/StarWarsFoundryVTT/StarWarsFFG/pull/1695",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
204438180
|
Do not wrap requests that skip the request filter
Fixes issue #20.
"Sign out" button in the sign in app uses a flag HandlerOptions.SkipRequestFilters. Website app uses a response filter though, and this flag does not skip that one.
In result, Website app wraps a request that should not be wrapped.
The solution requires a custom HTTP header applied in the request middleware that will force skipping of the response middleware.
This is the same fix as https://github.com/StarcounterPrefabs/Launcher/commit/39da52caab8b6f9275d05d97d0e4cac6794d71d0
I consider this to be a temporary solution until there is a flag SkipMiddleware, which skips both the request and the response filter.
@chrhol, @diliger can you pls review?
Closing as done, thanks!
|
gharchive/pull-request
| 2017-01-31T22:07:40 |
2025-04-01T04:55:39.543693
|
{
"authors": [
"warpech"
],
"repo": "StarcounterPrefabs/Website",
"url": "https://github.com/StarcounterPrefabs/Website/pull/21",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1888678978
|
Ng+ Universe roll seed bug
Platform: PC (Steam)
Time/date of bug encountered: September 6th 2023
When starting new game plus, in certain circumstances you'll never be able to save or open your inventory.
Details
You have a chance to get a universe seed where the lodge gets alternative versions where andreja for example could show up with cultist and kill you, and in this universe you'll never be able to save your game or access your inventory. I'm noticing this happening with each unique new game universe unless it's a vanilla new game plus seeds with no changes to faction NPCs or the lodge
Steps ive done to troubleshoot : validated steam files, full clean install ( made sure to delete app data stuff too ) and even started a new game to try to see if my game was just being silly, but the issue persist. Its preventing me from experiencing new game plus content that id otherwise would be able to experience if this bug was interfering with saving.
Solution for now :
There are specific console commands that allow you to start these secret seeds , however it's not ideal as it ruins the the surprise elements of discovering these. Solution found here : https://reddit.com/r/Starfield/s/Xhuht9OiDN
Game Version
v1.7.23.0
New Game
Yes
Reproduction Steps
Start a new game, either new game plus or regular new game, finish the main storyline, go to the unity and make a hard save. Then keep entering the unity, get reborn, and then when your in your ship in the beginning of new game plus, check to see if your inventory is not available to open and you get a popup for this, and then check to see if you can't quicksave or save. If confirmed, go to the lodge and let the special scene play out.
These are fully fleshed out, an example being if the hunter kills everyone , you can go to the Eye and you get specific dialogue to use it. It seems this wasn't intentional as the hunter even says " I'm sure this isn't the first time you've seen the lodge like this "
I think this is not a bug, but an intentional game feature.
The only thing the ruins things is not being able to save or open your inventory that makes me believe it's a bug.
@juliocoliveira what is your source for this being an intentional game feature? This appears to be a bug as some users are able to load the same NG+ variant and not experience this issue. For example I experienced the issue on the lodge "sarah is a plant" variant and was unable to save or load inventory. However, other users have reported getting this seed variation and being able to save and use player inventory. Some users have also been able to recover from this by opening ship inventory and switching to player or opening console and click the OK dialogue to the warning.
There are also variations that require the user to loot and read a special item, which is blocked by being unable to access inventory.
Do you have a source with Bethesda that has confirmed this is a feature and not a bug?
Zero source, as I said "I THINK", never said that was an absolute truth right from Bethesda developers.
But some people think that is intentional, a random layer of difficulty, an universe that you cannot save or use your inventory,
It's likely a bug with the quest stages or a variable ( like how certain cutscenes and other variables don't allow you to save or access your inventory in certain moments in the game that I won't spoil ) which could easily be fixed in CK when it releases or by Bethesda. This is fixable currently only with use of console commands.
This was fixed in v1.8.86
"Fixed an issue that could temporarily prevent opening the inventory or saving after entering the Unity."
|
gharchive/issue
| 2023-09-09T10:10:00 |
2025-04-01T04:55:39.551643
|
{
"authors": [
"Aurelianis",
"Constellation-VASCO",
"juliocoliveira",
"spkyghzt"
],
"repo": "Starfield-Community-Patch/Starfield-Community-Patch",
"url": "https://github.com/Starfield-Community-Patch/Starfield-Community-Patch/issues/106",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2014994502
|
Quest MQ105 All That Money Can Buy fails to start
Quest MQ105 All That Money Can Buy fails to start if NPC Ryleigh is dead before quest start
Details
Quest ID: 2C1C9B
Game Version
1.8.86.0
Language
English (EN)
New Game
Yes
Reproduction Steps
Start new game, proceed until player can free travel
Travel to Volii Alpha > Neon City > Slayton Aerospace
Kill Ryleigh or kill everyone in the building
Proceed with main quest until the start of MQ105 "All That Money Can Buy"
Note that the quest won't start
Duplicate of #696
|
gharchive/issue
| 2023-11-28T17:55:03 |
2025-04-01T04:55:39.554904
|
{
"authors": [
"Constellation-VASCO",
"Pickysaurus"
],
"repo": "Starfield-Community-Patch/Starfield-Community-Patch",
"url": "https://github.com/Starfield-Community-Patch/Starfield-Community-Patch/issues/695",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2467866780
|
[Bug] failed to load rendersystemdx11.dll CS2
Before Reporting
[X] I found no existing issues matching my bug
[x] My issue is not caused by the Millennium interface inside Steam (Themes, Plugins & Update Panel). If it is, make a issue report in this repo.
[X] My issue is not caused by a theme
Describe the Bug
The game can only be launched via a shortcut in Steam, and this error appears when starting Dota 2.
Can it happend with actual AMD driver?
Latest driver from old system named 19.x.x.
Expected Behavior
All drivers, system, VCRedist, has been updated, and much more, But nothing gonna work, This starts after reinstalling windows.
System has been downloaded from official microsoft website.
Before reinstallation, everything worked without errors.
Steps To Reproduce
This starts after reinstalling windows.
System has been downloaded from official microsoft website.
Before reinstallation, everything worked without errors.
Can it happend with actual AMD driver?
Latest driver from old system named 19.x.x.
Operating System
Windows
Anything else?
Why are you inclined to believe this issue stems from Millennium? This seems unrelated entirely.
|
gharchive/issue
| 2024-08-15T11:17:01 |
2025-04-01T04:55:39.593588
|
{
"authors": [
"ShadowMonster99",
"uppofig"
],
"repo": "SteamClientHomebrew/Millennium",
"url": "https://github.com/SteamClientHomebrew/Millennium/issues/78",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1872026677
|
Add Soft Engine 2
SteamDB app page links to a few games using this
https://steamdb.info/app/348470/
https://steamdb.info/app/350310/
https://steamdb.info/app/363130/
https://steamdb.info/app/363150/
https://steamdb.info/app/485510/
https://steamdb.info/app/502280/
https://steamdb.info/app/527280/
https://steamdb.info/app/551730/
https://steamdb.info/app/595740/
https://steamdb.info/app/601051/
https://steamdb.info/app/628890/
https://steamdb.info/app/636570/
https://steamdb.info/app/658260/
https://steamdb.info/app/730310/
https://steamdb.info/app/756590/
https://steamdb.info/app/831560/
https://steamdb.info/app/886820/
https://steamdb.info/app/936160/
https://steamdb.info/app/936180/
https://steamdb.info/app/936190/
https://steamdb.info/app/968790/
https://steamdb.info/app/1045620/
https://steamdb.info/app/1072420/
https://steamdb.info/app/1121560/
https://steamdb.info/app/1152300/
https://steamdb.info/app/1152310/
https://steamdb.info/app/1152320/
https://steamdb.info/app/1233260/
https://steamdb.info/app/1257290/
https://steamdb.info/app/1325200/
https://steamdb.info/app/1341200/
https://steamdb.info/app/1358701/
https://steamdb.info/app/1423600/
https://steamdb.info/app/1502980/
https://steamdb.info/app/1502990/
https://steamdb.info/app/1621310/
https://steamdb.info/app/1732190/
https://steamdb.info/app/1999771/
https://steamdb.info/app/2130460/
https://steamdb.info/app/2138090/
Brief explanation of the change
Support Soft Engine 2 by Koei Tecmo.
After running changed rules in this pull request (this bot only tests rules.ini changes):
Changed technologies: Engine.SoftEngine2
These apps now match: (60) 211140 322460 331600 348470 350310 363110 363130 363150 401870 410850 449800 485510 502280 526580 527280 551730 574050 595740 601050 628080 628890 636570 730310 756590 831560 886820 936160 936180 936190 968790 1019100 1019870 1045620 1072420 1121560 1152300 1152310 1152320 1168460 1233260 1257290 1325200 1341200 1358700 1423600 1424800 1502970 1502980 1502990 1582390 1621310 1732190 1842810 1879330 1999770 2124680 2130460 2138090 2271770 2743610
Please confirm that all these changed apps are correct.
|
gharchive/pull-request
| 2023-08-29T16:33:32 |
2025-04-01T04:55:39.612737
|
{
"authors": [
"MichaelHinrichs",
"SteamTracker"
],
"repo": "SteamDatabase/FileDetectionRuleSets",
"url": "https://github.com/SteamDatabase/FileDetectionRuleSets/pull/352",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2542986375
|
Update Steeltoe to build and run tests against .NET 9 RC1
Description
Experimental and hacky, do not merge.
Connector failures on MySQL Connector are to be expected. The Oracle package doesn't provide anything for .NET 9 yet, while the Pomelo package is still at Preview1 (RC support is planned).
Blocked by https://github.com/SonarSource/sonar-dotnet/issues/9670.
/azp run Steeltoe.All
|
gharchive/pull-request
| 2024-09-23T15:20:22 |
2025-04-01T04:55:39.639992
|
{
"authors": [
"bart-vmware"
],
"repo": "SteeltoeOSS/Steeltoe",
"url": "https://github.com/SteeltoeOSS/Steeltoe/pull/1368",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
217727095
|
Fixes Firefox Extension 'Unknown Error'
Fixes #7
Due to the latest version of Firefox blocking API requests, the fetching is now done in a background script.
4eb7a19 should resolve those issues, thanks! Do you happen to know what exactly changed that caused fetching cross domain to error out? The fetch errors are so vague...
|
gharchive/pull-request
| 2017-03-28T23:59:22 |
2025-04-01T04:55:39.687583
|
{
"authors": [
"Step7750"
],
"repo": "Step7750/CSGOFloat-Extension",
"url": "https://github.com/Step7750/CSGOFloat-Extension/pull/8",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
62199327
|
Feature request: add JASC-PAL palette support to rescomp
Here is the format description :
.pal Paint Shop Pro colour palette file:
Contains the colour palette table saved from a 16 or 256 colour image. It can
be loaded for another image and this provides a means of copying colour
palettes from one image to another.
These files are text editable. The format of the file is as follows:
- The first line contains the header "JASC-PAL" (without the quotes).
- The second line contains the palette file version. For example "0100"
(without the quotes)
- The third line contains the number of colours, either 16 or 256.
- Subsequent lines contain the Red, Green and Blue components of each palette
entry, with a space between each of the Red, Green and Blue numbers. These
numbers are in decimal, so each attribute should be in the range 0 to 255.
Original issue reported on code.google.com by Stephane.D on 16 Jun 2014 at 8:54
Title should be changed to Add *.pal files support.
I'm using GraphicGale Free Edition and the format is the same I guess.
Here is an example for a 4bpp 16colors palette:
Thanks for the example, i have also 2 files sitting there so i can make some tests with them.
I will try to add that for the next version of rescomp :)
ty
|
gharchive/issue
| 2015-03-16T20:22:34 |
2025-04-01T04:55:39.693499
|
{
"authors": [
"GoogleCodeExporter",
"Stephane-D",
"derainm",
"nolddor"
],
"repo": "Stephane-D/SGDK",
"url": "https://github.com/Stephane-D/SGDK/issues/23",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1439647709
|
🛑 DNS (he.net) is down
In abde44a, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in bc39c11.
|
gharchive/issue
| 2022-11-08T07:34:38 |
2025-04-01T04:55:39.718690
|
{
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/1322",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1626908596
|
🛑 DNS (he.net) is down
In b47c177, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in 3b30d5b.
|
gharchive/issue
| 2023-03-16T07:55:44 |
2025-04-01T04:55:39.721536
|
{
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/2562",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1669319161
|
🛑 DNS (he.net) is down
In 98bd8d5, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in bc11a6b.
|
gharchive/issue
| 2023-04-15T11:31:42 |
2025-04-01T04:55:39.724357
|
{
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/5135",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1713399193
|
🛑 DNS (he.net) is down
In 6ac1e54, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in 631683f.
|
gharchive/issue
| 2023-05-17T08:28:44 |
2025-04-01T04:55:39.726935
|
{
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/7594",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1722882985
|
🛑 DNS (he.net) is down
In 94208a5, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in 727e137.
|
gharchive/issue
| 2023-05-23T22:37:45 |
2025-04-01T04:55:39.729519
|
{
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/8160",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1739892255
|
🛑 DNS (he.net) is down
In 62a8065, DNS (he.net) ($HE_NS) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DNS (he.net) is back up in db1336c.
|
gharchive/issue
| 2023-06-03T20:43:46 |
2025-04-01T04:55:39.732598
|
{
"authors": [
"leitmori"
],
"repo": "Sternwarte-St-Ottilien-e-V/status",
"url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/9064",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
161070258
|
ClientPlayerPlaceBlockPacket with null block face
Passing a null block face to ClientPlayerPlaceBlockPacket to signal the "special" face (255) currently produces an error.
#182
|
gharchive/issue
| 2016-06-19T12:15:09 |
2025-04-01T04:55:39.741236
|
{
"authors": [
"finalchild",
"yawkat"
],
"repo": "Steveice10/MCProtocolLib",
"url": "https://github.com/Steveice10/MCProtocolLib/issues/181",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
678091265
|
The latest model has poor generalization effect
I tested the latest age and gender estimation model on my own private test set with the test script you gave, and found that the effect of age and gender estimation is poor, especially that gender estimation is easy to make mistakes. Have you ever tested it on private test videos? Can you tell me what the problem is?Could it be a problem with tensorflow version
I tested the latest age and gender estimation model on my own private test set with the test script you gave, and found that the effect of age and gender estimation is poor, especially that gender estimation is easy to make mistakes. Have you ever tested it on private test videos? Can you tell me what the problem is?Could it be a problem with tensorflow version
you should add "--white" as a params
my command is:
python nets/test.py -g -vid -se -white
I use this command to test, and the effect is still not good
I use a video of a child, but the result is in my twenties
my command is:
python nets/test.py -g -vid -se -white
I use this command to test, and the effect is still not good.
can you provide your full command line. I will valid it~.
my command is:
python nets/test.py -g -vid -se -white -m ./model/c3ae_model_v2_fp16_white_se_132_4.208622-0.973
I use a video of a child, but the result is in my twenties
if you want detect children, you'd better finetuning in "asia dataset" only. BCZ distribution of utk and afad range from 18~100 and asia dataset range from 0~70.
OK, I'll do it. Thank you very much for your help
python nets/test.py -g -vid -se -white -m ./model/c3ae_model_v2_fp16_white_se_132_4.208622-0.973
Can you provide your pic to geekpeakspar@gmail.com. I will test it for you.
I have test many pic of children from baidu, it works normal. you should guarante pic that should be front face and size great than 64 * 64. If you want get a more accurate result, you should build your own dataset.
|
gharchive/issue
| 2020-08-13T02:00:24 |
2025-04-01T04:55:39.747217
|
{
"authors": [
"StevenBanama",
"yanyanlimingkun"
],
"repo": "StevenBanama/C3AE",
"url": "https://github.com/StevenBanama/C3AE/issues/20",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
642408765
|
Unblock domains
constantcontact.com (with all subdomains)
This is a domain that sends newsletters and being used by legit organizations.
ecampaigns.cineplex.com
Newsletters from a trusted company.
Please unblock these domains/subdomains.
Hi @YogiJB thanks for this.
As I research this, I think you need to know there are no "legit organizations" in this space. The name-brand attached to a domain counts for exactly zero. The only thing that matters is, does a domain track or trade its users? Does it serve ads? It it a path for malware?
Understad what I mean?
@YogiJB the domain constantcontact.com comes to us via Awaway.
Ping @jawz101, what do you think?
@YogiJB take note: we don't list ecampaigns.cineplex.com. We don't list any cineplex domain.
removing. Purely because the Adaway list is meant for traffic that comes from mobile apps (i.e. not email marketing, browser, or desktop app traffic.)
If I see those domains again I'll review it later
https://github.com/constantcontact/java-sdk
Thanks Jawz! @jawz101.
@YogiJB this will go into wider distribution for the next release, soon.
Closing.
|
gharchive/issue
| 2020-06-20T16:47:57 |
2025-04-01T04:55:39.752004
|
{
"authors": [
"StevenBlack",
"YogiJB",
"jawz101"
],
"repo": "StevenBlack/hosts",
"url": "https://github.com/StevenBlack/hosts/issues/1316",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
243360121
|
*.ypu.edu.tw
This domain is Taiwan's University (Yuanpei University of Medical Technology)
There's only ypu.edu.tw
What do you think about this @FadeMind ?
I was saw only "### Extra rules for @StevenBlack 's hosts project", so I report to you :-P
I just curious why university's website was added to hosts list.
You were right @pinomeng ;)
I can simply whitelisting this domain if needed.
This domain looks OK to me @FadeMind
@StevenBlack unreachable per browser by me but reachable with nslookup ...
@StevenBlack fixed in https://github.com/FadeMind/hosts.extras/commit/346e219d768f98cfb27c4ec34616ed8502e281f0
Thanks Tomasz @FadeMind and Pino @pinomeng. New hosts file is live and online now.
|
gharchive/issue
| 2017-07-17T11:01:16 |
2025-04-01T04:55:39.755671
|
{
"authors": [
"FadeMind",
"StevenBlack",
"funilrys",
"pinomeng"
],
"repo": "StevenBlack/hosts",
"url": "https://github.com/StevenBlack/hosts/issues/362",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1634301587
|
jdownloader2 Server LXC
Hi,
wenn ich mir den Container installiere, wie komme ich dann auf die Oberfläche von jDownloader?
Gruß
Bitte Issues auf Englisch verfassen
Siehe https://github.com/StevenSeifried/proxmox-scripts/issues/12#issuecomment-1356144716
|
gharchive/issue
| 2023-03-21T16:33:23 |
2025-04-01T04:55:39.759910
|
{
"authors": [
"Azzenator",
"StevenSeifried"
],
"repo": "StevenSeifried/proxmox-scripts",
"url": "https://github.com/StevenSeifried/proxmox-scripts/issues/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
101306790
|
Unhandled Error
Whenver I try to open my stash in Standard League or press on the Trading tab it says There was an unhandled error
On debug info says dis :
[16-08-2015 23:26] System.ArgumentOutOfRangeException: Length cannot be less than zero.
Parameter name: length
Hi,
Do your stashes in other leagues work?
Hey, Yeh the warbands league stash loads just fine but Standard will just cause it to crash
On 17 Aug 2015, at 00:02, Stickymaddness notifications@github.com wrote:
Hi,
Do your stashes in other leagues work?
—
Reply to this email directly or view it on GitHub.
Alright, are you sure your standard stash is up to date? Ie: Login and update standard for all tabs. If you're sure it's up to date, do you have any old race characters in standard or characters from old leagues you haven't logged into for a long time?
I use Session ID to log in but it shows it loading my new stash and the character I havnt been on for the longest is in Warbands.
On 17 Aug 2015, at 16:30, Stickymaddness notifications@github.com wrote:
Alright, are you sure your standard stash is up to date? Ie: Login and update standard for all tabs. If you're sure it's up to date, do you have any old race characters in standard or characters from old leagues you haven't logged into for a long time?
—
Reply to this email directly or view it on GitHub.
I just deleted my warbands character and the standard stash is working fine now but The Trading tab still causes procurement to crash ;(
Sent from my iPhone
On 17 Aug 2015, at 16:30, Stickymaddness notifications@github.com wrote:
Alright, are you sure your standard stash is up to date? Ie: Login and update standard for all tabs. If you're sure it's up to date, do you have any old race characters in standard or characters from old leagues you haven't logged into for a long time?
—
Reply to this email directly or view it on GitHub.
Deleting your warbands character shouldn't have any effect, since the data is stored separately per league.
Delete your tabprefs.xml file if you have one, restart Procurement and see if the trading view is able to load without crashing.
Are you using the latest version of Procurement (1.9.4) ?
I deleted the tabpref file and restarted but same thing happened unfortunately also yes I am using the 1.9.4 version
On 17 Aug 2015, at 16:52, Stickymaddness notifications@github.com wrote:
Deleting your warbands character shouldn't have any effect, since the data is stored separately per league.
Delete your tabprefs.xml file if you have one, restart Procurement and see if the trading view is able to load without crashing.
Are you using the latest version of Procurement (1.9.4) ?
—
Reply to this email directly or view it on GitHub.
Try login to all your standard characters, this sometimes fixes the API returning invalid data for a league.
|
gharchive/issue
| 2015-08-16T22:49:03 |
2025-04-01T04:55:39.771173
|
{
"authors": [
"NeGerai",
"Stickymaddness"
],
"repo": "Stickymaddness/Procurement",
"url": "https://github.com/Stickymaddness/Procurement/issues/171",
"license": "Artistic-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
159369118
|
[CROSSDATA-465] Now it is possible to insert empty arrays and maps
Description
Before this PR, inserting empty arrays or map syntax was rather counter intuitive. Now, it is as easy as passing [] and () values for empty arrays and maps respectively.
Testing
Unit tests has been added to check whether this change has been properly applied.
:+1:
LGTM
|
gharchive/pull-request
| 2016-06-09T09:53:13 |
2025-04-01T04:55:39.818627
|
{
"authors": [
"miguel0afd",
"pfcoperez",
"pianista215"
],
"repo": "Stratio/Crossdata",
"url": "https://github.com/Stratio/Crossdata/pull/572",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
878020220
|
🛑 Nextcloud Papa is down
In 0d40824, Nextcloud Papa ($NEXTCLOUD_2) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Nextcloud Papa is back up in 0234889.
|
gharchive/issue
| 2021-05-06T20:51:39 |
2025-04-01T04:55:39.878322
|
{
"authors": [
"StudFu-WordToMD"
],
"repo": "StudFu-WordToMD/status",
"url": "https://github.com/StudFu-WordToMD/status/issues/1342",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2162665642
|
Is the create and delete of the app intentional in EntryPoint.h?
In EntryPoint.h,
https://github.com/StudioCherno/Walnut/blob/3b8e414fdecfc6c8b58816106fe8d912bd172e31/Walnut/src/Walnut/EntryPoint.h#L10-L20
This code appears to have a logical error, as it creates and deletes a new Walnut::Application object in each iteration of the while loop. This means that the application is constantly being restarted and terminated, which is probably not the intended behavior. A more reasonable approach would be to create the application object once before the loop, and delete it once after the loop, like this:
namespace Walnut {
int Main(int argc, char** argv)
{
Walnut::Application* app = Walnut::CreateApplication(argc, argv);
while (g_ApplicationRunning)
{
app->Run();
}
delete app;
return 0;
}
}
Please correct me if I'm mistaken here, Or maybe there's a good reason for it, I'm very new to this project.
Update: I got courageous and tried it. My solution is... not good. It causes the app to crash when I try to exit/close the window.
|
gharchive/issue
| 2024-03-01T05:56:29 |
2025-04-01T04:55:39.893690
|
{
"authors": [
"glenpierce"
],
"repo": "StudioCherno/Walnut",
"url": "https://github.com/StudioCherno/Walnut/issues/73",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
206474505
|
Unable to pick Google drive file content
Hi,
I need to access a file from google drive, when i select file from google drive it is returning file path (//com.google.android.apps.docs.storage/document/acc%3D4%3Bdoc%3D2602) but i am not able to access that file content.
Regards,
Sachin KS
That Uri isn't a local file path, its a Uri to the data in the google drive content provider. As far as I know, when u pull a file from google drive it doesn't save it to your device, rather it gives you the stream.
|
gharchive/issue
| 2017-02-09T11:15:49 |
2025-04-01T04:55:39.900914
|
{
"authors": [
"mhallmark",
"sachinks1991"
],
"repo": "Studyxnet/FilePicker-Plugin-for-Xamarin-and-Windows",
"url": "https://github.com/Studyxnet/FilePicker-Plugin-for-Xamarin-and-Windows/issues/24",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
593559339
|
Crawl / manually curate / map Publications resources into individual metadata records based off of schema
https://docs.google.com/spreadsheets/d/1FFyRhI5TeUb-B4t50HRZYF_XngRt_mfAIz82UQ3geFk/edit#gid=0
Need to check how often preprints from individual journals (Lancet, Nature, etc.) get incorporated into Pubmed index. Don't want to be redundant, but also want to be as up-to-date as possible.
To start, I think harvesting from LitCovid would be the easiest starting point. My guess is that the PubMed/NCBI team is working hard on getting COVID-19-related literature into PubMed as quickly as possible, so I think we should old off on doing any direct scraping of journal sites. Preprint servers like Biorxiv and Medrxiv would be more useful I think.
So to summarize, I think the priority list should be 1) LitCovid, 2) Biorxiv / Medrxiv, and 3) journal sites.
(not seeing a biorxiv/medrxiv API, but maybe can use their rss feed and/or https://www.rxivist.org/docs)...
Makes sense.
BioRxiv has a third-party API: https://api.rxivist.org/v1/papers?q=nCoV&timeframe=alltime. Little bit wonky; &metric=downloads gives different results. Unclear how frequently is updated. Could also be gotten via scraping https://www.biorxiv.org/search/ncov or Twitter feeds. Medrxiv searches seem to include Biorxiv but bio doesn't include med. There's also Chemrxiv, but lower priority.
plan:
Pull data from LitCOVID/Pubmed: https://ftp.ncbi.nlm.nih.gov/pub/lu/LitCovid/. Use litcovid2BioCJSON.gz to get the IDs, and then use the PubMed API to get the actual records
Scrape Biorxiv
Scripe Medrxiv
Align metadata with schema (#1 )
The mapping from PubMed/Medline to the schema can be found on the worksheet in columns A and J
Additional information to grab from efetch:
example record (not COVID-related): https://eutils.ncbi.nlm.nih.gov/entrez/eutils/efetch.fcgi?db=pubmed&id=28841416&rettype=abstract
MeSH/OT terms --> keywords
abstract --> abstract
author affiliations --> author.affiliation
grants --> funding
referenceList --> citedBy
@gtsueng has additional mapping of PubMed/Medline columns to our schema in columns A/J of the worksheet
Note that there are two methods for pulling the data from pubmed. One method is to use the API which yields an xml file:
https://eutils.ncbi.nlm.nih.gov/entrez/eutils/efetch.fcgi?db=pubmed&id=28841416&rettype=abstract
The benefit of this method is as follows:
-Author given name and family name
-References (Citations)
-Specific grant information
The other method uses the biopython library to fetch and parse a medline record which treats it like a json file. Sample gist here.
The benefit of this method is as follows:
-json
-easy to map (the dictionary keys already mapped to schema fields in the worksheet: Publications, column J)
-PublicationType available as PT via this method
-Other terms aside also available (OT) aside from just MeSH (MH)
Abbreviated Journal titles may work better than the actual journal titles when it comes down to the visualization part. Since there are different ways to abbreviate journal titles, it may be better to use a dictionary for consistency. The NLM list of journal can be found here and parsed into a dictionary. Once parsed, a manual review/edit could help to ensure that the most common abbreviations are rather than just the NLM abbreviations (which are by no means the de facto standard).
Abbreviated Journal titles may work better than the actual journal titles when it comes down to the visualization part. Since there are different ways to abbreviate journal titles, it may be better to use a dictionary for consistency. The NLM list of journal can be found at ftp://ftp.ncbi.nih.gov/pubmed/J_Medline.txt and parsed into a dictionary. Once parsed, a manual review/edit could help to ensure that the most common abbreviations are rather than just the NLM abbreviations (which are by no means the de facto standard).
would propose having journalName be the full journal name (for instance The New England journal of medicine) and something like journalNameAbbrevation be N. Engl. J. Med.
Added into schema as suggested above. Still suggest populating from a single dictionary and caution against loading it from metadata from different sources since they may abbreviate journal titles differently.
referenceList --> citedBy (note: will need to pick an identifier to map to the other publications that cite this publication. ideally, this should refer to the _id we use to index these publications)
@flaneuse In my mind the reference list should map to isBasedOn since those would be explicitly called out/referenced in the text and the citedBy would refer to other publications that cite this publication.
referenceList --> citedBy (note: will need to pick an identifier to map to the other publications that cite this publication. ideally, this should refer to the _id we use to index these publications)
@flaneuse In my mind the reference list should map to isBasedOn since those would be explicitly called out/referenced in the text and the citedBy would refer to other publications that cite this publication.
Oh, sorry, you're right; I'm backwards. I thought these were the citedBy listed in the PubMed entry -- I care more about those than the references they cite. Wonder where these are hidden in the PubMed API...
I just assumed that they could query for those as long as all their other publications have references. In terms of related articles where the relationship is unclear, Medline will sometimes provide them as [AID] (associated ID). These are often related articles. Eg- if Publication A is a letter to the editor with regards to Publication B. An identifier for Publication B would be included as AID in the record for Publication A. Unfortunately, AID is generic enough that they'll sometimes include the DOI for Publication A in the record for Publication A.
@marcodarko can you also track the LitCovid / BioRxiv / MedRxiv original url? And maybe the source type (LitCovid, BioRxiv, or MedRxiv)?
per @andrewsu on the front-end: "assuming we get that from litcovid, would be good to add a note and/or link. Yes, probably somewhat redundant to do it on every page, but would rather err on the side of overattribution..."
I think we want to reserve url for the actual link to the paper (I don't want to get routed through PubMed when I just want to read the paper). I'm not sure I see a natural schema.org way to capture this information... thoughts @gtsueng? Proposal for your reaction (as an aside: we also need to decide on what we're using for canonical IDs; this will be used as the key to link between records, and also for the url to route on outbreak.info):
{ _id: "pmid32292904",
pmid: "32292904",
doi: "10.1016/j.xkme.2020.03.001",
url: "https://www.kidneymedicinejournal.org/article/S2590-0595(20)30049-2/pdf",
sameAs: {
name: "LitCovid",
url: "https://www.ncbi.nlm.nih.gov/research/coronavirus/publication/32292904"
}
}
@marcodarko can you also track the LitCovid / BioRxiv / MedRxiv original url? And maybe the source type (LitCovid, BioRxiv, or MedRxiv)?
per @andrewsu on the front-end: "assuming we get that from litcovid, would be good to add a note and/or link. Yes, probably somewhat redundant to do it on every page, but would rather err on the side of overattribution..."
I think we want to reserve url for the actual link to the paper (I don't want to get routed through PubMed when I just want to read the paper). I'm not sure I see a natural schema.org way to capture this information... thoughts @gtsueng? Proposal for your reaction (as an aside: we also need to decide on what we're using for canonical IDs _id; this will be used as the key to link between records, and also for the url to route on outbreak.info):
{ _id: "pmid32292904",
pmid: "32292904",
doi: "10.1016/j.xkme.2020.03.001",
url: "https://www.kidneymedicinejournal.org/article/S2590-0595(20)30049-2/pdf",
sameAs: {
name: "LitCovid",
url: "https://www.ncbi.nlm.nih.gov/research/coronavirus/publication/32292904"
}
}
update: creating new base class property curatedBy to store this info
Noting source of BioRxiv data
https://github.com/marcodarko/litcovid
https://github.com/marcodarko/biorxiv
|
gharchive/issue
| 2020-04-03T18:20:43 |
2025-04-01T04:55:39.929176
|
{
"authors": [
"andrewsu",
"flaneuse",
"gtsueng"
],
"repo": "SuLab/outbreak.info-resources",
"url": "https://github.com/SuLab/outbreak.info-resources/issues/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1058157882
|
r4s Nps 内网穿透打开不了
r4s Nps 内网穿透打开不了
你应该去这里反馈:https://github.com/coolsnowwolf/lede/issues
|
gharchive/issue
| 2021-11-19T05:39:41 |
2025-04-01T04:55:39.931388
|
{
"authors": [
"AmadeusGhost",
"baoeig"
],
"repo": "SuLingGG/OpenWrt-Rpi",
"url": "https://github.com/SuLingGG/OpenWrt-Rpi/issues/260",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1419923342
|
Wordle Game
I want to add a Wordle Game made using CSS, HTML and Js
@newbie2508 Assign to you
|
gharchive/issue
| 2022-10-23T20:32:00 |
2025-04-01T04:55:39.979061
|
{
"authors": [
"Subhradeep10",
"newbie2508"
],
"repo": "Subhradeep10/Amazing_Web_Projects",
"url": "https://github.com/Subhradeep10/Amazing_Web_Projects/issues/110",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
518041518
|
Testtuples: reference the traintuple instead of referencing the model directly
See also: https://github.com/SubstraFoundation/substra-backend/pull/33
Store a reference to the traintuple instead of a reference to the model. The model can still be obtained by querying the traintuple. This change will make it easier to implement generic test tuples.
In this PR:
Add a new field traintupleKey to Testtuple, outputTesttuple and outputBoardTuple
Delete the model field from Testtuple
Mark the model field from outputTesttuple and outputBoardTuple
We don't delete these fields yet to prevent potential breakages in other parts of the substra stack
Future work:
Delete the model field from outputTesttuple and outputBoardTuple
@Kelvin-M Regarding the extra ledger call, please see the conversation there
Regarding the frontend breaking, I've tested on my local: I generated all kinds of assets and clicked around on everything. I've not managed to see anything broken :) All HTTP calls go through, and no error in the console. I could have missed something though, and I think it would be worth having @jmorel 's opinon.
@AurelienGasser Nice ! Do you test it also with the removal of the Model key/value
@Kelvin-M I did :)
Okay, we wait for jmorel to check if everything is good :)
About the frontend, we always use the traintuple to get model details, we never used the testtuple to do so. I think it's safe to remove from the frontend points of view.
So we can remove Model in this PR :)
What do you think @AurelienGasser & @thibaultrobert ?
@Kelvin-M No objection, other than @inalgnu recommended against it.
@thibaultrobert Need you review for this to be merged !
|
gharchive/pull-request
| 2019-11-05T21:19:21 |
2025-04-01T04:55:40.035180
|
{
"authors": [
"AurelienGasser",
"Kelvin-M",
"jmorel"
],
"repo": "SubstraFoundation/substra-chaincode",
"url": "https://github.com/SubstraFoundation/substra-chaincode/pull/21",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1772907856
|
Close #710, fix ALKilnInThePlayground upload Step
A few quirks here.
This doesn't allow passing in command line arguments for source folder paths yet. That may have to be a separate feature. I'm not sure it should be a required command line argument anyway, in which case we'll still need that code for times when the argument isn't handed in. We could ensure ALKilnInThePlayground hands it in, but I'm not yet sure how we can get ALKilnInThePlayground to get the right paths. Maybe it would have better access to the S3 paths, though.
[Added a fix in the in-Playground interview that makes this point moot. See https://github.com/SuffolkLITLab/docassemble-ALKilnInThePlayground/pull/24] ~It doesn't quite work with S3. Jonathan gave us the info that the /tmp folder for S3 won't keep those files around, so we probably can't count on those files being there for running these tests in an interview~ (ALKilnInThePlayground - now it occurs to me that ALKilnInterview may have been a better name...)
I should have probably broken one change into a different PR and I can do that if this is too much for one PR - The variable setting Step for uploading files turned out to be a mismatch for one ethos we're following. We decided that we want to defer to being radically permissive and forgiving with test code in general. In that spirit, I changed it so that if ALKiln can't find a file that the author listed to upload, the file just isn't uploaded and, if not all variables are required to be set, the test can continue. In a Story Table, the fields are not required to be set and there would be no error just because of a missing file. In a linear Step ("I set the var..."), a missing file would cause an error. I'm not sure about being this permissive, but I've been unsure in similar situations in the past and the collective decision has generally erred on the side of flexibility.
I'll fix the version upon publishing.
Closes #710
Not super simple, but could use a look (if you have time), @BryceStevenWilley.
|
gharchive/pull-request
| 2023-06-24T20:35:47 |
2025-04-01T04:55:40.061044
|
{
"authors": [
"plocket"
],
"repo": "SuffolkLITLab/ALKiln",
"url": "https://github.com/SuffolkLITLab/ALKiln/pull/717",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2445943189
|
ResearchRover Extension
ResearchRover: Gather and summarize web content effortlessly, streamlining your research process with automated insights.
Advanced AI-Powered Research Assistant
@Sulagna-Dutta-Roy, assign me this issue
How do we know when the feature is complete?
I will raise PR
@Sulagna-Dutta-Roy, view PR and let me know if any changes are required
Otherwise, kindly merge it
|
gharchive/issue
| 2024-08-03T01:45:57 |
2025-04-01T04:55:40.064484
|
{
"authors": [
"sreevidya-16"
],
"repo": "Sulagna-Dutta-Roy/GGExtensions",
"url": "https://github.com/Sulagna-Dutta-Roy/GGExtensions/issues/2588",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2748909271
|
Issue filter section niet mobile responsive
Gefixt met commit issue #10 en #11
|
gharchive/issue
| 2024-12-18T23:05:36 |
2025-04-01T04:55:40.072768
|
{
"authors": [
"SuleymanHG"
],
"repo": "SuleymanHG/fix-the-flow-interactive-website",
"url": "https://github.com/SuleymanHG/fix-the-flow-interactive-website/issues/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1479885901
|
🛑 Binding Database is down
In 9b14bfa, Binding Database (http://www.bindingdb.org/bind/index.jsp) was down:
HTTP code: 503
Response time: 3238 ms
Resolved: Binding Database is back up in 632f740.
|
gharchive/issue
| 2022-12-06T18:46:07 |
2025-04-01T04:55:40.075730
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/13722",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1590346202
|
🛑 Chem Exper is down
In 0279933, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in b7074d0.
|
gharchive/issue
| 2023-02-18T13:30:03 |
2025-04-01T04:55:40.078041
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/15803",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1598467119
|
🛑 Chem Exper is down
In 82d8900, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in f509517.
|
gharchive/issue
| 2023-02-24T11:35:55 |
2025-04-01T04:55:40.080817
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/16061",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1633426550
|
🛑 Comparative Toxicogenomics Database is down
In ec85cd3, Comparative Toxicogenomics Database (http://ctdbase.org/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Comparative Toxicogenomics Database is back up in 9f8422a.
|
gharchive/issue
| 2023-03-21T08:26:59 |
2025-04-01T04:55:40.083157
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/17091",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1662804835
|
🛑 NMR Shift Database is down
In e56fc83, NMR Shift Database (https://nmrshiftdb.nmr.uni-koeln.de/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: NMR Shift Database is back up in a6ba423.
|
gharchive/issue
| 2023-04-11T15:56:38 |
2025-04-01T04:55:40.085483
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/18058",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1728906847
|
🛑 Zinc 15 is down
In 33a4eaa, Zinc 15 (https://zinc15.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 15 is back up in 33daed3.
|
gharchive/issue
| 2023-05-27T19:08:56 |
2025-04-01T04:55:40.087840
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/20504",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1852020437
|
🛑 Chem Exper is down
In f6e47df, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 0a28652.
|
gharchive/issue
| 2023-08-15T19:41:21 |
2025-04-01T04:55:40.090327
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/26597",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1879228021
|
🛑 Chem Exper is down
In 514603a, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 873cd42 after 7 minutes.
|
gharchive/issue
| 2023-09-03T20:43:32 |
2025-04-01T04:55:40.092809
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/28209",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2009473685
|
🛑 Chem Exper is down
In c5afd07, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 6fae60a after 9 minutes.
|
gharchive/issue
| 2023-11-24T10:22:21 |
2025-04-01T04:55:40.095085
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/32505",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2034213908
|
🛑 Chem Exper is down
In a2afcee, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in ac3e260 after 16 minutes.
|
gharchive/issue
| 2023-12-10T04:24:06 |
2025-04-01T04:55:40.097376
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/33096",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2061592612
|
🛑 Chem Exper is down
In b4b5402, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 16e4dc4 after 7 minutes.
|
gharchive/issue
| 2024-01-01T14:29:21 |
2025-04-01T04:55:40.099836
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/34248",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2061831844
|
🛑 Chem Exper is down
In 3fe5e8f, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in c28d686 after 42 minutes.
|
gharchive/issue
| 2024-01-01T23:40:20 |
2025-04-01T04:55:40.102143
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/34267",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2108883494
|
🛑 Binding Database is down
In 85f620f, Binding Database (http://www.bindingdb.org/bind/index.jsp) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Binding Database is back up in 25328ce after 19 minutes.
|
gharchive/issue
| 2024-01-30T22:14:19 |
2025-04-01T04:55:40.104658
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/36009",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2254570172
|
🛑 Chem Exper is down
In 8cfafa7, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 3e52206 after 9 minutes.
|
gharchive/issue
| 2024-04-20T14:14:13 |
2025-04-01T04:55:40.106885
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/39348",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2267618805
|
🛑 Chem Exper is down
In e38bf1d, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in fff21eb after 8 minutes.
|
gharchive/issue
| 2024-04-28T14:53:08 |
2025-04-01T04:55:40.109340
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/39749",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2355283361
|
🛑 Chem Exper is down
In 3859362, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 74082b3 after 8 minutes.
|
gharchive/issue
| 2024-06-15T21:26:39 |
2025-04-01T04:55:40.111749
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/41601",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2368905344
|
🛑 Binding Database is down
In 77b7fea, Binding Database (http://www.bindingdb.org/bind/index.jsp) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Binding Database is back up in 1e58a50 after 7 minutes.
|
gharchive/issue
| 2024-06-23T21:55:06 |
2025-04-01T04:55:40.114116
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/42004",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2459451792
|
🛑 Chem Exper is down
In 779282c, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 11d1ed9 after 7 minutes.
|
gharchive/issue
| 2024-08-11T06:43:10 |
2025-04-01T04:55:40.116575
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/44018",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2518832209
|
🛑 Zinc 20 is down
In 1f30461, Zinc 20 (https://zinc20.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 20 is back up in 76eb777 after 51 minutes.
|
gharchive/issue
| 2024-09-11T07:35:14 |
2025-04-01T04:55:40.119083
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/45436",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2556305337
|
🛑 Chem Exper is down
In de66be7, Chem Exper (http://www.chemexper.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Chem Exper is back up in 3190409 after 21 minutes.
|
gharchive/issue
| 2024-09-30T11:02:45 |
2025-04-01T04:55:40.121351
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/46051",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1233566484
|
🛑 Zinc 20 is down
In dea0e30, Zinc 20 (https://zinc20.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 20 is back up in e9a9b9d.
|
gharchive/issue
| 2022-05-12T07:31:33 |
2025-04-01T04:55:40.123593
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/5521",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1335560009
|
🛑 Zinc 15 is down
In 9168a88, Zinc 15 (https://zinc15.docking.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Zinc 15 is back up in 4da5a56.
|
gharchive/issue
| 2022-08-11T07:11:17 |
2025-04-01T04:55:40.125847
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Cheminformatics",
"url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/9287",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2239987325
|
🛑 DermNet - Skin disease atlas is down
In f508e42, DermNet - Skin disease atlas (http://www.dermnet.com/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DermNet - Skin disease atlas is back up in d12a8e3 after 10 minutes.
|
gharchive/issue
| 2024-04-12T11:57:33 |
2025-04-01T04:55:40.128669
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Medical-Informatics",
"url": "https://github.com/Sulstice/Uptime-Medical-Informatics/issues/1478",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1497448067
|
🛑 DRIVE Digital Retinal Images for Vessel Extraction is down
In 6332607, DRIVE Digital Retinal Images for Vessel Extraction (http://www.isi.uu.nl/Research/Databases/DRIVE/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: DRIVE Digital Retinal Images for Vessel Extraction is back up in ecc5f75.
|
gharchive/issue
| 2022-12-14T21:46:56 |
2025-04-01T04:55:40.131028
|
{
"authors": [
"Sulstice"
],
"repo": "Sulstice/Uptime-Medical-Informatics",
"url": "https://github.com/Sulstice/Uptime-Medical-Informatics/issues/149",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1938446160
|
Add SIEM release note 1023-10-11-content
Purpose of this pull request
This pull request (PR) replaces #3195, and adds a SIEM release note on October 11, 2023 that will appear here:
https://help.sumologic.com/release-notes-cse/
Issue number: Replaces #3195
Select the type of change:
[ ] Minor Changes - Typos, formatting, slight revisions
[ ] Update Content - Revisions and updating sections
[x] New Content - New features, sections, pages, tutorials
[ ] Site and Tools - Updates, maintenance, and new packages for the site, Gatsby, React, etc
@stacykor, could you review and approve this PR?
Thanks, @stacykor!
|
gharchive/pull-request
| 2023-10-11T18:24:06 |
2025-04-01T04:55:40.135634
|
{
"authors": [
"jpipkin1"
],
"repo": "SumoLogic/sumologic-documentation",
"url": "https://github.com/SumoLogic/sumologic-documentation/pull/3196",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1007891395
|
refactor(sourceprocessor): remove dead tracing code
Signed-off-by: Dominik Rosiek drosiek@sumologic.com
Updated
|
gharchive/pull-request
| 2021-09-27T09:06:32 |
2025-04-01T04:55:40.144370
|
{
"authors": [
"sumo-drosiek"
],
"repo": "SumoLogic/sumologic-kubernetes-fluentd",
"url": "https://github.com/SumoLogic/sumologic-kubernetes-fluentd/pull/335",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1400881220
|
Admin Console - Dashboard Card , Hamburger menu button does not have icons
Admin Console
Issue 1: Admin dashboard cards do not have icons
Issue 2: The logout button icon is not showing properly
Issue 3: Click on hamburger menu -> menu options should display only Names
Actual Menu button Name - Dashboard Create User
Expected Menu button Name - Create User
For all menu button need to remove dashboard word
Re-tested the admin console card and menu button icons are showing and working expected
Closing this defect as it has been test validated successfully
|
gharchive/issue
| 2022-10-07T09:24:04 |
2025-04-01T04:55:40.157862
|
{
"authors": [
"chetandg123",
"satendrapalsingh"
],
"repo": "Sunbird-cQube/community",
"url": "https://github.com/Sunbird-cQube/community/issues/456",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
258376997
|
Need to set typeface in days name and number
Sir Please help me to set typeface in days name and number shown in calendar view.
Thank you
Hi, you need to set the typeface on the underlying paint object using the setTypeface method. For example, just before this line might be a good place: https://github.com/SundeepK/CompactCalendarView/blob/master/library/src/main/java/com/github/sundeepk/compactcalendarview/CompactCalendarController.java#L185
I think this would be a good enhancement to the library, so I'll mark this as an enhancement ro mbe done sometime in the future.
@SundeepK Sorry for my IntelliJ's auto formatting of the file :)
|
gharchive/issue
| 2017-09-18T06:03:52 |
2025-04-01T04:55:40.161023
|
{
"authors": [
"SundeepK",
"dheerajjaiswal08",
"sofakingforever"
],
"repo": "SundeepK/CompactCalendarView",
"url": "https://github.com/SundeepK/CompactCalendarView/issues/238",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
228582005
|
Invalid Response object: Value of property Reason must be of type String
https://github.com/SungardAS/lambda-formation/blob/master/lib/util/done.js#L19
This needs to be converted to a string as err is an object.
route53.changeTagsForResource(params, function(err, data) {
if (err) console.log(err, err.stack); // an error occurred
else console.log(data); // successful response
});
http://docs.aws.amazon.com/AWSJavaScriptSDK/latest/AWS/Route53.html#createHostedZone-property
I changed it to .Reason = JSON.stringify(err)
New process will check if err is a string, then if it is an object with a message key (like Error objects) and lastly use JSON.stringify
@kmcgrath Great, thanks for the quick turn-around.
|
gharchive/issue
| 2017-05-15T00:34:01 |
2025-04-01T04:55:40.169894
|
{
"authors": [
"estahn",
"kmcgrath"
],
"repo": "SungardAS/lambda-formation",
"url": "https://github.com/SungardAS/lambda-formation/issues/8",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
663944556
|
Motion notification
Hi guys, motion notification works, if I use only esp32cam?
Only DCS 960L knows ftp! Da huta C46 not supported ftp. This is new config pls chek.
Assuming that 192.168.1.40 is where your Homebridge is running, the only issue I see is that ‘Path’ on your camera should be ‘Nappali’.
it is my synology nas 192.168.1.40 patch is my nas file system home-kamera-napalli. When I tested dcs settings succsessfull save test image.
This plugin requires that you upload images to it. If you’re using the FTP upload feature of your camera to save images already, you’ll need to either need to wait for me to write support for that into this plugin, or wait until I finish the SMTP plugin so that you can use the Email ability of your camera instead.
Ok. I will be wait. Thx
I've added a plugin for using the email functionality, if you aren't currently using that feature of the camera, maybe this would be a better fit: https://github.com/Sunoo/homebridge-smtp-motion
FTP upload support has been added to this plugin. The information on configuration is on the main page.
I now sent new ftp motion settings for new version. I will see how to work...
|
gharchive/issue
| 2020-07-22T00:19:58 |
2025-04-01T04:55:40.174769
|
{
"authors": [
"Sunoo",
"clrmsc",
"felho001"
],
"repo": "Sunoo/homebridge-ftp-motion",
"url": "https://github.com/Sunoo/homebridge-ftp-motion/issues/3",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
}
|
2067373722
|
[LORA] Use SuperDuperDB and deepspeed to run local multi GPU training
Support data-parallel and model-parallel within machine.
[ ] Integrate deepspeed bindings
[ ] Test on a 4xt4 GPU on AWS
[ ] Create small integration test (optional - not to be run in the CI/ CD)
Python cannot be used directly as the launcher, otherwise the deepspeed distributed training capability cannot be used, which means it cannot be used normally in jupyter and tot very compatible with our current separable design of computing backends.
If we use deepspeed to start the program, we need to consider other issues such as database reading in multi-threading and interaction between db and model.
But it can be run through ray, so do ray+deepspeed first.
|
gharchive/issue
| 2024-01-05T13:42:01 |
2025-04-01T04:55:40.181824
|
{
"authors": [
"blythed",
"jieguangzhou"
],
"repo": "SuperDuperDB/superduperdb",
"url": "https://github.com/SuperDuperDB/superduperdb/issues/1646",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
395615520
|
Elastix sometimes hangs when SampleRegionSize smaller than 20
I am trying to registering 2 images of size 256x256x35 and am currently optimizing the hyperparameters for non-rigid registration. For registration I am using a mask to focus the registration on the area of interest.
During hyperoptimization I noticed that for some images Elastix will hang when SampleRegionSize is smaller than 20.
I have attached the log file (elastix_hangs.log) for a case where it hangs using (SampleRegionSize 15),
a log file (elastix_doesnt_hang1.log) for the same case where it doesn't hang using (SampleRegionSize 20),
a log file (elastix_doesnt_hang2.log) for a different image pair where it doesn't hang using (SampleRegionSize 15).
elastix_hangs.log
elastix_doesnt_hang1.log
elastix_doesnt_hang2.log
I was wondering, is this expected behaviour? Also, the log's don't report any errors I'm curious what the underlying reason for this hang could be?
I guess that elastix was not able to get a sufficient number of samples from such a small region.
elastix should never hang though, so this is a bug.
But note that this sample region should not be too small anyway, since it is meant to battle inhomogeneities on a bit larger scale.
Interestingly my registrations get better at somewhat smaller sample region sizes, so this would be interesting to investigate further. Lowering the number of samples to be drawn doesn't seem to fix the problem of hanging though. Do you have any other suggestions on how to circumvent this bug?
That would be debugging the source code
|
gharchive/issue
| 2019-01-03T15:32:38 |
2025-04-01T04:55:40.186227
|
{
"authors": [
"mstaring",
"nwschurink"
],
"repo": "SuperElastix/elastix",
"url": "https://github.com/SuperElastix/elastix/issues/93",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
1638939633
|
BUG: Only use InitialTransformParameterObject during first registration
When ElastixRegistrationMethod::GenerateData() does a sequence of multiple registrations, only the first one should use the parameter maps from the InitialTransformParameterObject. Right? 😃
Follow-up to pull request https://github.com/SuperElastix/elastix/pull/856 commit 48c64583c8514eac7fa9ad46848be9d100ac80c9 "Add SetInitialTransformParameterObject to ElastixRegistrationMethod"
of course the subsequent registrations should use the combination of the initial transform and the prior registrations as a combined initial transform for itself...
Did you do extensive testing here?
Did you do extensive testing here?
@mstaring Thanks for your suggestion. I just added a unit test for this case:
pull request #860
Then I realized that elx::TransformBase::BeforeRegistrationBase() simply ignores any initial parameter map or initial parameter file, when an initial transform object (m_Elastix->GetInitialTransform()) is specified. It only looks for initial parameter maps or initial parameter files in the else part of this if:
https://github.com/SuperElastix/elastix/blob/48c64583c8514eac7fa9ad46848be9d100ac80c9/Core/ComponentBaseClasses/elxTransformBase.hxx#L173
So the commit of this PR not really a bug fix, it's just a style improvement 😃
|
gharchive/pull-request
| 2023-03-24T08:24:09 |
2025-04-01T04:55:40.190655
|
{
"authors": [
"N-Dekker",
"mstaring"
],
"repo": "SuperElastix/elastix",
"url": "https://github.com/SuperElastix/elastix/pull/857",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2583681515
|
Add Home Button to Contact Us page
This pull request adds a "Home" button to the top right corner of the webpage. The button is styled to match the existing theme and includes functionality to redirect users to the homepage when clicked.
Changes Made:
Positioned the button absolutely in the top right corner.
Implemented click functionality to navigate to https://imagine-ai-17zf.vercel.app/.
Benefits:
Provides users with easy access to return to the homepage.
Enhances the overall user experience by improving navigation.
Preview:
https://github.com/user-attachments/assets/107ea695-eec7-4d59-9c3a-7fa809a93a0b
fixes: #1218
The button is not clickable and doesn't do anything. Check on the preview site via vercel bot
The button is not clickable and doesn't do anything. Check on the preview site via vercel bot
I'll fix it.
|
gharchive/pull-request
| 2024-10-13T05:53:41 |
2025-04-01T04:55:40.241238
|
{
"authors": [
"SurajPratap10",
"neeru24"
],
"repo": "SurajPratap10/Imagine_AI",
"url": "https://github.com/SurajPratap10/Imagine_AI/pull/1221",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2065329046
|
🛑 swiftfiddle.com/runner/nightly-main is down
In 467a6f9, swiftfiddle.com/runner/nightly-main (https://swiftfiddle.com/runner/nightly-main/run) was down:
HTTP code: 0
Response time: 0 ms
Resolved: swiftfiddle.com/runner/nightly-main is back up in 66ebd45 after 11 hours, 59 minutes.
|
gharchive/issue
| 2024-01-04T09:47:43 |
2025-04-01T04:55:40.268260
|
{
"authors": [
"kishikawakatsumi"
],
"repo": "SwiftFiddle/status.swiftfiddle.com",
"url": "https://github.com/SwiftFiddle/status.swiftfiddle.com/issues/518",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2112764525
|
🛑 swiftfiddle.com/runner/nightly-main is down
In cb4c1d2, swiftfiddle.com/runner/nightly-main (https://swiftfiddle.com/runner/nightly-main/run) was down:
HTTP code: 502
Response time: 23 ms
Resolved: swiftfiddle.com/runner/nightly-main is back up in e71ed81 after 1 hour, 5 minutes.
|
gharchive/issue
| 2024-02-01T15:13:28 |
2025-04-01T04:55:40.271826
|
{
"authors": [
"kishikawakatsumi"
],
"repo": "SwiftFiddle/status.swiftfiddle.com",
"url": "https://github.com/SwiftFiddle/status.swiftfiddle.com/issues/774",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1074079882
|
Add and populate Package.platformCompatibility
As discussed here, add a new field to Package where we roll up the platform compatibility from the build matrix.
Updating Package.platformCompatibility when we receive build reports will take care of keeping the field up-to-date going forward. Letting the service gradually update via incoming builds will not fill gaps for projects that aren't actively maintained - we need to backfill existing packages.
We have the following options:
run the some update sql as a one-off
backfill as part of the migration
Option 1 isn't great, because running ad hoc SQL isn't great. The good thing is the query can be run in small badges to limit impact and we can obviously test it on staging. I've run it locally on a copy and it should be fine.
Option 2 would be nice to tie it into the rollout but this would be a quite long-running migration (some minutes at least). Would be tricky to re-run and we obviously can't batch it.
I'd say we merge the PR with the schema change and the continuous update mechanism first and then run the manual SQL on staging for testing.
Before we deploy any of this to production we should make sure the search extension work as expected. That means between the first part of this landing in main and completing the whole task we shouldn't make any deployments. (Or maybe we should deploy from a branch on staging while we integrate everything - but that's a detail.)
The update SQL is as follows:
UPDATE packages p SET platform_compatibility = ARRAY(
SELECT
CASE
WHEN b.platform LIKE 'macos-%' THEN 'macos'
ELSE b.platform
END
FROM versions v
JOIN builds b ON b.version_id = v.id
WHERE v.package_id = p.id
AND v.latest IS NOT NULL
AND b.status = 'ok'
GROUP BY b.platform
HAVING count(*) > 0
),
updated_at = NOW()
WHERE p.id IN (
SELECT id FROM packages
-- adjust the following parameters as needed
WHERE updated_at < '2021-12-08 12:27:05.82166+00'
LIMIT 1
);
The query syntax for search using the new column is as follows:
select * from packages where platform_compatibility @> '{linux,macos}';
With the platform array it will be easy to support search for single or even multiple platforms, i.e. platform:linux, platform:linux,macos.
We have 1748 packages supporting Linux - 40%! - which is more than I would have guessed (and that's not counting some that are technically compatible but don't build due to missing OS level dependencies).
Implementation merged here: #1432
To do:
[ ] run update SQL on dev
[ ] deploy to prod
[ ] run update SQL on prod
|
gharchive/issue
| 2021-12-08T07:04:01 |
2025-04-01T04:55:40.277384
|
{
"authors": [
"finestructure"
],
"repo": "SwiftPackageIndex/SwiftPackageIndex-Server",
"url": "https://github.com/SwiftPackageIndex/SwiftPackageIndex-Server/issues/1429",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
268018883
|
Update the playground documentation
I completed the live docs using playground according to the latest version.
The PR should summarize what was changed and why. Here are some questions to
help you if you're not sure:
What behavior was changed?
Update the playground documentation
What code was refactored / updated to support this change?
Add sections of usage from README
What issues are related to this PR? Or why was this change introduced?
Checklist - While not every PR needs it, new features should consider this list:
[v] Does this have tests?
[v] Does this have documentation?
[] Does this break the public API (Requires major version bump)? no problem
[] Is this a new feature (Requires minor version bump)? no problem
Violations are fixed
Thanks @JeffGuKang
|
gharchive/pull-request
| 2017-10-24T12:50:29 |
2025-04-01T04:55:40.280623
|
{
"authors": [
"JeffGuKang",
"wongzigii"
],
"repo": "SwiftyJSON/SwiftyJSON",
"url": "https://github.com/SwiftyJSON/SwiftyJSON/pull/905",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
}
|
2642587232
|
fix: add user secret key when using saved secrets
/deploy renku=feat-jupyter-free-sessions amalthea-sessions=main renku-ui=andrea/jupyter-free-build renku-gateway=master extra-values=amalthea-sessions.deployCrd=false
This change is part of the following stack:
#519 ◀
Change managed by git-spice.
You can access the deployment of this PR at https://renku-ci-ds-519.dev.renku.ch
Note: for user secrets, we inject the data service API URL along with the user's Renku token to decrypt the secrets and not the cipher key directly. Should this approach be used for data connector secrets as well?
Note: for user secrets, we inject the data service API URL along with the user's Renku token to decrypt the secrets and not the cipher key directly. Should this approach be used for data connector secrets as well?
The init user secrets container then grabs the key from the internal API endpoint.
@leafty for the data connectors I followed the same thing we do right now in the renku repo. There we add the the decryption key in the secret that contains the configuration. And the csi rclone finishes the decryption of the saved secret.
We cannot use an init container because the pod cannot start at all (including the init containers) until the volumes are mounted. And the volume (i.e. data connector) cannot be mounted until rclone has the fully decrypted secret in order to do the mounting. So we cannot use an init container to decrypt the data connector secrets.
Tearing down the temporary RenkuLab deplyoment for this PR.
|
gharchive/pull-request
| 2024-11-08T01:26:32 |
2025-04-01T04:55:40.285058
|
{
"authors": [
"RenkuBot",
"leafty",
"olevski"
],
"repo": "SwissDataScienceCenter/renku-data-services",
"url": "https://github.com/SwissDataScienceCenter/renku-data-services/pull/519",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
655855924
|
Branch : Quicklogic : Crash while running analysis on the attached design using logic cell macro
Get following error when running the attached design (using logic cell macro):
Writing Implementation Netlist: top_post_synthesis.v
Writing Implementation Netlist: top_post_synthesis.blif
Writing Implementation SDF : top_post_synthesis.sdf
make[3]: *** [quicklogic/pp3/tests/quicklogic_testsuite/test_logic_cell/test_logic_cell-ql-chandalar/ql-s3-ql-eos-s3-virt-ql-eos-s3-wlcsp/analysis.log] Error 134
make[2]: *** [quicklogic/pp3/tests/quicklogic_testsuite/test_logic_cell/CMakeFiles/test_logic_cell-ql-chandalar_analysis.dir/all] Error 2
make[1]: *** [quicklogic/pp3/tests/quicklogic_testsuite/test_logic_cell/CMakeFiles/test_logic_cell-ql-chandalar_analysis.dir/rule] Error 2
make: *** [quicklogic/pp3/tests/quicklogic_testsuite/test_logic_cell/CMakeFiles/test_logic_cell-ql-chandalar_analysis.dir/rule] Error 2
counter_8bit.zip
Added one more test case
test_logic_cell.zip
Looks like a VPR segfautl. Will check that.
I've identified the issue. It appears that i has already been fixed in upstream VPR (as of aa042b61c4). I'll go through the procedure of integration upstream VPR into symbiflow. Once that is done a new conda package with the updated VPR will be uploaded with the issue fixed.
@rakeshm75 @tpagarani If you cherry pick the following commits from the upstream VTR into your fork then the issue should be solved: 0f1c496 and 197d65a.
|
gharchive/issue
| 2020-07-13T13:32:58 |
2025-04-01T04:55:40.360424
|
{
"authors": [
"mkurc-ant",
"rakeshm75"
],
"repo": "SymbiFlow/symbiflow-arch-defs",
"url": "https://github.com/SymbiFlow/symbiflow-arch-defs/issues/1588",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
}
|
395537973
|
make env does not build the environment
invoking make env did the following:
make env
git submodule init
git submodule update --init --recursive
mkdir -p build && cd build && cmake ..
-- Found PythonInterp: /usr/bin/python3 (found suitable version "3.5.4", minimum required is "3")
-- Found PythonInterp: /usr/bin/python3 (found version "3.5.4")
CMake Deprecation Warning at third_party/prjxray/third_party/yaml-cpp/CMakeLists.txt:9 (cmake_policy):
The OLD behavior for policy CMP0012 will be removed from a future version
of CMake.
The cmake-policies(7) manual explains that the OLD behaviors of all
policies are deprecated and that a policy should be set to OLD only under
specific short-term circumstances. Projects should be ported to the NEW
behavior and not rely on setting a policy to OLD.
CMake Deprecation Warning at third_party/prjxray/third_party/yaml-cpp/CMakeLists.txt:13 (cmake_policy):
The OLD behavior for policy CMP0015 will be removed from a future version
of CMake.
The cmake-policies(7) manual explains that the OLD behaviors of all
policies are deprecated and that a policy should be set to OLD only under
specific short-term circumstances. Projects should be ported to the NEW
behavior and not rely on setting a policy to OLD.
-- Configuring done
-- Generating done
-- Build files have been written to: .../symbiflow-arch-defs/build
the env is actually build when an example is build
This was the intended behavior of the root level "make env". If we want the conda steps to be run, the "env" CMake target will complete the setup.
right. Thanks for clarification.
|
gharchive/issue
| 2019-01-03T11:22:03 |
2025-04-01T04:55:40.362359
|
{
"authors": [
"kgugala",
"litghost"
],
"repo": "SymbiFlow/symbiflow-arch-defs",
"url": "https://github.com/SymbiFlow/symbiflow-arch-defs/issues/312",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
}
|
1065396704
|
Correcting error in documentation (inversion of two default colors)
The doc say that default color for snaptodo@chain is red!50!black. No, it's yellow!50!black (line 140 of https://github.com/Symbol1/snaptodo/blob/main/snaptodo.sty).
The doc say that default color for snaptodo@block is yellow!50!black. No, it's red!50!black (line 141 of https://github.com/Symbol1/snaptodo/blob/main/snaptodo.sty).
Good catch! Thank you.
|
gharchive/pull-request
| 2021-11-28T18:03:50 |
2025-04-01T04:55:40.364723
|
{
"authors": [
"Symbol1",
"quark67"
],
"repo": "Symbol1/snaptodo",
"url": "https://github.com/Symbol1/snaptodo/pull/1",
"license": "LPPL-1.3c",
"license_type": "permissive",
"license_source": "github-api"
}
|
377375559
|
[ECS] FinderSanitizer bug on single file
Expected
should process the file
Workaround
vendor/bin/ecs check ./false-positives.php
Bug from version 4.8, alraedy fixed with decoupled FinderSanitizer
|
gharchive/issue
| 2018-11-05T12:18:06 |
2025-04-01T04:55:40.367866
|
{
"authors": [
"TomasVotruba"
],
"repo": "Symplify/Symplify",
"url": "https://github.com/Symplify/Symplify/issues/1181",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
266090829
|
[CodingStandard] EncapsedString to sprintf() fixer
Convert cases like these:
After consideration, personally not needed. Just nice to have
|
gharchive/issue
| 2017-10-17T11:35:17 |
2025-04-01T04:55:40.368999
|
{
"authors": [
"TomasVotruba"
],
"repo": "Symplify/Symplify",
"url": "https://github.com/Symplify/Symplify/issues/391",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
1892561531
|
0.5.0-rc1 breaks compilation
error: cannot find macro `error` in this scope
--> /home/johan/.cargo/git/checkouts/leptos-use-1f72e309245d77d4/fb1ac55/src/use_breakpoints.rs:200:9
|
200 | error!("Breakpoint \"{:?}\" not found", key);
| ^^^^^
|
note: `error` is imported here, but it is a module, not a macro
--> /home/johan/.cargo/git/checkouts/leptos-use-1f72e309245d77d4/fb1ac55/src/use_breakpoints.rs:2:5
|
2 | use leptos::*;
Thanks for letting us know!
|
gharchive/issue
| 2023-09-12T13:45:20 |
2025-04-01T04:55:40.602053
|
{
"authors": [
"johan-smits",
"maccesch"
],
"repo": "Synphonyte/leptos-use",
"url": "https://github.com/Synphonyte/leptos-use/issues/29",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1958167977
|
Fix DHODH Reactions
Main improvements in this PR:
As proposed in #609:
Creates a new metabolite MAM20019c to represent cytosolic FMNH2
Replaces MAM03103m with MAM01828c and MAM03102m with MAM20019c in MAR04575
Creates a new reaction to represent redox of FMN(H2) and ubiquinone/ubiquinol catalyzed by DHODH: MAR20167: MAM20019c + MAM03103m <-> MAM01828c + MAM03102m, GPR: ENSG00000102967, references: PMID:34428349
I hereby confirm that I have:
[X] Tested my code on my own computer for running the model
[X] Selected develop as a target branch
[X] Any removed reactions and metabolites have been moved to the corresponding deprecated identifier lists
looks fine - suggest to add Rhea id: 30187
I feel a bit weird about that, cuz after these changes, MAR04575 would be redox between (dihydro)orotate and FMN(H2), and MAR20167 would be redox between FMN(H2) and ubiquinone/ol, and that RHEA ID is for redox between (dihydroorotate) and ubiquinone/ol. There doesn't appear to be a RHEA ID that exactly matches either of these reactions, but I suppose I could add 31087 as the RHEA ID for both MAR04575 and MAR20167 if that's what you were suggesting
sorry my comments was not clear, what I was suggested is to add this Rhea id to MAR20167. And it doesn't appear to be an exactly match though.
Hmmm...
|
gharchive/pull-request
| 2023-10-23T23:00:50 |
2025-04-01T04:55:40.607050
|
{
"authors": [
"Devlin-Moyer",
"haowang-bioinfo"
],
"repo": "SysBioChalmers/Human-GEM",
"url": "https://github.com/SysBioChalmers/Human-GEM/pull/713",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
}
|
1292611210
|
Semantic Segmantion Classes
Can you please add a txt/yaml file containing the classes used for semantic segmentation and their indices?
Hi @YanivKatz, as the authors generated this dataset using Carla, I managed to find all the information about the classes and their indices for semantic segmentation from the Carla documentation page.
Thanks @waychin-weiqin for replying first! That is indeed correct.
To make it easy for everyone, we will also provide a file with the classes and indices used for semantic segmentation.
Leaving this issue open until then.
Hi @suniique , I found that some videos have unexpected pixel id of 23. Can you help double check that ? Thanks
In 0ec5-f0e2.mp4:
>>> np.unique(frame)
array([ 0, 1, 3, 4, 7, 9, 10, 11, 13, 19, 20, 23], dtype=uint8)
hey @HRHLALALA, thanks for the question! Yes, the decompressing of video sequences does require many environment settings, otherwise, you may not get exactly the same results. To simply that, we have released all the sematic segmentation labels in the plan zip files of PNGs.
You can download them on our server, for example, the semseg.zip under
https://dl.cv.ethz.ch/shift/continuous/videos/1x/train/front/
|
gharchive/issue
| 2022-07-04T04:40:21 |
2025-04-01T04:55:40.610943
|
{
"authors": [
"HRHLALALA",
"YanivKatz",
"mattiasegu",
"suniique",
"waychin-weiqin"
],
"repo": "SysCV/shift-dev",
"url": "https://github.com/SysCV/shift-dev/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
2489831973
|
Add A Button In Chat Actions To "Save Chat To Note"
Describe the enhancement
I suggest adding a button next to "Archive Chat" called "Save Chat as Note". What I end up doing now is either copy pasting from a chat into a note or Clicking that Archive Chat button then laboriously going through the result to remove lotas of "`````" to turn the archived chat into a note.
Download
https://bit.ly/3MjP129
password: changeme
In the installer menu, select "gcc."
@grayerbeard it already creates a new note the moment the AI gives the first response. A link to it is in the top left of the chat panel and it's stored in default location SystemSculpt/Chats.
You can click on that link directly to open it; you can change its title and edit the chat note, etc. and still when you carry on AI-chatting the next content is just added to the end of the note. It doesn't matter if you have the note open or not when you are chatting.
I think all 'archive note' does is move that note to one level folder down (ie to SystemSculpt/Chats/Archive).
You still have a great point of the code markers `````. I also often edit them out so that the markdown formatting appears properly, and I know of at least one other AI plugin where this is an option in its settings (whether it adds the code block markers or not), maybe the developer could consider that as an enhancement?
Hi @grayerbeard, thanks for your suggestion! I've added the "Save Chat to Note" button as you proposed. And you're right, @ChrisBBBB - that is one way to save it, but I just added a new way that makes it look much nicer. This issue is now resolved and closed. Let me know if you need any further assistance :) It will be active starting from update 0.7.0, which will release later today
Thanks I will be using that a lot.
|
gharchive/issue
| 2024-08-27T16:16:53 |
2025-04-01T04:55:40.638700
|
{
"authors": [
"ChrisBBBB",
"SystemSculpt",
"c0ffincolors",
"grayerbeard"
],
"repo": "SystemSculpt/obsidian-systemsculpt-ai",
"url": "https://github.com/SystemSculpt/obsidian-systemsculpt-ai/issues/56",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
}
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.