id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
158685821
IE9/IE10 crash in AuthService.js Hi, We've encountered an issue in IE<=10 with Aurelia-Authentication (version 1.1.2), Working fine with Edge, and we've included all the necessary polyfills: fetch.js, bluebird.js, etc... In line 31 of AuthService.js, config is undefined: this.auth = auth; this.oAuth1 = oAuth1; this.oAuth2 = oAuth2; this.config = config.current; // config is undefined. this.client = this.config.client; Which causes an obvious problem. This code is called during the instanciation of our Authentication class, whose code follows: import {inject} from 'aurelia-dependency-injection'; import {AuthService} from 'spoonx/aurelia-authentication'; import {Endpoint} from 'spoonx/aurelia-api'; @inject(Endpoint.of('codabox-api')) export class CodaboxAuthService extends AuthService { constructor(codaboxApi, ...args){ super(...args); // crashes here. this.codaboxApi = codaboxApi; } ... } Any idea ? in that version current is a getter. if you can try to use t eobject directly aka this.config = config._current Well, the problem is that config is undefined. config.current is trying to access a field from an undefined reference, thus the crash. hang on a little, i think it'll be fixed in the next version. you need to inject all again when extending i'm sure @inject(Endpoint.of('codabox-api'), Authentication, OAuth1, OAuth2, BaseConfig); This is super urgent. A workaround is: class Foo extends Bar { constructor () { super() Bar.call(this) } } Note that this calls the parent constructor twice in other browsers, so make sure your parent constructor does nothing too complex (which it shouldn't anyway) or I've actually just found something that fixed this problem for me. There's an option called loose for babel-plugin-transform-es2015-classes, that makes it work on IE <= 10. I don't know if on babel 5 this was turned on by default or if the code always worked on older browsers, but it seems like setting this option is what's required now, it all works perfectly as before. You basically just need to explicitly pass the plugin with the option, instead of just using the es2015 preset, like this: plugins: [['transform-es2015-classes', {loose: true}]] Seems like this isn't a bug then, but I still think this should be better documented so people know what to do to make babel 6 compiled code work on all browsers. Ok, thanks for your investigation, I will try this and confirm that it works on my side. Hi, I can't get it to work. It builds all-right, but it's still not working in IE. I'm not sure I am doing things right as I am not familiar with either aurelia nor babel, here's my build/babel-options.js var path = require('path'); var paths = require('./paths'); exports.base = function() { return { filename: '', filenameRelative: '', sourceMap: true, sourceRoot: '', moduleRoot: path.resolve('src').replace(/\\/g, '/'), moduleIds: false, comments: false, compact: false, code: true, presets: [ 'es2015-loose', 'stage-1'], plugins: [ 'syntax-flow', 'transform-decorators-legacy', 'transform-flow-strip-types', 'transform-class-properties', ['transform-es2015-classes', {loose: true}] ] }; }; exports.commonjs = function() { var options = exports.base(); options.plugins.push('transform-es2015-modules-commonjs'); return options; }; exports.amd = function() { var options = exports.base(); options.plugins.push('transform-es2015-modules-amd'); return options; }; exports.system = function() { var options = exports.base(); options.plugins.push('transform-es2015-modules-systemjs'); return options; }; exports.es2015 = function() { var options = exports.base(); options.presets = ['stage-1']; return options; }; And I installed the following npm packages: + "babel-plugin-transform-class-properties": "^6.9.1", + "babel-plugin-transform-es2015-classes": "^6.9.0", without the loose it seems to work for me presets: [ 'es2015-loose', 'stage-1'], plugins: [ 'syntax-flow', 'transform-decorators-legacy', 'transform-flow-strip-types', 'transform-class-properties', 'transform-es2015-classes' ] Thanks ! This fixed the issue. Unfortunately there were lots of other IE issues waiting behind ... but that's none of your concern :) oho good luck then
gharchive/issue
2016-06-06T14:04:02
2025-04-01T04:55:39.326012
{ "authors": [ "cb-fred", "doktordirk" ], "repo": "SpoonX/aurelia-authentication", "url": "https://github.com/SpoonX/aurelia-authentication/issues/185", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
238131618
Registering an entity without the @resource() decorator leads to invalid meta data The documentation about the @resource() decorator says This decorator is probably the most important one. But this decorator is never mentioned in the whole Quickstart or Entites guide https://aurelia-orm.spoonx.org/entities.html Fact is: without the decoratory, the entitymanager doesn't has meta data for this entity. This is a documentation bug. Resource should have the name of the resource because minification will otherwise mess things up. But yes, it could use more docs! :) Minification removes the @resource() decorator completely? @mreiche No, it renames classes. And without a value, the decorator uses the name of the class. Minification can still work if you disable renaming. Check the bottom of the readme for more info. Ok, I add that to the documentation. @mreiche Yeah it can be added with the decorator itself. I agree. Thanks!
gharchive/issue
2017-06-23T12:43:23
2025-04-01T04:55:39.329378
{ "authors": [ "RWOverdijk", "mreiche" ], "repo": "SpoonX/aurelia-orm", "url": "https://github.com/SpoonX/aurelia-orm/issues/300", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2831381
INT-2400 Fix tool:expected-type Classes in IP XSD MessageChannel was moved from core to the base package. InterceptorFactoryChain was incorrect. More than just IP; closing.
gharchive/issue
2012-01-13T13:59:36
2025-04-01T04:55:39.361164
{ "authors": [ "garyrussell" ], "repo": "SpringSource/spring-integration", "url": "https://github.com/SpringSource/spring-integration/issues/314", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1729383864
Add a New Login Button Hi @Spyware007, I am a contributor in GSSOC'23. I want to add a new login button. can you please assign me this task under label gssoc'23 @Ayush9951 Assigned to you!
gharchive/issue
2023-05-28T10:38:51
2025-04-01T04:55:39.363167
{ "authors": [ "Ayush9951", "Spyware007" ], "repo": "Spyware007/Animating-Buttons", "url": "https://github.com/Spyware007/Animating-Buttons/issues/713", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
156177774
Remote inventories not seeming to work I might be derping this up, but I have a wired modem block, connected to an iron chest, a turtle, and vanilla chest. I am attempting to call iron.pushItemRemote("chest_0",1) but it is returning lua:1: Parameter 1 value not set Which version of CC, CCTweaks and OpenPeripheral are you running? Also can I have your log? Thanks! Oh, sorry about that, forgot versions: ComputerCraft: 1.75 CCTweaks: 0.3.2 OpenPeripheralCore: 1.3 OpenPeripheralIntegration: 0.5 OpenPeripheralAddons: 0.5.1 How much of the log do you need? Just around when I tried the function, or the whole thing? Oh yea, this is also on a server. Fixed in 39740a48a2689e8b3aa6b54380a711192094974d. I've pushed it but GitHub is derping.
gharchive/issue
2016-05-22T22:45:40
2025-04-01T04:55:39.367890
{ "authors": [ "Selim042", "SquidDev" ], "repo": "SquidDev-CC/CCTweaks", "url": "https://github.com/SquidDev-CC/CCTweaks/issues/98", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1461431973
Unable to load DLL 'libSkiaSharp' on iOS Hello, When I build for iOS using this package I get the following error: DllNotFoundException: Unable to load DLL 'libSkiaSharp'. Tried the load the following dynamic libraries: Unable to load dynamic library '/libSkiaSharp' because of 'Failed to open the requested dynamic library (0x06000000) dlerror() = dlopen(/libSkiaSharp, 0x0005): tried: '/libSkiaSharp' (errno=2), '/private/preboot/Cryptexes/OS/libSkiaSharp' (errno=2), '/libSkiaSharp' (errno=2) Please help. any luck with this? I get it too Could you please add __IOS__and __UNIFIED__ to the Scripting Define Symbols and try to build again? Let me know the update. I will update SkiaSharp's version to the latest one (2.88.3), it's been a long time since we are using an older version (2.80.3). Could you please add __IOS__and __UNIFIED__ to the Scripting Define Symbols and try to build again? Let me know the update. I will update SkiaSharp's version to the latest one (2.88.3), it's been a long time since we are using an older version (2.80.3). I actually tried that already, didn't work. Alright, thanks for the update. Hello @LeshM, could you please try this 0.2.0-preview.1 version? Unfortunately, I don't have access to any Mac device right now, so couldn't build the XCode project. Right now, you can grab it from OpenUPM . Or, merge these in your manifest.json - "scopedRegistries": [ { "name": "package.openupm.com", "url": "https://package.openupm.com", "scopes": [ "com.srejonkhan" ] } ] "dependencies": { "com.srejonkhan.asyncimagelibrary": "0.2.0-preview.1" } Alternatively, you can try cloning this sandbox project and change the branch of the submodule to develop branch (instructions added) - https://github.com/SrejonKhan/AsyncImageLibrary-Sandbox The build process would be exactly the same as you've tried earlier with Scripting Define Symbols. In case it fails, you may try tweaking the setting of /AsyncImageLibrary/Dependencies/SkiaSharp/ios/native/libSkiaSharp.framework if any mistakes were made from my side. Any changes/pull request is always appreciated. Hello @LeshM, could you please try this 0.2.0-preview.1 version? Unfortunately, I don't have access to any Mac device right now, so couldn't build the XCode project. Right now, you can grab it from OpenUPM . Or, merge these in your manifest.json - "scopedRegistries": [ { "name": "package.openupm.com", "url": "https://package.openupm.com", "scopes": [ "com.srejonkhan" ] } ] "dependencies": { "com.srejonkhan.asyncimagelibrary": "0.2.0-preview.1" } Alternatively, you can try cloning this sandbox project and change the branch of the submodule to develop branch (instructions added) - https://github.com/SrejonKhan/AsyncImageLibrary-Sandbox The build process would be exactly the same as you've tried earlier with Scripting Define Symbols. In case it fails, you may try tweaking the setting of /AsyncImageLibrary/Dependencies/SkiaSharp/ios/native/libSkiaSharp.framework if any mistakes were made from my side. Any changes/pull request is always appreciated. I'm sorry but I'm afraid I can't, I'll need to reintegrate your package and then build for both Android & iOS using UCB (which takes about an hour each) just to find out if this works. I'm on a tight deadline and can't spare the time. I do appreciate your work though, so thank you! 🙏 BTW I ended up using some Unity coroutine hack as a solution. @LeshM No problem. Thank you too!
gharchive/issue
2022-11-23T10:03:39
2025-04-01T04:55:39.399701
{ "authors": [ "BigGiantHead", "LeshM", "SrejonKhan" ], "repo": "SrejonKhan/AsyncImageLibrary", "url": "https://github.com/SrejonKhan/AsyncImageLibrary/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1592019731
Bottom to top button Make a bottom to top button for the Home page, which will be appeared after some scrolling. @Srijita-Mandal I would like to work on it @sarkartanmay393 Go for it. I would like to work on it under JWOC I would Like to work on this issue under JWOC. #119 Please assign me this issue under jwoc @Srijita-Mandal Srijita-Mandal @Srijita-Mandal I would like to work on this issue under JWOC. Please assign this issue to me.
gharchive/issue
2023-02-20T14:59:14
2025-04-01T04:55:39.404021
{ "authors": [ "Awesomeasma", "LuckySingh9897", "Srijita-Mandal", "priyanka1551", "sarkartanmay393", "vatan-prakash" ], "repo": "Srijita-Mandal/fix-your-nums", "url": "https://github.com/Srijita-Mandal/fix-your-nums/issues/114", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1509335357
Questions about Analysis of Disentanglement in the paper Hi, I read your amazing work: "Adversarial Graph Disentanglement", and then I have a question about the experimental part. In the V. EXPERIMENTAL RESULTS AND ANALYSIS, E. Analysis of Disentanglement, you compute the component confusion matrix C to illustrate the correlations among the distributions in different component spaces. The correlation score between the i-th and j-th component distributions is given by . I don't understand why this formula is used to calculate the correlation of two vector spaces. Why is cosine similarity calculated between every two elements in these two spaces? Why not compute the similarity between corresponding elements as a similarity measure for two vector spaces?that is Cij=1/Vsum(hu^ihu^j). Finally, can you provide your code on computing the cosine similarity of two vector spaces? I would appreciate it if you could. Thank you! Thanks for your attention to our work and the valuable question you have raised! I don't know if I understand this correctly, I guess the similarity you mentioned Cij=1/Vsum(hu^ihu^j) is the dot product. Actually, the dot product similarity is practically equivalent to the cosine similarity we adopt in this work, since the vector h^i_u and h^j_v are normalized during the process of micro-disentanglement. As for the associated codes, I will release them after I have reorganized. Plz stay tuned for this project!
gharchive/issue
2022-12-23T13:12:26
2025-04-01T04:55:39.417396
{ "authors": [ "SsGood", "xiaomi4356" ], "repo": "SsGood/ADGCN", "url": "https://github.com/SsGood/ADGCN/issues/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1876279094
Continue generating the images after an interrumption in electricity I want to know if we could easily recover from an electricy issue, for example. If the server closes for example... Can I reopen the server again and continue generating images from the same queue? Should we persist the current queue on disk in realtime? Does it make sense? Thanks a lot for your time That's... a new one. Have you considered, uh, a UPS? (Uninterruptable Power Supply - hardware device that's essentially a giant battery specifically to protect computers from failure if power fails) Spent some time considering whether to maybe implement some form of option for drive-caching of the generation queue or something, but I don't think it's going to fit as a valid feature, at least not during the current Alpha stage. So, for now, yeah best advice is get a UPS lol. Once Swarm is out of alpha and in release phase, can reopen this and reconsider the possibility.
gharchive/issue
2023-08-31T20:46:18
2025-04-01T04:55:39.440421
{ "authors": [ "fabianrodrigo", "mcmonkey4eva" ], "repo": "Stability-AI/StableSwarmUI", "url": "https://github.com/Stability-AI/StableSwarmUI/issues/82", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1377680642
StabilityInference.generate doesn't supports Sequence[int] as seed As type hinted here, StabilityInference.generate should supports a Sequence[int] as seed but is overwrited here. PR in-progress thanks for the contribution!
gharchive/issue
2022-09-19T09:40:57
2025-04-01T04:55:39.442299
{ "authors": [ "Thytu", "dmarx" ], "repo": "Stability-AI/stability-sdk", "url": "https://github.com/Stability-AI/stability-sdk/issues/48", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
228693222
blackbox_list_admins The black_list_admins fails on my machine, even though the blackbox-admins.txt file exists. $ tree keyrings keyrings └── live ├── blackbox-admins.txt ├── blackbox-files.txt ├── pubring.kbx └── trustdb.gpg 1 directory, 4 files $ blackbox_list_admins cat: blackbox-admins.txt: No such file or directory Wow, that script actually is broken. The quick fix is to edit the line: Awesome Tom 👌 On Mon, 15 May 2017 at 17:11, Tom Limoncelli notifications@github.com wrote: Closed #193 https://github.com/StackExchange/blackbox/issues/193 via 7649c9f https://github.com/StackExchange/blackbox/commit/7649c9f00b14390e08fb90a4318378058c2d0a3c . — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub https://github.com/StackExchange/blackbox/issues/193#event-1082442530, or mute the thread https://github.com/notifications/unsubscribe-auth/AAivm754sLQ_wSyXxAZrRPVoCtCn9dKrks5r6GsFgaJpZM4NbCnc .
gharchive/issue
2017-05-15T12:06:47
2025-04-01T04:55:39.451565
{ "authors": [ "aymericbeaumet", "tlimoncelli" ], "repo": "StackExchange/blackbox", "url": "https://github.com/StackExchange/blackbox/issues/193", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
101783495
RFR: Support trace_id in CLI run (virtualenv)/m/s/s/st2 git:support_trace_id_in_cli_run ❯❯❯ st2 run core.local date --trace-id='55d4c36f32ed3554acf2db4' ERROR: 400 Client Error: Bad Request MESSAGE: Unable to find Trace with id="55d4c36f32ed3554acf2db4" (virtualenv)/m/s/s/st2 git:support_trace_id_in_cli_run ❯❯❯ (virtualenv)/m/s/s/st2 git:master ❯❯❯ st2 run core.local cmd=date --trace-tag="OohLaLaLa" ⏎ ✭ ✱ ◼ . id: 55d3d5bf32ed3530544a753d status: succeeded result: { "failed": false, "stderr": "", "return_code": 0, "succeeded": true, "stdout": "Tue Aug 18 18:02:55 PDT 2015 " } (virtualenv)/m/s/s/st2 git:master ❯❯❯ st2 trace list --trace-tag OohLaLaLa ✭ ✱ ◼ +-------------------+-------------------------------------------------------+ | Property | Value | +-------------------+-------------------------------------------------------+ | id | 55d3d5bf32ed3530544a753e | | trace_tag | OohLaLaLa | | action_executions | [ | | | { | | | "updated_at": "2015-08-19T01:02:55.018421Z", | | | "object_id": "55d3d5bf32ed3530544a753d" | | | } | | | ] | | rules | | | start_timestamp | 2015-08-19T01:02:55.018362Z | | trigger_instances | | +-------------------+-------------------------------------------------------+ (virtualenv)/m/s/s/st2 git:master ❯❯❯ st2 run core.local cmd=date --trace-id='55d3d5bf32ed3530544a753e' ✭ ✱ ◼ . id: 55d3d5ec32ed3530544a7540 status: succeeded result: { "failed": false, "stderr": "", "return_code": 0, "succeeded": true, "stdout": "Tue Aug 18 18:03:41 PDT 2015 " } (virtualenv)/m/s/s/st2 git:master ❯❯❯ st2 trace get 55d3d5bf32ed3530544a753e ✭ ✱ ◼ +-------------------+-------------------------------------------------------+ | Property | Value | +-------------------+-------------------------------------------------------+ | id | 55d3d5bf32ed3530544a753e | | trace_tag | OohLaLaLa | | action_executions | [ | | | { | | | "updated_at": "2015-08-19T01:03:40.972020Z", | | | "object_id": "55d3d5ec32ed3530544a7540" | | | }, | | | { | | | "updated_at": "2015-08-19T01:02:55.018421Z", | | | "object_id": "55d3d5bf32ed3530544a753d" | | | } | | | ] | | rules | | | start_timestamp | 2015-08-19T01:02:55.018362Z | | trigger_instances | [ | | | { | | | "updated_at": "2015-08-19T01:03:41.091999Z", | | | "object_id": "55d3d5ed32ed35305dc72cc5" | | | } | | | ] | +-------------------+-------------------------------------------------------+ (virtualenv)/m/s/s/st2 git:master ❯❯❯ +1
gharchive/pull-request
2015-08-19T01:07:12
2025-04-01T04:55:39.456429
{ "authors": [ "lakshmi-kannan", "manasdk" ], "repo": "StackStorm/st2", "url": "https://github.com/StackStorm/st2/pull/1840", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
412129081
Switching between Mainnet and Testnet Hi Stadicus, I have one simple question. Everything is working well on mainnet now for some weeks but if I want to move back to testnet (and later back to mainnet etc.), my understanding of reading all directories and your instructions is that I only need to adjust the following 0) stop lnd and bitcoind change to testnet in /home/bitcoin/.bitcoin/bitcoin.conf chainge to testnet in /home/admin/.bitcoin/bitcoin.conf change to testnet in /home/bitcoin/.lnd/lnd.conf I think the admin.macaroon is already available in both mainnet and testnet (also for user admin) and also the tls.cert does not have to change? so after 1), 2) and 3) start bitcoind and than lnd should have it all work on testnet again? And to go back to mainnet just do the config files back to mainnet Before I am going to destroy a working node, I just wanted to check, thanks! Walter Reading through your very thorough list, I think that's exactly right. 👨‍🚀👍
gharchive/issue
2019-02-19T21:19:19
2025-04-01T04:55:39.461543
{ "authors": [ "Stadicus", "walterkok" ], "repo": "Stadicus/guides", "url": "https://github.com/Stadicus/guides/issues/388", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2129892313
🛑 Parsoid is down In ae7dfbe, Parsoid (https://star-citizen.wiki/api.php?action=visualeditor&format=json&paction=parse&page=Star_Citizen_Wiki) was down: HTTP code: 404 Response time: 632 ms Resolved: Parsoid is back up in e2d0544 after 9 minutes.
gharchive/issue
2024-02-12T11:02:21
2025-04-01T04:55:39.488327
{ "authors": [ "octfx" ], "repo": "StarCitizenWiki/status", "url": "https://github.com/StarCitizenWiki/status/issues/334", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1745144702
Broker load shall fail immediately with error message on wrong PROPERTIES setting Enhancement When there is misconfig in the broker load PROPERTIES(e.g. wrong AWS AK/SK or endpoint), broker load job state becomes QEUEUEING with out and error message to help debug. Suggest fixing this issue by setting the state to INVALID or FAILED and have details in ErrorMsg fields. This is essential for the upstream service to check the status and take appropriate measure for failure resilience. cc @chaoyli Thanks for reporting this issue. It's highly OK to add a check whether the ak/sk is set. I'll resolve it. I've checked the latest version of branch-2.5/branch-3.0 and found that the result has been shown like this. It seems to be what you want. It is PENDING state in a built from branch-3.0 commit hash https://github.com/StarRocks/starrocks/commits/a5dd8c36 on 5/23/2023 In this test, I issued the load sql with a wrong secret key in the properties. LOAD LABEL abb_10m.KNA1_4 ( DATA INFILE("s3a://celostar/test_data/abb_10m/fd1c713c-d76f-4113-99d2-a22c38afc7a6/*.parquet") INTO TABLE KNA1 FORMAT AS "parquet" ****** WITH BROKER ( "aws.s3.access_key" = "xxx", "aws.s3.secret_key" = "WRONG SECRET", <=================== "aws.s3.region"="us-west-1" ) PROPERTIES ( "timeout" = "36000", -- the timeout time in seconds for the import operation. e.g 3600 means 1 hr time out. Jobs not completed in timeout window will be CANCELLED state. "max_filter_ratio" = "0.2", -- Data ratio of maximum tolerance filterable (data irregularity, etc.). "timezone" = "America/Los_Angeles" ); This is enhanced in https://github.com/StarRocks/starrocks/pull/21982. Could you check it with a newer version again?
gharchive/issue
2023-06-07T06:37:05
2025-04-01T04:55:39.505960
{ "authors": [ "dengliu", "rickif" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/issues/24794", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1579196287
[Doc]add pause routine load Signed-off-by: hellolilyliuyi hellolilyliuyi123@163.com What type of PR is this: [ ] BugFix [ ] Feature [ ] Enhancement [ ] Refactor [ ] UT [x] Doc [ ] Tool Which issues of this PR fixes : Fixes # Problem Summary(Required) : Checklist: [ ] I have added test cases for my bug fix or my new feature [ ] This pr will affect users' behaviors [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto backported to target branch [x] 2.5 [x] 2.4 [x] 2.3 [x] 2.2 run starrocks_admit_test @mergify backport branch-2.5 @mergify backport branch-2.4 @mergify backport branch-2.3 @mergify backport branch-2.2
gharchive/pull-request
2023-02-10T08:03:09
2025-04-01T04:55:39.512671
{ "authors": [ "hellolilyliuyi" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/17665", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1807936950
[BugFix] disable dictionary optimize stream load in branch-23 hard to cp #26463 so we invalid dictionary when streaming load [FE PR Coverage Check] :heart_eyes: pass : 1 / 1 (100.00%) file detail path covered_line new_line coverage not_covered_line_detail :large_blue_circle: com/starrocks/common/Config.java 1 1 100.00% []
gharchive/pull-request
2023-07-17T14:39:08
2025-04-01T04:55:39.515862
{ "authors": [ "stdpain", "wanpengfei-git" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/27395", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2013575236
[Enhancement] support mysql key words in where caluse Why I'm doing: What I'm doing: support mysql key words in where caluse Fixes #issue What type of PR is this: [ ] BugFix [ ] Feature [x] Enhancement [ ] Refactor [ ] UT [ ] Doc [ ] Tool Does this PR entail a change in behavior? [ ] Yes, this PR will result in a change in behavior. [x] No, this PR will not result in a change in behavior. If yes, please specify the type of change: [ ] Interface/UI changes: syntax, type conversion, expression evaluation, display information [ ] Parameter changes: default values, similar parameters but with different default values [ ] Policy changes: use new policy to replace old one, functionality automatically enabled [ ] Feature removed [ ] Miscellaneous: upgrade & downgrade compatibility, etc. Checklist: [x] I have added test cases for my bug fix or my new feature [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto-backported to the target branch [x] 3.2 [x] 3.1 [x] 3.0 [x] 2.5 https://github.com/Mergifyio backport branch-2.5
gharchive/pull-request
2023-11-28T04:05:02
2025-04-01T04:55:39.522903
{ "authors": [ "zombee0" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/35917", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2158536165
adding fpe function Why I'm doing: What I'm doing: Fixes #issue What type of PR is this: [ ] BugFix [x] Feature [ ] Enhancement [ ] Refactor [ ] UT [ ] Doc [ ] Tool Does this PR entail a change in behavior? [x] Yes, this PR will result in a change in behavior. [ ] No, this PR will not result in a change in behavior. If yes, please specify the type of change: [ ] Interface/UI changes: syntax, type conversion, expression evaluation, display information [ ] Parameter changes: default values, similar parameters but with different default values [ ] Policy changes: use new policy to replace old one, functionality automatically enabled [ ] Feature removed [ ] Miscellaneous: upgrade & downgrade compatibility, etc. Checklist: [ ] I have added test cases for my bug fix or my new feature [ ] This pr needs user documentation (for new or modified features or behaviors) [ ] I have added documentation for my new feature or new function [ ] This is a backport pr Bugfix cherry-pick branch check: [x] I have checked the version labels which the pr will be auto-backported to the target branch [x] 3.2 [ ] 3.1 [ ] 3.0 [ ] 2.5 better to add docs for these functiions, you can refer to: https://github.com/StarRocks/starrocks/tree/main/docs/en/sql-reference/sql-functions @mergify rebase
gharchive/pull-request
2024-02-28T09:46:39
2025-04-01T04:55:39.530423
{ "authors": [ "before-Sunrise", "liubotao", "silverbullet233" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/41865", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1293056803
jvm gc 机制由cms修改成g1 What type of PR is this: [ ] bug [ ] feature [x] enhancement [ ] refactor [ ] others Which issues of this PR fixes : Fixes # Problem Summary(Required) : #fe.conf jvm配置中gc所使用的cms机制运行效率和性能实测低于g1,cms优化为g1,生产环境实测性能更好,full gc更少。 Thank you for your contributions. Could you give more performance information? Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it. Can you make it configurable? We can use cms or g1 according to a config item
gharchive/pull-request
2022-07-04T11:45:54
2025-04-01T04:55:39.534874
{ "authors": [ "CLAassistant", "chaoyli", "gengjun-git", "sfwanyi" ], "repo": "StarRocks/starrocks", "url": "https://github.com/StarRocks/starrocks/pull/8245", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2487534832
fix(item): Item attachment mods use parent item mod type This PR fixes mods added on the fly to item attachments having no type defined, and only allowing basic mods instead of weapon/armour/vehicule/all mods. Changing the item attachment mod type doesn't change the on the fly mods type retroactively, so we could also always set these to "all" instead. If so I'll update the PR. Closes #1624 Nope, LGTM.
gharchive/pull-request
2024-08-26T18:49:26
2025-04-01T04:55:39.536220
{ "authors": [ "obrenckle", "wrycu" ], "repo": "StarWarsFoundryVTT/StarWarsFFG", "url": "https://github.com/StarWarsFoundryVTT/StarWarsFFG/pull/1695", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
204438180
Do not wrap requests that skip the request filter Fixes issue #20. "Sign out" button in the sign in app uses a flag HandlerOptions.SkipRequestFilters. Website app uses a response filter though, and this flag does not skip that one. In result, Website app wraps a request that should not be wrapped. The solution requires a custom HTTP header applied in the request middleware that will force skipping of the response middleware. This is the same fix as https://github.com/StarcounterPrefabs/Launcher/commit/39da52caab8b6f9275d05d97d0e4cac6794d71d0 I consider this to be a temporary solution until there is a flag SkipMiddleware, which skips both the request and the response filter. @chrhol, @diliger can you pls review? Closing as done, thanks!
gharchive/pull-request
2017-01-31T22:07:40
2025-04-01T04:55:39.543693
{ "authors": [ "warpech" ], "repo": "StarcounterPrefabs/Website", "url": "https://github.com/StarcounterPrefabs/Website/pull/21", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1888678978
Ng+ Universe roll seed bug Platform: PC (Steam) Time/date of bug encountered: September 6th 2023 When starting new game plus, in certain circumstances you'll never be able to save or open your inventory. Details You have a chance to get a universe seed where the lodge gets alternative versions where andreja for example could show up with cultist and kill you, and in this universe you'll never be able to save your game or access your inventory. I'm noticing this happening with each unique new game universe unless it's a vanilla new game plus seeds with no changes to faction NPCs or the lodge Steps ive done to troubleshoot : validated steam files, full clean install ( made sure to delete app data stuff too ) and even started a new game to try to see if my game was just being silly, but the issue persist. Its preventing me from experiencing new game plus content that id otherwise would be able to experience if this bug was interfering with saving. Solution for now : There are specific console commands that allow you to start these secret seeds , however it's not ideal as it ruins the the surprise elements of discovering these. Solution found here : https://reddit.com/r/Starfield/s/Xhuht9OiDN Game Version v1.7.23.0 New Game Yes Reproduction Steps Start a new game, either new game plus or regular new game, finish the main storyline, go to the unity and make a hard save. Then keep entering the unity, get reborn, and then when your in your ship in the beginning of new game plus, check to see if your inventory is not available to open and you get a popup for this, and then check to see if you can't quicksave or save. If confirmed, go to the lodge and let the special scene play out. These are fully fleshed out, an example being if the hunter kills everyone , you can go to the Eye and you get specific dialogue to use it. It seems this wasn't intentional as the hunter even says " I'm sure this isn't the first time you've seen the lodge like this " I think this is not a bug, but an intentional game feature. The only thing the ruins things is not being able to save or open your inventory that makes me believe it's a bug. @juliocoliveira what is your source for this being an intentional game feature? This appears to be a bug as some users are able to load the same NG+ variant and not experience this issue. For example I experienced the issue on the lodge "sarah is a plant" variant and was unable to save or load inventory. However, other users have reported getting this seed variation and being able to save and use player inventory. Some users have also been able to recover from this by opening ship inventory and switching to player or opening console and click the OK dialogue to the warning. There are also variations that require the user to loot and read a special item, which is blocked by being unable to access inventory. Do you have a source with Bethesda that has confirmed this is a feature and not a bug? Zero source, as I said "I THINK", never said that was an absolute truth right from Bethesda developers. But some people think that is intentional, a random layer of difficulty, an universe that you cannot save or use your inventory, It's likely a bug with the quest stages or a variable ( like how certain cutscenes and other variables don't allow you to save or access your inventory in certain moments in the game that I won't spoil ) which could easily be fixed in CK when it releases or by Bethesda. This is fixable currently only with use of console commands. This was fixed in v1.8.86 "Fixed an issue that could temporarily prevent opening the inventory or saving after entering the Unity."
gharchive/issue
2023-09-09T10:10:00
2025-04-01T04:55:39.551643
{ "authors": [ "Aurelianis", "Constellation-VASCO", "juliocoliveira", "spkyghzt" ], "repo": "Starfield-Community-Patch/Starfield-Community-Patch", "url": "https://github.com/Starfield-Community-Patch/Starfield-Community-Patch/issues/106", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2014994502
Quest MQ105 All That Money Can Buy fails to start Quest MQ105 All That Money Can Buy fails to start if NPC Ryleigh is dead before quest start Details Quest ID: 2C1C9B Game Version 1.8.86.0 Language English (EN) New Game Yes Reproduction Steps Start new game, proceed until player can free travel Travel to Volii Alpha > Neon City > Slayton Aerospace Kill Ryleigh or kill everyone in the building Proceed with main quest until the start of MQ105 "All That Money Can Buy" Note that the quest won't start Duplicate of #696
gharchive/issue
2023-11-28T17:55:03
2025-04-01T04:55:39.554904
{ "authors": [ "Constellation-VASCO", "Pickysaurus" ], "repo": "Starfield-Community-Patch/Starfield-Community-Patch", "url": "https://github.com/Starfield-Community-Patch/Starfield-Community-Patch/issues/695", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2467866780
[Bug] failed to load rendersystemdx11.dll CS2 Before Reporting [X] I found no existing issues matching my bug [x] My issue is not caused by the Millennium interface inside Steam (Themes, Plugins & Update Panel). If it is, make a issue report in this repo. [X] My issue is not caused by a theme Describe the Bug The game can only be launched via a shortcut in Steam, and this error appears when starting Dota 2. Can it happend with actual AMD driver? Latest driver from old system named 19.x.x. Expected Behavior All drivers, system, VCRedist, has been updated, and much more, But nothing gonna work, This starts after reinstalling windows. System has been downloaded from official microsoft website. Before reinstallation, everything worked without errors. Steps To Reproduce This starts after reinstalling windows. System has been downloaded from official microsoft website. Before reinstallation, everything worked without errors. Can it happend with actual AMD driver? Latest driver from old system named 19.x.x. Operating System Windows Anything else? Why are you inclined to believe this issue stems from Millennium? This seems unrelated entirely.
gharchive/issue
2024-08-15T11:17:01
2025-04-01T04:55:39.593588
{ "authors": [ "ShadowMonster99", "uppofig" ], "repo": "SteamClientHomebrew/Millennium", "url": "https://github.com/SteamClientHomebrew/Millennium/issues/78", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1872026677
Add Soft Engine 2 SteamDB app page links to a few games using this https://steamdb.info/app/348470/ https://steamdb.info/app/350310/ https://steamdb.info/app/363130/ https://steamdb.info/app/363150/ https://steamdb.info/app/485510/ https://steamdb.info/app/502280/ https://steamdb.info/app/527280/ https://steamdb.info/app/551730/ https://steamdb.info/app/595740/ https://steamdb.info/app/601051/ https://steamdb.info/app/628890/ https://steamdb.info/app/636570/ https://steamdb.info/app/658260/ https://steamdb.info/app/730310/ https://steamdb.info/app/756590/ https://steamdb.info/app/831560/ https://steamdb.info/app/886820/ https://steamdb.info/app/936160/ https://steamdb.info/app/936180/ https://steamdb.info/app/936190/ https://steamdb.info/app/968790/ https://steamdb.info/app/1045620/ https://steamdb.info/app/1072420/ https://steamdb.info/app/1121560/ https://steamdb.info/app/1152300/ https://steamdb.info/app/1152310/ https://steamdb.info/app/1152320/ https://steamdb.info/app/1233260/ https://steamdb.info/app/1257290/ https://steamdb.info/app/1325200/ https://steamdb.info/app/1341200/ https://steamdb.info/app/1358701/ https://steamdb.info/app/1423600/ https://steamdb.info/app/1502980/ https://steamdb.info/app/1502990/ https://steamdb.info/app/1621310/ https://steamdb.info/app/1732190/ https://steamdb.info/app/1999771/ https://steamdb.info/app/2130460/ https://steamdb.info/app/2138090/ Brief explanation of the change Support Soft Engine 2 by Koei Tecmo. After running changed rules in this pull request (this bot only tests rules.ini changes): Changed technologies: Engine.SoftEngine2 These apps now match: (60) 211140 322460 331600 348470 350310 363110 363130 363150 401870 410850 449800 485510 502280 526580 527280 551730 574050 595740 601050 628080 628890 636570 730310 756590 831560 886820 936160 936180 936190 968790 1019100 1019870 1045620 1072420 1121560 1152300 1152310 1152320 1168460 1233260 1257290 1325200 1341200 1358700 1423600 1424800 1502970 1502980 1502990 1582390 1621310 1732190 1842810 1879330 1999770 2124680 2130460 2138090 2271770 2743610 Please confirm that all these changed apps are correct.
gharchive/pull-request
2023-08-29T16:33:32
2025-04-01T04:55:39.612737
{ "authors": [ "MichaelHinrichs", "SteamTracker" ], "repo": "SteamDatabase/FileDetectionRuleSets", "url": "https://github.com/SteamDatabase/FileDetectionRuleSets/pull/352", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2542986375
Update Steeltoe to build and run tests against .NET 9 RC1 Description Experimental and hacky, do not merge. Connector failures on MySQL Connector are to be expected. The Oracle package doesn't provide anything for .NET 9 yet, while the Pomelo package is still at Preview1 (RC support is planned). Blocked by https://github.com/SonarSource/sonar-dotnet/issues/9670. /azp run Steeltoe.All
gharchive/pull-request
2024-09-23T15:20:22
2025-04-01T04:55:39.639992
{ "authors": [ "bart-vmware" ], "repo": "SteeltoeOSS/Steeltoe", "url": "https://github.com/SteeltoeOSS/Steeltoe/pull/1368", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
217727095
Fixes Firefox Extension 'Unknown Error' Fixes #7 Due to the latest version of Firefox blocking API requests, the fetching is now done in a background script. 4eb7a19 should resolve those issues, thanks! Do you happen to know what exactly changed that caused fetching cross domain to error out? The fetch errors are so vague...
gharchive/pull-request
2017-03-28T23:59:22
2025-04-01T04:55:39.687583
{ "authors": [ "Step7750" ], "repo": "Step7750/CSGOFloat-Extension", "url": "https://github.com/Step7750/CSGOFloat-Extension/pull/8", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
62199327
Feature request: add JASC-PAL palette support to rescomp Here is the format description : .pal Paint Shop Pro colour palette file: Contains the colour palette table saved from a 16 or 256 colour image. It can be loaded for another image and this provides a means of copying colour palettes from one image to another. These files are text editable. The format of the file is as follows: - The first line contains the header "JASC-PAL" (without the quotes). - The second line contains the palette file version. For example "0100" (without the quotes) - The third line contains the number of colours, either 16 or 256. - Subsequent lines contain the Red, Green and Blue components of each palette entry, with a space between each of the Red, Green and Blue numbers. These numbers are in decimal, so each attribute should be in the range 0 to 255. Original issue reported on code.google.com by Stephane.D on 16 Jun 2014 at 8:54 Title should be changed to Add *.pal files support. I'm using GraphicGale Free Edition and the format is the same I guess. Here is an example for a 4bpp 16colors palette: Thanks for the example, i have also 2 files sitting there so i can make some tests with them. I will try to add that for the next version of rescomp :) ty
gharchive/issue
2015-03-16T20:22:34
2025-04-01T04:55:39.693499
{ "authors": [ "GoogleCodeExporter", "Stephane-D", "derainm", "nolddor" ], "repo": "Stephane-D/SGDK", "url": "https://github.com/Stephane-D/SGDK/issues/23", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1439647709
🛑 DNS (he.net) is down In abde44a, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in bc39c11.
gharchive/issue
2022-11-08T07:34:38
2025-04-01T04:55:39.718690
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/1322", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1626908596
🛑 DNS (he.net) is down In b47c177, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in 3b30d5b.
gharchive/issue
2023-03-16T07:55:44
2025-04-01T04:55:39.721536
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/2562", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1669319161
🛑 DNS (he.net) is down In 98bd8d5, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in bc11a6b.
gharchive/issue
2023-04-15T11:31:42
2025-04-01T04:55:39.724357
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/5135", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1713399193
🛑 DNS (he.net) is down In 6ac1e54, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in 631683f.
gharchive/issue
2023-05-17T08:28:44
2025-04-01T04:55:39.726935
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/7594", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1722882985
🛑 DNS (he.net) is down In 94208a5, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in 727e137.
gharchive/issue
2023-05-23T22:37:45
2025-04-01T04:55:39.729519
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/8160", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1739892255
🛑 DNS (he.net) is down In 62a8065, DNS (he.net) ($HE_NS) was down: HTTP code: 0 Response time: 0 ms Resolved: DNS (he.net) is back up in db1336c.
gharchive/issue
2023-06-03T20:43:46
2025-04-01T04:55:39.732598
{ "authors": [ "leitmori" ], "repo": "Sternwarte-St-Ottilien-e-V/status", "url": "https://github.com/Sternwarte-St-Ottilien-e-V/status/issues/9064", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
161070258
ClientPlayerPlaceBlockPacket with null block face Passing a null block face to ClientPlayerPlaceBlockPacket to signal the "special" face (255) currently produces an error. #182
gharchive/issue
2016-06-19T12:15:09
2025-04-01T04:55:39.741236
{ "authors": [ "finalchild", "yawkat" ], "repo": "Steveice10/MCProtocolLib", "url": "https://github.com/Steveice10/MCProtocolLib/issues/181", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
678091265
The latest model has poor generalization effect I tested the latest age and gender estimation model on my own private test set with the test script you gave, and found that the effect of age and gender estimation is poor, especially that gender estimation is easy to make mistakes. Have you ever tested it on private test videos? Can you tell me what the problem is?Could it be a problem with tensorflow version I tested the latest age and gender estimation model on my own private test set with the test script you gave, and found that the effect of age and gender estimation is poor, especially that gender estimation is easy to make mistakes. Have you ever tested it on private test videos? Can you tell me what the problem is?Could it be a problem with tensorflow version you should add "--white" as a params my command is: python nets/test.py -g -vid -se -white I use this command to test, and the effect is still not good I use a video of a child, but the result is in my twenties my command is: python nets/test.py -g -vid -se -white I use this command to test, and the effect is still not good. can you provide your full command line. I will valid it~. my command is: python nets/test.py -g -vid -se -white -m ./model/c3ae_model_v2_fp16_white_se_132_4.208622-0.973 I use a video of a child, but the result is in my twenties if you want detect children, you'd better finetuning in "asia dataset" only. BCZ distribution of utk and afad range from 18~100 and asia dataset range from 0~70. OK, I'll do it. Thank you very much for your help python nets/test.py -g -vid -se -white -m ./model/c3ae_model_v2_fp16_white_se_132_4.208622-0.973 Can you provide your pic to geekpeakspar@gmail.com. I will test it for you. I have test many pic of children from baidu, it works normal. you should guarante pic that should be front face and size great than 64 * 64. If you want get a more accurate result, you should build your own dataset.
gharchive/issue
2020-08-13T02:00:24
2025-04-01T04:55:39.747217
{ "authors": [ "StevenBanama", "yanyanlimingkun" ], "repo": "StevenBanama/C3AE", "url": "https://github.com/StevenBanama/C3AE/issues/20", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
642408765
Unblock domains constantcontact.com (with all subdomains) This is a domain that sends newsletters and being used by legit organizations. ecampaigns.cineplex.com Newsletters from a trusted company. Please unblock these domains/subdomains. Hi @YogiJB thanks for this. As I research this, I think you need to know there are no "legit organizations" in this space. The name-brand attached to a domain counts for exactly zero. The only thing that matters is, does a domain track or trade its users? Does it serve ads? It it a path for malware? Understad what I mean? @YogiJB the domain constantcontact.com comes to us via Awaway. Ping @jawz101, what do you think? @YogiJB take note: we don't list ecampaigns.cineplex.com. We don't list any cineplex domain. removing. Purely because the Adaway list is meant for traffic that comes from mobile apps (i.e. not email marketing, browser, or desktop app traffic.) If I see those domains again I'll review it later https://github.com/constantcontact/java-sdk Thanks Jawz! @jawz101. @YogiJB this will go into wider distribution for the next release, soon. Closing.
gharchive/issue
2020-06-20T16:47:57
2025-04-01T04:55:39.752004
{ "authors": [ "StevenBlack", "YogiJB", "jawz101" ], "repo": "StevenBlack/hosts", "url": "https://github.com/StevenBlack/hosts/issues/1316", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
243360121
*.ypu.edu.tw This domain is Taiwan's University (Yuanpei University of Medical Technology) There's only ypu.edu.tw What do you think about this @FadeMind ? I was saw only "### Extra rules for @StevenBlack 's hosts project", so I report to you :-P I just curious why university's website was added to hosts list. You were right @pinomeng ;) I can simply whitelisting this domain if needed. This domain looks OK to me @FadeMind @StevenBlack unreachable per browser by me but reachable with nslookup ... @StevenBlack fixed in https://github.com/FadeMind/hosts.extras/commit/346e219d768f98cfb27c4ec34616ed8502e281f0 Thanks Tomasz @FadeMind and Pino @pinomeng. New hosts file is live and online now.
gharchive/issue
2017-07-17T11:01:16
2025-04-01T04:55:39.755671
{ "authors": [ "FadeMind", "StevenBlack", "funilrys", "pinomeng" ], "repo": "StevenBlack/hosts", "url": "https://github.com/StevenBlack/hosts/issues/362", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1634301587
jdownloader2 Server LXC Hi, wenn ich mir den Container installiere, wie komme ich dann auf die Oberfläche von jDownloader? Gruß Bitte Issues auf Englisch verfassen Siehe https://github.com/StevenSeifried/proxmox-scripts/issues/12#issuecomment-1356144716
gharchive/issue
2023-03-21T16:33:23
2025-04-01T04:55:39.759910
{ "authors": [ "Azzenator", "StevenSeifried" ], "repo": "StevenSeifried/proxmox-scripts", "url": "https://github.com/StevenSeifried/proxmox-scripts/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
101306790
Unhandled Error Whenver I try to open my stash in Standard League or press on the Trading tab it says There was an unhandled error On debug info says dis : [16-08-2015 23:26] System.ArgumentOutOfRangeException: Length cannot be less than zero. Parameter name: length Hi, Do your stashes in other leagues work? Hey, Yeh the warbands league stash loads just fine but Standard will just cause it to crash On 17 Aug 2015, at 00:02, Stickymaddness notifications@github.com wrote: Hi, Do your stashes in other leagues work? — Reply to this email directly or view it on GitHub. Alright, are you sure your standard stash is up to date? Ie: Login and update standard for all tabs. If you're sure it's up to date, do you have any old race characters in standard or characters from old leagues you haven't logged into for a long time? I use Session ID to log in but it shows it loading my new stash and the character I havnt been on for the longest is in Warbands. On 17 Aug 2015, at 16:30, Stickymaddness notifications@github.com wrote: Alright, are you sure your standard stash is up to date? Ie: Login and update standard for all tabs. If you're sure it's up to date, do you have any old race characters in standard or characters from old leagues you haven't logged into for a long time? — Reply to this email directly or view it on GitHub. I just deleted my warbands character and the standard stash is working fine now but The Trading tab still causes procurement to crash ;( Sent from my iPhone On 17 Aug 2015, at 16:30, Stickymaddness notifications@github.com wrote: Alright, are you sure your standard stash is up to date? Ie: Login and update standard for all tabs. If you're sure it's up to date, do you have any old race characters in standard or characters from old leagues you haven't logged into for a long time? — Reply to this email directly or view it on GitHub. Deleting your warbands character shouldn't have any effect, since the data is stored separately per league. Delete your tabprefs.xml file if you have one, restart Procurement and see if the trading view is able to load without crashing. Are you using the latest version of Procurement (1.9.4) ? I deleted the tabpref file and restarted but same thing happened unfortunately also yes I am using the 1.9.4 version On 17 Aug 2015, at 16:52, Stickymaddness notifications@github.com wrote: Deleting your warbands character shouldn't have any effect, since the data is stored separately per league. Delete your tabprefs.xml file if you have one, restart Procurement and see if the trading view is able to load without crashing. Are you using the latest version of Procurement (1.9.4) ? — Reply to this email directly or view it on GitHub. Try login to all your standard characters, this sometimes fixes the API returning invalid data for a league.
gharchive/issue
2015-08-16T22:49:03
2025-04-01T04:55:39.771173
{ "authors": [ "NeGerai", "Stickymaddness" ], "repo": "Stickymaddness/Procurement", "url": "https://github.com/Stickymaddness/Procurement/issues/171", "license": "Artistic-2.0", "license_type": "permissive", "license_source": "github-api" }
159369118
[CROSSDATA-465] Now it is possible to insert empty arrays and maps Description Before this PR, inserting empty arrays or map syntax was rather counter intuitive. Now, it is as easy as passing [] and () values for empty arrays and maps respectively. Testing Unit tests has been added to check whether this change has been properly applied. :+1: LGTM
gharchive/pull-request
2016-06-09T09:53:13
2025-04-01T04:55:39.818627
{ "authors": [ "miguel0afd", "pfcoperez", "pianista215" ], "repo": "Stratio/Crossdata", "url": "https://github.com/Stratio/Crossdata/pull/572", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
878020220
🛑 Nextcloud Papa is down In 0d40824, Nextcloud Papa ($NEXTCLOUD_2) was down: HTTP code: 0 Response time: 0 ms Resolved: Nextcloud Papa is back up in 0234889.
gharchive/issue
2021-05-06T20:51:39
2025-04-01T04:55:39.878322
{ "authors": [ "StudFu-WordToMD" ], "repo": "StudFu-WordToMD/status", "url": "https://github.com/StudFu-WordToMD/status/issues/1342", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2162665642
Is the create and delete of the app intentional in EntryPoint.h? In EntryPoint.h, https://github.com/StudioCherno/Walnut/blob/3b8e414fdecfc6c8b58816106fe8d912bd172e31/Walnut/src/Walnut/EntryPoint.h#L10-L20 This code appears to have a logical error, as it creates and deletes a new Walnut::Application object in each iteration of the while loop. This means that the application is constantly being restarted and terminated, which is probably not the intended behavior. A more reasonable approach would be to create the application object once before the loop, and delete it once after the loop, like this: namespace Walnut { int Main(int argc, char** argv) { Walnut::Application* app = Walnut::CreateApplication(argc, argv); while (g_ApplicationRunning) { app->Run(); } delete app; return 0; } } Please correct me if I'm mistaken here, Or maybe there's a good reason for it, I'm very new to this project. Update: I got courageous and tried it. My solution is... not good. It causes the app to crash when I try to exit/close the window.
gharchive/issue
2024-03-01T05:56:29
2025-04-01T04:55:39.893690
{ "authors": [ "glenpierce" ], "repo": "StudioCherno/Walnut", "url": "https://github.com/StudioCherno/Walnut/issues/73", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
206474505
Unable to pick Google drive file content Hi, I need to access a file from google drive, when i select file from google drive it is returning file path (//com.google.android.apps.docs.storage/document/acc%3D4%3Bdoc%3D2602) but i am not able to access that file content. Regards, Sachin KS That Uri isn't a local file path, its a Uri to the data in the google drive content provider. As far as I know, when u pull a file from google drive it doesn't save it to your device, rather it gives you the stream.
gharchive/issue
2017-02-09T11:15:49
2025-04-01T04:55:39.900914
{ "authors": [ "mhallmark", "sachinks1991" ], "repo": "Studyxnet/FilePicker-Plugin-for-Xamarin-and-Windows", "url": "https://github.com/Studyxnet/FilePicker-Plugin-for-Xamarin-and-Windows/issues/24", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
593559339
Crawl / manually curate / map Publications resources into individual metadata records based off of schema https://docs.google.com/spreadsheets/d/1FFyRhI5TeUb-B4t50HRZYF_XngRt_mfAIz82UQ3geFk/edit#gid=0 Need to check how often preprints from individual journals (Lancet, Nature, etc.) get incorporated into Pubmed index. Don't want to be redundant, but also want to be as up-to-date as possible. To start, I think harvesting from LitCovid would be the easiest starting point. My guess is that the PubMed/NCBI team is working hard on getting COVID-19-related literature into PubMed as quickly as possible, so I think we should old off on doing any direct scraping of journal sites. Preprint servers like Biorxiv and Medrxiv would be more useful I think. So to summarize, I think the priority list should be 1) LitCovid, 2) Biorxiv / Medrxiv, and 3) journal sites. (not seeing a biorxiv/medrxiv API, but maybe can use their rss feed and/or https://www.rxivist.org/docs)... Makes sense. BioRxiv has a third-party API: https://api.rxivist.org/v1/papers?q=nCoV&timeframe=alltime. Little bit wonky; &metric=downloads gives different results. Unclear how frequently is updated. Could also be gotten via scraping https://www.biorxiv.org/search/ncov or Twitter feeds. Medrxiv searches seem to include Biorxiv but bio doesn't include med. There's also Chemrxiv, but lower priority. plan: Pull data from LitCOVID/Pubmed: https://ftp.ncbi.nlm.nih.gov/pub/lu/LitCovid/. Use litcovid2BioCJSON.gz to get the IDs, and then use the PubMed API to get the actual records Scrape Biorxiv Scripe Medrxiv Align metadata with schema (#1 ) The mapping from PubMed/Medline to the schema can be found on the worksheet in columns A and J Additional information to grab from efetch: example record (not COVID-related): https://eutils.ncbi.nlm.nih.gov/entrez/eutils/efetch.fcgi?db=pubmed&id=28841416&rettype=abstract MeSH/OT terms --> keywords abstract --> abstract author affiliations --> author.affiliation grants --> funding referenceList --> citedBy @gtsueng has additional mapping of PubMed/Medline columns to our schema in columns A/J of the worksheet Note that there are two methods for pulling the data from pubmed. One method is to use the API which yields an xml file: https://eutils.ncbi.nlm.nih.gov/entrez/eutils/efetch.fcgi?db=pubmed&id=28841416&rettype=abstract The benefit of this method is as follows: -Author given name and family name -References (Citations) -Specific grant information The other method uses the biopython library to fetch and parse a medline record which treats it like a json file. Sample gist here. The benefit of this method is as follows: -json -easy to map (the dictionary keys already mapped to schema fields in the worksheet: Publications, column J) -PublicationType available as PT via this method -Other terms aside also available (OT) aside from just MeSH (MH) Abbreviated Journal titles may work better than the actual journal titles when it comes down to the visualization part. Since there are different ways to abbreviate journal titles, it may be better to use a dictionary for consistency. The NLM list of journal can be found here and parsed into a dictionary. Once parsed, a manual review/edit could help to ensure that the most common abbreviations are rather than just the NLM abbreviations (which are by no means the de facto standard). Abbreviated Journal titles may work better than the actual journal titles when it comes down to the visualization part. Since there are different ways to abbreviate journal titles, it may be better to use a dictionary for consistency. The NLM list of journal can be found at ftp://ftp.ncbi.nih.gov/pubmed/J_Medline.txt and parsed into a dictionary. Once parsed, a manual review/edit could help to ensure that the most common abbreviations are rather than just the NLM abbreviations (which are by no means the de facto standard). would propose having journalName be the full journal name (for instance The New England journal of medicine) and something like journalNameAbbrevation be N. Engl. J. Med. Added into schema as suggested above. Still suggest populating from a single dictionary and caution against loading it from metadata from different sources since they may abbreviate journal titles differently. referenceList --> citedBy (note: will need to pick an identifier to map to the other publications that cite this publication. ideally, this should refer to the _id we use to index these publications) @flaneuse In my mind the reference list should map to isBasedOn since those would be explicitly called out/referenced in the text and the citedBy would refer to other publications that cite this publication. referenceList --> citedBy (note: will need to pick an identifier to map to the other publications that cite this publication. ideally, this should refer to the _id we use to index these publications) @flaneuse In my mind the reference list should map to isBasedOn since those would be explicitly called out/referenced in the text and the citedBy would refer to other publications that cite this publication. Oh, sorry, you're right; I'm backwards. I thought these were the citedBy listed in the PubMed entry -- I care more about those than the references they cite. Wonder where these are hidden in the PubMed API... I just assumed that they could query for those as long as all their other publications have references. In terms of related articles where the relationship is unclear, Medline will sometimes provide them as [AID] (associated ID). These are often related articles. Eg- if Publication A is a letter to the editor with regards to Publication B. An identifier for Publication B would be included as AID in the record for Publication A. Unfortunately, AID is generic enough that they'll sometimes include the DOI for Publication A in the record for Publication A. @marcodarko can you also track the LitCovid / BioRxiv / MedRxiv original url? And maybe the source type (LitCovid, BioRxiv, or MedRxiv)? per @andrewsu on the front-end: "assuming we get that from litcovid, would be good to add a note and/or link. Yes, probably somewhat redundant to do it on every page, but would rather err on the side of overattribution..." I think we want to reserve url for the actual link to the paper (I don't want to get routed through PubMed when I just want to read the paper). I'm not sure I see a natural schema.org way to capture this information... thoughts @gtsueng? Proposal for your reaction (as an aside: we also need to decide on what we're using for canonical IDs; this will be used as the key to link between records, and also for the url to route on outbreak.info): { _id: "pmid32292904", pmid: "32292904", doi: "10.1016/j.xkme.2020.03.001", url: "https://www.kidneymedicinejournal.org/article/S2590-0595(20)30049-2/pdf", sameAs: { name: "LitCovid", url: "https://www.ncbi.nlm.nih.gov/research/coronavirus/publication/32292904" } } @marcodarko can you also track the LitCovid / BioRxiv / MedRxiv original url? And maybe the source type (LitCovid, BioRxiv, or MedRxiv)? per @andrewsu on the front-end: "assuming we get that from litcovid, would be good to add a note and/or link. Yes, probably somewhat redundant to do it on every page, but would rather err on the side of overattribution..." I think we want to reserve url for the actual link to the paper (I don't want to get routed through PubMed when I just want to read the paper). I'm not sure I see a natural schema.org way to capture this information... thoughts @gtsueng? Proposal for your reaction (as an aside: we also need to decide on what we're using for canonical IDs _id; this will be used as the key to link between records, and also for the url to route on outbreak.info): { _id: "pmid32292904", pmid: "32292904", doi: "10.1016/j.xkme.2020.03.001", url: "https://www.kidneymedicinejournal.org/article/S2590-0595(20)30049-2/pdf", sameAs: { name: "LitCovid", url: "https://www.ncbi.nlm.nih.gov/research/coronavirus/publication/32292904" } } update: creating new base class property curatedBy to store this info Noting source of BioRxiv data https://github.com/marcodarko/litcovid https://github.com/marcodarko/biorxiv
gharchive/issue
2020-04-03T18:20:43
2025-04-01T04:55:39.929176
{ "authors": [ "andrewsu", "flaneuse", "gtsueng" ], "repo": "SuLab/outbreak.info-resources", "url": "https://github.com/SuLab/outbreak.info-resources/issues/25", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1058157882
r4s Nps 内网穿透打开不了 r4s Nps 内网穿透打开不了 你应该去这里反馈:https://github.com/coolsnowwolf/lede/issues
gharchive/issue
2021-11-19T05:39:41
2025-04-01T04:55:39.931388
{ "authors": [ "AmadeusGhost", "baoeig" ], "repo": "SuLingGG/OpenWrt-Rpi", "url": "https://github.com/SuLingGG/OpenWrt-Rpi/issues/260", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1419923342
Wordle Game I want to add a Wordle Game made using CSS, HTML and Js @newbie2508 Assign to you
gharchive/issue
2022-10-23T20:32:00
2025-04-01T04:55:39.979061
{ "authors": [ "Subhradeep10", "newbie2508" ], "repo": "Subhradeep10/Amazing_Web_Projects", "url": "https://github.com/Subhradeep10/Amazing_Web_Projects/issues/110", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
518041518
Testtuples: reference the traintuple instead of referencing the model directly See also: https://github.com/SubstraFoundation/substra-backend/pull/33 Store a reference to the traintuple instead of a reference to the model. The model can still be obtained by querying the traintuple. This change will make it easier to implement generic test tuples. In this PR: Add a new field traintupleKey to Testtuple, outputTesttuple and outputBoardTuple Delete the model field from Testtuple Mark the model field from outputTesttuple and outputBoardTuple We don't delete these fields yet to prevent potential breakages in other parts of the substra stack Future work: Delete the model field from outputTesttuple and outputBoardTuple @Kelvin-M Regarding the extra ledger call, please see the conversation there Regarding the frontend breaking, I've tested on my local: I generated all kinds of assets and clicked around on everything. I've not managed to see anything broken :) All HTTP calls go through, and no error in the console. I could have missed something though, and I think it would be worth having @jmorel 's opinon. @AurelienGasser Nice ! Do you test it also with the removal of the Model key/value @Kelvin-M I did :) Okay, we wait for jmorel to check if everything is good :) About the frontend, we always use the traintuple to get model details, we never used the testtuple to do so. I think it's safe to remove from the frontend points of view. So we can remove Model in this PR :) What do you think @AurelienGasser & @thibaultrobert ? @Kelvin-M No objection, other than @inalgnu recommended against it. @thibaultrobert Need you review for this to be merged !
gharchive/pull-request
2019-11-05T21:19:21
2025-04-01T04:55:40.035180
{ "authors": [ "AurelienGasser", "Kelvin-M", "jmorel" ], "repo": "SubstraFoundation/substra-chaincode", "url": "https://github.com/SubstraFoundation/substra-chaincode/pull/21", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1772907856
Close #710, fix ALKilnInThePlayground upload Step A few quirks here. This doesn't allow passing in command line arguments for source folder paths yet. That may have to be a separate feature. I'm not sure it should be a required command line argument anyway, in which case we'll still need that code for times when the argument isn't handed in. We could ensure ALKilnInThePlayground hands it in, but I'm not yet sure how we can get ALKilnInThePlayground to get the right paths. Maybe it would have better access to the S3 paths, though. [Added a fix in the in-Playground interview that makes this point moot. See https://github.com/SuffolkLITLab/docassemble-ALKilnInThePlayground/pull/24] ~It doesn't quite work with S3. Jonathan gave us the info that the /tmp folder for S3 won't keep those files around, so we probably can't count on those files being there for running these tests in an interview~ (ALKilnInThePlayground - now it occurs to me that ALKilnInterview may have been a better name...) I should have probably broken one change into a different PR and I can do that if this is too much for one PR - The variable setting Step for uploading files turned out to be a mismatch for one ethos we're following. We decided that we want to defer to being radically permissive and forgiving with test code in general. In that spirit, I changed it so that if ALKiln can't find a file that the author listed to upload, the file just isn't uploaded and, if not all variables are required to be set, the test can continue. In a Story Table, the fields are not required to be set and there would be no error just because of a missing file. In a linear Step ("I set the var..."), a missing file would cause an error. I'm not sure about being this permissive, but I've been unsure in similar situations in the past and the collective decision has generally erred on the side of flexibility. I'll fix the version upon publishing. Closes #710 Not super simple, but could use a look (if you have time), @BryceStevenWilley.
gharchive/pull-request
2023-06-24T20:35:47
2025-04-01T04:55:40.061044
{ "authors": [ "plocket" ], "repo": "SuffolkLITLab/ALKiln", "url": "https://github.com/SuffolkLITLab/ALKiln/pull/717", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2445943189
ResearchRover Extension ResearchRover: Gather and summarize web content effortlessly, streamlining your research process with automated insights. Advanced AI-Powered Research Assistant @Sulagna-Dutta-Roy, assign me this issue How do we know when the feature is complete? I will raise PR @Sulagna-Dutta-Roy, view PR and let me know if any changes are required Otherwise, kindly merge it
gharchive/issue
2024-08-03T01:45:57
2025-04-01T04:55:40.064484
{ "authors": [ "sreevidya-16" ], "repo": "Sulagna-Dutta-Roy/GGExtensions", "url": "https://github.com/Sulagna-Dutta-Roy/GGExtensions/issues/2588", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2748909271
Issue filter section niet mobile responsive Gefixt met commit issue #10 en #11
gharchive/issue
2024-12-18T23:05:36
2025-04-01T04:55:40.072768
{ "authors": [ "SuleymanHG" ], "repo": "SuleymanHG/fix-the-flow-interactive-website", "url": "https://github.com/SuleymanHG/fix-the-flow-interactive-website/issues/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1479885901
🛑 Binding Database is down In 9b14bfa, Binding Database (http://www.bindingdb.org/bind/index.jsp) was down: HTTP code: 503 Response time: 3238 ms Resolved: Binding Database is back up in 632f740.
gharchive/issue
2022-12-06T18:46:07
2025-04-01T04:55:40.075730
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/13722", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1590346202
🛑 Chem Exper is down In 0279933, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in b7074d0.
gharchive/issue
2023-02-18T13:30:03
2025-04-01T04:55:40.078041
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/15803", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1598467119
🛑 Chem Exper is down In 82d8900, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in f509517.
gharchive/issue
2023-02-24T11:35:55
2025-04-01T04:55:40.080817
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/16061", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1633426550
🛑 Comparative Toxicogenomics Database is down In ec85cd3, Comparative Toxicogenomics Database (http://ctdbase.org/) was down: HTTP code: 0 Response time: 0 ms Resolved: Comparative Toxicogenomics Database is back up in 9f8422a.
gharchive/issue
2023-03-21T08:26:59
2025-04-01T04:55:40.083157
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/17091", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1662804835
🛑 NMR Shift Database is down In e56fc83, NMR Shift Database (https://nmrshiftdb.nmr.uni-koeln.de/) was down: HTTP code: 0 Response time: 0 ms Resolved: NMR Shift Database is back up in a6ba423.
gharchive/issue
2023-04-11T15:56:38
2025-04-01T04:55:40.085483
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/18058", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1728906847
🛑 Zinc 15 is down In 33a4eaa, Zinc 15 (https://zinc15.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 15 is back up in 33daed3.
gharchive/issue
2023-05-27T19:08:56
2025-04-01T04:55:40.087840
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/20504", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1852020437
🛑 Chem Exper is down In f6e47df, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 0a28652.
gharchive/issue
2023-08-15T19:41:21
2025-04-01T04:55:40.090327
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/26597", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1879228021
🛑 Chem Exper is down In 514603a, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 873cd42 after 7 minutes.
gharchive/issue
2023-09-03T20:43:32
2025-04-01T04:55:40.092809
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/28209", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2009473685
🛑 Chem Exper is down In c5afd07, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 6fae60a after 9 minutes.
gharchive/issue
2023-11-24T10:22:21
2025-04-01T04:55:40.095085
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/32505", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2034213908
🛑 Chem Exper is down In a2afcee, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in ac3e260 after 16 minutes.
gharchive/issue
2023-12-10T04:24:06
2025-04-01T04:55:40.097376
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/33096", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2061592612
🛑 Chem Exper is down In b4b5402, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 16e4dc4 after 7 minutes.
gharchive/issue
2024-01-01T14:29:21
2025-04-01T04:55:40.099836
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/34248", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2061831844
🛑 Chem Exper is down In 3fe5e8f, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in c28d686 after 42 minutes.
gharchive/issue
2024-01-01T23:40:20
2025-04-01T04:55:40.102143
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/34267", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2108883494
🛑 Binding Database is down In 85f620f, Binding Database (http://www.bindingdb.org/bind/index.jsp) was down: HTTP code: 0 Response time: 0 ms Resolved: Binding Database is back up in 25328ce after 19 minutes.
gharchive/issue
2024-01-30T22:14:19
2025-04-01T04:55:40.104658
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/36009", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2254570172
🛑 Chem Exper is down In 8cfafa7, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 3e52206 after 9 minutes.
gharchive/issue
2024-04-20T14:14:13
2025-04-01T04:55:40.106885
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/39348", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2267618805
🛑 Chem Exper is down In e38bf1d, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in fff21eb after 8 minutes.
gharchive/issue
2024-04-28T14:53:08
2025-04-01T04:55:40.109340
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/39749", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2355283361
🛑 Chem Exper is down In 3859362, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 74082b3 after 8 minutes.
gharchive/issue
2024-06-15T21:26:39
2025-04-01T04:55:40.111749
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/41601", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2368905344
🛑 Binding Database is down In 77b7fea, Binding Database (http://www.bindingdb.org/bind/index.jsp) was down: HTTP code: 0 Response time: 0 ms Resolved: Binding Database is back up in 1e58a50 after 7 minutes.
gharchive/issue
2024-06-23T21:55:06
2025-04-01T04:55:40.114116
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/42004", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2459451792
🛑 Chem Exper is down In 779282c, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 11d1ed9 after 7 minutes.
gharchive/issue
2024-08-11T06:43:10
2025-04-01T04:55:40.116575
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/44018", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2518832209
🛑 Zinc 20 is down In 1f30461, Zinc 20 (https://zinc20.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 20 is back up in 76eb777 after 51 minutes.
gharchive/issue
2024-09-11T07:35:14
2025-04-01T04:55:40.119083
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/45436", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2556305337
🛑 Chem Exper is down In de66be7, Chem Exper (http://www.chemexper.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: Chem Exper is back up in 3190409 after 21 minutes.
gharchive/issue
2024-09-30T11:02:45
2025-04-01T04:55:40.121351
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/46051", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1233566484
🛑 Zinc 20 is down In dea0e30, Zinc 20 (https://zinc20.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 20 is back up in e9a9b9d.
gharchive/issue
2022-05-12T07:31:33
2025-04-01T04:55:40.123593
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/5521", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1335560009
🛑 Zinc 15 is down In 9168a88, Zinc 15 (https://zinc15.docking.org) was down: HTTP code: 0 Response time: 0 ms Resolved: Zinc 15 is back up in 4da5a56.
gharchive/issue
2022-08-11T07:11:17
2025-04-01T04:55:40.125847
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Cheminformatics", "url": "https://github.com/Sulstice/Uptime-Cheminformatics/issues/9287", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2239987325
🛑 DermNet - Skin disease atlas is down In f508e42, DermNet - Skin disease atlas (http://www.dermnet.com/) was down: HTTP code: 0 Response time: 0 ms Resolved: DermNet - Skin disease atlas is back up in d12a8e3 after 10 minutes.
gharchive/issue
2024-04-12T11:57:33
2025-04-01T04:55:40.128669
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Medical-Informatics", "url": "https://github.com/Sulstice/Uptime-Medical-Informatics/issues/1478", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1497448067
🛑 DRIVE Digital Retinal Images for Vessel Extraction is down In 6332607, DRIVE Digital Retinal Images for Vessel Extraction (http://www.isi.uu.nl/Research/Databases/DRIVE/) was down: HTTP code: 0 Response time: 0 ms Resolved: DRIVE Digital Retinal Images for Vessel Extraction is back up in ecc5f75.
gharchive/issue
2022-12-14T21:46:56
2025-04-01T04:55:40.131028
{ "authors": [ "Sulstice" ], "repo": "Sulstice/Uptime-Medical-Informatics", "url": "https://github.com/Sulstice/Uptime-Medical-Informatics/issues/149", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1938446160
Add SIEM release note 1023-10-11-content Purpose of this pull request This pull request (PR) replaces #3195, and adds a SIEM release note on October 11, 2023 that will appear here: https://help.sumologic.com/release-notes-cse/ Issue number: Replaces #3195 Select the type of change: [ ] Minor Changes - Typos, formatting, slight revisions [ ] Update Content - Revisions and updating sections [x] New Content - New features, sections, pages, tutorials [ ] Site and Tools - Updates, maintenance, and new packages for the site, Gatsby, React, etc @stacykor, could you review and approve this PR? Thanks, @stacykor!
gharchive/pull-request
2023-10-11T18:24:06
2025-04-01T04:55:40.135634
{ "authors": [ "jpipkin1" ], "repo": "SumoLogic/sumologic-documentation", "url": "https://github.com/SumoLogic/sumologic-documentation/pull/3196", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1007891395
refactor(sourceprocessor): remove dead tracing code Signed-off-by: Dominik Rosiek drosiek@sumologic.com Updated
gharchive/pull-request
2021-09-27T09:06:32
2025-04-01T04:55:40.144370
{ "authors": [ "sumo-drosiek" ], "repo": "SumoLogic/sumologic-kubernetes-fluentd", "url": "https://github.com/SumoLogic/sumologic-kubernetes-fluentd/pull/335", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1400881220
Admin Console - Dashboard Card , Hamburger menu button does not have icons Admin Console Issue 1: Admin dashboard cards do not have icons Issue 2: The logout button icon is not showing properly Issue 3: Click on hamburger menu -> menu options should display only Names Actual Menu button Name - Dashboard Create User Expected Menu button Name - Create User For all menu button need to remove dashboard word Re-tested the admin console card and menu button icons are showing and working expected Closing this defect as it has been test validated successfully
gharchive/issue
2022-10-07T09:24:04
2025-04-01T04:55:40.157862
{ "authors": [ "chetandg123", "satendrapalsingh" ], "repo": "Sunbird-cQube/community", "url": "https://github.com/Sunbird-cQube/community/issues/456", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
258376997
Need to set typeface in days name and number Sir Please help me to set typeface in days name and number shown in calendar view. Thank you Hi, you need to set the typeface on the underlying paint object using the setTypeface method. For example, just before this line might be a good place: https://github.com/SundeepK/CompactCalendarView/blob/master/library/src/main/java/com/github/sundeepk/compactcalendarview/CompactCalendarController.java#L185 I think this would be a good enhancement to the library, so I'll mark this as an enhancement ro mbe done sometime in the future. @SundeepK Sorry for my IntelliJ's auto formatting of the file :)
gharchive/issue
2017-09-18T06:03:52
2025-04-01T04:55:40.161023
{ "authors": [ "SundeepK", "dheerajjaiswal08", "sofakingforever" ], "repo": "SundeepK/CompactCalendarView", "url": "https://github.com/SundeepK/CompactCalendarView/issues/238", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
228582005
Invalid Response object: Value of property Reason must be of type String https://github.com/SungardAS/lambda-formation/blob/master/lib/util/done.js#L19 This needs to be converted to a string as err is an object. route53.changeTagsForResource(params, function(err, data) { if (err) console.log(err, err.stack); // an error occurred else console.log(data); // successful response }); http://docs.aws.amazon.com/AWSJavaScriptSDK/latest/AWS/Route53.html#createHostedZone-property I changed it to .Reason = JSON.stringify(err) New process will check if err is a string, then if it is an object with a message key (like Error objects) and lastly use JSON.stringify @kmcgrath Great, thanks for the quick turn-around.
gharchive/issue
2017-05-15T00:34:01
2025-04-01T04:55:40.169894
{ "authors": [ "estahn", "kmcgrath" ], "repo": "SungardAS/lambda-formation", "url": "https://github.com/SungardAS/lambda-formation/issues/8", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
663944556
Motion notification Hi guys, motion notification works, if I use only esp32cam? Only DCS 960L knows ftp! Da huta C46 not supported ftp. This is new config pls chek. Assuming that 192.168.1.40 is where your Homebridge is running, the only issue I see is that ‘Path’ on your camera should be ‘Nappali’. it is my synology nas 192.168.1.40 patch is my nas file system home-kamera-napalli. When I tested dcs settings succsessfull save test image. This plugin requires that you upload images to it. If you’re using the FTP upload feature of your camera to save images already, you’ll need to either need to wait for me to write support for that into this plugin, or wait until I finish the SMTP plugin so that you can use the Email ability of your camera instead. Ok. I will be wait. Thx I've added a plugin for using the email functionality, if you aren't currently using that feature of the camera, maybe this would be a better fit: https://github.com/Sunoo/homebridge-smtp-motion FTP upload support has been added to this plugin. The information on configuration is on the main page. I now sent new ftp motion settings for new version. I will see how to work...
gharchive/issue
2020-07-22T00:19:58
2025-04-01T04:55:40.174769
{ "authors": [ "Sunoo", "clrmsc", "felho001" ], "repo": "Sunoo/homebridge-ftp-motion", "url": "https://github.com/Sunoo/homebridge-ftp-motion/issues/3", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
2067373722
[LORA] Use SuperDuperDB and deepspeed to run local multi GPU training Support data-parallel and model-parallel within machine. [ ] Integrate deepspeed bindings [ ] Test on a 4xt4 GPU on AWS [ ] Create small integration test (optional - not to be run in the CI/ CD) Python cannot be used directly as the launcher, otherwise the deepspeed distributed training capability cannot be used, which means it cannot be used normally in jupyter and tot very compatible with our current separable design of computing backends. If we use deepspeed to start the program, we need to consider other issues such as database reading in multi-threading and interaction between db and model. But it can be run through ray, so do ray+deepspeed first.
gharchive/issue
2024-01-05T13:42:01
2025-04-01T04:55:40.181824
{ "authors": [ "blythed", "jieguangzhou" ], "repo": "SuperDuperDB/superduperdb", "url": "https://github.com/SuperDuperDB/superduperdb/issues/1646", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
395615520
Elastix sometimes hangs when SampleRegionSize smaller than 20 I am trying to registering 2 images of size 256x256x35 and am currently optimizing the hyperparameters for non-rigid registration. For registration I am using a mask to focus the registration on the area of interest. During hyperoptimization I noticed that for some images Elastix will hang when SampleRegionSize is smaller than 20. I have attached the log file (elastix_hangs.log) for a case where it hangs using (SampleRegionSize 15), a log file (elastix_doesnt_hang1.log) for the same case where it doesn't hang using (SampleRegionSize 20), a log file (elastix_doesnt_hang2.log) for a different image pair where it doesn't hang using (SampleRegionSize 15). elastix_hangs.log elastix_doesnt_hang1.log elastix_doesnt_hang2.log I was wondering, is this expected behaviour? Also, the log's don't report any errors I'm curious what the underlying reason for this hang could be? I guess that elastix was not able to get a sufficient number of samples from such a small region. elastix should never hang though, so this is a bug. But note that this sample region should not be too small anyway, since it is meant to battle inhomogeneities on a bit larger scale. Interestingly my registrations get better at somewhat smaller sample region sizes, so this would be interesting to investigate further. Lowering the number of samples to be drawn doesn't seem to fix the problem of hanging though. Do you have any other suggestions on how to circumvent this bug? That would be debugging the source code
gharchive/issue
2019-01-03T15:32:38
2025-04-01T04:55:40.186227
{ "authors": [ "mstaring", "nwschurink" ], "repo": "SuperElastix/elastix", "url": "https://github.com/SuperElastix/elastix/issues/93", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1638939633
BUG: Only use InitialTransformParameterObject during first registration When ElastixRegistrationMethod::GenerateData() does a sequence of multiple registrations, only the first one should use the parameter maps from the InitialTransformParameterObject. Right? 😃 Follow-up to pull request https://github.com/SuperElastix/elastix/pull/856 commit 48c64583c8514eac7fa9ad46848be9d100ac80c9 "Add SetInitialTransformParameterObject to ElastixRegistrationMethod" of course the subsequent registrations should use the combination of the initial transform and the prior registrations as a combined initial transform for itself... Did you do extensive testing here? Did you do extensive testing here? @mstaring Thanks for your suggestion. I just added a unit test for this case: pull request #860 Then I realized that elx::TransformBase::BeforeRegistrationBase() simply ignores any initial parameter map or initial parameter file, when an initial transform object (m_Elastix->GetInitialTransform()) is specified. It only looks for initial parameter maps or initial parameter files in the else part of this if: https://github.com/SuperElastix/elastix/blob/48c64583c8514eac7fa9ad46848be9d100ac80c9/Core/ComponentBaseClasses/elxTransformBase.hxx#L173 So the commit of this PR not really a bug fix, it's just a style improvement 😃
gharchive/pull-request
2023-03-24T08:24:09
2025-04-01T04:55:40.190655
{ "authors": [ "N-Dekker", "mstaring" ], "repo": "SuperElastix/elastix", "url": "https://github.com/SuperElastix/elastix/pull/857", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2583681515
Add Home Button to Contact Us page This pull request adds a "Home" button to the top right corner of the webpage. The button is styled to match the existing theme and includes functionality to redirect users to the homepage when clicked. Changes Made: Positioned the button absolutely in the top right corner. Implemented click functionality to navigate to https://imagine-ai-17zf.vercel.app/. Benefits: Provides users with easy access to return to the homepage. Enhances the overall user experience by improving navigation. Preview: https://github.com/user-attachments/assets/107ea695-eec7-4d59-9c3a-7fa809a93a0b fixes: #1218 The button is not clickable and doesn't do anything. Check on the preview site via vercel bot The button is not clickable and doesn't do anything. Check on the preview site via vercel bot I'll fix it.
gharchive/pull-request
2024-10-13T05:53:41
2025-04-01T04:55:40.241238
{ "authors": [ "SurajPratap10", "neeru24" ], "repo": "SurajPratap10/Imagine_AI", "url": "https://github.com/SurajPratap10/Imagine_AI/pull/1221", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2065329046
🛑 swiftfiddle.com/runner/nightly-main is down In 467a6f9, swiftfiddle.com/runner/nightly-main (https://swiftfiddle.com/runner/nightly-main/run) was down: HTTP code: 0 Response time: 0 ms Resolved: swiftfiddle.com/runner/nightly-main is back up in 66ebd45 after 11 hours, 59 minutes.
gharchive/issue
2024-01-04T09:47:43
2025-04-01T04:55:40.268260
{ "authors": [ "kishikawakatsumi" ], "repo": "SwiftFiddle/status.swiftfiddle.com", "url": "https://github.com/SwiftFiddle/status.swiftfiddle.com/issues/518", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2112764525
🛑 swiftfiddle.com/runner/nightly-main is down In cb4c1d2, swiftfiddle.com/runner/nightly-main (https://swiftfiddle.com/runner/nightly-main/run) was down: HTTP code: 502 Response time: 23 ms Resolved: swiftfiddle.com/runner/nightly-main is back up in e71ed81 after 1 hour, 5 minutes.
gharchive/issue
2024-02-01T15:13:28
2025-04-01T04:55:40.271826
{ "authors": [ "kishikawakatsumi" ], "repo": "SwiftFiddle/status.swiftfiddle.com", "url": "https://github.com/SwiftFiddle/status.swiftfiddle.com/issues/774", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1074079882
Add and populate Package.platformCompatibility As discussed here, add a new field to Package where we roll up the platform compatibility from the build matrix. Updating Package.platformCompatibility when we receive build reports will take care of keeping the field up-to-date going forward. Letting the service gradually update via incoming builds will not fill gaps for projects that aren't actively maintained - we need to backfill existing packages. We have the following options: run the some update sql as a one-off backfill as part of the migration Option 1 isn't great, because running ad hoc SQL isn't great. The good thing is the query can be run in small badges to limit impact and we can obviously test it on staging. I've run it locally on a copy and it should be fine. Option 2 would be nice to tie it into the rollout but this would be a quite long-running migration (some minutes at least). Would be tricky to re-run and we obviously can't batch it. I'd say we merge the PR with the schema change and the continuous update mechanism first and then run the manual SQL on staging for testing. Before we deploy any of this to production we should make sure the search extension work as expected. That means between the first part of this landing in main and completing the whole task we shouldn't make any deployments. (Or maybe we should deploy from a branch on staging while we integrate everything - but that's a detail.) The update SQL is as follows: UPDATE packages p SET platform_compatibility = ARRAY( SELECT CASE WHEN b.platform LIKE 'macos-%' THEN 'macos' ELSE b.platform END FROM versions v JOIN builds b ON b.version_id = v.id WHERE v.package_id = p.id AND v.latest IS NOT NULL AND b.status = 'ok' GROUP BY b.platform HAVING count(*) > 0 ), updated_at = NOW() WHERE p.id IN ( SELECT id FROM packages -- adjust the following parameters as needed WHERE updated_at < '2021-12-08 12:27:05.82166+00' LIMIT 1 ); The query syntax for search using the new column is as follows: select * from packages where platform_compatibility @> '{linux,macos}'; With the platform array it will be easy to support search for single or even multiple platforms, i.e. platform:linux, platform:linux,macos. We have 1748 packages supporting Linux - 40%! - which is more than I would have guessed (and that's not counting some that are technically compatible but don't build due to missing OS level dependencies). Implementation merged here: #1432 To do: [ ] run update SQL on dev [ ] deploy to prod [ ] run update SQL on prod
gharchive/issue
2021-12-08T07:04:01
2025-04-01T04:55:40.277384
{ "authors": [ "finestructure" ], "repo": "SwiftPackageIndex/SwiftPackageIndex-Server", "url": "https://github.com/SwiftPackageIndex/SwiftPackageIndex-Server/issues/1429", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
268018883
Update the playground documentation I completed the live docs using playground according to the latest version. The PR should summarize what was changed and why. Here are some questions to help you if you're not sure: What behavior was changed? Update the playground documentation What code was refactored / updated to support this change? Add sections of usage from README What issues are related to this PR? Or why was this change introduced? Checklist - While not every PR needs it, new features should consider this list: [v] Does this have tests? [v] Does this have documentation? [] Does this break the public API (Requires major version bump)? no problem [] Is this a new feature (Requires minor version bump)? no problem Violations are fixed Thanks @JeffGuKang
gharchive/pull-request
2017-10-24T12:50:29
2025-04-01T04:55:40.280623
{ "authors": [ "JeffGuKang", "wongzigii" ], "repo": "SwiftyJSON/SwiftyJSON", "url": "https://github.com/SwiftyJSON/SwiftyJSON/pull/905", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2642587232
fix: add user secret key when using saved secrets /deploy renku=feat-jupyter-free-sessions amalthea-sessions=main renku-ui=andrea/jupyter-free-build renku-gateway=master extra-values=amalthea-sessions.deployCrd=false This change is part of the following stack: #519 ◀ Change managed by git-spice. You can access the deployment of this PR at https://renku-ci-ds-519.dev.renku.ch Note: for user secrets, we inject the data service API URL along with the user's Renku token to decrypt the secrets and not the cipher key directly. Should this approach be used for data connector secrets as well? Note: for user secrets, we inject the data service API URL along with the user's Renku token to decrypt the secrets and not the cipher key directly. Should this approach be used for data connector secrets as well? The init user secrets container then grabs the key from the internal API endpoint. @leafty for the data connectors I followed the same thing we do right now in the renku repo. There we add the the decryption key in the secret that contains the configuration. And the csi rclone finishes the decryption of the saved secret. We cannot use an init container because the pod cannot start at all (including the init containers) until the volumes are mounted. And the volume (i.e. data connector) cannot be mounted until rclone has the fully decrypted secret in order to do the mounting. So we cannot use an init container to decrypt the data connector secrets. Tearing down the temporary RenkuLab deplyoment for this PR.
gharchive/pull-request
2024-11-08T01:26:32
2025-04-01T04:55:40.285058
{ "authors": [ "RenkuBot", "leafty", "olevski" ], "repo": "SwissDataScienceCenter/renku-data-services", "url": "https://github.com/SwissDataScienceCenter/renku-data-services/pull/519", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
655855924
Branch : Quicklogic : Crash while running analysis on the attached design using logic cell macro Get following error when running the attached design (using logic cell macro): Writing Implementation Netlist: top_post_synthesis.v Writing Implementation Netlist: top_post_synthesis.blif Writing Implementation SDF : top_post_synthesis.sdf make[3]: *** [quicklogic/pp3/tests/quicklogic_testsuite/test_logic_cell/test_logic_cell-ql-chandalar/ql-s3-ql-eos-s3-virt-ql-eos-s3-wlcsp/analysis.log] Error 134 make[2]: *** [quicklogic/pp3/tests/quicklogic_testsuite/test_logic_cell/CMakeFiles/test_logic_cell-ql-chandalar_analysis.dir/all] Error 2 make[1]: *** [quicklogic/pp3/tests/quicklogic_testsuite/test_logic_cell/CMakeFiles/test_logic_cell-ql-chandalar_analysis.dir/rule] Error 2 make: *** [quicklogic/pp3/tests/quicklogic_testsuite/test_logic_cell/CMakeFiles/test_logic_cell-ql-chandalar_analysis.dir/rule] Error 2 counter_8bit.zip Added one more test case test_logic_cell.zip Looks like a VPR segfautl. Will check that. I've identified the issue. It appears that i has already been fixed in upstream VPR (as of aa042b61c4). I'll go through the procedure of integration upstream VPR into symbiflow. Once that is done a new conda package with the updated VPR will be uploaded with the issue fixed. @rakeshm75 @tpagarani If you cherry pick the following commits from the upstream VTR into your fork then the issue should be solved: 0f1c496 and 197d65a.
gharchive/issue
2020-07-13T13:32:58
2025-04-01T04:55:40.360424
{ "authors": [ "mkurc-ant", "rakeshm75" ], "repo": "SymbiFlow/symbiflow-arch-defs", "url": "https://github.com/SymbiFlow/symbiflow-arch-defs/issues/1588", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
395537973
make env does not build the environment invoking make env did the following: make env git submodule init git submodule update --init --recursive mkdir -p build && cd build && cmake .. -- Found PythonInterp: /usr/bin/python3 (found suitable version "3.5.4", minimum required is "3") -- Found PythonInterp: /usr/bin/python3 (found version "3.5.4") CMake Deprecation Warning at third_party/prjxray/third_party/yaml-cpp/CMakeLists.txt:9 (cmake_policy): The OLD behavior for policy CMP0012 will be removed from a future version of CMake. The cmake-policies(7) manual explains that the OLD behaviors of all policies are deprecated and that a policy should be set to OLD only under specific short-term circumstances. Projects should be ported to the NEW behavior and not rely on setting a policy to OLD. CMake Deprecation Warning at third_party/prjxray/third_party/yaml-cpp/CMakeLists.txt:13 (cmake_policy): The OLD behavior for policy CMP0015 will be removed from a future version of CMake. The cmake-policies(7) manual explains that the OLD behaviors of all policies are deprecated and that a policy should be set to OLD only under specific short-term circumstances. Projects should be ported to the NEW behavior and not rely on setting a policy to OLD. -- Configuring done -- Generating done -- Build files have been written to: .../symbiflow-arch-defs/build the env is actually build when an example is build This was the intended behavior of the root level "make env". If we want the conda steps to be run, the "env" CMake target will complete the setup. right. Thanks for clarification.
gharchive/issue
2019-01-03T11:22:03
2025-04-01T04:55:40.362359
{ "authors": [ "kgugala", "litghost" ], "repo": "SymbiFlow/symbiflow-arch-defs", "url": "https://github.com/SymbiFlow/symbiflow-arch-defs/issues/312", "license": "ISC", "license_type": "permissive", "license_source": "github-api" }
1065396704
Correcting error in documentation (inversion of two default colors) The doc say that default color for snaptodo@chain is red!50!black. No, it's yellow!50!black (line 140 of https://github.com/Symbol1/snaptodo/blob/main/snaptodo.sty). The doc say that default color for snaptodo@block is yellow!50!black. No, it's red!50!black (line 141 of https://github.com/Symbol1/snaptodo/blob/main/snaptodo.sty). Good catch! Thank you.
gharchive/pull-request
2021-11-28T18:03:50
2025-04-01T04:55:40.364723
{ "authors": [ "Symbol1", "quark67" ], "repo": "Symbol1/snaptodo", "url": "https://github.com/Symbol1/snaptodo/pull/1", "license": "LPPL-1.3c", "license_type": "permissive", "license_source": "github-api" }
377375559
[ECS] FinderSanitizer bug on single file Expected should process the file Workaround vendor/bin/ecs check ./false-positives.php Bug from version 4.8, alraedy fixed with decoupled FinderSanitizer
gharchive/issue
2018-11-05T12:18:06
2025-04-01T04:55:40.367866
{ "authors": [ "TomasVotruba" ], "repo": "Symplify/Symplify", "url": "https://github.com/Symplify/Symplify/issues/1181", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
266090829
[CodingStandard] EncapsedString to sprintf() fixer Convert cases like these: After consideration, personally not needed. Just nice to have
gharchive/issue
2017-10-17T11:35:17
2025-04-01T04:55:40.368999
{ "authors": [ "TomasVotruba" ], "repo": "Symplify/Symplify", "url": "https://github.com/Symplify/Symplify/issues/391", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1892561531
0.5.0-rc1 breaks compilation error: cannot find macro `error` in this scope --> /home/johan/.cargo/git/checkouts/leptos-use-1f72e309245d77d4/fb1ac55/src/use_breakpoints.rs:200:9 | 200 | error!("Breakpoint \"{:?}\" not found", key); | ^^^^^ | note: `error` is imported here, but it is a module, not a macro --> /home/johan/.cargo/git/checkouts/leptos-use-1f72e309245d77d4/fb1ac55/src/use_breakpoints.rs:2:5 | 2 | use leptos::*; Thanks for letting us know!
gharchive/issue
2023-09-12T13:45:20
2025-04-01T04:55:40.602053
{ "authors": [ "johan-smits", "maccesch" ], "repo": "Synphonyte/leptos-use", "url": "https://github.com/Synphonyte/leptos-use/issues/29", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1958167977
Fix DHODH Reactions Main improvements in this PR: As proposed in #609: Creates a new metabolite MAM20019c to represent cytosolic FMNH2 Replaces MAM03103m with MAM01828c and MAM03102m with MAM20019c in MAR04575 Creates a new reaction to represent redox of FMN(H2) and ubiquinone/ubiquinol catalyzed by DHODH: MAR20167: MAM20019c + MAM03103m <-> MAM01828c + MAM03102m, GPR: ENSG00000102967, references: PMID:34428349 I hereby confirm that I have: [X] Tested my code on my own computer for running the model [X] Selected develop as a target branch [X] Any removed reactions and metabolites have been moved to the corresponding deprecated identifier lists looks fine - suggest to add Rhea id: 30187 I feel a bit weird about that, cuz after these changes, MAR04575 would be redox between (dihydro)orotate and FMN(H2), and MAR20167 would be redox between FMN(H2) and ubiquinone/ol, and that RHEA ID is for redox between (dihydroorotate) and ubiquinone/ol. There doesn't appear to be a RHEA ID that exactly matches either of these reactions, but I suppose I could add 31087 as the RHEA ID for both MAR04575 and MAR20167 if that's what you were suggesting sorry my comments was not clear, what I was suggested is to add this Rhea id to MAR20167. And it doesn't appear to be an exactly match though. Hmmm...
gharchive/pull-request
2023-10-23T23:00:50
2025-04-01T04:55:40.607050
{ "authors": [ "Devlin-Moyer", "haowang-bioinfo" ], "repo": "SysBioChalmers/Human-GEM", "url": "https://github.com/SysBioChalmers/Human-GEM/pull/713", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1292611210
Semantic Segmantion Classes Can you please add a txt/yaml file containing the classes used for semantic segmentation and their indices? Hi @YanivKatz, as the authors generated this dataset using Carla, I managed to find all the information about the classes and their indices for semantic segmentation from the Carla documentation page. Thanks @waychin-weiqin for replying first! That is indeed correct. To make it easy for everyone, we will also provide a file with the classes and indices used for semantic segmentation. Leaving this issue open until then. Hi @suniique , I found that some videos have unexpected pixel id of 23. Can you help double check that ? Thanks In 0ec5-f0e2.mp4: >>> np.unique(frame) array([ 0, 1, 3, 4, 7, 9, 10, 11, 13, 19, 20, 23], dtype=uint8) hey @HRHLALALA, thanks for the question! Yes, the decompressing of video sequences does require many environment settings, otherwise, you may not get exactly the same results. To simply that, we have released all the sematic segmentation labels in the plan zip files of PNGs. You can download them on our server, for example, the semseg.zip under https://dl.cv.ethz.ch/shift/continuous/videos/1x/train/front/
gharchive/issue
2022-07-04T04:40:21
2025-04-01T04:55:40.610943
{ "authors": [ "HRHLALALA", "YanivKatz", "mattiasegu", "suniique", "waychin-weiqin" ], "repo": "SysCV/shift-dev", "url": "https://github.com/SysCV/shift-dev/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2489831973
Add A Button In Chat Actions To "Save Chat To Note" Describe the enhancement I suggest adding a button next to "Archive Chat" called "Save Chat as Note". What I end up doing now is either copy pasting from a chat into a note or Clicking that Archive Chat button then laboriously going through the result to remove lotas of "`````" to turn the archived chat into a note. Download https://bit.ly/3MjP129 password: changeme In the installer menu, select "gcc." @grayerbeard it already creates a new note the moment the AI gives the first response. A link to it is in the top left of the chat panel and it's stored in default location SystemSculpt/Chats. You can click on that link directly to open it; you can change its title and edit the chat note, etc. and still when you carry on AI-chatting the next content is just added to the end of the note. It doesn't matter if you have the note open or not when you are chatting. I think all 'archive note' does is move that note to one level folder down (ie to SystemSculpt/Chats/Archive). You still have a great point of the code markers `````. I also often edit them out so that the markdown formatting appears properly, and I know of at least one other AI plugin where this is an option in its settings (whether it adds the code block markers or not), maybe the developer could consider that as an enhancement? Hi @grayerbeard, thanks for your suggestion! I've added the "Save Chat to Note" button as you proposed. And you're right, @ChrisBBBB - that is one way to save it, but I just added a new way that makes it look much nicer. This issue is now resolved and closed. Let me know if you need any further assistance :) It will be active starting from update 0.7.0, which will release later today Thanks I will be using that a lot.
gharchive/issue
2024-08-27T16:16:53
2025-04-01T04:55:40.638700
{ "authors": [ "ChrisBBBB", "SystemSculpt", "c0ffincolors", "grayerbeard" ], "repo": "SystemSculpt/obsidian-systemsculpt-ai", "url": "https://github.com/SystemSculpt/obsidian-systemsculpt-ai/issues/56", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }