id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
357687855
Configure Intake to cache files in the current directory From @jbednar In most cases I'd rather have the files associated with the notebook or project where I'm working, not dumped into an ever-increasing global cache heap. I would be nice to have a simple flag for configuring intake (something intake.local_cache=True) to store cached items in a subdirectory like .intake_cache/ in the current directory. I hadn't thought of this usage model, but I guess it would work in some cases. I would definitely not want to make it the default behavior, as these cache directories are both invisible and hard for a CLI to discover without scanning your whole hard drive. This sounds pretty reasonable to me. The CLI would get file locations from the metadata, which would be small and remain in the (global) config directory. I would suggest config['cache_dir'] = "datadir" could do this (better names accepted). Yeah, the CLI could look at the metadata which should avoid the need to scan the hard drive. I would definitely not want to make it the default behavior, as these cache directories are both invisible and hard for a CLI to discover without scanning your whole hard drive. It might make sense to make the directory not be invisible, e.g. ./intake_cache/; there are pros and cons to doing that. But there are problems with both approaches, i.e. putting cache files in cwd and in the home directory; in one case it's clear what would be affected by deleting that file, but hard to find all of them (though one can do find ~ -name intake_cache), and in the other it's easy to find all of them but hard to tell what any particular file is and whether it can be deleted without a major download cost in your future. Yeah, the CLI could look at the metadata which should avoid the need to scan the hard drive. That makes sense -- the metadata should be tiny in any reasonable case, and presumably no one should object to having that in their homedir.
gharchive/issue
2018-09-06T14:41:41
2025-04-01T04:54:51.668196
{ "authors": [ "jbednar", "martindurant", "mmccarty", "seibert" ], "repo": "ContinuumIO/intake", "url": "https://github.com/ContinuumIO/intake/issues/151", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
674557982
POST api/users/token/:id I still need Zyleaf to make the logic that creates the token. I only setup the boilerplate for it. Should we change it to no 7 day thing This is ready for merge
gharchive/pull-request
2020-08-06T19:44:00
2025-04-01T04:54:51.676928
{ "authors": [ "CreeperPlanet26" ], "repo": "Cool-Bot-List/Cool-Bot-List", "url": "https://github.com/Cool-Bot-List/Cool-Bot-List/pull/55", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1323559002
⚠️ Samsung has degraded performance In 6452b5b, Samsung (https://samsung.com) experienced degraded performance: HTTP code: 200 Response time: 3027 ms Resolved: Samsung performance has improved in 32825fa.
gharchive/issue
2022-07-31T19:49:14
2025-04-01T04:54:51.724542
{ "authors": [ "CoopPlayzz-Bot" ], "repo": "CoopPlayzz/Webstatus-electronicbrands", "url": "https://github.com/CoopPlayzz/Webstatus-electronicbrands/issues/2301", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1228726060
⚠️ Samsung has degraded performance In f34ad1d, Samsung (https://samsung.com) experienced degraded performance: HTTP code: 200 Response time: 2181 ms Resolved: Samsung performance has improved in 7502ec0.
gharchive/issue
2022-05-07T21:23:14
2025-04-01T04:54:51.727348
{ "authors": [ "CoopPlayzz-Bot" ], "repo": "CoopPlayzz/Webstatus-electronicbrands", "url": "https://github.com/CoopPlayzz/Webstatus-electronicbrands/issues/827", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1317841472
⚠️ Fortnite has degraded performance In cdfa796, Fortnite (https://fortnite.com) experienced degraded performance: HTTP code: 200 Response time: 1599 ms Resolved: Fortnite performance has improved in 0f955ba.
gharchive/issue
2022-07-26T07:50:35
2025-04-01T04:54:51.729828
{ "authors": [ "CoopPlayzz-Bot" ], "repo": "CoopPlayzz/Webstatus-games", "url": "https://github.com/CoopPlayzz/Webstatus-games/issues/1380", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1514466272
⚠️ Fortnite has degraded performance In 7b65845, Fortnite (https://fortnite.com) experienced degraded performance: HTTP code: 200 Response time: 1556 ms Resolved: Fortnite performance has improved in af47af9.
gharchive/issue
2022-12-30T11:47:19
2025-04-01T04:54:51.731749
{ "authors": [ "CoopPlayzz-Bot" ], "repo": "CoopPlayzz/Webstatus-games", "url": "https://github.com/CoopPlayzz/Webstatus-games/issues/2718", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1344787821
🛑 The Coop Computer [By Coop] is down In 0890bb7, The Coop Computer [By Coop] (http:/$COOPWEBSITE/) was down: HTTP code: 0 Response time: 0 ms Resolved: The Coop Computer [By Coop] is back up in ef14337.
gharchive/issue
2022-08-19T18:41:43
2025-04-01T04:54:51.733275
{ "authors": [ "CoopPlayzz-Bot" ], "repo": "CoopPlayzz/Webstatus-other", "url": "https://github.com/CoopPlayzz/Webstatus-other/issues/181", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
164285212
Dismiss parent view controller after picking the time Hello, I wonder if it is possible to dismiss the picker and the parent view controller at the same time. I experimented with putting dismissViewControllerAnimated in the Done action but it does not dismiss the parent - only RMDateSelectionViewController disappears. Is there any way it could be done? Hi there, I think you can do this by calling dismissViewControllerAnimated twice. However, this might be bad style :D. Best regards, Roland Thank you! This is how I overcome it for now but I wonder if there a "cleaner" way to do it.
gharchive/issue
2016-07-07T11:11:58
2025-04-01T04:54:51.735864
{ "authors": [ "CooperRS", "dafdafydd" ], "repo": "CooperRS/RMDateSelectionViewController", "url": "https://github.com/CooperRS/RMDateSelectionViewController/issues/70", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1145169463
Sample notebook reports errors Hi, many thanks for the development of the great package! I'm trying to run this sample notebook Predicting Conversations Gone Awry With Convokit on Google Colab here. I did no modifications except for the first chunk I added ! pip -q install convokit ! pip uninstall spacy -y ! pip install -U spacy==3.1.4 !python -m spacy download en_core_web_sm However, an error occurred in the second cell from the bottom: TypeError: __init__() takes from 1 to 2 positional arguments but 3 were given. Would it be possible for you to point out how to correct the error? Many thanks! Running prediction task for feature set politeness_strategies Generating labels... Computing paired features... Using 38 features Running leave-one-page-out prediction... --------------------------------------------------------------------------- RemoteTraceback Traceback (most recent call last) RemoteTraceback: """ Traceback (most recent call last): File "/usr/lib/python3.7/multiprocessing/pool.py", line 121, in worker result = (True, func(*args, **kwds)) File "/usr/lib/python3.7/multiprocessing/pool.py", line 44, in mapstar return list(map(*args)) File "<ipython-input-37-de914fca85cc>", line 11, in run_pred_single base_clf = Pipeline([("scaler", StandardScaler()), ("featselect", SelectPercentile(f_classif, 10)), ("logreg", LogisticRegression(solver='liblinear'))]) TypeError: __init__() takes from 1 to 2 positional arguments but 3 were given """ The above exception was the direct cause of the following exception: TypeError Traceback (most recent call last) [<ipython-input-38-9704095ec82e>](https://localhost:8080/#) in <module>() 4 for combo in feature_combos: 5 combo_names.append("+".join(combo).replace("_", " ")) ----> 6 accuracy = run_pipeline(combo) 7 accs.append(accuracy) 8 results_df = pd.DataFrame({"Accuracy": accs}, index=combo_names) 6 frames [<ipython-input-37-de914fca85cc>](https://localhost:8080/#) in run_pipeline(feature_set) 97 y = labeled_pairs_df.first_convo_toxic.values 98 print("Running leave-one-page-out prediction...") ---> 99 accuracy, coefs, scores, hyperparams, pvalue = run_pred(X, y, feature_names, labeled_pairs_df.page_id) 100 print("Accuracy:", accuracy) 101 print("p-value: %.4e" % pvalue) [<ipython-input-37-de914fca85cc>](https://localhost:8080/#) in run_pred(X, y, fnames, groups) 33 34 with Pool(os.cpu_count()) as p: ---> 35 prediction_results = p.map(partial(run_pred_single, X=X, y=y), splits) 36 37 fselect_pvals_all = [] [/usr/lib/python3.7/multiprocessing/pool.py](https://localhost:8080/#) in map(self, func, iterable, chunksize) 266 in a list that is returned. 267 ''' --> 268 return self._map_async(func, iterable, mapstar, chunksize).get() 269 270 def starmap(self, func, iterable, chunksize=None): [/usr/lib/python3.7/multiprocessing/pool.py](https://localhost:8080/#) in get(self, timeout) 655 return self._value 656 else: --> 657 raise self._value 658 659 def _set(self, i, obj): [/usr/lib/python3.7/multiprocessing/pool.py](https://localhost:8080/#) in worker() 119 job, i, func, args, kwds = task 120 try: --> 121 result = (True, func(*args, **kwds)) 122 except Exception as e: 123 if wrap_exception and func is not _helper_reraises_exception: [/usr/lib/python3.7/multiprocessing/pool.py](https://localhost:8080/#) in mapstar() 42 43 def mapstar(args): ---> 44 return list(map(*args)) 45 46 def starmapstar(args): [<ipython-input-37-de914fca85cc>](https://localhost:8080/#) in run_pred_single() 9 y_train, y_test = y[train_idx], y[test_idx] 10 ---> 11 base_clf = Pipeline([("scaler", StandardScaler()), ("featselect", SelectPercentile(f_classif, 10)), ("logreg", LogisticRegression(solver='liblinear'))]) 12 clf = GridSearchCV(base_clf, {"logreg__C": [10**i for i in range(-4,4)], "featselect__percentile": list(range(10, 110, 10))}, cv=3) 13 TypeError: __init__() takes from 1 to 2 positional arguments but 3 were given Hi @Raychanan, It appears that this is caused by a change to scikit-learn's SelectPercentile class in the 1.x scikit-learn release. I've committed an updated version of the notebook to deal with this change. The change is small, so if you don't want to re-upload the notebook to colab from scratch, you can simply change one line in your existing colab notebook. Find the following line: base_clf = Pipeline([("scaler", StandardScaler()), ("featselect", SelectPercentile(f_classif, 10)), ("logreg", LogisticRegression(solver='liblinear'))]) And change it to: base_clf = Pipeline([("scaler", StandardScaler()), ("featselect", SelectPercentile(score_func=f_classif, percentile=10)), ("logreg", LogisticRegression(solver='liblinear'))]) That should resolve the error! This helps a lot! Thanks so much!
gharchive/issue
2022-02-20T23:40:55
2025-04-01T04:54:51.743164
{ "authors": [ "Raychanan", "jpwchang" ], "repo": "CornellNLP/Cornell-Conversational-Analysis-Toolkit", "url": "https://github.com/CornellNLP/Cornell-Conversational-Analysis-Toolkit/issues/145", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2444478726
Jlou/feat analyse variables Description of the PR: Jira ticket ref : Jira ticket In order to easily change variables values, a refactoring of values has been made. The majority of hardcoded values in child modules now use variables and all default variables values have been moved to root variables. Variables for the remote module Cosmo-Tech/platform-core/cosmotech have been created, as part of the refactoring of the corresponding repository. All root variables have been sorted into module-related variables.tf files Format fixed, but I can't rename branch to azure
gharchive/pull-request
2024-08-02T09:10:21
2025-04-01T04:54:51.775420
{ "authors": [ "JoelLourenco" ], "repo": "Cosmo-Tech/terraform-kubernetes-cosmotech-tenant", "url": "https://github.com/Cosmo-Tech/terraform-kubernetes-cosmotech-tenant/pull/24", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
958075660
[Nomination] giansalex for telegram bot Who are you nominating? @giansalex (Giancarlos Salas) What are you nominating them for? His incredible work on Juno Telegram Bot and creation of the code that allows a user to upload a wasm file on chain Please provide links to their work. https://github.com/Highlander-maker/juno-tg-bot Nice. Would be cool if it supported token gated groups. I.E. own a certain amount of cw20 tokens to gain access to a telegram group. interesting... So smart contract that checks the balance of a wallet that knows how much of the token that can gain access ? interesting... So smart contract that checks the balance of a wallet that knows how much of the token that can gain access ? You shouldn't need to write any custom smart contact. Can be handled entirely by the bot. Basically, have the user connect their wallet, get the balance from the cw20 token contract. If the user's amount is larger than the minimum set in bot config, then they get access to the token gated channel. you should have also nominated yourself @Highlander-maker. you have spent a great amount of time on the project and started it also. @JakeHartnell ok Jake. Let me look in to it. :) @JakeHartnell pretty sure omniflix have this working already for telegram: juno.omniflix.com @JakeHartnell pretty sure omniflix have this working already for telegram: juno.omniflix.com If they open source it, will nominate them. They deserve some $JUNO for that. Does theirs work with cw20s? This sounds like https://collab.land/ @giansalex has been one of the most outstanding contributors during Hack Juno. :clap: :clap: :clap: This ticket is a duplicate of his main nomination #13. So closing, but will link to this issue from there.
gharchive/issue
2021-08-02T12:07:31
2025-04-01T04:54:51.779367
{ "authors": [ "Highlander-maker", "JakeHartnell", "blockcreators", "nullmames" ], "repo": "CosmosContracts/hack-juno", "url": "https://github.com/CosmosContracts/hack-juno/issues/15", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
53181656
Problem with Fed Node installer from develop branch 2014-12-31 12:18:20,353|DEBUG: RUNNING COMMAND: cd /home/xcp/counterpartyd_build && git pull origin develop From https://github.com/CounterpartyXCP/counterpartyd_build * branch develop -> FETCH_HEAD *** Please tell me who you are. Run git config --global user.email "you@example.com" git config --global user.name "Your Name" to set your account's default identity. Omit --global to set the identity only in this repository. fatal: unable to auto-detect email address (got 'root@ubuntu.(none)') 2014-12-31 12:18:22,020|ERROR: Command failed: 'cd /home/xcp/counterpartyd_build && git pull origin develop' I thnk so...it's a public repo so it shouldn't ask you for that, unless it thinks you are somehow committing...I'll close this and if it's an issue again, reopen it Ah, yes, now I know why - prolly because of https://github.com/CounterpartyXCP/counterwallet/issues/627
gharchive/issue
2014-12-31T12:24:36
2025-04-01T04:54:51.854748
{ "authors": [ "rippler", "robby-dermody" ], "repo": "CounterpartyXCP/counterpartyd_build", "url": "https://github.com/CounterpartyXCP/counterpartyd_build/issues/172", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
126182427
Angular 2.0 I believe this plugin must be made in angular 2.0 too. I am planing to recreate it or add code to this project branch, when angular 2 get the release. Is there someone who are interesting in this? +1 +1 I will start this soon, when i do that i will show the link here. +1. Have you started work on this? Yes but process it stopped. I will continue with it when angular 2 is released. We have tried to make small app and we saw that ng2 is not ready. Any update on this ? I think no. @CrackerakiUA did you do any research on Angular 2.0? @rickerd same as before, my team tried couple projects on ng2. They are changing things, we didn't like it. We decide to have a look again after 6 months. This repo is dead. Please move to the other project: https://github.com/CrackerakiUA/ui-cropper
gharchive/issue
2016-01-12T13:44:40
2025-04-01T04:54:51.885325
{ "authors": [ "CrackerakiUA", "Mobius77x", "camposcristianeze", "cleever", "mpmlj", "rickerd" ], "repo": "CrackerakiUA/ngImgCropFullExtended", "url": "https://github.com/CrackerakiUA/ngImgCropFullExtended/issues/48", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1673335968
"get writable" errors with "Logon failure" in some cases Using the command get writable sometimes fails with "Logon failure". I do not know exactly what triggers this but I get this error when authenticating using regular low priviledge accounts but not when authenticating using DAs so this may be related to privileges or lack there of. In the screenshot you can see that the credentials for domainuser1 are correct but that the command get writable still fails when authenticating using the same account. FYI. The account domainuser1 does not have write rights of any sort to any object: Just to be certain, using the command get writable will list objects the authenticating account has GenericAll or GenericWrite rights on, correct? Perhaps WriteDACL too? @jsdhasfedssad The username is different between your two commands @embargod Please open your own issue, it's a not related to the current issue Ooops. You are correct. My bad! What about the privileges? Will get writable list objects the authenticating account has GenericAll or GenericWrite rights on?Perhaps WriteDACL too? It will output you what you can write on the object or child if you can create a child to the object. If you want to know exactly what you can write on the object, use --detail. E.g. if it outputs a WRITE permission on nTSecurityDescriptor it means you have WRITEDACL . If you want a Security Descriptor way of permissions you can alway use get object --resolve-sd This is an interesting feature. Would you say it can partially replace Bloodhound/Sharphound? At least when it comes to identify and abuse edges that grants you write access? Is there a way of using get writable in combination with --detail on a single object? Not only filtered by type. It seems some attributes cannot be written to regardless of access rights. For example the attribute badPwdCount. BloodyAD is not showing that I have write access to that but I wanted to try it anyway using a DA account. Are some attributes "protected"? I get the below error when trying the --resolve-sd command. The account I use is a DA: It can help to identify interesting rights yes. Especially the write child which is not part of bloodhound edges if I'm not mistaken. But of course it doesn't replace the powerful graph db used by bloodhound. You can't do it for a single object or use get object with attributes allowedAttributesEffective or allowedChildClassesEffective. You could also use resolve-sd. For your error can you add a print(self.control_enum) before l.254 in utils.py and show me the output? @jsdhasfedssad can you confirm your issue is fixed? otype filtering on user should be fixed too
gharchive/issue
2023-04-18T15:09:02
2025-04-01T04:54:51.892145
{ "authors": [ "CravateRouge", "jsdhasfedssad" ], "repo": "CravateRouge/bloodyAD", "url": "https://github.com/CravateRouge/bloodyAD/issues/37", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1227134393
CASMCMS-7876: Update BOS cli for v2 Summary and Scope This adds the BOS v2 api to the cli. It also adds a new way to inject code into the generator so that it's possible to access data at a point that is not otherwise possible. This is necessary in order to setup a new bulk update endpoint for the cli. Issues and Related PRs Resolves CASMCMS-7876 Testing Tested on: Hela and others Test description: We have been using this to call all the commands while testing BOS v2 Risks and Mitigations None Pull Request Checklist [X] Version number(s) incremented, if applicable [X] Copyrights updated [X] License file intact [X] Target branch correct [X] CHANGELOG.md updated [X] Testing is appropriate and complete, if applicable [X] HPC Product Announcement prepared, if applicable This needs to have the lint failures in the build fixed.
gharchive/pull-request
2022-05-05T20:16:37
2025-04-01T04:54:51.896968
{ "authors": [ "erl-hpe", "rbak-hpe" ], "repo": "Cray-HPE/craycli", "url": "https://github.com/Cray-HPE/craycli/pull/50", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1418665161
CASMTRIAGE-4425: Incorrect node command - Update the desired configuration for all NCNs. Description Mistyped option in command needed a capitlization. Checklist Before Merging [x] If I added any command snippets, the steps they belong to follow the prompt conventions (see example). [x] If I added a new directory, I also updated .github/CODEOWNERS with the corresponding team in Cray-HPE. [x] My commits or Pull-Request Title contain my JIRA information, or I do not have a JIRA. /backport release/1.3
gharchive/pull-request
2022-10-21T17:44:44
2025-04-01T04:54:51.899828
{ "authors": [ "ndavidson-hpe" ], "repo": "Cray-HPE/docs-csm", "url": "https://github.com/Cray-HPE/docs-csm/pull/2702", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1597649753
CASMTRIAGE-4986 1.4 : Modify the VCS per flight service check Description Resolves CASMTRIAGE-4986 for CSM 1.4 The previous VCS pre flight service check does not work if other services such as COS are not also installed. The new VCS pre flight service check will instead check that the service is able to return the api version Checklist [ ] If I added any command snippets, the steps they belong to follow the prompt conventions (see example). [ ] If I added a new directory, I also updated .github/CODEOWNERS with the corresponding team in Cray-HPE. [x] My commits or Pull-Request Title contain my JIRA information, or I do not have a JIRA. /backport main
gharchive/pull-request
2023-02-23T23:09:05
2025-04-01T04:54:51.902874
{ "authors": [ "kimjensen-hpe" ], "repo": "Cray-HPE/docs-csm", "url": "https://github.com/Cray-HPE/docs-csm/pull/3341", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2114443804
🛑 sneak.page is down In f6771f9, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in 383610c after 11 minutes.
gharchive/issue
2024-02-02T09:05:09
2025-04-01T04:54:51.905233
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/2118", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2145794972
🛑 sneak.page is down In f8aa170, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in 3ef6d5a after 8 minutes.
gharchive/issue
2024-02-21T04:21:25
2025-04-01T04:54:51.907961
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/2438", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2406826551
🛑 sneak.page is down In 6363376, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in 905d45a after 20 minutes.
gharchive/issue
2024-07-13T09:54:34
2025-04-01T04:54:51.910222
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/4881", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2616458170
🛑 sneak.page is down In 52bff8d, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in 2812fe7 after 16 minutes.
gharchive/issue
2024-10-27T09:42:45
2025-04-01T04:54:51.912459
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/6297", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1986439694
🛑 sneak.page is down In 9a241cb, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in c99dc2f after 8 minutes.
gharchive/issue
2023-11-09T21:36:14
2025-04-01T04:54:51.914702
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/694", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2011772882
🛑 sneak.page is down In d2f8a03, sneak.page (https://sneak.page) was down: HTTP code: 0 Response time: 0 ms Resolved: sneak.page is back up in bed5bfc after 8 minutes.
gharchive/issue
2023-11-27T08:39:22
2025-04-01T04:54:51.917162
{ "authors": [ "CrazyMarvin" ], "repo": "Crazy-Marvin/upptime", "url": "https://github.com/Crazy-Marvin/upptime/issues/997", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
940923572
To-Do Extension Pull Request Checklist Go through the check boxes given below and make sure you mark the relevant ones. To check put x inside [ ] eg. [x] Extension Is your project an [x] Extension [ ] Application Given below are the mandatory requirements we expect from your project. Your Pull Request will be reviewed only if all the minimum requirements relevant to your project are satisfied. [x] Added readme.md to my project folder with relevant information. [x] I have provided the demo video link (showing the functionality in action) or the public deployment link of my extension of an application within the readme.md (if Applicable). [x] I have explained the process in a readable manner within readme.md. Terms and Conditions Plagiarism check from Team Crio will be solid and if your code is found to be plagiarized, the team has every right to reject your Pull Request and take necessary actions. [x] I can assure you that this is my own contribution and I did not involve in any kind of plagiarism activities. Buckle up and wait till you receive communication from the review team. Keep a tab on this pull request page for comments on your project. Any necessary suggestions will also be provided here. All the best :) Once your PR gets merged, take some time ⏲️, celebrate 🥳 and share it with the community on Slack. To-Do Extension Hey @BhavikSojitra , I can see that there are many to-do list chrome extensions. We are sorry to say that we are not moving forward with this PR as the deadline is over I have already submitted my project two days from the deadline so there is nothing no issue with the deadline.
gharchive/pull-request
2021-07-09T16:51:11
2025-04-01T04:54:51.998695
{ "authors": [ "BhavikSojitra", "jainaayush01" ], "repo": "Crio-WFH/Chrome-extensions", "url": "https://github.com/Crio-WFH/Chrome-extensions/pull/32", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1335513707
🛑 cris book is down In a026bd9, cris book (http://cristo.top/) was down: HTTP code: 0 Response time: 0 ms Resolved: cris book is back up in 49f7dd7.
gharchive/issue
2022-08-11T06:11:40
2025-04-01T04:54:52.001206
{ "authors": [ "CristoMonte" ], "repo": "CristoMonte/uptime-demo", "url": "https://github.com/CristoMonte/uptime-demo/issues/204", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2026307753
Feature/stereotyped mappings Add stereotype on datasource to enable using Mappings for different purposes besides MDDE modeling. For example when there is horizontal and vertical lineage defined in the model @willem-xbreeze I had a look at the changes, which seem ok this way. I only thought you probably wanted to force the stereotype. Because things like model checks, form, etc. in the BaseClassifierMapping and BaseStructuralFeatureMapping metaclass (https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/tree/main/decomposed/extensions/MDDE_LDM/Profile/BaseClassifierMapping & https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/tree/feature/stereotyped_mappings/decomposed/extensions/MDDE_LDM/Profile/BaseStructuralFeatureMapping) should all also be moved into a Criterea on the DataSource stereotype. Right? Otherwise all behavior modelling on the BaseClassifierMapping and BaseStructuralFeatureMapping will be applied to all mappings. Based on your story on being able to use mappings for different purposed won't work I think. @willem-xbreeze I had a look at the changes, which seem ok this way. I only thought you probably wanted to force the stereotype. Because things like model checks, form, etc. in the BaseClassifierMapping and BaseStructuralFeatureMapping metaclass (https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/tree/main/decomposed/extensions/MDDE_LDM/Profile/BaseClassifierMapping & https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/tree/feature/stereotyped_mappings/decomposed/extensions/MDDE_LDM/Profile/BaseStructuralFeatureMapping) should all also be moved into a Criterea on the DataSource stereotype. Right? Otherwise all behavior modelling on the BaseClassifierMapping and BaseStructuralFeatureMapping will be applied to all mappings. Based on your story on being able to use mappings for different purposed won't work I think. Latest checkin made the following changes: Moved all MDDE content in the extension under BaseClassifierMapping and BaseStructuralFeatureMapping to a new criterion. Note: This gives a lot of changes in Git which are only relocations of existing items in the extension Changed the initialize event handler under BaseStructuralFeatureMapping (since event handlers cannot be moved under criterion) so it only runs on appropriate attribute mappings Made the export of mappings conditional, the MDDE XML export only includes MDDE mappings Added an example model with business concepts and a local extension in the ExampleDWH model to illustrate how this can work with an additional form for "business concept mappings" The changes look fine, what I am missing is a change in the add new mapping method, where you would set the stereotype of the new data source to mdde_DataSource. Or how is this taken care of for new users of the extension that the stereotype is set on creation of the first mapping? The criterion allows datasource stereotype to be empty. But i have added a check now in the new mapping method to make sure that the datasource has a valid stereotype, either empty or mdde_datasource. See commit https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/pull/14/commits/5a269365f8d102aa46f54b66b6be73c27b70a535
gharchive/pull-request
2023-12-05T14:15:48
2025-04-01T04:54:52.015899
{ "authors": [ "harmen-xb", "willem-xbreeze" ], "repo": "CrossBreezeNL/PowerDesigner-MDDE-Extension", "url": "https://github.com/CrossBreezeNL/PowerDesigner-MDDE-Extension/pull/14", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
627771670
exception on release im using xamarin forms, its works fine on debug mode but when i generate the apk file and install on device, it fails with the exception: The google sign in could not completed it's process correctly have you added the sha-1 of the release key ? have you added the sha-1 of the release key ? i mean my release apk works fine but when i try google login shows that exception, the app works fine only the google login fails What exception are you getting? What exception are you getting? after it shows account for pick up, the exception is: The google sign in could not completed it's process correctly in debug mode have no problem It seems you have an issue with release configuration did you add the sha1 of the keystore you are using for release in firebase console portal? On Sat, May 30, 2020, 8:22 PM Alexis Vidal notifications@github.com wrote: What exception are you getting? after it shows account for pick up, the exception is: The google sign in could not completed it's process correctly in debug mode have no problem — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/CrossGeeks/GoogleClientPlugin/issues/57#issuecomment-636402174, or unsubscribe https://github.com/notifications/unsubscribe-auth/AATOAJ6NU67XTKXVY6I376LRUGPNRANCNFSM4NOYN4ZA . excuse can you help with these configuration Are you using a local keystore to sign the apk on release or using PlayStore auto-signing? On Sat, May 30, 2020, 8:32 PM Alexis Vidal notifications@github.com wrote: excuse can you help with these configuration — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/CrossGeeks/GoogleClientPlugin/issues/57#issuecomment-636403056, or unsubscribe https://github.com/notifications/unsubscribe-auth/AATOAJZNS7ZMOOLO4D2CF4LRUGQRBANCNFSM4NOYN4ZA . im using a local keystore(ah-doc) i already create on firebase console an android application with the sha1 key To get the SHA1 for the release keystore: keytool -exportcert -list -v -alias -keystore On Sat, May 30, 2020, 8:42 PM Alexis Vidal notifications@github.com wrote: im using a local keystore(ah-doc) i already create on firebase console an android application with the sha1 key — You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/CrossGeeks/GoogleClientPlugin/issues/57#issuecomment-636403845, or unsubscribe https://github.com/notifications/unsubscribe-auth/AATOAJ4UGBIMCIZT462YRJDRUGRWXANCNFSM4NOYN4ZA .
gharchive/issue
2020-05-30T16:45:48
2025-04-01T04:54:52.026515
{ "authors": [ "alexisVidal29", "husseinshaib1", "rdelrosario" ], "repo": "CrossGeeks/GoogleClientPlugin", "url": "https://github.com/CrossGeeks/GoogleClientPlugin/issues/57", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
316669852
Fix Raxx.body spec Raxx.body was String.t(). Changes it to binary. String.t() implies it's a string. It may be a string, but could just as easily be a raw binary that's not a string. Well you learn something new everyday. This looks good, but as I know people who run dialyzer over raxx projects I still think this needs a mention in the changelog. Perhaps I should also add something to an ISSUE_TEMPLATE saying about a changlog. Also if chunked responses/requests can have binary data then should the spec of Raxx.Data should be changed? yeah I think this also needs changing. https://github.com/CrowdHailer/raxx/blob/master/lib/raxx.ex#L233 Although I am curious to see if dialyzer is smart enough to pick it up
gharchive/pull-request
2018-04-23T05:51:46
2025-04-01T04:54:52.032196
{ "authors": [ "CrowdHailer", "varnerac" ], "repo": "CrowdHailer/raxx", "url": "https://github.com/CrowdHailer/raxx/pull/109", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2565211424
falcon-linux-sensor pre-start job failed because it can't find systemctl to restart the service When using Crowdstrike v1.3 and other service broker tiles on Cloud Foundry, we tried to remove the old aws broker v1.4.20 tile but delete-errand failed due to pre-start job falcon-linux-sensor. failed. We re-ran the errand manually with keep-alive then ssh on the VM and found below error message in the /var/vcap/sys/log/falcon-linux-sensor/falcon-installer.log - 2024/10/03 05:41:09 Error restarting Falcon sensor service: Could not find systemctl: /usr/bin/systemctl It looks like the falcon sensor installation worked but just failed because it can't find systemctl to restart the service, however it's still installed, so when we re-ran the errand manually with the same vm this time it passed and succeeded (because falcon sensor was already installed...). Could you please fix this issue with Crowdstrike in case it might impact other errands? Thanks. Hello, Can you provide the OS and OS version that the errand was run on? Also, can you provide the PATH to systemctl on the VM? Hi @redhatrises , The stemcell is light-bosh-aws-xen-hvm-ubuntu-xenial-go_agent-621.969.tgz, OS: ubuntu-xenial, Version: 621.969. PATH to systemctl on the VM is /usr/bin/systemctl. Thanks for providing that info. This is fixed in the installer and will be fixed in the next release. Closing.
gharchive/issue
2024-10-04T00:44:58
2025-04-01T04:54:52.035746
{ "authors": [ "fiona8953", "redhatrises" ], "repo": "CrowdStrike/falcon-boshrelease", "url": "https://github.com/CrowdStrike/falcon-boshrelease/issues/16", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
799807285
Error reading integer. Unexpected token: StartObject. Path 'data.accounts_active', line 1, position 1601. Just started failing within the last few hours. Looks like something may have changed with the Reddit API? Or maybe it's just an intermittent error? I'm using the latest pre-release (v2.0.0-CI00075) Code: var subreddit = await new Reddit().GetSubredditAsync("/r/videos"); Exception: JsonReaderException: Error reading integer. Unexpected token: StartObject. Path 'data.accounts_active', line 1, position 1601. at Newtonsoft.Json.JsonReader.ReadAsInt32() at Newtonsoft.Json.JsonReader.ReadForType(JsonContract contract, Boolean hasConverter) at Newtonsoft.Json.Serialization.JsonSerializerInternalReader.PopulateObject(Object newObject, JsonReader reader, JsonObjectContract contract, JsonProperty member, String id) at Newtonsoft.Json.Serialization.JsonSerializerInternalReader.Populate(JsonReader reader, Object target) at Newtonsoft.Json.JsonSerializer.PopulateInternal(JsonReader reader, Object target) at RedditSharp.Helpers.PopulateObject(JToken json, Object obj) at RedditSharp.Things.Thing.Populate(JToken json) at RedditSharp.Things.Thing.Parse(IWebAgent agent, JToken json) at RedditSharp.Things.Thing.Parse[T](IWebAgent agent, JToken json) at RedditSharp.Helpers.d__01.MoveNext() at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Runtime.CompilerServices.TaskAwaiter1.GetResult() at WatchNow.SourceManagement.d__13.MoveNext() in C:\Users\derek\source\repos\WatchNow\WatchNow\SourceManagement.cs:line 161 Is this still ongoing? This previously occurred when Reddit broke some shit and started returning a null or missing the json for it entirely. Seems to be working now, I guess? I guess it was some sort of intermittent error. Odd that it was accounts_active in the error, but probably is the first key parsed. accounts_active is the actual JSON key they screw up every once in a while. I've complained about it before but apparently it's regressed yet again.
gharchive/issue
2021-02-02T23:16:44
2025-04-01T04:54:52.046367
{ "authors": [ "CrustyJew", "derekantrican" ], "repo": "CrustyJew/RedditSharp", "url": "https://github.com/CrustyJew/RedditSharp/issues/237", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2182975308
Devise unlimited user's development API key and one user for production key. When attaching a development key in the SDK there must be some sort of determination when the key get sent with the OS information if it is a development key or a production key. User should be able to connect multiple SDK's with the development key and only one instance of the production key unless they are paying for a higher subscription (this should be coded in another ticket). User's should be able to see all operating system information of the developers currently using the SDK in development. (To be coded in another ticket). This was written and pushed already. When storing the information in cache upon sending it to the API. if (existingCacheInformation != null) { JsonSerializerOptions options = new JsonSerializerOptions() { PropertyNameCaseInsensitive = false, }; OSInfoRedisEntry cacheInformation = JsonSerializer.Deserialize<OSInfoRedisEntry>(existingCacheInformation, options); // TODO: perform other checks besides IP address and Operating System // also perform check based upon the API key. if (!cacheInformation.IsApiKeyProd) { result = new OkObjectResult(new { message = "Enjoy using your development key" }); }
gharchive/issue
2024-03-13T02:05:19
2025-04-01T04:54:52.058724
{ "authors": [ "WingZer0o" ], "repo": "Cryptographic-API-Services/cas-api", "url": "https://github.com/Cryptographic-API-Services/cas-api/issues/40", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1306946252
Add "Pause" update stage to Generic This is not for requesting support for new games/applications To do this you should go to https://github.com/CubeCoders/AMPTemplates and first attempt to build a configuration yourself - otherwise you can request a template from this repo. Feature Request Feature Information: It would be useful to have a configurable "Pause" update stage. This would pause the update stages for a specified number of seconds. One use case is the new Stationeers server build. The updated draft template includes start/wait/stop stages to generate the server files. However the console output that appears then is different to subsequent starts. If a console line that is present on all starts is selected as the app ready regex, then on first start the stop stage kicks in too quickly, because the server does some additional work after that point on first start. This causes the stop stage to not work correctly. This would be avoided if a pause could be introduced after the wait update stage, that operates once the regex is matched and the server is otherwise considered to be started. This might also help in some cases where a start stage seems to hang if run immediately after a set executable stage. I confirm: [x] that I have searched for an existing feature request matching the description. This could also help the case (eg Vintage Story) where an executable fails to be started if that update stage occurs immediately after the server files are unzipped/extracted.
gharchive/issue
2022-07-17T00:20:03
2025-04-01T04:54:52.072033
{ "authors": [ "Greelan" ], "repo": "CubeCoders/AMP", "url": "https://github.com/CubeCoders/AMP/issues/684", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1669482751
SteamCMD Steam Guard Issues Bug Report System Information Operating System - Ubuntu 22.04 LTS AMP version and build date - v2.4.3.2, built 31/03/2023 20:24 Which AMP release stream you're using - Bleeding I confirm: [x] that I have searched for an existing bug report for this issue. [x] that I am using the latest available version of AMP. [x] that my operating system is up-to-date. Symptoms What are you trying to do? Update a Steam game for the first time. What are you expecting to happen? AMP should appropriately handle the Steam Guard code. What is actually happening? ('Nothing' is not an acceptable answer!) The Steam Guard prompt happens twice and hangs if "Cancel" is pressed. Reproduction If you setup a game like Assetto Corsa Compatizione which requires login for the first time, AMP prompts for the Steam Guard code as expected. If you press "Cancel" the process just hangs. If you type the code correctly AMP accepts it but then prompts again with the code pre-filled. It seems AMP is using the regex of Steam Guard which is in the request from SteamCMD and the response as seen below: This computer has not been authenticated for your account using Steam Guard. Please check your email for the message from Steam, and enter the Steam Guard code from that message. Steam Guard code:OK Heh... I fixed this one a while ago.
gharchive/issue
2023-04-15T16:55:32
2025-04-01T04:54:52.076965
{ "authors": [ "IceOfWraith" ], "repo": "CubeCoders/AMP", "url": "https://github.com/CubeCoders/AMP/issues/870", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
491259988
Select All doesnt appear onInit Hi and thanks for this wonderful module. When the module is loading (ngOnInit) for the first time, the Select All is not visible. Only if i select All the Items in the list the Select All and Unselect All appears. When i go to see the demo its always there. The other problem, as u can see on the picture, is an Input select number under the checkbox of Unselect all. I didn't put the [(ngModel)]="selectedItems" because angular said that it is depreciated on version 7 and 8 My code on .html : <form [formGroup]="selectsForm" class="d-md-flex flex-md-row"> <div class="form-group mr-3"> <label class="">Sélectionnez un ou plusieurs secteurs</label> <angular2-multiselect class="form-control" [data]="dropdownList" [settings]="dropdownSettings" (onSelect)="onItemSelect($event)" (onSelectAll)="onSelectAll($event)" formControlName="secteurAct" > </angular2-multiselect> </div> </form> My code on .ts import { Component, OnInit } from '@angular/core'; import { Subscription } from 'rxjs/internal/Subscription'; import { FormBuilder, FormGroup, Validators } from '@angular/forms'; //IMPORT DES SERVICES POUR LES ANNONCES import { AnnoncesService } from "./../../services/annonces.service"; import { DepartementService } from "./../../services/departement.service"; //IMPORT DU MODELE POUR LES SECTEUR ACTIVITES import { SecteurActivite } from "./../../models/secteurActivite.model"; @Component({ selector: 'app-home', templateUrl: './home.component.html', styleUrls: ['./home.component.scss'] }) export class HomeComponent implements OnInit { //Propriétés pour la création et la subscription aux données FireStore listSecteurAct: SecteurActivite[]; annoncesSubscription: Subscription; isFetchingData:boolean = false; //Propriétés des Forms dropdownList = []; selectedItems = []; dropdownSettings = {}; selectsForm: FormGroup; //Constructeur constructor( private annnonceService: AnnoncesService, private formBuilder: FormBuilder ) { } //---------- NGONINIT -------------------------// ngOnInit() { this.initForm(); this.isFetchingData = true; //----------- CHARGEMENT DES DONNÉES A LA COMPILATION DU COMPOSANT -------------------------------// this.annoncesSubscription = this.annnonceService.getSecteurActivite().subscribe(actionArray => { this.isFetchingData = false this.listSecteurAct = actionArray.map(item => { return { id : item.payload.doc.id, ...item.payload.doc.data() } as SecteurActivite; }) //On affecte au champt select les valeurs récupérées dans la BDD //La méthode Array.prototype.sort() permet de trier par ordre croissant ou décroissant un tableau ou un //tableau d'objets this.dropdownList = this.listSecteurAct.sort(function(a, b) { var nameA = a.title.toUpperCase(); // ignore upper and lowercase var nameB = b.title.toUpperCase(); // ignore upper and lowercase if (nameA < nameB) { return -1; } if (nameA > nameB) { return 1; } // titles must be equal return 0; }); }); //----------- END CHARGEMENT DES DONNÉES -------------------------------// //---------- CHARGEMENT DES CHAMPS SELECTS -----------------------------// //Data du champ select secteur activite this.dropdownList = []; this.selectedItems = []; //Options du module ng-multiselect-dropdown voir doc su NPM ou fichier ReadMe this.dropdownSettings = { singleSelection: false, primaryKey: 'id', labelKey: 'title', selectAllText: 'Toute la France', unSelectAllText: 'Tout supprimer', text:"Secteur d\'activité", enableSearchFilter:false }; //---------- END CHARGEMENT DES CHAMPS SELECTS -------------------------// }//----------END NGONINIT -------------------------// onItemSelect(item: any) { console.log(item); } onSelectAll(items: any) { console.log(items); } //Initialisation du formulaire (ReactiveFormsModule méthode) initForm() { this.selectsForm = this.formBuilder.group({ secteurAct: ['', Validators.required] }); } }//END CLASS HOME COMPONENT I had the wrong version sorry. I install the 4.6.3
gharchive/issue
2019-09-09T18:40:27
2025-04-01T04:54:52.081921
{ "authors": [ "jeremie74" ], "repo": "CuppaLabs/angular2-multiselect-dropdown", "url": "https://github.com/CuppaLabs/angular2-multiselect-dropdown/issues/421", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2740150483
🛑 lemmy.best is down In 98c4df7, lemmy.best (https://lemmy.best) was down: HTTP code: 500 Response time: 11804 ms Resolved: lemmy.best is back up in 4c3c129 after 6 minutes.
gharchive/issue
2024-12-14T21:25:47
2025-04-01T04:54:52.084853
{ "authors": [ "Curly-Mo" ], "repo": "Curly-Mo/lemmy.best-upptime", "url": "https://github.com/Curly-Mo/lemmy.best-upptime/issues/925", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
387976939
Some grammar, spelling fixes It looks like I set up my last PR too early! The name of the nightwatch example had become out-of-date in the README. Thanks again :+1:
gharchive/pull-request
2018-12-05T22:38:22
2025-04-01T04:54:52.085833
{ "authors": [ "JDemler", "barbeque" ], "repo": "CurrySoftware/elm-datepicker", "url": "https://github.com/CurrySoftware/elm-datepicker/pull/18", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
903149495
MCD SE not unable to launch since the Hidden Depths update Yeah, they changed the encryption keys again. Next release (version 1.3.5) will work. I will make sure it gets published next week sometime. how close is it to being done There is a way around this, if you only need to edit item power levels and/or character level, or item rarity, then you can just switch install location. Simply uninstall, and make sure you (1) back up your character folder and (2) make sure the keep save file is turned on in the uninstall screen, then reinstall in a non-default location (I used my primary games folder) and then, when the editor fails to find MCD, simply press no game files (forget what the actual button is called) and find your character folder by going to File > Open > your save game. You cannot create items, change enchant slots, or anything like that, but it will work for most things. There is a way around this, if you only need to edit item power levels and/or character level, or item rarity, then you can just switch install location. Simply uninstall, and make sure you (1) back up your character folder and (2) make sure the keep save file is turned on in the uninstall screen, then reinstall in a non-default location (I used my primary games folder) and then, when the editor fails to find MCD, simply press no game files (forget what the actual button is called) and find your character folder by going to File > Open > your save game. You cannot create items, change enchant slots, or anything like that, but it will work for most things. ^ This is true ^ Yeah, they changed the encryption keys again. Next release (version 1.3.5) will work. I will make sure it gets published next week sometime. can't wait!!! Going to place some new armor properties onto my all-in-one gilded opluent armor Yeah, they changed the encryption keys again. Next release (version 1.3.5) will work. I will make sure it gets published next week sometime. But when... Yeah, they changed the encryption keys again. Next release (version 1.3.5) will work. I will make sure it gets published next week sometime. But when... You actually can make your Opulent Armor gilded right now by doing the method I mentioned above, as long as you have the right enchantments. If not, then you can look for an armor piece that has Luck of the Sea on it, max out that enchant, and restock merchants until the restock cost is more than 10k emeralds. This works super well, and I've gotten about ~50 uniques just from this. Seeing as you can set your player level (to get more enchant points) and set your player emerald and gold count (for restocking and buying items), with this limited version, I'd say its worth it. (side note, before this, I had never gotten a verdant robe, but with luck of the sea and high emerald count, I now have 3. So yea.) But when will it come out? From what I understand, finding the key can be time consuming, but I don't really know cause I'm not coding this. On a side note, this program still works for the most part with this bug using the method that I mentioned, but as Cutflame said, it should be coming out this week. this is the AES key, @CutFlame : 0x0A01138AC7553ECEA7E9E257B5A45FA6665679D360B270278AE5D5367B172E40 There is a way around this, if you only need to edit item power levels and/or character level, or item rarity, then you can just switch install location. Simply uninstall, and make sure you (1) back up your character folder and (2) make sure the keep save file is turned on in the uninstall screen, then reinstall in a non-default location (I used my primary games folder) and then, when the editor fails to find MCD, simply press no game files (forget what the actual button is called) and find your character folder by going to File > Open > your save game. You cannot create items, change enchant slots, or anything like that, but it will work for most things. I've tried this and It just closes when I click "OK" on the error box, tried running as admin and all the ins and outs, can't get to this "no game files" section as what I said earlier, it just closes itself after closing the error. It did the same thing for me, however my install of MCD has been messed up for a long time because I transferred my save data (all of it) from another computer to my new one back when MCD had its own launcher. Did the error change from looking like the picture at the top of this thread to something like this? Yeah it looks like that, when I click "OK" it just closes the entire thing, i'm just wanting to set my gear up to the current max power level but none of my old editors are working either. When the program will be updated When the program will be updated I have no Idea. He said this week New version is released (1.3.6)
gharchive/issue
2021-05-27T02:54:48
2025-04-01T04:54:52.100193
{ "authors": [ "CutFlame", "Jaypamieukomnom", "Maksman2000", "PhoenixMC-0", "Viinyl", "lordgecko10", "megaenderdragon" ], "repo": "CutFlame/MCDSaveEdit", "url": "https://github.com/CutFlame/MCDSaveEdit/issues/58", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1968915068
Communication protocol I've been thinking about how to make sure that C# objects can be easily understood by the Arduino, and I figured the easiest way would be to not use objects at all. Thus, I propose the following communication standard, which would ensure the datastream sent over BLE can be easily received, stored, and processed: comms protocol v1.txt I looked a bit at the way C# handles serial communication, and I think it's gonna be easier if we instead do it like this: Each event is a 32 bit number, first bit representing the target setting (or override signal for the sync time signal), the other bits representing the 31 bit integer - just add a leading zero to get a standard 32 bit one, we won't ever need more than 17 bits anyway. So instead of a proper protocol we just create our own standard?
gharchive/issue
2023-10-30T17:48:33
2025-04-01T04:54:52.110975
{ "authors": [ "CyanideData2", "Wojtas1PL" ], "repo": "CyanideData2/Sun-O-Matic", "url": "https://github.com/CyanideData2/Sun-O-Matic/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
423670559
Split building and loading the model This can be useful if you want to build the BERT model from given config, but don't want to load pretrained weights. Pull Request Test Coverage Report for Build 97 5 of 5 (100.0%) changed or added relevant lines in 1 file are covered. No unchanged relevant lines lost coverage. Overall coverage remained the same at 100.0% Totals Change from base Build 96: 0.0% Covered Lines: 340 Relevant Lines: 340 💛 - Coveralls Thanks for the PR~ I've published a new version. Great, that was fast, thanks!
gharchive/pull-request
2019-03-21T10:52:32
2025-04-01T04:54:52.122054
{ "authors": [ "CyberZHG", "coveralls", "nsmetanin" ], "repo": "CyberZHG/keras-bert", "url": "https://github.com/CyberZHG/keras-bert/pull/37", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1518875526
为何选择了iptables iptables已经是过去时了,为何不选择它的继任者nftables? CFW 设计的初衷是兼容所有 Linux 平台,无论您使用的是发行版,或是部分嵌入式系统,您总能在支持 iptables 的平台中原生运行 CFW。
gharchive/issue
2023-01-04T12:16:04
2025-04-01T04:54:52.122963
{ "authors": [ "Cyberbolt", "saintak" ], "repo": "Cyberbolt/cfw", "url": "https://github.com/Cyberbolt/cfw/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1564581293
Feature request: Committed use discounts for RHEL images Compute Engine committed use discounts are now also available for Red Hat Enterprise Linux (RHEL) image licenses. Add: [x] Committed use discounts for RHEL images [x] Committed use discounts for RHEL for SAP images Source: https://cloud.google.com/compute/disks-image-pricing#premiumimages Commit: https://github.com/Cyclenerd/google-cloud-pricing-cost-calculator/commit/a245e9612ab64a6b3913c6cb625731d1ede9cfba New release: https://github.com/Cyclenerd/google-cloud-pricing-cost-calculator/releases/tag/v1.1.0
gharchive/issue
2023-01-31T15:54:03
2025-04-01T04:54:52.126652
{ "authors": [ "Cyclenerd" ], "repo": "Cyclenerd/google-cloud-pricing-cost-calculator", "url": "https://github.com/Cyclenerd/google-cloud-pricing-cost-calculator/issues/53", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2343018014
Error handling when returning undefined Issue type: :heavy_plus_sign: Feature request Short description: Functions that return undefined (e.g. function f() { return undefined; }, or function f() {} ) give a rather verbose error as follows when apply0ed; It would be desireable to either have: a dedicated error for a function returning undefined, rather than a translation error occuring (e.g. 'error: function returned undefined instead of a value'); or for an explicit 'undefined' singleton to be the output instead of erroring, allowing for logic such as equals(f(), undefined) on the InDy side. Thanks for the suggestion!
gharchive/issue
2024-06-10T06:54:15
2025-04-01T04:54:52.135136
{ "authors": [ "met4000", "rubensworks" ], "repo": "CyclopsMC/IntegratedScripting", "url": "https://github.com/CyclopsMC/IntegratedScripting/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
217062008
Trying to serve html to browser on tacopie server I am trying to serve up some html content to a browser over a tacopie server. This was my message handler - where I tried both sync and async writes and trying to close the connection when done. Neither works, maybe because I dont understand how the write/disconnect works. Any ideas? thank you! void on_new_message(const std::shared_ptr<tacopie::tcp_client>& client, const tacopie::tcp_client::read_result& res) { if (res.success) { std::string resp = "this is the response"; std::string fullResp = "HTTP/ 1.1 200 OK\r\nContent-Type: type/html\r\nConnection: Closed\r\n Content-Length: " + std::to_string(resp.length()) + std::string("\r\n\r\n") + resp; std::vector<char> buf(fullResp.begin(), fullResp.end()); client->get_socket().send(buf, fullResp.length()); //client->async_write({ buf, nullptr }); client->disconnect(); } else { info("WebClient disconnected"); client->disconnect(); } } Hi, Sorry for the time to reply, I finally could have a look. I tried your code and it first worked well over curl and netcat, but apparently not with a browser (I tried chrome and safari). So my guess was that there is an issue with the content of your request. I double checked with the expected HTTP format, and there is indeed an issue in your response. HTTP/ 1.1 There is an extra space after the / \r\n Content-Length: : There is an extra space Before the Content When I changed your output into: HTTP/1.1 200 OK\r\nContent-Type: type/html\r\nConnection: Closed\r\nContent-Length: it worked perfectly fine :) I close this issue, but feel free to re-open it if necessary :) Best BTW, my final code is the following one: void on_new_message(const std::shared_ptr<tacopie::tcp_client>& client, const tacopie::tcp_client::read_result& res) { if (res.success) { std::cout << "write" << std::endl; std::string resp = "this is the response"; std::string fullResp = "HTTP/1.1 200 OK\r\nContent-Type: type/html\r\nConnection: Closed\r\nContent-Length: " + std::to_string(resp.length()) + std::string("\r\n\r\n") + resp; std::vector<char> buf(fullResp.begin(), fullResp.end()); client->async_write({buf, [=](tacopie::tcp_client::write_result&) { std::cout << "write cb" << std::endl; client->disconnect(); }}); } else { std::cout << "disconnect" << std::endl; client->disconnect(); } }
gharchive/issue
2017-03-26T16:03:41
2025-04-01T04:54:52.139203
{ "authors": [ "Cylix", "vivekvrao" ], "repo": "Cylix/tacopie", "url": "https://github.com/Cylix/tacopie/issues/6", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
233052029
6.0适配问题 发现当在滚动布局下,垂直放两个recycleView时,为避免与滚动布局的滑动冲突,同时把两个recycleView的滑动事件禁止后(mRecyclerView.setNestedScrollingEnabled(false););滑动是比较流畅啦,但是在6.0版本下,放在上面的recycleView会显示不全。不过,在6.0版本以下,没问题。希望,大神后续能分析一下。 测试环境变量只是系统版本的区别吗? 请描述,版本,复现操作,使用代码。 你看下是不是这个问题?https://github.com/CymChad/BaseRecyclerViewAdapterHelper/issues/1140
gharchive/issue
2017-06-02T01:21:03
2025-04-01T04:54:52.141629
{ "authors": [ "CymChad", "cervy", "dacaoyuan" ], "repo": "CymChad/BaseRecyclerViewAdapterHelper", "url": "https://github.com/CymChad/BaseRecyclerViewAdapterHelper/issues/1174", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
265155563
结合粘性库stickyheadersrecyclerview后导致上拉加载无效 原本你们的加载是在最后一个item的时候就开始自动加载, 结合了该库后会导致滑动到最后一个item后没有触发自动加载回调... 不知道是你们还是他的库的冲突...? 你们在"下拉加载"的时候没有做headview , 没有出现问题, 在上拉加载的时候出现,不知道是不是因为上拉加载的FootView与该库对item创建的HeadView冲突的原因? 另外再问下 , 下拉加载这个需求在对网络断开或者请求失败的情况下 怎么跟上拉加载一样有一个error的view来重新加载? 第一个问题虽然可以把预加载的位置提前几个来规避,但是快速滑动下还是不行 第一个问题虽然可以把预加载的位置提前几个来规避,但是如果快速滑动的话还是会出现问题,请问这样的issue可以解决吗? 1.最好提供demo 2.没看懂 3.可以下拉失败的时候清空data并设置EmptyView @1109993488 原本上拉加载是在最后一个item的时候触发onLoadMoreRequested的,因为结合了该库之后,无法主动触发 , 只能设置setPreLoadNumber来强制去让其触发. demo 我找找接口看怎么弄个出来... 结合使用的时候加载更多布局应该不显示吧?如果显示了肯定会回调onLoadMoreRequested,你是否重写了getItemCount? @1109993488 刚刚试了一下 , 重写了getItemCount ,直接return mData.size() 的话好像是没问题了..不过那个那个加载完成的footview 就没有显示了... getItemCount 最好不要重写,里面控制显示header、footer、加载更多 @1109993488 (・(ェ)・)...我以为是说需要重写 , 之前是没有重写会导致issue标题的问题 , 重写后请求正常情况是没问题, 如果出现错误或者网络断开 , 那个Error的点击加载更多就无法显示了吧? 如果重写了getItemCount 数量会有问题,可能不会显示加载更多, 加载更多出现的错误或没有网络可以调loadMoreFail显示加载失败的布局, 如果下拉刷新出现错误或无网 可以通过清空data数据并且设置EmptyView显示自定义的错误界面 1.不是下拉刷新 我上面说的是下拉加载 就是UpFetch的情况下 如果下拉出现错误, 怎么跟上拉一样有个显示加载失败的布局? 2.那我现在是要重写getItemCount吗? 然后再针对网络异常的情况的时候,进行显示加载失败的布局么? 1.现在下拉加载没有加载状态这个功能 2.你可以把加载失败布局放到header里 @1109993488 你好, https://github.com/zofnk/RecyclerViewLoadMoreTest 我弄了demo 你可以看下吗?就是还原了下场景会出现的问题 你这是滑动到底部就异常了 @Override public long getHeaderId(int position) { final int dataPosition = position - getHeaderLayoutCount(); if (dataPosition >= mData.size()) { return -1; } return mData.get(position).hashCode(); } 这里应该控制下position,position可能会越界 @1109993488 对呀 这个问题是要怎么解决...?Orz.. 0.0上面不是给你列出了解决办法么 喔喔...我还以为是问题出现的地方..今天找这个问题找得有点头痛... 非常感谢!!!!!~么么哒~
gharchive/issue
2017-10-13T03:02:22
2025-04-01T04:54:52.148087
{ "authors": [ "1109993488", "zofnk" ], "repo": "CymChad/BaseRecyclerViewAdapterHelper", "url": "https://github.com/CymChad/BaseRecyclerViewAdapterHelper/issues/1633", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
215642201
fragment切换,页面不显示任何内容的问题 fragment恢复可见时,这时候adapter != null && data有数据,这时候调用adapter.newData(data);,页面不显示任何内容,问题找到了,但没有解决思路... 问题找到了? /js,replace时每次都new Fragment能解决现象,但不知道会不会影响性能。 请谅解一个菜鸟求真之路,没找到问题? ╮(╯▽╰)╭,还是把这个问题关了吧
gharchive/issue
2017-03-21T06:37:46
2025-04-01T04:54:52.149860
{ "authors": [ "CymChad", "ping999" ], "repo": "CymChad/BaseRecyclerViewAdapterHelper", "url": "https://github.com/CymChad/BaseRecyclerViewAdapterHelper/issues/890", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2265326064
🛑 Servicio de impuestos interno is down In fd073fb, Servicio de impuestos interno (https://homer.sii.cl) was down: HTTP code: 403 Response time: 368 ms Resolved: Servicio de impuestos interno is back up in 8f3f3dd after 13 minutes.
gharchive/issue
2024-04-26T09:10:08
2025-04-01T04:54:52.152677
{ "authors": [ "fbarriosCL" ], "repo": "CymaSuite/upptime", "url": "https://github.com/CymaSuite/upptime/issues/1944", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
105121191
Merge remote-tracking branch 'upstream/stable' into merge_stable Conflicts: src/rt/profilegc.d Auto-merge toggled on
gharchive/pull-request
2015-09-06T22:42:50
2025-04-01T04:54:52.164459
{ "authors": [ "9rnsr", "MartinNowak" ], "repo": "D-Programming-Language/druntime", "url": "https://github.com/D-Programming-Language/druntime/pull/1382", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
91677728
Build/Run with DMD on Windows with a username with white spaces Hi! I'm trying to build a DlangIDE with dub and dmd on Windows 7 Pro, but show me an error like that: Error: cannot read file C:\Users\My.d When my realy path is like that: *C:\Users\My Username* Searching in internet I found that my problem is my username because it has white spaces, but never found how can I change my dub's local repository. Anybody can help me with that? Can you try out the latest beta (0.9.24-beta.1, installer, zip) and see if it works there? This is most likely the same root cause as #540. Ok, I'll try and I'll let you know. Great!!! That's all!!! Thank you!!! I'm experiencing essentially the same bug, and I can confirm the beta fixes the bug.
gharchive/issue
2015-06-29T02:50:17
2025-04-01T04:54:52.168085
{ "authors": [ "ajvincent", "rro4785", "s-ludwig" ], "repo": "D-Programming-Language/dub", "url": "https://github.com/D-Programming-Language/dub/issues/610", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
230284455
Update projectServer.c Add ThreadQueue, DeleteQueue, Mutex so bad 👎
gharchive/pull-request
2017-05-22T04:53:52
2025-04-01T04:54:52.168902
{ "authors": [ "D4isy", "junog115" ], "repo": "D4isy/ScreenCaptureNServer", "url": "https://github.com/D4isy/ScreenCaptureNServer/pull/3", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1932072451
Autogenerate argparser to execute Driver from command line Is your feature request related to a problem? Please describe. When deploying scripts in SageMaker or VertexAI, configuration needs to be passed via CLI / argparse. This requires writing tedious argparse code, which often ends up being poorly typed, documented, and maintained. Describe the solution you'd like Automatically generate argparser (or OmegaConf/Hydra) for Hamilton nodes of the instantiated driver. Each parsing argument could include the type and docstring (if not a top node). It is possible to resolve what should be in inputs and in overrides. final_vars could also be specified. import argparse class Driver: ... def with_cli(self): parser = argparse.ArgumentParser(prog="HamiltonCLI", description="Generated CLI") for n in self.graph.get_nodes(): parser.add_argument(f"--{n.name}") self.args = parser.parse_args() def resolve_args(self): inputs, overrides = resolve_node_value(self.args) action, kwargs = resolve_action(self.args) # visualize, execute if action == "execute": self.execute(inputs=inputs, overrides=overrides, **kwargs) if __name__ == "__main__": dr = ( driver.Builder() .with_module(transforms) .with_cli() # include arguments to limit supported operations (e.g., execute only) .build() ) Alternative A simpler and more explicit approach could be passing two list of strings to expose input and override nodes. This prevents having a CLI flooded with irrelevant args if __name__ == "__main__": dr = ( driver.Builder() .with_module(transforms) .with_cli(inputs=list(), overrides=list()) .build() ) The nodes supported would be limited to primitives that can be expressed on the command line. One challenge is properly coercing args, which are all strings, into the correct Hamilton type. This could be done efficiently with Pydantics yep -- https://typer.tiangolo.com/ could perhaps make this simpler? Another clean pattern could be to decorate functions with @cli and then .with_cli() collects that when building the driver. However, it wouldn't be possible to annotate top level nodes Another clean pattern could be to decorate functions with @cli and then .with_cli() collects that when building the driver. However, it wouldn't be possible to annotate top level nodes Yeah without instantiating a driver and knowing the requested outputs we wouldn't know what they are. But that doesn't mean we couldn't have something dynamic... or alternatively we just have a command line that creates a CLI file for a given driver set up...
gharchive/issue
2023-10-08T21:24:07
2025-04-01T04:54:52.180512
{ "authors": [ "skrawcz", "zilto" ], "repo": "DAGWorks-Inc/hamilton", "url": "https://github.com/DAGWorks-Inc/hamilton/issues/448", "license": "BSD-3-Clause-Clear", "license_type": "permissive", "license_source": "github-api" }
220268338
use of DKPro-Wrapper as a server (or at least with stdin and stdout is it possible to use the DKPro-Wrapper as a server (similar to coreNLP) or at least as continously running, expecting text on stdin and producing stdout, without closing. The startup time is a heavy overhead if the pipelline is started for each small snippet of text as it occurs and i would prefer not to collect all input and then treat it in a single file. thank you! it is possible to run the pipeline on a folder of text or xml files, the wrapper will then create a filename.csv for each file in the given output folder. Just pass a directory to the -input option. There are currently no plans on our side to implement a server mode. (Running a full analysis on texts of notable length takes quite a while, so there's a certain risk of timeouts) A real streaming mode implementation isn't possible since the document must be loaded completely before further analysis steps begin. It might be possible to implement something that could accept multiple documents on stdin separated by sth like \0 …
gharchive/issue
2017-04-07T16:46:10
2025-04-01T04:54:52.182824
{ "authors": [ "andrewufrank", "thvitt" ], "repo": "DARIAH-DE/DARIAH-DKPro-Wrapper", "url": "https://github.com/DARIAH-DE/DARIAH-DKPro-Wrapper/issues/29", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2355509163
Deprecated "Programming Related" link The autocode library got shut down https://discordresources.com/resources/tools/programming#autocode https://discordresources.com/resources/moderation#a-guide-to-civil-conversation this one is broken too This is an open project. Anyone can do PR to fix these issues :)
gharchive/issue
2024-06-16T07:40:33
2025-04-01T04:54:52.186145
{ "authors": [ "JantsoP", "Pdzly" ], "repo": "DC-Resources-Wiki/Discord-Resources-Wiki", "url": "https://github.com/DC-Resources-Wiki/Discord-Resources-Wiki/issues/20", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
699128758
cannot see ShapeVariationAnalyzer module @jcfr @allemangD @bpaniagua Hi ShapeVariationAnalyzer is not working/cannot find module after installing SlicerCMF. Thanks! Per the log in https://github.com/DCBIA-OrthoLab/SlicerCMF/issues/18, the issue seems to be that sklearn is not configured properly for Mac, and this causes errors when it is imported by shapepca, which is imported by ShapeVariationAnalysis. The same happens if only ShapeVariationAnalyzer is installed via the extension manager (without SlicerCMF). Here is a full log from my machine (using yesterday's Slicer preview release) Slicer_29370_20200917_132211_450.log When I build ShapeVariationAnalyzer myself, no such error occurs. $ git clone https://github.com/DCBIA-OrthoLab/ShapeVariationAnalyzer.git $ cd ShapeVariationAnalyzer $ mkdir build $ cd build $ cmake -DShapeVariationAnalyzer_BUILD_SLICER_EXTENSION=ON -DSlicer_DIR=~/src/Slicer/build/Slicer-build .. $ ninja $ inner-build/SlicerWithShapeVariationAnalyzer This tells me there may be some issue in the factory machine causing sklearn not to be configured correctly. @jcfr, do you have any thoughts on this? If I understand the superbuild correctly, scikit-learn is installed via pip in the default configuration. https://github.com/DCBIA-OrthoLab/ShapeVariationAnalyzer/blob/master/SuperBuild/External_python-ShapeVariationAnalyzer-requirements.cmake#L43-L54 If I manually install via pip with the same hashes, I get an error that the hash does not match: ERROR: THESE PACKAGES DO NOT MATCH THE HASHES FROM THE REQUIREMENTS FILE. If you have updated the package versions, please update the hashes. Otherwise, examine the package contents carefully; someone may have tampered with them. scikit-learn==0.23.1 from https://files.pythonhosted.org/packages/ea/47/078b5ef83ccff1def48c29158be998375321397be53af6dc10f22c6dc08d/scikit_learn-0.23.1-cp38-cp38-macosx_10_9_x86_64.whl#sha256=0c3464e46ef8bd4f1bfa5c009648c6449412c8f7e9b3fc0c9e3d800139c48827 (from -r req.txt (line 1)): Expected sha256 e585682e37f2faa81ad6cd4472fff646bf2fd0542147bec93697a905db8e6bd2 Expected or 058d213092de4384710137af1300ed0ff030b8c40459a6c6f73c31ccd274cc39 Expected or e9879ba9e64ec3add41bf201e06034162f853652ef4849b361d73b0deb3153ad Got 0c3464e46ef8bd4f1bfa5c009648c6449412c8f7e9b3fc0c9e3d800139c48827 This is confusing to me, since the superbuild succeeds locally without issue and cdash does not report any such issue. The issue doesn't seem to affect the Linux build of Slicer or the extension manager. Here is the log from my Ubuntu 20 machine after installing the SlicerCMF extension bundle. ShapeVariationAnalyzer and the Population Analysis module load without issue. Slicer_29363_20200917_135032_863.log Do you mind testing in your macbook? The above logs are from my macbook. Everything works correctly if I build the extension myself, but loading from the extension manager causes the problem. This makes me think there is some issue with the factory machine. In the second comment I was just noting that the issue is never present for me on linux, to emphasize I think this issue is specific to the mac factory machine and not the others, although I can't confirm for Windows. Thanks for checking, David! @sjh26 can you please check in any issues in the macOS factory? If I manually install via pip with the same hashes, I get an error that the hash does not match: @allemangD and I looked at this, and python 3.6 (version also used in Slicer) need to be used to ensure the hashes match. It looks like the .dylibs directory containing libomp.dylib is properly packaged in the .tar.gz file but there is any extracting the content of the archive into its final destination. This has been fixed in https://github.com/Slicer/Slicer/pull/5307 @bpaniagua @allemangD @jcfr ShapeVariationAnalyzer does not show and instead in the SlicerCMF index when we click on ShapeVariationAnalyzer a module Population Analysis starts that does not seem to allow computing group averages or group differences , and is quite different from SVA. Not sure how to report this issue and will further clarify with Bea in June. @luciacev The fix for this bug required a change to Slicer, so the stable version from last year we had told you to use () will not contain that fix. I've confirmed that SVA loads correctly on Mac in the latest stable version (4.11.20210226). As best I can tell, all the other modules work correctly there, too. I would recommend we try to use that 2021 version moving forward. Regarding https://github.com/DCBIA-OrthoLab/SlicerCMF/issues/13, SPHARM-PDM is available on Mac in that version, but not RigidAlignment. I'm working on fixing that next. I've made notes about these in the weekly updates document, and will send that out to the team this Friday. Downloads for that revision here. Closing as the issue is fixed in the current stable release 4.11.20210226.
gharchive/issue
2020-09-11T10:15:03
2025-04-01T04:54:52.197442
{ "authors": [ "allemangD", "bpaniagua", "jcfr", "luciacev" ], "repo": "DCBIA-OrthoLab/ShapeVariationAnalyzer", "url": "https://github.com/DCBIA-OrthoLab/ShapeVariationAnalyzer/issues/54", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1746150800
Images link for first two manuscripts in "Browse Chants" page lead to a page that never loads To reproduce: Go to "Browse Chants" page; Click on the "Images" link for either A-Gu 29 or A-Gu 30; A new tab opens, but the url is weird and the site never loads. Demo (New vs. Old comparison): https://github.com/DDMAL/CantusDB/assets/83373378/fa8da2fc-0d77-4e7b-bb9a-73a384936b68 I've clicked on random other manuscript images and they all load fine. From my casual random selection, it seems to be just these two that are badly behaved. That weird URL shows up elsewhere on the internet, so I'm assuming the library changed it at some point; I think it was still floating around on OldCantus last October, so it should go away once the data is updated (but will be worth checking.) OldCantus links to the images at literature.at : http://www.literature.at/viewer.alo?objid=1138&page=9&viewmode=fullscreen&rotate=&scale=1 It would actually be much better to link to the university's images: https://unipub.uni-graz.at/obvugrscript/content/pageview/6789708 , both because the URL is less ugly and also because the literature.at ones are in a completely bonkers order (which will be relevant to serving them on CU: @dchiller take note!) But maybe that is a problem for Cantus to solve and not DDMAL! We can't do too much about other sites not loading well, but if there are different urls, this is a data issue that we can expect to be resolved once we sync the data from OldCantus to New. @annamorphism, I'm assigning you to this, in case you want to change the urls for these sources as you suggested (either do it on OldCantus in the next few days so we can sync from there, or wait for the full switchover and make the change both on Staging and Production). Please close this once you're done (or whenever you feel it's appropriate) As there is nothing for NewCantus developers to do about this issue, I'm closing it.
gharchive/issue
2023-06-07T15:19:32
2025-04-01T04:54:52.211894
{ "authors": [ "JoyfulGen", "annamorphism", "jacobdgm" ], "repo": "DDMAL/CantusDB", "url": "https://github.com/DDMAL/CantusDB/issues/712", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1813248830
🛑 vicmap-hydro-rest-api is down In d3b2258, vicmap-hydro-rest-api (https://services6.arcgis.com/GB33F62SbDxJjwEL/arcgis/rest/services/Vicmap_Hydro/FeatureServer) was down: HTTP code: 411 Response time: 218 ms Resolved: vicmap-hydro-rest-api is back up in d3ab180.
gharchive/issue
2023-07-20T06:23:38
2025-04-01T04:54:52.239489
{ "authors": [ "alex-vic-geo" ], "repo": "DELWP-DTV/santa", "url": "https://github.com/DELWP-DTV/santa/issues/454", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
287784375
Add "unit" tests. src/test/scala and src/test/resources currently empty Needs basic tests and example image data. Fixed in commit e48f8d2ca0096e52491dc5c5d93268604878829e of GIF git repository.
gharchive/issue
2018-01-11T13:23:01
2025-04-01T04:54:52.244233
{ "authors": [ "mbeckerle" ], "repo": "DFDLSchemas/GIF", "url": "https://github.com/DFDLSchemas/GIF/issues/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
701790337
Check if find is down only once on providers page Context The list of available courses on Apply (https://www.apply-for-teacher-training.service.gov.uk/candidate/providers) is a bit slow according to Skylight: https://www.skylight.io/app/applications/t8bEzG0cuIkd/recent/6h/endpoints/CandidateInterface::ContentController%23providers?responseType=html Changes proposed in this pull request Query is find is down only once. This saves a lot of SiteSetting queries and calculation, and might solve the performance issues with this page. Guidance to review Link to Trello card https://trello.com/c/x0OxYLae/2143-speed-up-the-providers-page-in-the-candidate-ui Things to check [x] This code does not rely on migrations in the same Pull Request [x] If this code includes a migration adding or changing columns, it also backfills existing records for consistency [x] API release notes have been updated if necessary [x] New environment variables have been added to the Azure config @davidgisbey it was my bad - I introduced the SiteSetting lookup in the timetable, which made this slow. @tvararu should have mentioned that, but it's a class method, so memoizing it would persist across requests and not change if today's date or the cycle schedule changes.
gharchive/pull-request
2020-09-15T09:59:03
2025-04-01T04:54:52.248951
{ "authors": [ "tijmenb" ], "repo": "DFE-Digital/apply-for-teacher-training", "url": "https://github.com/DFE-Digital/apply-for-teacher-training/pull/2902", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2225998783
[1514] Part 1 add unique index to Vender name Context Currently, we store and validate the Vendor name as an enum on the Vendor model. There are problems with this: Adding a new Vender requires a code change. This is also a potential security risk as all our vendors are located in our public repo. We want to replace the enum with basic presence and uniqueness validation. Changes proposed in this pull request Migration to add a unique index to vendor name. I've checked in production that all of our vendors have unique names as it is (there are only 6). Once this has been merged in, I will open the PR here which actually gets rid of the enum on the model Guidance to review Link to Trello card https://trello.com/c/jMnBxRIb Things to check [x] If the code removes any existing feature flags, a data migration has also been added to delete the entry from the database [x] This code does not rely on migrations in the same Pull Request [x] If this code includes a migration adding or changing columns, it also backfills existing records for consistency [ ] If this code adds a column to the DB, decide whether it needs to be in analytics yml file or analytics blocklist [ ] API release notes have been updated if necessary [ ] If it adds a significant user-facing change, is it documented in the CHANGELOG? [ ] Required environment variables have been updated added to the Azure KeyVault Tech debt: Remove Enum of Vendor names
gharchive/pull-request
2024-04-04T16:28:30
2025-04-01T04:54:52.254988
{ "authors": [ "elceebee" ], "repo": "DFE-Digital/apply-for-teacher-training", "url": "https://github.com/DFE-Digital/apply-for-teacher-training/pull/9262", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
365912714
Fixed socket exception Context SocketException Changes proposed in this pull request Exhausted socket Guidance to review Same vein as https://github.com/DFE-Digital/search-and-compare-ui/pull/186 Craig David bug causing exceptions - 21 seconds... timeout
gharchive/pull-request
2018-10-02T14:02:09
2025-04-01T04:54:52.268982
{ "authors": [ "defong", "timabell" ], "repo": "DFE-Digital/search-and-compare-ui", "url": "https://github.com/DFE-Digital/search-and-compare-ui/pull/187", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1075693897
[TEVA-3546] Allow education support roles to be apply via TV Jira ticket URL https://dfedigital.atlassian.net/browse/TEVA-3546 Screenshots of UI changes: @cesidio I've got jobseeker open on left and hiring staff on right. I still see the Professional status question for a Teaching Assistant role I don't see this for SENDCo or education support though Look good, section no longer there for Teaching Assistant
gharchive/pull-request
2021-12-09T15:20:13
2025-04-01T04:54:52.272667
{ "authors": [ "bencmitchell", "cesidio" ], "repo": "DFE-Digital/teaching-vacancies", "url": "https://github.com/DFE-Digital/teaching-vacancies/pull/4407", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1402273733
[Bug]: 导出的UIGF Json版本中gacha_type这一字段与UIGF标准不一致 Windows 版本 22621.521 Snap Hutao 版本 1.1.8.0 发生了什么? 导出的json中每个list里面的对象中gacha_type字段为NoviceWish、PermanentWish、AvatarEventWish2、AvatarEventWish、WeaponEventWish中的一种,而不是预期的100、200、301、302、400,这导致了胡桃导出的json导入到snap.genshin或者其他工具时可能会报错。 你期望发生的行为? 一个当前版本胡桃导出的json片段示例: { "uigf_gacha_type": "200", "gacha_type": "PermanentWish", "item_id": "", "count": "1", "time": "2020-11-07 00:41:41", "name": "魔导绪论", "item_type": "武器", "rank_type": "3", "id": 1000000000000000080 } 实际上应该是: { "uigf_gacha_type": "200", "gacha_type": "200", "item_id": "", "count": "1", "time": "2020-11-07 00:41:41", "name": "魔导绪论", "item_type": "武器", "rank_type": "3", "id": 1000000000000000080 } 相关的崩溃日志 位于 %HOMEPATH%/Documents/Hutao/Log.db 无 我确认已在表单中附上了充足的补充说明以帮助开发人员确定问题 [X] 是 我确认该问题是一个新问题,没有其他人已经提出相同的问题 [X] 是 我确认该问题没有在文档中解释 [X] 是 导出功能尚未制作完成
gharchive/issue
2022-10-09T13:23:17
2025-04-01T04:54:52.310392
{ "authors": [ "CHxCOOH", "Lightczx" ], "repo": "DGP-Studio/Snap.Hutao", "url": "https://github.com/DGP-Studio/Snap.Hutao/issues/101", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1594164235
Set up jekyll build teston PRs We should probably test that the build doesn't fail before merging. Set this up. In theory, when we create PRs where, a check should run to see if the PR breaks the jekyll build.
gharchive/issue
2023-02-21T22:13:28
2025-04-01T04:54:52.312443
{ "authors": [ "smythp" ], "repo": "DHRI-Curriculum/Dhrift-Homepage", "url": "https://github.com/DHRI-Curriculum/Dhrift-Homepage/issues/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
184943820
Exchange XML examples for renew domain The examples was originally lifted from the RFCs to support the developers of the extension of the service, these examples should be exchanged for real working examples lifted from the actual service so all details are captured. The examples for this particular command have been updated: ref: https://github.com/DK-Hostmaster/epp-service-specification#renew-domain
gharchive/issue
2016-10-24T20:18:53
2025-04-01T04:54:52.362892
{ "authors": [ "jonasbn" ], "repo": "DK-Hostmaster/epp-service-specification", "url": "https://github.com/DK-Hostmaster/epp-service-specification/issues/8", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
922757145
Respect PointBehaviour for tiglWingGetSegmentEtaXsi Description I had to change the class structure / method signature slightly to allow to pass, how GetEtaXsi is evaluated. Unfortunately, GetEtaXsi on the paramter surfaces is almost a factor of thousand slower, which could be a problem for the VTK export with metadata. Therefore I decided to use the old eta / xsi interpretation for the exports for now. closes #808 How Has This Been Tested? To validate, I changed the consistency test, added a performance test and added a function to tiglviewer to play around with the function. Screenshots, that help to understand the changes(if applicable): Checklist: [x] A test for the new functionality was added. [ ] All tests run without failure. [x] The new code complies with the TiGL style guide. [ ] New classes have been added to the Python interface. [ ] API changes were documented properly in tigl.h. Codecov Report Merging #811 (1833e84) into master (e05dfb1) will increase coverage by 0.07%. The diff coverage is 85.93%. @@ Coverage Diff @@ ## master #811 +/- ## ========================================== + Coverage 62.17% 62.25% +0.07% ========================================== Files 291 291 Lines 23438 23466 +28 ========================================== + Hits 14572 14608 +36 + Misses 8866 8858 -8 Impacted Files Coverage Δ src/wing/CCPACSWing.h 100.00% <ø> (ø) src/wing/CCPACSWingSegment.h 50.00% <ø> (ø) src/wing/CCPACSWing.cpp 74.55% <46.15%> (+0.99%) :arrow_up: src/wing/CCPACSWingSegment.cpp 82.64% <95.55%> (+1.65%) :arrow_up: src/api/tigl.cpp 54.39% <100.00%> (-0.03%) :arrow_down: src/geometry/CTiglTriangularizer.cpp 86.90% <100.00%> (+0.07%) :arrow_up: src/wing/CCPACSWingComponentSegment.cpp 50.09% <100.00%> (-0.19%) :arrow_down:
gharchive/pull-request
2021-06-16T15:22:12
2025-04-01T04:54:52.379753
{ "authors": [ "codecov-commenter", "rainman110" ], "repo": "DLR-SC/tigl", "url": "https://github.com/DLR-SC/tigl/pull/811", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2014876095
Suggestion to include nbviewer-link in readme The experience of looking at Jupyter Notebooks on GitHub is unfortunately not the best. It might be a good idea to include the following link in the Readme: https://nbviewer.org/github/DLR-terrabyte/demo-bids23/blob/main/demo.ipynb Good call @maawoo. Thanks!
gharchive/issue
2023-11-28T16:50:58
2025-04-01T04:54:52.381475
{ "authors": [ "johntruckenbrodt", "maawoo" ], "repo": "DLR-terrabyte/demo-bids23", "url": "https://github.com/DLR-terrabyte/demo-bids23/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1044887627
Update synth_t structure Update synth_t struct according to current FPGA structure, which sums to a size of 1244 bytes: package shape_pkg; typedef enum logic [7:0]{ SAWTOOTH = 0, SQUARE, SIN, PIANO } wave_shape; endpackage package protocol_pkg; //---------------------------------------------------------------------------------------------/ // Definition of protocol for communication between MCU and FPGA //---------------------------------------------------------------------------------------------/ import shape_pkg::*; typedef struct packed { logic [31:0] gain /* FIXED POINT */; logic [31:0] duration; } envelope_t; typedef struct packed { logic [31:0] freq /* FIXED POINT */; logic [31:0] velocity; envelope_t [0:`ENVELOPE_LEN-1] envelopes; wave_shape shape; logic [7:0] cmds; } wavegen_t; // typedef struct packed { // logic [31:0] delay; // logic [31:0] feedback; // } echo_t; typedef struct packed { logic signed [31:0] balance /* FIXED POINT */; } pan_t; // typedef struct packed { // logic [31:0] playback_volume; // logic [7:0] playback_speed; // logic [7:0] cmds; // } looper_t; typedef struct packed { // Tau values are delays. 0, 1, 2 and 3 are for comb filters, selected in a 1:1.5 range where // the smallest determines the initial delay. Tau 4 and 5 are for all-pass filters, usually // set to 5ms and 1.7 ms. Tau is computed as time [ms] * sample rate // Gain values are always < 1.0. 0, 1, 2 and 3 are for comb filters = 10 ^ (-3 t / T), where // T is the reverb time. logic signed [31:0][0:5] tau; logic signed [31:0][0:6] gain /* FIXED POINT */; } reverb_t; typedef struct packed { wavegen_t [0:`N_OSCILLATORS-1] wave_gens; logic [31:0] master_volume; reverb_t reverb; pan_t pan; // looper_t looper; // echo_t echo; } synth_t; endpackage Possible changes to save bandwidth: typedef struct packed { logic [7:0] gain /* FIXED POINT */; logic [7:0] duration; } envelope_t; Takes size from 1244 bytes to 524 bytes.
gharchive/issue
2021-11-04T15:26:19
2025-04-01T04:54:52.385263
{ "authors": [ "akberg" ], "repo": "DMPRO2021-Audio/mcu", "url": "https://github.com/DMPRO2021-Audio/mcu/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
338494380
Configure Plan readability through branding.yml Currently, plans are readable by a super_admin, an admin of any of the organizations to which co-owners belong, or to any user with a role in the plan. For some services, it is not wanted for org-admins to be able to view the full text of any plan within their organization, and this should be restricted to super-admins, or org-admins with requested reviews. Note that a user requesting a review adds a role for the associated org-admin users on the plan. We should add a configuration option for whether or not org-admins can view plans from users of their org. This will affect whether the plans at /org_admin/plans have active links to the full DMP content or not. Where all plans can't be viewed by org admins, they will only have access to plans where a specific review has been requested (see notification panel above plans table) i tested as a super admin and org admin for UCSF and noted some weird behaviors: as a super admin I see the notifications panel w/many plans awaiting feedback (screenshot 1). i can also click the links to open each plan below and view the content. as an org admin I don't see the notifications panel, but I should (screenshot 2). when i click the links to open the plans in the main table i receive an error msg (screenshot 3) and am taken to My Dashboard. desired behavior is to make the default configuration to give Org Admins authorization to open and view plans. I think this was a requirement for UK unis in fact, and is the current behavior. Org Admins should always see plans awaiting feedback in the Notifications panel regardless of the configuration setting. if the tool is configured to prevent Org Admins from viewing the contents of all plans, the error message should not redirect them to My Dashboard. instead they should stay on the Plans page. Super Admin UCSF view of plans page Org Admin UCSF view of plans page Error message when I click the link to view a plan as an Org Admin default configuration looks good now. will just need to keep an eye on whether newly privileged admins can see a complete list of plans awaiting feedback. closing issue.
gharchive/issue
2018-07-05T09:25:56
2025-04-01T04:54:52.390846
{ "authors": [ "stephaniesimms", "xsrust" ], "repo": "DMPRoadmap/roadmap", "url": "https://github.com/DMPRoadmap/roadmap/issues/1661", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1214808100
Conditional questions causing plans to disapear Plans that include a conditional question are disappearing. For example, a user can successfully create a plan from a template including a conditional question, but when they go back to view the plan, the plan will disappear from the page. We temporarily fixed this issue by removing the conditional question from the template, but I believe this bug will be affecting all plans that include conditional questions. @raycarrick-ed This seems to be related to issue I noticed when dealing with a bug for Maastricht. The plan sections container continues to have style="display:none". Will investigate. @mariapraetzellis, @briri Not yet identified cause, but working on it. As I have an example which fails. Just proved in broken case cause this rendering cuases everything in content_for to be surrouned by a display:none block. <%= render partial: 'phases/edit_plan_answers', locals: { plan: plan, phase: phase, answers: answers, readonly: readonly, base_template_org: base_template_org, guidance_presenter: guidance_presenter, } %> Thanks @briri. A cursory test shows commenting out hide questions fixes issue. Will tomorrow dig deeper as suggested in your comment. I think @raycarrick-ed added this to the latest release, so I will close out here. @pherterich I can't see it in https://github.com/DigitalCurationCentre/roadmap/commits/deploy/dmponline?before=99d780716fca202f19a6dcb7da40ce62edb52e76+35&branch=deploy%2Fdmponline&qualified_name=refs%2Fheads%2Fdeploy%2Fdmponline It is possible has been patched by @raycarrick It is patched and part of release 3.1.1 proper
gharchive/issue
2022-04-25T17:15:40
2025-04-01T04:54:52.395305
{ "authors": [ "johnpinto1", "mariapraetzellis", "pherterich" ], "repo": "DMPRoadmap/roadmap", "url": "https://github.com/DMPRoadmap/roadmap/issues/3163", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
365847975
Fix form validations on Orgs form Fixes #1860 . Changes proposed in this PR: Update .gitignore file Add validations for contact_name field Remove validations for links fields I think we may need to restore some JS validation for the links (used here and on the template details page when the org is a funder). The links aren't required, but when one is provided it must have both a URL and Text. The URL should be validated to ensure that its a valid URL. Lets refactor that afterward (along with the CSS changes you've made). Orgs don't update their links very often. We'll have to add a cleanup task though to capture any unvalidated links that get through between releases.
gharchive/pull-request
2018-10-02T11:08:34
2025-04-01T04:54:52.397757
{ "authors": [ "Bodacious", "briri" ], "repo": "DMPRoadmap/roadmap", "url": "https://github.com/DMPRoadmap/roadmap/pull/1902", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1987022956
Add test case 7.5 for MEASUREMENTS. fix: #117 fix: #112 please show the test result. please show the test result.
gharchive/pull-request
2023-11-10T06:48:55
2025-04-01T04:54:52.405089
{ "authors": [ "Xiaohanjlll", "jyao1" ], "repo": "DMTF/SPDM-Responder-Validator", "url": "https://github.com/DMTF/SPDM-Responder-Validator/pull/118", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
210498492
[PP] Give an URI to periods Historical periods should have an URI in order to avoid blank and duplicate nodes in the graph. Query: select distinct ?label COUNT(?s) where { ?s a ecrm:E4_Period; ecrm:P1_is_identified_by ?label . } GROUP BY ?label Proposal: <http://data.doremus.org/period/UUID>. This can also be aligned with dbpedia, e.g. <http://dbpedia.org/resource/17th_century> What about <http://data.doremus.org/period/17th_century> ? Why not! But would you be able to name any period you may encounter in the data? It is easy when we talk about plain century but in the data, I can see "20 ème siècle - 2 ème moitié" or "Périodes historiques" or "Révolution française". In short, I think it would be better to provide a UUID to identify the period and add a proper label and sameAs links to wikipedia/dbpedia entities when they exist
gharchive/issue
2017-02-27T14:25:38
2025-04-01T04:54:52.431414
{ "authors": [ "pasqLisena", "rtroncy" ], "repo": "DOREMUS-ANR/marc2rdf", "url": "https://github.com/DOREMUS-ANR/marc2rdf/issues/40", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1977071745
[Registry Mobile][WP theme] Remove sections from mobile version "About the Digital Public Good Standard" Button Acceptance Criteria: URL: https://digitalpublicgoods.net/registry/ "About the Digital Public Good Standard" is visible on desktop Set screen size to mobile such as iPhone SE, and "About the Digital Public Good Standard" text and link should disappear.
gharchive/issue
2023-11-04T00:18:56
2025-04-01T04:54:52.452277
{ "authors": [ "law909", "skylark1848" ], "repo": "DPGAlliance/publicgoods-scripts", "url": "https://github.com/DPGAlliance/publicgoods-scripts/issues/239", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2584021711
🛑 SSO Unhas is down In 26beb46, SSO Unhas (https://sso.unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: SSO Unhas is back up in 5dc3464 after 5 minutes.
gharchive/issue
2024-10-13T14:43:30
2025-04-01T04:54:52.465897
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/11520", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2007433670
🛑 SSO Unhas is down In f39b477, SSO Unhas (https://sso.unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: SSO Unhas is back up in 57c1b92 after 5 minutes.
gharchive/issue
2023-11-23T03:36:42
2025-04-01T04:54:52.468233
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/4634", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2041105061
🛑 Fakultas Hukum is down In 98fdd1b, Fakultas Hukum (https://lawfaculty.unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: Fakultas Hukum is back up in 7a20ad9 after 11 minutes.
gharchive/issue
2023-12-14T07:29:55
2025-04-01T04:54:52.470728
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/5125", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2175355778
🛑 Fakultas Kedokteran is down In 75ecbc6, Fakultas Kedokteran (https://med.unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: Fakultas Kedokteran is back up in 9daaee0 after 13 hours, 37 minutes.
gharchive/issue
2024-03-08T06:20:05
2025-04-01T04:54:52.473044
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/6829", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2392064106
🛑 LP2M is down In 27436da, LP2M (https://lp2m.unhas.ac.id) was down: HTTP code: 0 Response time: 0 ms Resolved: LP2M is back up in f3f7d91 after 15 minutes.
gharchive/issue
2024-07-05T07:56:34
2025-04-01T04:54:52.475522
{ "authors": [ "aisprayogi" ], "repo": "DSITD-Universitas-Hasanuddin/uptime-checker", "url": "https://github.com/DSITD-Universitas-Hasanuddin/uptime-checker/issues/9403", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
179222759
DS-2623 backport set description in the upload step files Backport for #1013 that fix the set description in the upload file and add the possibility to setting up the description for multiple files. I tested both with simple UploadStep than with UploadWithEmbargoStep and it work fine! I tested simple UploadStep and UploadWithEmbargoStep on my DSpace-CRIS 5.5 and it work fine! Great Work! This is just a backport. Code changes look bigger as they are caused by changed indentation. +1 by review and test. Looks good. Since this has gotten some testing, I'm OK with this going into the 5.6 release. +1 as well, tested it works
gharchive/pull-request
2016-09-26T13:04:59
2025-04-01T04:54:52.485927
{ "authors": [ "abollini", "jpiscanc", "lap82", "pnbecker", "tdonohue" ], "repo": "DSpace/DSpace", "url": "https://github.com/DSpace/DSpace/pull/1531", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
473254355
implement upload bitstream to archived item Support for uploading a bitstream to an archived item (outside the workspace/workflow) The contract can be found at https://github.com/DSpace/Rest7Contract/pull/61 A sample json call is: curl 'http://10.211.55.20:8080/rest/api/core/items/d30de96b-1e76-40ae-8ef9-ab426b6f9763/bitstreams' -XPOST -H 'Content-Type: multipart/form-data' -H "Authorization: $authrization2" -F "file=@Desktop/test-pdf.pdf" -F 'properties={ "metadata": { "dc.description": [ { "value": "example file", "language": null, "authority": null, "confidence": -1, "place": 0 } ]}, "bundleName": "TEST" };type=application/json' The corresponding output is: { "id" : "f31a304e-9c7b-4809-99e8-e4940956e50c", "uuid" : "f31a304e-9c7b-4809-99e8-e4940956e50c", "name" : "test-pdf.pdf", "handle" : null, "metadata" : { "dc.description" : [ { "value" : "example file", "language" : null, "authority" : null, "confidence" : -1, "place" : 0 } ], "dc.title" : [ { "value" : "test-pdf.pdf", "language" : null, "authority" : null, "confidence" : -1, "place" : 0 } ] }, "bundleName" : "TEST", "sizeBytes" : 6519, "checkSum" : { "checkSumAlgorithm" : "MD5", "value" : "d8d81bf9a24c3aa511b9f670db45d13a" }, "sequenceId" : 17, "type" : "bitstream", "_links" : { "content" : { "href" : "http://10.211.55.20:8080/rest/api/core/bitstreams/f31a304e-9c7b-4809-99e8-e4940956e50c/content" } }, "_embedded" : { "format" : { "id" : 4, "shortDescription" : "Adobe PDF", "description" : "Adobe Portable Document Format", "mimetype" : "application/pdf", "supportLevel" : 0, "internal" : false, "extensions" : null, "type" : "bitstreamformat" } } } This PR is ready to be reviewed again This is at +2, and it's already been updated per today's discussion (around making sequence ID read only). Merging.
gharchive/pull-request
2019-07-26T09:02:32
2025-04-01T04:54:52.489887
{ "authors": [ "benbosman", "tdonohue" ], "repo": "DSpace/DSpace", "url": "https://github.com/DSpace/DSpace/pull/2473", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2204527289
Convert the patches from being hard-coded to having their own file format. The easiest way for both creating new patches and reading them might be a zip file, with multiple files inside it describing each patch and containing the code to modify. What about just using regular C# files, where it will modify the code via functions? I thought about that, and that could be a good option, but that would require the person to know C# to use. I was thinking something along the lines of a few configuration files and gml code, that way creating mods doesn't require you to learn two languages + UndertaleModLib. So the patchscript would basically just be a file with patch's config and gml code. I have another question, will there be some sort kf "mods"? Like a collection of multiple patches. Or patches are already considered mods? Patchscript would be for the config files, not sure how that's gonna look yet but it's going to be fairly simple to write. I was planning on "mods" just being a zip file with the multiple patches in them (since writing one file would be ridiculous for large mods), and the patcher would look for every config file that existed in that zip file and load it accordingly. That means patchscript would be as a config for the name of the patch, description, etc. And maybe the GML file wity the actual code. I don't really think it's worth it storing a single line of configuration in a different file. Maybe so, but many of the patches within cirQOL right now modify multiple files at a time, or replace multiple different sections in a file. The config would also be used to specify the regions of a file to find and/or replace, and how to go about doing it. (Plus, it's just cleaner to separate the config from the actual code once a mod gets big enough). As a solution to this, maybe making it possible to insert a file as a string in patchscript may be wise, since then the code to insert/replace can just be a string input that can be linked to a file, or it can just be placed in the same file as the config if you need it.
gharchive/issue
2024-03-24T19:50:49
2025-04-01T04:54:52.493785
{ "authors": [ "DT-is-not-available", "qaptivator" ], "repo": "DT-is-not-available/circloO-patcher", "url": "https://github.com/DT-is-not-available/circloO-patcher/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1872836764
Check HiveSQL chore syntax Refer Doc https://cwiki.apache.org/confluence/display/Hive/LanguageManual Todo List [ ] support missing syntax [ ] correct incorrect syntax [ ] unit tests @HaydenOrz HIVESQL 和 SPARKSQL 有跟引擎确认版本么,之前TRINOSQL 每个版本差异都比较大 Hive SQL 文档中废弃的语法会有记录有,所以直接按最新的文档来, Spark SQL 还没有确认 目前只关心是否有废弃语法,现在的做法是按最新的来,并且兼容废弃语法
gharchive/issue
2023-08-30T03:37:29
2025-04-01T04:54:52.496557
{ "authors": [ "HaydenOrz", "mumiao" ], "repo": "DTStack/dt-sql-parser", "url": "https://github.com/DTStack/dt-sql-parser/issues/143", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
622499602
Add Emote#isAvailable() method Pull Request Etiquette [x] I have checked the PRs for upcoming features/bug fixes. [x] I have read the contributing guidelines. Changes [ ] Internal code [x] Library interface (affecting end-user code) [ ] Documentation [ ] Other: _____ Closes Issue: NaN Description Adds Emote#isAvailable() method which tells if the emote is available or not. The Emoji structure has a available field that indicates if the emote is available for the users, depending on the current boost tier of the Guild. Will this also get updated once the emote becomes available again or becomes unavailable? Shouldn't there be an event for that or discord doesn't send it? I'm not sure about the details of this. The only thing I do know is, that this is part of the emoji/emote object that Discord sends. My guess is, that the emotes will update when the Boost level of the Discord changes either in positive or negative direction. Would it be reasonable to add a isAvailable() method to Emote related Events? Considering that on an EmoteAddEvent, this would always be true, since you can't add emotes when you lack the required boost level. For this PR to continue, we need to find out if the GUILD_EMOJIS_UPDATE event is fired when boost level changes such that some emojis are no longer available. From what I could do the docs not mention anything like that(?) Should I perhaps ask for clarification on the API-docs GitHub? You could ask in the docs repo, you could also test it out yourself. https://github.com/discord/discord-api-docs/issues/1725#issuecomment-640947828 Looks like to trigger an event, but won't be guaranteed. I currently think about how this could be added properly. My thought was about having an EmoteUpdateX event of some sort that would update its availability, but this would probably require a updateEmote method in the EntityBuilder for what I know. Since update events seem unreliable according to night, should this now be pretty much everything to add. I perhaps should add a notice that the isAvailable() boolean might not be updated due to the EmoteUbdateEvent being not reliable on this and that Discord might not send an event in the first place?
gharchive/pull-request
2020-05-21T13:28:43
2025-04-01T04:54:52.506015
{ "authors": [ "Andre601", "DV8FromTheWorld", "caneleex" ], "repo": "DV8FromTheWorld/JDA", "url": "https://github.com/DV8FromTheWorld/JDA/pull/1303", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
684996189
Fix inconsistency with MarkdownSanitizer Pull Request Etiquette [x] I have checked the PRs for upcoming features/bug fixes. [x] I have read the contributing guidelines. Changes [x] Internal code [ ] Library interface (affecting end-user code) [ ] Documentation [ ] Other: _____ Closes Issue: NaN Description A couple of weeks ago, I submitted a pull request that attempted to resolve the issue causing symbols to not be sanitized correctly inside of quotes. Unfortunately, I found that putting a symbol in front of a quote causes my previous fix to not work correctly. Example: > _Hello \n> _World ->\\> \\_Hello \n> \\_World This fixes the issue by changing the regex to allow for quotes that contain just spaces. While Discord removes these extra spaces itself, they are still able to generate quotes (a bit of a weird fix). BEFORE > ​ (As stated before, these are still valid quotes) AFTER \> ​ If there are any more issues related to this in the future, it might be better just to redo how quotes are handled entirely. The client doesn't render this as a quote: The WYSIWYG editor does this because you might enter more content to indicate that it will be in a quote.
gharchive/pull-request
2020-08-24T21:53:07
2025-04-01T04:54:52.512320
{ "authors": [ "MinnDevelopment", "Owen1212055" ], "repo": "DV8FromTheWorld/JDA", "url": "https://github.com/DV8FromTheWorld/JDA/pull/1379", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
577523007
Error Append services error: list index out of range an idea why the error occurs when connecting to the box Hello. Unfortunately, too little information. Have you edited the settings in other programs or through the web interface? To start with, show the output in the download dialog. Then it is desirable to get the following files from the 'data/enigma2/profile name' folder: lamedb(lamedb5 if exist) and bouquets.tv, bouquets.radio. Or (and) best the last zip archive from the 'data/enigma2/backup/profile name' name folder! For exact paths look in settings. At the moment I can only assume that the problem when reading lamedb or lamedb5 (has experimental support and has not been fully tested)! Hello. Unfortunately, too little information. Have you edited the settings in other programs or through the web interface? To start with, show the output in the download dialog. Then it is desirable to get the following files from the 'data/enigma2/profile name' folder: lamedb(lamedb5 if exist) and bouquets.tv, bouquets.radio. Or (and) best the last zip archive from the 'data/enigma2/backup/profile name' folder! For exact paths look in settings. At the moment I can only assume that the problem when reading lamedb or lamedb5 (has experimental support and has not been fully tested)! Hello, After using the editor about three times i got the same error. I can't 'load' my saved channellist, can't 'ftp' the channellist into the program. "list index out of range" Pity, it sure has potential. greetings tom Hi, Which version are you using? Try to run the program from the terminal as described above and show the output. Better yet, simply archive the current folder of that profile where the error occurs(or better the entire data folder) and send it to me. Best Hello, first of all, thank you for your quick response. Second, I'm sorry. My problem is solved. I had an IP address conflict in my network and one of my receivers acted strange. The error is gone now. Have a nice weekend. grtzz tom Same problem here. The network works fine. FTP-Transfer -> Downloading Bouquets (or anything else) -> downloading -> Done. and PopUp list index out of range. Clicking ok. The window closed -> No list Hi. First of all, install the latest version from here to the Application folder. Then we delete the previous program settings. We give the command in the terminal: rm -R .config/demon-editor/ Run app from also from the terminal: cd /Applications/DemonEditor.app/Contents/MacOS/ ./DemonEditor If app starts, then set up your profile for the receiver and try to download. If an error occurs, then you need to lay out the output from the terminal, as well as those files that you downloaded from the receiver, as described above. In addition, you can run the program with logging enabled (with the -l key, described above). The demon-editor.log should be created in the root of the folder. Hi. First of all, install the latest version from here to the Application folder. Then we delete the previous program settings. We give the command in the terminal: rm -R .config/demon-editor/ Run app from also from the terminal: cd /Applications/DemonEditor.app/Contents/MacOS/ ./DemonEditor If app starts, then set up your profile for the receiver and try to download. If an error occurs, then you need to lay out the output from the terminal, as well as those files that you downloaded from the receiver, as described above. In addition, you can run the program with logging enabled (with the -l key, described above). The demon-editor.log should be created in the root of the folder. Hi. First of all, install the latest version from here to the Application folder. Then we delete the previous program settings. We give the command in the terminal: rm -R .config/demon-editor/ Run app also from the terminal: cd /Applications/DemonEditor.app/Contents/MacOS/ ./DemonEditor If app starts, then set up your profile for the receiver and try to download. If an error occurs, then you need to lay out the output from the terminal, as well as those files that you downloaded from the receiver, as described above. In addition, you can run the program with logging enabled (with the -l key, described above). The demon-editor.log should be created in the root of the folder. Hi. First of all, install the latest version from here to the Application folder. Then we delete the previous program settings. We give the command in the terminal: rm -R .config/demon-editor/ Run app also from the terminal: cd /Applications/DemonEditor.app/Contents/MacOS/ ./DemonEditor If app starts, then set up your profile for the receiver and try to download. If an error occurs, then you need to lay out the output from the terminal, as well as those files that you downloaded from the receiver, as described above. In addition, you can run the program with logging enabled (with the -l key, described above). The demon-editor.log should be created in the root of the folder. Hi. First of all, install the latest version from here to the Application folder. Then we delete the previous program settings. We give the command in the terminal: rm -R .config/demon-editor/ Run app also from the terminal: cd /Applications/DemonEditor.app/Contents/MacOS/ ./DemonEditor If app starts, then set up your profile for the receiver and try to download. If an error occurs, then you need to lay out the output from the terminal, as well as those files that you downloaded from the receiver, as described above. In addition, you can run the program with logging enabled (with the -l key, described above). The demon-editor.log should be created in the root of the folder. Thank you so much for your support. It's totally crazy. I unzipped the zip file with another program. And now I have an app that opens immediately with a double click. Unbelievable Should I make preparations (Terminal) now or can I start the app right away? Is there a update feature for upcomming versions or must i install every time a new version I am not familiar with these editors and all the features of a box. I haven't dealt with that so far. So, i can now tryout what it can do for me. Best and thanks again Maik Am 7. Mai 2020, 17:28 +0200 schrieb Dmitriy Yefremov notifications@github.com: Hi. First of all, install the latest version from here to the Application folder. Then we delete the previous program settings. We give the command in the terminal: rm -R .config/demon-editor/ Run app also from the terminal: cd /Applications/DemonEditor.app/Contents/MacOS/ ./DemonEditor If app starts, then set up your profile for the receiver and try to download. If an error occurs, then you need to lay out the output from the terminal, as well as those files that you downloaded from the receiver, as described above. In addition, you can run the program with logging enabled (with the -l key, described above). The demon-editor.log should be created in the root of the folder. — You are receiving this because you commented. Reply to this email directly, view it on GitHub, or unsubscribe. It’s me again But the issue ist still there Cheers, Maik Am 7. Mai 2020, 17:28 +0200 schrieb Dmitriy Yefremov notifications@github.com: Hi. First of all, install the latest version from here to the Application folder. Then we delete the previous program settings. We give the command in the terminal: rm -R .config/demon-editor/ Run app also from the terminal: cd /Applications/DemonEditor.app/Contents/MacOS/ ./DemonEditor If app starts, then set up your profile for the receiver and try to download. If an error occurs, then you need to lay out the output from the terminal, as well as those files that you downloaded from the receiver, as described above. In addition, you can run the program with logging enabled (with the -l key, described above). The demon-editor.log should be created in the root of the folder. — You are receiving this because you commented. Reply to this email directly, view it on GitHub, or unsubscribe. Hi Since the program is in the process of active development, there is no possibility of automatic updating yet. In addition, if problems arise, you must have a certain level of knowledge, at least know what a terminal is. If this is all ok, then I can try to help you in this discussion. If not, then I will delete this discussion later, and you can write to me on the email indicated in my profile. And in the presence of free time I can try to help you. Hi Since the program is in the process of active development, there is no possibility (for macOS) of automatic updating yet. In addition, if problems arise, you must have a certain level of knowledge, at least know what a terminal is. If this is all ok, then I can try to help you in this discussion. If not, then I will delete this discussion later, and you can write to me on the email indicated in my profile. And in the presence of free time I can try to help you. Thanks a lot. Sounds good. Best Maik Am 7. Mai 2020, 19:26 +0200 schrieb Dmitriy Yefremov notifications@github.com: Hi Since the program is in the process of active development, there is no possibility (for macOS) of automatic updating yet. In addition, if problems arise, you must have a certain level of knowledge, at least know what a terminal is. If this is all ok, then I can try to help you in this discussion. If not, then I will delete this discussion later, and you can write to me on the email indicated in my profile. And in the presence of free time I can try to help you. — You are receiving this because you commented. Reply to this email directly, view it on GitHub, or unsubscribe. Previous conversation is cleared, due to weak intersection with discussion. If such an error occurs, please do the following (for macOS version): Install the latest version from here to the Application folder . Allow the launch of the program, if necessary, in the security settings! Then remove the previous program settings. Give the command in the terminal: rm -R .config/demon-editor/ Run app also from the terminal (macOS): cd /Applications/DemonEditor.app/Contents/MacOS/ ./DemonEditor For Linux: ./start.py in the root of app folder, or demon-editor if the program is installed from a *.deb package or PPA repository. After app starts, then set up your profile for the receiver and try to download. If an error occurs, then lay out the output from the terminal, as well as those files that you downloaded from the receiver, as described above. In addition, you can run the program with logging enabled (with the -l option, described above). The demon-editor.log should be created in the root of the folder. The following information is also desirable: The build date of the program (maсOS) or version (brаnch). The used enigma2 image. Have the settings been edited in other programs. Previous conversation is cleared, due to weak intersection with discussion. Previous conversation is cleared, due to weak intersection with discussion. Perhaps this fix will help to prevent this error. Thanks again. Don't stress, I'm in no hurry with this whole thing. It's not easy. With the terminal (new shell) in Catalina, it doesn't fit everything. Especially because I can't see in which directory I have to be right now to enter the commands Anyway. I have deleted corresponding data / folders in my user directory. Download from the receiver works. But then it becomes difficult. Closing the window not possible. No idea why. Then i start screenflow to capture (screencast) whats going on. Suddenly, the Demon Editor window is ready to close … Crazy Using: MacOS Catalina Version 10.15.4 (19E266) Enigma 2, Octagon SF4008, Image OpenATV 6.4 I have no editor because nothing works with Catalina Cheers Maik Am 8. Mai 2020, 14:16 +0200 schrieb Dmitriy Yefremov notifications@github.com: Previous conversation is cleared, due to weak intersection with discussion. If such an error occurs, please do the following (for macOS version): Install the latest version from here to the Application folder . Allow the launch of the program, if necessary, in the security settings! Then remove the previous program settings. Give the command in the terminal: rm -R .config/demon-editor/ Run app also from the terminal (macOS): cd /Applications/DemonEditor.app/Contents/MacOS/ ./DemonEditor For Linux: ./start.py in the root of app folder, or demon-editor if the program is installed from a *.deb package or PPA repository. After app starts, then set up your profile for the receiver and try to download. If an error occurs, then lay out the output from the terminal, as well as those files that you downloaded from the receiver, as described above. In addition, you can run the program with logging enabled (with the -l option, described above). The demon-editor.log should be created in the root of the folder. The following information is also desirable: The build date of the program (maсOS) or version (brаnch). The used enigma2 image. Have the settings been edited in other programs. — You are receiving this because you commented. Reply to this email directly, view it on GitHub, or unsubscribe. But then it becomes difficult. Closing the window not possible. No idea why. Then i start screenflow to capture (screencast) whats going on. Suddenly, the Demon Editor window is ready to close … Crazy Hi, Maik Please create, if you can, for this case, a separate issue. I will close it! Thanks! But then it becomes difficult. Closing the window not possible. No idea why. Then i start screenflow to capture (screencast) whats going on. Suddenly, the Demon Editor window is ready to close … Hi, Maik Please create, if you can, for this case, a separate issue. I will close it! Thanks! Hi, I returned the older build. Try with it. Best But then it becomes difficult. Closing the window not possible. Hi, Maik Please try with the new build. If it does not help, create a separate issue for this case. Thanks!
gharchive/issue
2020-03-08T16:00:39
2025-04-01T04:54:52.559519
{ "authors": [ "DYefremov", "MarkusH12", "mischugo", "tommiedepommie" ], "repo": "DYefremov/DemonEditor", "url": "https://github.com/DYefremov/DemonEditor/issues/12", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
395481114
indicator not working for any tab other than the first one if you run the example in an android device, you'll notice that the bottom indicator for the active tab will show for the initial tab but if you change tab the indicator will stop appearing. Hi @bakaa99 - so sorry I missed your issue. Can you provide any more details? We are using this successfully with the indicator moving along. Here is our basic implementation if it is helpful: // index state is used by the TabViewAnimated to determine the active tab _handleIndexChange = (index: number) => { this.props.updateLastActivity(); const tabState = { ...this.state.tabState, index }; this.setState({ tabState }, () => { this.props.onTabChange(this.state.tabState.routes[index].title); }); }; _renderLabelFactory = (props: TabScreenSceneRenderPropsType): TabBarCallbackType => ( ({ route }: TabScreenSceneType): Element<*> => { const index = props.navigationState.routes.indexOf(route); const inputRange = props.navigationState.routes.map((x: TabScreenRouteType, i: number): number => i); const outputRange = inputRange.map((inputIndex: number): string => (inputIndex === index ? Colors.lightBlue : Colors.dhsWhite)); const color = props.position.interpolate({ inputRange, outputRange }); return <Animated.Text style={[styles.label, { color }]}>{route.title}</Animated.Text>; } ); _renderIconFactory = (props: TabScreenSceneRenderPropsType): TabBarCallbackType => ( ({ route }: TabScreenSceneType): Element<*> => { const index = props.navigationState.routes.indexOf(route); const inputRange = props.navigationState.routes.map((x: TabScreenRouteType, i: number): number => i); const outputRange = inputRange.map((inputIndex: number): string => (inputIndex === index ? Colors.lightBlue : Colors.dhsWhite)); const color = props.position.interpolate({ inputRange, outputRange }); const AnimatedIcon = Animated.createAnimatedComponent(Icon); return <AnimatedIcon name={route.icon} size={30} style={[styles.icon, { color }]} />; } ); _renderTabs = (sceneProps: SceneRendererProps<TabScreenRouteType>): Element<*> => { const landscape = this.props.screenProps.orientation && this.props.screenProps.orientation.split('-')[0] === 'LANDSCAPE'; const SelectedTabBar = landscape ? TabBarVertical : TabBar; return ( <SelectedTabBar {...sceneProps} renderLabel={this._renderLabelFactory(sceneProps)} renderIcon={this._renderIconFactory(sceneProps)} style={styles.tabbar} tabStyle={styles.tab} indicatorStyle={styles.indicator} scrollEnabled /> ); }; render(): Element<*> { /* $FlowFixMe orientation is a maybe type - we should probably just require it */ const landscape = this.props.screenProps.orientation.split('-')[0] === 'LANDSCAPE'; const SelectedTabView = landscape ? TabViewVertical : TabView; const tabBarPosition = landscape ? 'top' : 'bottom'; // 'top' is same as 'left' in vertical tabs const initialLayout = { width: 600, height: 400 }; const sceneMap = SceneMap(this.state.scenes); return ( <SelectedTabView initialLayout={initialLayout} renderTabBar={this._renderTabs} style={styles.container} navigationState={this.state.tabState} renderScene={sceneMap} onIndexChange={this._handleIndexChange} swipeEnabled={Platform.OS !== 'web'} tabBarPosition={tabBarPosition} /> ); }
gharchive/issue
2019-01-03T07:41:50
2025-04-01T04:54:52.566903
{ "authors": [ "DaKaZ", "bakaa99" ], "repo": "DaKaZ/react-native-vertical-tab-view", "url": "https://github.com/DaKaZ/react-native-vertical-tab-view/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
545944425
Calling API I tried to call a basic API with a request via HttpClient. I am getting the following errors: blazor.webassembly.js:1 WASM: Unhandled exception rendering component: p.printErr @ blazor.webassembly.js:1 blazor.webassembly.js:1 WASM: System.NotImplementedException: Cannot invoke method because it was wiped. See stack trace for details. p.printErr @ blazor.webassembly.js:1 blazor.webassembly.js:1 WASM: at System.Net.Http.HttpClientHandler.Dispose (System.Boolean disposing) <0x3bf5ca0 + 0x00004> in :0 p.printErr @ blazor.webassembly.js:1 blazor.webassembly.js:1 WASM: at System.Net.Http.HttpMessageHandler.Dispose () <0x3bf5b50 + 0x00008> in :0 I am essentially calling this function under the Counter.razor sample file right after the increment button. I've tested the API with Postman so I know it isn't the API. public async Task GetDataLoaded() { RequestModel request = new RequestModel(); request.TargetAmount = 5000; //create post data var postData = JsonConvert.SerializeObject(request); using (HttpClientHandler handler = new HttpClientHandler()) { handler.AutomaticDecompression = DecompressionMethods.GZip | DecompressionMethods.Deflate; using (HttpClient client = new HttpClient(handler, false)) { //client.DefaultRequestHeaders.Add("<keyname>", "<value>"); client.DefaultRequestHeaders.Accept.Add(new MediaTypeWithQualityHeaderValue("application/json")); // Wrap our JSON inside a StringContent which then can be used by the HttpClient class var httpContent = new StringContent(postData, Encoding.UTF8, "application/json"); var responseFromServer = await client.PostAsync("<personal api URL>", httpContent).Result.Content.ReadAsStringAsync(); return responseFromServer; } } } I never encountered any kind of error of this type before, maybe something similar on Edge browser when using some Polyfills in the past. First of all: Is your sample working on a regular Blazor sample project ? Not the BlazorMobile default one. Is the Blazor version used on your BlazorMobile project similar to the one mentioned in the doc ? It may be important, as i don't have personnally updated the Blazor package since some times, and if there is some breaking changes on Microsoft side on the project Startup / Service registration, it must be done on the BlazorMobile template project too. Is your exception occuring on all Browser ? I never had any problem at the moment (and using it for a client), with BlazorMobile if you respect exactly the requirements mentionned on the documentation. @Daddoon Good suggestion on trying it on a regular Blazor project. I tried it on the regular template in VS2019 and got the same errors. It must be something about using HttpClient and HttpClientHandler. Yes it is all client side. I am running Blazor WASM. Seems like it is related to Blazor WASM and something with CORS. I tried enabling everything I could on the web service end but still getting the same error. For your template you're essentially taking the Blazor WASM and putting it into the Android/iOS container right? I think you are using the HttpClient API wrongly with Blazor. As stated in the doc, you must inject the HttpClient object from your page / dependency services, not construct a custom one by yourself as there is some specific implementation made with BCL, depending your are using the server or WASM version. Actually the WASM HttpClient for Blazor is a special implementation of the library, hidden by the BCL process. Take note of the documentation: https://docs.microsoft.com/en-us/aspnet/core/blazor/call-web-api?view=aspnetcore-3.1 In my opinion your problem should go away if instanciated from an injection on a page or BlazorComponent. Note also that you only need to get the reference of this object once in lifecycle of you application, so you may cache it on a internal helper of your, to call it everywhere when needed. Otherwise, the CORS issue is variable depending the browser security default parameters. Note also that if you are using HTTPS, you may have some blocking issue on Edge / UWP if the certificate is not really certified (like auto signed or expired). I didn't mention before but I tried it with DI before and I am getting the same results. CORS is fully enabled in the web service deployment and on Azure app service. Below is the same function with the HttpClient injected public async Task GetDataLoaded() { RequestModel request = new RequestModel(); request.TargetAmount = 5000; //create post data var postData = JsonConvert.SerializeObject(request); //client.DefaultRequestHeaders.Add("<keyname>", "<value>"); Http.DefaultRequestHeaders.Accept.Add(new MediaTypeWithQualityHeaderValue("application/json")); // Wrap our JSON inside a StringContent which then can be used by the HttpClient class var httpContent = new StringContent(postData, Encoding.UTF8, "application/json"); var responseFromServer = await (await Http.PostAsync("<api url>", httpContent)).Content.ReadAsStringAsync(); return responseFromServer; } Sorry, but as the problem is occuring on Blazor too, i think it's a Blazor issue, not a BlazorMobile one. Try with a simple Http request to determine what is the issue, maybe it's a Blazor / Mono WASM issue. Feel free to ask help on the official Blazor issues feed here: https://github.com/dotnet/aspnetcore/issues Closing.
gharchive/issue
2020-01-06T21:17:55
2025-04-01T04:54:52.585170
{ "authors": [ "Daddoon", "ajhena" ], "repo": "Daddoon/BlazorMobile", "url": "https://github.com/Daddoon/BlazorMobile/issues/153", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2633971906
Wrong realtime lighting for vertex-lit surfaces With vertex lighting, the wrong diffuse color is used for realtime lights. Instead of the basic diffuse texture, the result of multiplying the diffuse with the vertex color is used. This comes from the line diffuse *= var_Color; in lightMapping_fp.glsl. So dynamic lights are practically impossible to see on dimly lit vertex-lit surfaces such as parts of the station15 evergreen trees. I guess it's structured this way because vertex lighting is a type of "rgbGen", at least in explicitly specified vertex lighting shaders where the syntax is rgbGen vertex. In principle, any rgbGen operation should be included in the diffuse. But obviously this is the wrong idea for vertex lighting, so we should try to detect the vertex lighting case and exclude that calculation from the diffuse used for lighting purposes. I guess the criteria for detecting vertex lighting can be rgbgen vertex + is bsp surface + no lightmap, in a lightmap stage. Isn't vertex lighting supposed to be part of diffuse? If not, how is it expected to contribute to the final colour? Diffuse color multiplied by vertex color. Vertex color is a light color serving the same function as lightmaps.
gharchive/issue
2024-11-04T22:20:27
2025-04-01T04:54:52.588021
{ "authors": [ "VReaperV", "slipher" ], "repo": "DaemonEngine/Daemon", "url": "https://github.com/DaemonEngine/Daemon/issues/1415", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
936348550
When setting in_nograb to 1, input is not grabbed but pointer is hidden when focusing When setting in_nograb to 1, input is not grabbed but pointer is hidden, maybe that's unwanted? Also, the pointer move still affects the camera. Enabling in_nograb may also be useful when doing benchmark: in that specific use case, not hiding pointers and not being used to move the camera may be wanted. When setting in_nograb to 1, input is not grabbed but pointer is hidden, maybe that's unwanted? After re-reading myself I had to reproduce to understand why I've said… I mean the desktop pointer is hidden like if it was captured. Also, the pointer move still affects the camera. The only thing in_nograb seems to do is to not recenter the pointer. I don't know if there is a bug, but the behaviour is a bit puzzling. With in_nograb enabled when the game is on the main menu, the main menu displays the in-game mouse pointer, so there is a continuity, when entering the window the desktop pointer disappears and the in game pointer appear, when leaving the window the in game pointer disappears and the desktop pointer reappears. But outside of a menu, in a loaded game level, when entering the window the mouse pointer disappears and will only reappear when leaving the windows again. I assume this was fixed by 4e9095477beab0a310958b6513a9a16ef3b0bd4c.
gharchive/issue
2021-07-04T00:20:15
2025-04-01T04:54:52.591436
{ "authors": [ "illwieckz", "slipher" ], "repo": "DaemonEngine/Daemon", "url": "https://github.com/DaemonEngine/Daemon/issues/496", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
655040384
Client: Improvements for false-positives handling This should be implemented to complete #238 #241 When reading reports/false-positives list from server: sort items by severity,ID In markFalsePositives/unmarkFalsePositives sum up the changes and add confirmation "Are you really sure?" Documentation of new features (use proper options - new: unmarkFalsePositves/markFalsePositves/interactiveMarkFalsePositves/interactiveUnmarkFalsePositves) increase test coverage of Go tests Sven Dolderer sven.dolderer@daimler.com, Daimler TSS GmbH, imprint Standard false positive commands are already documented in #323 Will close this issue, because: When reading reports/false-positives list from server: sort items by severity,ID -> currently they come already sorted - open a new issued once they come unsorted from the server. In markFalsePositives/unmarkFalsePositives sum up the changes and add confirmation "Are you really sure?" -> Postponing, because this is not essential (open a new issue if needed) Documentation of new features (use proper options - new: unmarkFalsePositves/markFalsePositves/interactiveMarkFalsePositves/interactiveUnmarkFalsePositves) -> already done: see https://daimler.github.io/sechub/latest/sechub-client.html increase test coverage of Go tests -> already done along with the other changes
gharchive/issue
2020-07-10T21:15:30
2025-04-01T04:54:52.603237
{ "authors": [ "de-jcup", "sven-dmlr" ], "repo": "Daimler/sechub", "url": "https://github.com/Daimler/sechub/issues/285", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1000833358
PDS Cluster Monitoring Status Exception Problem While running a cluster of 3 PDS members and requesting the status of the cluster, exceptions occur on the PDS members: com.daimler.sechub.pds.PDSJSONConverterException: Was not able to convert JSON string to class com.daimler.sechub.pds.monitoring.PDSClusterMember object … Caused by: com.fasterxml.jackson.databind.exc.UnrecognizedPropertyException: Unrecognized field "uuid" (class com.daimler.sechub.pds.job.PDSJob), not marked as ignorable (8 known properties: "started", "owner", "created", "ended", "result", "serverId", "state", "jsonConfiguration"]) at [Source: (byte[])"{"hostname":"708ac9247372","ip":"10.2.5.3","port":8444,"heartBeatTimestamp":"2021-09-20T09:24:24.719885","executionState":{"queueMax":10,"jobsInQueue":1,"entries":[{"jobUUID":"4e1c67d8-21fc-4066-a39b-a8ac245b0d82","done":false,"canceled":false,"job":{"owner":"admin","serverId":"PRODUCT_CLUSTER","created":"2021-09-20T09:17:32.188512","started":"2021-09-20T09:22:07.725691","jsonConfiguration":"{\"sechubJobUUID\":\"b6196fb6-62e8-44f8-99f2-4a74daf2090d\",\"apiVersion\":\"1.0\",\"productId\":\"PDS_P"[truncated 98 bytes]; line: 1, column: 557] (through reference chain: com.daimler.sechub.pds.monitoring.PDSClusterMember["executionState"]->com.daimler.sechub.pds.execution.PDSExecutionStatus["entries"]->java.util.ArrayList[0]->com.daimler.sechub.pds.execution.PDSExecutionJobInQueueStatusEntry["job"]->com.daimler.sechub.pds.job.PDSJob["uuid"]) At first no exception occurs while requesting the monitoring_status using sechub-developertools/scripts/pds-api.sh monitoring_status and the cluster has three members: { "jobs": { "CREATED": 3, "READY_TO_START": 0, "QUEUED": 0, "RUNNING": 1, "CANCEL_REQUESTED": 0, "CANCELED": 0, "FAILED": 0, "DONE": 8 }, "members": [ { "hostname": "708ac9247372", "ip": "10.2.5.3", "port": 8444, "heartBeatTimestamp": "2021-09-20T09:17:24.449131", "executionState": { "queueMax": 10, "jobsInQueue": 0, "entries": [] } }, { "hostname": "708ac9247372", "ip": "10.2.5.3", "port": 8444, "heartBeatTimestamp": "2021-09-20T08:27:29.945557", "executionState": { "queueMax": 10, "jobsInQueue": 0, "entries": [] } }, { "hostname": "6b200213fce3", "ip": "10.2.5.5", "port": 8444, "heartBeatTimestamp": "2021-09-20T09:17:24.441473", "executionState": { "queueMax": 10, "jobsInQueue": 0, "entries": [] } } ] } After a couple more requests of the monitoring_status, exceptions in the PDS cluster members are thrown. For example, first in member 2, afterwards member 3 and finally in member 1. Once exceptions occurred in all members, only one member will be shown in the monitoring status: { "jobs": { "CREATED": 1, "READY_TO_START": 0, "QUEUED": 0, "RUNNING": 3, "CANCEL_REQUESTED": 0, "CANCELED": 0, "FAILED": 0, "DONE": 12 }, "members": [ { "hostname": "708ac9247372", "ip": "10.2.5.3", "port": 8444, "heartBeatTimestamp": "2021-09-20T08:27:29.945557", "executionState": { "queueMax": 10, "jobsInQueue": 0, "entries": [] } } ] } This is the same as #804 - with some additional information. The problem seems to be that there comes no more an heartbeat when a PDS Job was inside the queue and the deserialization problem from #804 happens... So even when #804 happens the heartbeat must still work!
gharchive/issue
2021-09-20T11:01:26
2025-04-01T04:54:52.608464
{ "authors": [ "Jeeppler", "de-jcup" ], "repo": "Daimler/sechub", "url": "https://github.com/Daimler/sechub/issues/798", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
285232021
Use in Nuxt.js How can I use your plugin in nuxt.js? I tried the following but it doesn't recognize this.popmotion.tween. mounted() { if (process.browser) { require('~/plugins/popmotion') const container = this.popmotion.styler(document.querySelector('section')); const formElements = document.querySelector('.animatedimage'); const scaleanim = this.popmotion.tween({ from: { scale: .7 }, to: { scale: 1 }, duration: 1000 }) scaleanim.start(container.set); } }, @digisocialnet i got it working via import * as popmotion from 'popmotion' i can see popmotion is attached but not tested more :D
gharchive/issue
2017-12-31T00:24:31
2025-04-01T04:54:52.610096
{ "authors": [ "cannap", "digisocialnet" ], "repo": "DakshMiglani/vue-popmotion", "url": "https://github.com/DakshMiglani/vue-popmotion/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1168931575
Turn ON/OFF Device in Rainbow mode Plug-in does a great job controlling devices that are in static mode (on/off/change colors), but I have no luck with the ones set in rainbow mode. When powering on or off it does nothing but a quick blink. I don't know if this is a limitation of openrgb or this plug-in. Can it be implemented? Thanks for this useful plug-in! I see what you mean. Thanks for the report. After some quick tests of possible solutions, I think, yes, I should be able to make this better. I'll work on it. Thank you very much!! Hello Sir, any news on this matter? Hey, I did some of the work needed for this update a while back, but I've haven't been able to get back to it and finish it yet. I haven't forgotten about this, and I'll finish it when I get the chance. Just haven't been able to prioritize this lately. I'm still busy all this month, but I should be able to revisit this in August. Hi there. I've just published version 1.4.0 with these changes implemented. If you could test it out and let me know if everything works well for you, that would be great. thanks for taking the time to look at this! I updated to 1.4 and tested it with mixed results. 2 of my devices set to rainbow mode seems to work correctly as expected when turning off and on: Adata d60g ram sticks Adata s40g ssd 2 other of my devices set to rainbow mode do weird things when turning off and on: Rog strix LC cooler doesn't respond to off when in rainbow mode Asus Motherboard lights and zones do respond to off command, but when turning on there are set to red color by default and not rainbow mode Hmm, that's going to be tough to debug without having those devices. Could you list all of the available modes that OpenRGB has as options for those devices? Asus Rog Strix LC modes: I just realized that I don't use rainbow mode for the Asus motherboard, I use direct mode. Changing to rainbow mode and tested on/off and it does work correctly. However the reset to red is happening when in direct mode and turning off and on. Asus Motherboard modes: I'm also wondering if the motherboard's zones can be displayed as separate accessories to control individual colors of each zone? Asus Motherboard zones: Do you know if you had the issue of the motherboard lights turning red before you updated the plugin to version 1.4.0? Or is that new behavior? I had the issue before 1.4.0
gharchive/issue
2022-03-14T21:31:09
2025-04-01T04:54:52.620151
{ "authors": [ "DallasHoff", "arturoovf" ], "repo": "DallasHoff/homebridge-openrgb", "url": "https://github.com/DallasHoff/homebridge-openrgb/issues/5", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2263418839
Create new label for gen-express-app package! Since the project has its regular users and there is an opportunity to extend this package to more users, why don't we rename the package “egen”? This will help users remember the package better and increase the user audience. What do you think about this idea? Nice, but the package name "egen" has already been used (npm). 😢 How about "egn"? This name tag is free and so easy to remember!)
gharchive/issue
2024-04-25T12:06:24
2025-04-01T04:54:52.621697
{ "authors": [ "Dalufishe", "the-az-dev" ], "repo": "Dalufishe/gen-express-app", "url": "https://github.com/Dalufishe/gen-express-app/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2326738888
.WithParameterValidation() does not work with [FromRoute] parameters //Should fail when id does not equal "hello" or "world". Expect 400 response builder.MapGet("/test/{id}", async ([FromRoute, RegularExpression("^(hello|world)$")] string id) => { //Currently returns "hello world" regardless of validation return Results.Ok("hello world"); }).WithParameterValidation(); Validating parameters that are simple types is not supported currently. This is similar to how ModelValidation in ASP.NET Core MVC works. You need to accept parameters that are complex types whose properties have the validation attributes on them. You can combine a complex type with AsParameters to achieve what you want: // Will fail when id does not equal "hello" or "world". Expect 400 response builder.MapGet("/test/{id}", async ([AsParameters] TestReuqest request) => { // Will only return if properties of `request` are valid return Results.Ok("hello world"); }).WithParameterValidation(); public class TestRequest { [FromRoute, RegularExpression("^(hello|world)$")] public string? Id { get; set; } }
gharchive/issue
2024-05-31T00:05:25
2025-04-01T04:54:52.624888
{ "authors": [ "DamianEdwards", "JCKortlang" ], "repo": "DamianEdwards/MinimalApis.Extensions", "url": "https://github.com/DamianEdwards/MinimalApis.Extensions/issues/46", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1094618685
request.path has been replaced by request.url.pathname I'm getting this error when calling the signIn() function using the latest version of SvelteKit. Same, related to https://stackoverflow.com/questions/70534899/sveltekit-error-request-query-has-been-replaced-by-request-url-searchparams @krg-digital looks like the author is not very active, have you found a workaround yet? Same, related to https://stackoverflow.com/questions/70534899/sveltekit-error-request-query-has-been-replaced-by-request-url-searchparams Yeah, stumbled upon that thread when investigating the error. I'm definitely no expert when it comes to the inner-workings of SvelteKit, but I'd guess these lines of code is where the issues lies... and that TODO might hint at what needs to be done? Hey everyone! Sorry for the inactivity, SK-Auth has now been upgraded to v0.4.0 thanks to #58 by @lulzneko which should resolve these issues.
gharchive/issue
2022-01-05T18:11:29
2025-04-01T04:54:52.632785
{ "authors": [ "Dan6erbond", "Kelvijn", "krg-digital" ], "repo": "Dan6erbond/sk-auth", "url": "https://github.com/Dan6erbond/sk-auth/issues/59", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2754864078
bug: default is not shown unless actual default value is truthy # cli.py from dataclasses import dataclass, field from typing import Annotated from typing_extensions import Doc import cappa @cappa.command(name="hello") @dataclass(kw_only=True) class Command: config: Annotated[ dict, cappa.Arg( short="-c", long=True, show_default="hello", ), Doc("Configuration file"), ] = field(default_factory=dict) def __call__(self): print(self.config) if __name__ == "__main__": cappa.invoke(Command) $ python cli.py Usage: hello [-c CONFIG] [-h] [--completion COMPLETION] Options [-c, --config CONFIG] Configuration file Help [-h, --help] Show this message and exit. [--completion COMPLETION] Use --completion generate to print shell-specific completion source. Valid options: generate, complete. If I change the annotation to dict | None and the default value to None, same result. I have to actually set a truthy default value like "hey" for the default to be shown. Expected output $ python cli.py Usage: hello [-c CONFIG] [-h] [--completion COMPLETION] Options [-c, --config CONFIG] Configuration file (Default: hello) Help [-h, --help] Show this message and exit. [--completion COMPLETION] Use --completion generate to print shell-specific completion source. Valid options: generate, complete. It's because default.default is None when I set None as default, and Empty when I set a default factory, in if not self.show or default.default in (None, Empty): (DefaultFormatter.format_default()). I can see why it's not immediate to fix: if a string is provided as show_default, but it contains the {default} placeholder, then yeah we might not want to show None. But if it's a string without the {default} placeholder, I think we can assume it should always be shown? So something like if "{default}" in self.format: ...? I wonder if we shouldn't assume that if the user provided something to show_default, it's that the default value is not none or empty, and therefore we should always show the default. So, using a sentinel for DefaultFormatter.format instead of "{default}". WDYT? Or maybe field factories could be seen as something else than Empty :shrug: That wouldn't solve the case for None though. log_path: An[ str | None, cappa.Arg(short="-P", long=True, group=_GROUP_GLOBAL, propagate=True, show_default="standard error"), Doc("Write log messages to this file path."), ] = None Lmk what you think of that ^. Not using any of your suggested alternatives 😬 tl;dr coerce None/Empty default values to "" format them if the resultant formatted string is "", dont show the default because it would be (Default: ) Where item 3 is potentially fixes other cases in which an empty string formatted default would look broken. I havent much thought about what if you want '' to be clear about empty string, but that wouldn't have displayed as such today either, and i'm not certain whether repr anything is a good idea or not. Yes!! Clever! Seems perfect 🙂
gharchive/issue
2024-12-22T22:02:06
2025-04-01T04:54:52.642924
{ "authors": [ "DanCardin", "pawamoy" ], "repo": "DanCardin/cappa", "url": "https://github.com/DanCardin/cappa/issues/190", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1088787609
Create python-package-conda.yml trying to set up CI testing in GitHub Well, this doesn't seem to work. Not sure what the deal is. Needed to have conda install scipy and numpy and mpl. I think that one is supposed to use an "environment.yml" file to list these requirements, but I wasn't sure where this file was supposed to be located, so I just included these in the workflows file. Success! Basic tests are now being run.
gharchive/pull-request
2021-12-26T14:55:21
2025-04-01T04:54:52.645026
{ "authors": [ "DanHickstein" ], "repo": "DanHickstein/pyNLSE", "url": "https://github.com/DanHickstein/pyNLSE/pull/29", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
315019898
Webassembly 3.4.1 version on npm doesn't randomize weights Output of npm list: -- jsnet@3.4.1 Weights in the network [2, 3, 1] (XOR example) at second layer stay the same: [ { bias: 1, weights: [ -0.4384047377812256, 0.71476953983941 ] }, { bias: 1, weights: [ 0.836493930022522, 0.7845843352178457 ] }, { bias: 1, weights: [ -0.8790343793408859, -1.049725288141738 ] } ] This produces the same results in 5 (or any number of) epochs training. Interestingly, the javascript version does not have this problem. Sorry for the delayed reply, I just tried this out, and I wasn't actually able to replicate it. What I did was npm installed v3.4.1, navigated to node_modules/jsnet/nodejsDemo.js, and ran it, printing out the weights in layer [1] in a callback as console.log(net.layers[1].neurons.map(n => n.weights)). Is this the same as what you did? If this is still an issue, and you made any changes, could you perhaps share what they were? If there were no other changes, then this is somewhat strange, and will need further exploring. Did the inference look plausible, at the end of the training, for you? I'm using Windows, also, so that's probably not it. And both versions are loaded, and executed after each other, to show speed comparison, but are not otherwise interlinked. Are the values at the end correct-ish for XOR? For example: If they are, then the network trained, so the values should have changed, but if not, then, as though learning rate was 0, the weight values for some reason don't get updated. Are you using node at least v8?
gharchive/issue
2018-04-17T11:34:31
2025-04-01T04:54:52.650856
{ "authors": [ "DanRuta", "Mitiko" ], "repo": "DanRuta/jsNet", "url": "https://github.com/DanRuta/jsNet/issues/56", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
465028832
listen to only message thx for greate machine. i really appreciate. but i have one issue. ex) /// @listen_to('') def to_duty(message): print(message) .... i want only listen_to NEW MESSAGE, NOT THREAD. how can i solve this problem? You can't do it via the api, but you can look for thread_ts in the internal message event object like so: if 'thread_ts' in msg._msg_event; # Do your stuff here pass I have created a pull request #286 with this feature.
gharchive/issue
2019-07-08T02:39:21
2025-04-01T04:54:52.658275
{ "authors": [ "davidolrik", "hyowoo" ], "repo": "DandyDev/slack-machine", "url": "https://github.com/DandyDev/slack-machine/issues/214", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }