id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
930683236
Sample Execution Error Hello. I am trying to execute the first example code of AI-Feynman, following its GitHub repository procedure. I have done $ pip install aifeynman, and executed the below codes at the REPL. import aifeynman aifeynman.get_demos("example_data") # Download examples from server aifeynman.run_aifeynman("./example_data/", "example1.txt", 60, "14ops.txt", polyfit_deg=3, NN_epochs=500) The code has been started, but it output an error message "STOP DEATH ERROR: missing file args.dat" with the following stdout statements. Checking for brute force + Trying to solve mysteries with brute force... Trying to solve ./example_data/example1.txt_train STOP DEATH ERROR: missing file args.dat Checking for brute force * Trying to solve mysteries with brute force... Trying to solve ./example_data/example1.txt_train STOP DEATH ERROR: missing file args.dat Checking polyfit Pareto frontier in the current branch: Complexity # MDL Loss # Expression 820.58 23.51 -3.97366916299281e-9*x0**3 - 4.98792375183732e-6*x0**2*x1 - 2.34615280130395e-5*x0**2*x2 + 0.000207395709218903*x0**2 - 0.000127172549695894*x0*x1**2 - 0.0297271185372981*x0*x1*x2 + 0.253671372975734*x0*x1 + 0.00425288419012183*x0*x2**2 - 0.0277197179641274*x0*x2 + 0.932555887619613*x0 - 0.000233584852340369*x1**3 - 0.0590672539598628*x1**2*x2 + 0.504778926577637*x1**2 + 0.0253291848989736*x1*x2**2 - 0.131143431894439*x1*x2 - 0.651859763950773*x1 - 0.00362777965859679*x2**3 + 0.0321967297940855*x2**2 - 0.0913890606187051*x2 + 0.614004068741101 Checking for brute force + Trying to solve mysteries with brute force... Trying to solve results/mystery_world_squared/example1.txt_train STOP DEATH ERROR: missing file args.dat Checking for brute force * (These lines above are repeated...) My environment was constructed on an AWS VM based on the Deep Learning AMI of Ubuntu 18.04, and its pre-installed conda environment with py36 & torch was activated. I would appreciate it if you could tell me how to solve this problem. Thank you very much. Are you sure it is not a problem of rights ? maybe open up the file permissions with chmod (see §4 of https://towardsdatascience.com/ai-feynman-2-0-learning-regression-equations-from-data-3232151bd929 )
gharchive/issue
2021-06-26T12:13:27
2025-04-01T04:55:35.807325
{ "authors": [ "koskr", "sansho-asakura" ], "repo": "SJ001/AI-Feynman", "url": "https://github.com/SJ001/AI-Feynman/issues/47", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
949448051
Zoom in/out buttons are not working The buttons are not working and neither is executing the command "MenuId.Zoom.DetailMinus" nor "MenuId.Zoom.DetailPlus" working. The issue was fixed in 25ee3a38e20fc54dd94f113f139e453093b1331a
gharchive/issue
2021-07-21T08:14:52
2025-04-01T04:55:35.954608
{ "authors": [ "dsn27" ], "repo": "SOFAgh/CADability", "url": "https://github.com/SOFAgh/CADability/issues/40", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2740280453
🛑 VEDI TEST is down In d96d32d, VEDI TEST (https://vedi.test.cordoba.gob.ar/) was down: HTTP code: 503 Response time: 519 ms Resolved: VEDI TEST is back up in 09aacc8 after 26 minutes.
gharchive/issue
2024-12-15T04:57:20
2025-04-01T04:55:35.957059
{ "authors": [ "SOPORTEAYIGROUP" ], "repo": "SOPORTEAYIGROUP/STATUSVEDI", "url": "https://github.com/SOPORTEAYIGROUP/STATUSVEDI/issues/538", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2286787545
feat: change verified badge color, position and added tooltip Mudanças no Verified Badge Mudança de cor de vermelho pra azul Mudança de posição do início para o final do nome do abrigo Adicionado a funcionalidade de tooltip ao passar com o mouse ou clicar em cima do ícone de verificado Criado um componente WithTooltip para encapsular componentes dentro de tooltips de maneira reutilizável Criado um componente VerifiedBadge com o ícone de verificado e com a estilização específica de cor azul Antes: Agora: Satisfaz os requisitos. Parabéns pelo trabalho.
gharchive/pull-request
2024-05-09T02:45:24
2025-04-01T04:55:35.960014
{ "authors": [ "filipepacheco", "rhuam" ], "repo": "SOS-RS/frontend", "url": "https://github.com/SOS-RS/frontend/pull/16", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1505178726
Nom efunsio na el salito cuamdo le doi a le'spasio nose ue que pas a pero cuand o le doi al epsalsio no salta parese n un yanderedev!! de sos que asen los juegos asiu de mal, es desir no fufa no ase su fnisopn, de exo cua,do camino con las patas me trpieso i ago fiuuuuuuummmm i ja no camina mals parese nikocado abogado pues igaul. Creo que este pryecyto se merese un 3/10 cofdigo de creador de fortnite: TheWillysex. grasisa pro ver esta rresenya-. 🧱🦝💫🍙🫕🐸😱🧱🐐😍🌳📣🍙💫🦝 creo que ste senior tiene rason porke le uelen las patas a chetos hmmmmm chetos pandillaaaaaa 😍😍😍😍😍
gharchive/issue
2022-12-20T20:04:17
2025-04-01T04:55:35.982614
{ "authors": [ "BinarySandia04", "Sussysep" ], "repo": "SRRafael8/WormsGame", "url": "https://github.com/SRRafael8/WormsGame/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
786976016
Added AWS temporary access token example. Added AWS temporary access token example. Requires mapping AWS roles to PrivX roles to PrivX admin. See PrivX documentation. Added some error handling. Invalid, merged with another request.
gharchive/pull-request
2021-01-15T15:22:03
2025-04-01T04:55:35.984712
{ "authors": [ "jjheinon" ], "repo": "SSHcom/privx-sdk-for-python", "url": "https://github.com/SSHcom/privx-sdk-for-python/pull/13", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1989723756
Assigned robot trajectories sometimes don't make sense Sometimes our robots are being assigned (or are following) seemingly nonsense trajectories. It's unclear whether this is due to a bug in the path planner, robot assignment, or simply the UI being wrong. Working on tests for the path planner to see if I can find out what problems we have on this branch: https://github.com/SSL-A-Team/software/tree/dev/christian/path_tests
gharchive/issue
2023-11-13T01:32:04
2025-04-01T04:55:35.986327
{ "authors": [ "fourpenny" ], "repo": "SSL-A-Team/software", "url": "https://github.com/SSL-A-Team/software/issues/176", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
253119214
quotable-printable error Feature request or bug Bug, but not sure if code bug or user bug If a bug, what did you expect to happen? Perfection ? If a bug, what happened? If a bug, list steps to reproduce bugs. If a bug, did you do these steps? [x] Download and use the lastest stable version [x] See if the issue has already been reported [x] Debug If a feature request, what do you want to be added or changed? If a feature request, is this feature already in a pull request? If a feature request, do you know anyone who can help? Side notes(Read then del this chunk) I am reading a message that contains a url with = signs in it, and I'm getting a quotable printable error from it. I tried other text lines, and anything in the message that contains = followed by \S causes the same problem. Messages coming from Apple Mail.app set to plain text. http://amdflames.org/scripts/news_article_c.php?id=862&s=1PQrV13145 What I get: http://amdflames.org/scripts/news_article_c.php?id�2&s=1QrV13145 And Try setting c=abc+def comes out Try setting c�c+def Also, from PHP: Notice: Unknown: Invalid quoted-printable sequence: =PQrV13145 Not sure how to resolve this. The message array shows encoding=4, but not sure if that's the problem. [text] => SSilence\ImapClient\SubtypeBody Object ( [charset] => us-ascii [_structure:SSilence\ImapClient\Section:private] => stdClass Object ( [type] => 0 [encoding] => 4 [ifsubtype] => 1 [subtype] => PLAIN [ifdescription] => 0 [ifid] => 0 [lines] => 45 [bytes] => 1362 [ifdisposition] => 0 [ifdparameters] => 0 [ifparameters] => 1 [parameters] => Array ( [0] => stdClass Object ( [attribute] => charset [value] => us-ascii ) ) ) The raw message does in fact show encoding is quoted printable. --Apple-Mail=_6A494713-DB75-4D0E-9ED7-9A45342F18F6 Content-Transfer-Encoding: quoted-printable Content-Type: text/plain; charset=us-ascii It seems another web service has the same issue: https://github.com/mailwatch/MailWatch/issues/702 Also may have to do with apple mail... They arent to nice :) Experimented some more... attached some code, output and email source code, if it helps. qp problem.txt any ETA on the fix for this? thanks! I get the same error: quotable-printable error The script I use in PHP 7.2.3 is: $emails = $imap->getMessagesByCriteria('SINCE "2018-01-01"'); foreach($emails as $email) { echo $email->header->subject; echo '<br>'; } ``
gharchive/issue
2017-08-26T19:59:53
2025-04-01T04:55:35.999963
{ "authors": [ "gr0g", "halojoy", "jorygun", "xXXIMMATTHEWXXx" ], "repo": "SSilence/php-imap-client", "url": "https://github.com/SSilence/php-imap-client/issues/189", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
407581863
Generate PDF again from documention (with Sphinx) There was a request to have PDF documentation for printing. I'll try to get to this soon. (I've also closed the previous PDF issue: #332.) Much appreciated for those of us who read from paper, not flickering computer screens. :+1: I've also closed the previous PDF issue: #332. Must have been the oldest still open ticket we've had ... @grantrostig I've managed to build a PDF of our documentation, albeit with quite a few warnings (and ignored errors). The output does look reasonable to me though. Does this look useful to you? I'd like to avoid going down the rabbit hole of trying to fix latex build errors unless needed. I'm afraid I'd never come out of it again. If this looks useful I'd enable it but without any guarantees that it will work in the future. I might also enable building single-page html documentation, and a man page (would anyone use that?). They would essentially come for free. @msimberg that looks very nice! single-page html would be great, but I don't think man pages are of any value nowadays. @msimberg , I will print and review the content (part of it) in the next few days. Thank you. @hkaiser , @msimberg , I definitely want the man pages, both in print on paper sometimes, and via the command line always. Man pages are the only complete and correct (if not too terse) documentation in a world of failed attempts at alternate documentation systems on the command line. Man pages has been serving Unix programmers for 20+ years, may they live forever (as a last resort when at the shell prompt. :) @grantrostig thanks for having a look. Note that the manpages would just be exactly the same content as in the html or pdf documentation. Probably too verbose if anything. @msimberg , well the PDF document is very nice and I have found no "formatting" problems at all yet. But it is not a set of "man pages". The important "boost" c++ library for instance does not offer man pages for programming libraries. But the gcc compiler does for its "executable commands", such as the compile "gcc" or "g++". Perhaps that would be a standard to eventually implement for binaries such as "hpxcc"? Not a high priority though. @msimberg , well the PDF document is very nice and I have found no "formatting" problems at all yet. But it is not a set of "man pages". No objections there. It's very much a least effort manpage. The important "boost" c++ library for instance does not offer man pages for programming libraries. But the gcc compiler does for its "executable commands", such as the compile "gcc" or "g++". Perhaps that would be a standard to eventually implement and create man pages for file formats (if any) and for binaries such as "hpxcc"? Not a high priority though in my opinion. Yes, that would definitely be nice, but you know how it is with time and priorities. I think this is low priority also for us.
gharchive/issue
2019-02-07T08:10:20
2025-04-01T04:55:36.021201
{ "authors": [ "grantrostig", "hkaiser", "msimberg" ], "repo": "STEllAR-GROUP/hpx", "url": "https://github.com/STEllAR-GROUP/hpx/issues/3670", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
851428511
Small fixes and improvements to CUDA/MPI polling Return a status enum from the polling functions to indicate if the scheduler should call the idle callback (useful if polling is the only thing done on a separate pool, and the pool would otherwise call the idle callback even if there's polling to be done). The CUDA polling function used to spin waiting for events to complete. As far as I can tell this was unintentional (@biddisco? the iterator was not incremented on cudaErrorNotReady). Avoid quadratic worst-case complexity in the CUDA polling function in terms of the number of active futures (if every future is ready every iteration would shift the remaining elements to the front when calling vector::erase). Use CUDA host pinned memory in the cuda_future test to actually enable asynchronous data transfers. Could you explain the rationale for these changes, please? Specifically, I would like to understand this: https://github.com/STEllAR-GROUP/hpx/pull/5277/files#diff-93e12b963ad29614127c32b33e22c54428ac32e8c435de5d9be363b284722c62R980-R984 Is it the particular implementation, the motivation, or both that you're wondering about? Could you explain the rationale for these changes, please? Specifically, I would like to understand this: https://github.com/STEllAR-GROUP/hpx/pull/5277/files#diff-93e12b963ad29614127c32b33e22c54428ac32e8c435de5d9be363b284722c62R980-R984 Is it the particular implementation, the motivation, or both that you're wondering about? Mostly the motivation... Could you explain the rationale for these changes, please? Specifically, I would like to understand this: https://github.com/STEllAR-GROUP/hpx/pull/5277/files#diff-93e12b963ad29614127c32b33e22c54428ac32e8c435de5d9be363b284722c62R980-R984 Is it the particular implementation, the motivation, or both that you're wondering about? Mostly the motivation... This came up in DLA-Future where they're attempting to use the MPI polling integration for their communication. One configuration is to create a separate thread pool with a dedicated worker thread only for polling, and enabling polling only on that thread pool. However, if the thread pool is really dedicated only polling it means that it never gets any tasks to execute and with the default settings it starts suspending itself (for up to a second at a time) even though there are CUDA events/MPI requests to poll for since the scheduler knows nothing about what state the polling vectors/queues are in. With this change the scheduling loop won't sleep if the polling function claims that it still has something to poll for. We've discussed an alternative solution as well where a dedicated polling thread would be created (without the scheduling loop) for polling, but that's a larger change, so this works as an intermediate solution until we look at that. @msimberg thanks for the explanation! I messed up my review, by looking at each commit individually. I'd like to rerun the octotiger/kokkkos tests hat we did with tthe dedicated polling pool to see if there's any noticable difference. The missing iterator increment might have caused a noticable slowdown on one stream if for some reason an event was held up. Thanks @biddisco for having a look. Yes, agreed that it might make a difference. It may actually also improve timings in some situations by not returning to executing normal tasks, but only testing will tell... @G-071 this might be interesting for you as well, especially if you can easily rerun the tests we did earlier with this branch/master once this is merged.
gharchive/pull-request
2021-04-06T13:31:59
2025-04-01T04:55:36.029585
{ "authors": [ "hkaiser", "msimberg" ], "repo": "STEllAR-GROUP/hpx", "url": "https://github.com/STEllAR-GROUP/hpx/pull/5277", "license": "BSL-1.0", "license_type": "permissive", "license_source": "github-api" }
451860604
ImportError: cannot import name region ImportError: cannot import name region @ Duplicate of #16. Please refer to INSTALL.md thanks,when it offer the training code? @lb1100
gharchive/issue
2019-06-04T08:22:40
2025-04-01T04:55:36.084058
{ "authors": [ "chenbolinstudent", "lb1100" ], "repo": "STVIR/pysot", "url": "https://github.com/STVIR/pysot/issues/28", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
364805084
openstack-ardana: don't reuse workspace to cleanup heat stacks The openstack-ardana-heat Jenkins job builds triggered as post build actions to clean up heat stacks must not reuse the workspace from their parent jobs because they are running asynchronously and the parent workspace may no longer be available. This error has affected at least one job run so far: https://ci.suse.de/job/openstack-ardana-heat/61/ Those parameters from the job and their use in the Jenkinsfile should also be removed. Those parameters from the job and their use in the Jenkinsfile should also be removed. Thanks @JanZerebecki . That would be reasonable, except those parameters are still used elsewhere in the openstack-ardana-vcloud job, in the stage that creates the stack. That stage does reuse the workspace, so we can't remove those parameters just yet.
gharchive/pull-request
2018-09-28T09:12:28
2025-04-01T04:55:36.100908
{ "authors": [ "JanZerebecki", "stefannica" ], "repo": "SUSE-Cloud/automation", "url": "https://github.com/SUSE-Cloud/automation/pull/2814", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
99644830
use the cluster VIP hostname, not the first node in the cluster Untested - don't merge! In HA mode we should always go through haproxy, otherwise we'll fail to access horizon at very least (this was breaking mkcloud tests with hacloud=1). Have manually triggered an mkcloud run with hacloud=1; should start soon... Tested on c19 and it works. Let's see if https://ci.suse.de/job/openstack-mkcloud/6096/ succeeds ... Well it didn't succeed but it certainly got further thanks to this PR. So IMHO this is ready to be merged. Superseded by #477.
gharchive/pull-request
2015-08-07T13:12:28
2025-04-01T04:55:36.103533
{ "authors": [ "aspiers" ], "repo": "SUSE-Cloud/automation", "url": "https://github.com/SUSE-Cloud/automation/pull/475", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
930173722
New token New TIP-3 token support Dropped TIP-3 by public key support Support for all known dictionaries for seed phrases Support 24 words seed phrase Support TIP-3 transfer by multisig address 0.0.5 ready
gharchive/pull-request
2021-06-25T13:25:44
2025-04-01T04:55:36.162807
{ "authors": [ "lailune" ], "repo": "SVOIcom/browser-extension", "url": "https://github.com/SVOIcom/browser-extension/pull/26", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
935236796
Mockup User-interface Dashboard - latest There are already some issues open related to this feature: #12, #14, #15, #107. Here I just collected notable information from previous issues and start to mock up this feature. In general, this feature should provide user-interface dashboard that helps users to notice their metadata submission progress, and get sense of what other metadata they should/can provide. In addition, this feature is also able to validate users' metadata and display some information to facilitate their data curation process. I will start to mock up this feature by splitting three tabs and below is some details for this feature: First of all, the feature should be optional for users. One of the reasons could be because extracting information from manifest may require downloading the files. It could cause much time to run for users who have a large number of files. To achieve this, we can initially hide the feature and let users use button to control. The feature is tab panel with three tabs: "selected template", "uploaded files", "metadata validation" (I'm bad at naming). The tab panel should be hidden in the second tab, select your dataset, of app. ["selected template"] :label:: this tab will demonstrate what metadata are required for the selected template :bulb: checklist + simple network ["uploaded files"] :label:: this tab will show all required metadata for uploaded manifests in the selected project :bulb: tree ["metadata validation"] :label:: this tab will validate users' uploaded metadata (provide button to initiate) :bulb: tubular data table whether metadata is out of data, aka, the metadata come with any errors from validateModelManifest we could also add more validation results similar with what current dashboard has Besides, we could add some valueBox to show overall stats about users progress in each project, such as total number of uploaded manifests; total progress % ( this will required a fixed number to avoid confusion and I will just use total number of templates in schema.json for now). I will create different PRs for each tab. Please feel free to leave suggestions here or in the front-end meeting. @milen-sage @sujaypatil96 Above docs will be updated once we are working on to finalize the dashboard. Acceptance criteria: dependent on the pr #181 In PR, rater systems is higher priority than the parallelization Blocked by this error with latest setup in shinyServer: Error: package or namespace load failed for ‘networkD3’ in dyn.load(file, DLLpath = DLLpath, ...): unable to load shared object '/home/rchai/R/x86_64-pc-linux-gnu-library/3.6/igraph/libs/igraph.so': libglpk.so.40: cannot open shared object file: No such file or directory Need one more week to make dashboard work again -> fix ^ issue and potential conflicts @ychae I am still refactoring the codes of dashboard. I will update the progress this week. After the discussion with @milen-sage today, I will do some follow-up bug fixes and improvements on the validation tab. Next step for sprint 16: testing all projects for HTAN work for all three of the dashboard's tabs. @milen-sage I have tested all HTAN projects in the demo, while most projects take <30s to load dashboard, HTAPP and TNP-TMA take ~ 6 min and ~1 min respectively. We will need to further test when cross-manifest is incorporated - which will not be benefited from parallelization. It's great to hear that the dashboard works across the board for these projects though! Once we test cross-validation in the context of HTAN, following your PR merge, we'll be likely ready to release the feature. We should have this PR unblocked next week. So very likely the next release of the DCA will contain the dashboard - hopefully in the next week or so 🥳
gharchive/issue
2021-07-01T21:43:42
2025-04-01T04:55:36.227383
{ "authors": [ "milen-sage", "rrchai", "ychae" ], "repo": "Sage-Bionetworks/data_curator", "url": "https://github.com/Sage-Bionetworks/data_curator/issues/174", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
890322285
Question: Mapping between Many-to-Many relation Hi, I have been using your repository as a guide to develop an application and I must say your work is amazing. I've been trying to develop a library management system and whenever I find myself confused, I try to approach a solution similar to your work. However, I am having difficulties in something you haven't covered yet. I am trying to define a many-to-many relationship between OrmEntities ( in my case book and author ). In my domain entities, I have a book entity that has as attribute an array of author entity, so when I am trying to map the props of book to Orm props, I find myself needing to map also the props of author ( which has already its own mapper ). What I have implemented currently is that I use the author mapper in the book mapper to map the props of the author. I was wondering how should I approach mapping between entities. Is calling a mapper inside another mapper the correct approach? Thank you and looking forward for your answer! Hey, thanks for kind words. What I have implemented currently is that I use the author mapper in the book mapper to map the props of the author. That's seems like the most obvious solution and I've been using the same approach for many-to-many relationships. So just keep using that unless you figure out something better :)
gharchive/issue
2021-05-12T17:28:39
2025-04-01T04:55:36.328793
{ "authors": [ "MahdiTurki", "Sairyss" ], "repo": "Sairyss/domain-driven-hexagon", "url": "https://github.com/Sairyss/domain-driven-hexagon/issues/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2748384243
fix: Add retry limits and improve error handling for JSON extraction Add comprehensive rate limit handling across API providers This PR implements robust rate limit handling across all API providers used in the AI-Scientist framework, addressing the continuous retry issue (#155). Changes Add RateLimitHandler class for centralized rate limit management Implement provider-specific request queues and locks Add proper error handling and logging for rate limit events Extend backoff patterns to all API providers (OpenAI, Anthropic, Google, xAI) Add user feedback during rate limiting Add configurable minimum request intervals per provider Implementation Details Created new rate_limit.py module for rate limit handling Added provider-specific rate limit detection Implemented request queuing mechanism Added comprehensive logging for debugging Extended backoff patterns with proper error type detection Testing The changes have been tested by: Verifying rate limit detection for different providers Testing backoff behavior with simulated rate limits Checking proper queue management Validating logging output Impact These changes make the system more robust by: Preventing continuous retries on rate limits Providing better error messages and logging Managing request rates across different providers Improving overall stability of API interactions Fixes #155 Link to Devin run: https://app.devin.ai/sessions/2ec43d6fe7a84849a348753167e5a895 Hey, thanks for your contributions! I am unable to see your Devin run right now without paying for access first, any way around this? I am unable to see your Devin run right now without paying for access first, any way around this? Sadly none yet.
gharchive/pull-request
2024-12-18T17:30:36
2025-04-01T04:55:36.335706
{ "authors": [ "conglu1997", "erkinalp" ], "repo": "SakanaAI/AI-Scientist", "url": "https://github.com/SakanaAI/AI-Scientist/pull/162", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1689462556
🛑 Nostr Relay - wss://knostr.neutrine.com is down In e5a593b, Nostr Relay - wss://knostr.neutrine.com (https://knostr.neutrine.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Nostr Relay - wss://knostr.neutrine.com is back up in dcae075.
gharchive/issue
2023-04-29T08:57:51
2025-04-01T04:55:36.338605
{ "authors": [ "Sakhalinfox" ], "repo": "Sakhalinfox/orangepilldevuptime", "url": "https://github.com/Sakhalinfox/orangepilldevuptime/issues/1748", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2135869350
알림 여부에 따라 홈화면의 알림 아이콘 다르게 보여주기 🚀 목표 알림 여부에 따라 홈화면의 알림 아이콘 다르게 보여주기 ⏱️ 할일 [x] 에셋 추가 및, 홈화면에서 알림 api 호출 -> 결과값에 따른 이미지 분기 💬 참고사항 수정완료
gharchive/issue
2024-02-15T07:59:26
2025-04-01T04:55:36.340029
{ "authors": [ "FirstDo" ], "repo": "Sal-Mal/salmal-iOS", "url": "https://github.com/Sal-Mal/salmal-iOS/issues/83", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
26917747
Household Mailing Street Line Breaks Issue by kbromer from Friday Apr 26, 2013 at 13:22 GMT Originally opened as https://github.com/SalesforceFoundation/Households/issues/29 The Household mailing street field does not support line breaks and multiple line addresses. Likely need to convert the Mailing Street field from straight Text to a TextArea @kbromer @jlantz any idea why this was reopened and the wontfix label removed by mrbelevedere? did we do this, or some automation accidentally do this. I'd think we'd close this since I don't believe we will fix it in the old address object.
gharchive/issue
2014-02-04T21:38:09
2025-04-01T04:55:36.343822
{ "authors": [ "davidhabib", "mrbelvedere" ], "repo": "SalesforceFoundation/Cumulus", "url": "https://github.com/SalesforceFoundation/Cumulus/issues/320", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
198208606
Display user friendly errors Critical Changes Changes Errors will display in a notification toast Issues Closed Fixes #271 Error display: @davidjray Could you please also add an screenshot of how it looks in Classic? Thanks! @ceiroa Good idea! Here it is in Classic: @ceiroa The display of the messages as text at the bottom was the default behavior. Lightning Out (for VisualForce) was adding those messages as text by default. I am listening to and acting upon the system event "aura:systemError" and showing a error message inside a container that has the ID "auraErrorMessage". Aura automatically appends the error message to that DIV. By using the system message event the default behavior is bypassed since the framework is aware I am subscribed to "aura:systemError".
gharchive/pull-request
2016-12-31T00:23:38
2025-04-01T04:55:36.347248
{ "authors": [ "ceiroa", "davidjray" ], "repo": "SalesforceFoundation/HEDAP", "url": "https://github.com/SalesforceFoundation/HEDAP/pull/361", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
169346240
Enforce order of fields in "author" in package.json This order: https://github.com/sindresorhus/np/blob/9ea8f5aa9f05ef6e2c236977141eb8b2bc693e5d/package.json#L8-L10 name, email, url. Same goes for maintainers if a maintainers list is found? Yup, and contributors. Any improvements on the error message? Tried a couple of things but this was the best I could come up with. Looks great
gharchive/issue
2016-08-04T10:48:55
2025-04-01T04:55:36.387150
{ "authors": [ "SamVerschueren", "sindresorhus" ], "repo": "SamVerschueren/clinton", "url": "https://github.com/SamVerschueren/clinton/issues/39", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1502112775
ISIL Conversion: Jump target not found in method The exception for this uses the message: Failed to convert to ISIL. Reason: Jump target not found in method. Ruh roh It occurs when the IsilBuilder can't find a match in the InstructionAddressMap while fixing jumps. Should be fixed as of #318
gharchive/issue
2022-12-19T00:48:00
2025-04-01T04:55:36.417211
{ "authors": [ "SamboyCoding", "ds5678" ], "repo": "SamboyCoding/Cpp2IL", "url": "https://github.com/SamboyCoding/Cpp2IL/issues/178", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
99607401
Is there any plan to support the ESP8266? The ESP8266 by Espressif-Systems is currently the (by far) cheapest System-on-chip for IoT-purposes with around 40k of available RAM and usually >500kB of flash-memory. With nodeMCU there already exists a firmware that sports a LUA-Interpreter so I would really love to see a JS-interpreter doing the same. Is this possible at all, and is there any available Documentation on the steps needed to make jerryscript run on this Hardware? @usefulthink IMHO, JerryScript can running on ESP8266. But, first we have to check up the ESP8266 by ourselves. Please reference the (https://github.com/Samsung/iotjs/issues/136) also. Thanks! i'm curious as well, as it seems is @maxogden - do you have a timeframe on when you'd be able to try it on an esp8266? @JerrySievert , We already ordered the ESP8266 but not delivered yet. So, it's hard to estimate the schedule now. I will keep update the progress here. To whom interested, I'm now working to make JerryScript run in ESP8266. First thing is getting used to the build system. \o/ Thank you so much! that's great news! /cc @maxogden, @kenichi Hello, I've uploaded initial development branch esp8266-dev for ESP-01 board. There are some patches to the SDK that may be not correct. Please add comments. First commit just blinks LED from the JS. APIs not just for testing. I'll try to bind network API so that we can use it from the JS codes. But I am not sure where to start from, so anyone who knows please leave a comment. About the memory usage... JerryHeap is set to 20KB and task stack size is 8K and free system heap memory is about 10K. Awesomeness @seanshpark . I'm on Mac OS and will try and figure out how to get the ESP SDK working and report back if I succeed @maxogden , Nice ! Howdy ... I am a keen ESP8266 developer and have a free book on the subject available here ... http://neilkolban.com/tech/esp8266/ I would love to offer any free assistance I can on getting JerryScript running on the ESP8266 and ESP32. I can be reached by email at kolban1@kolban.com @nkolban do you have a source for esp32's? @nkolban , Wow, glad to hear from you. I've read the book and it helped a lot!. I've made a branch embedding-dev, previously esp8266-dev but now merged also with mbed, and JerryScript itself runs in ESP8266. I was working to enable network, in my personal repo, but have little time to do this while working with others. @JerrySievert Earlier this year I worked hard and created a free book on the ESP8266. I listed this in a previous post. Each month I continued to update it. The good folks at Espressif were pleased with the result and I have offered to update it and/or re-write it for the ESP32. As such, they have invited me to their beta programme. I have no idea when I'll receive an instance but when I do, I'll be using it for working on a PDF of notes for us all to build upon. So ... to your question ... I have no source other than Espressif themselves and I don't expect to receive more than one and that would be for my study to allow me to write the PDF notes. When it arrives, I plan to go dark on it for months and simply work on the write-up for release when it itself becomes generally available. @seanshpark I'll be delighted to participate in any way I can. I assisted with porting the Espruino open source JavaScript to ESP8266 and that is working well, am also tinkering with Smart.js JavaScript .... so I'm growing more and more knowledge on what it takes to get JavaScript going on these ESP8266s. If you don't have time just now to work any further on the ESP8266 and there is no-one else contributing, then I'll step back until such time as work on it should restart. I'm keen to collaborate ... but collaboration is the operative words here. I'll be delighted to bring any and all ESP8266 skills and experience to the table, but I don't know anything yet about the internals of JerryScript ... so would be looking to be part of the community and not so much an active leader. @nkolban , first thing, thank you for writing down your kind thoughts. As english is not my first language writings may seem dry and somethings it's hard for to express my thoughts and emotions. YES it would be a big help if you can guide me and others with JerryScript running on ESP8266 and may be yourself with the codes. Current status is that just running JerryScript in ESP8266 is somewhat done, and it would be an honor if you can look the codes here and give comments. Also wrote some summary I've done with in here and here. It would be wonderful if you can give some notes about it. But after that I had some problems with enabling network, but it was my AP that was working improperly and now I'm waiting for the new one. I've tried also at home and connection itself is ok, but not sure my test codes was proper or not. I've reference nodemcu codes. So if possible can you give some informations how to implement this? I'm thinking of exposing JS APIs so that users can control AP connection with JavaScript, or maybe is this a bad idea? Currently I cannot find my code. May be at home... The next step I'm thinking is more difficult cause as memory is short I can't make IoT.js run so have to build-up more compact JS APis with TCP, UDP and HTTP or maybe CoAP... Hello, JerryScript on ESP8266 now in master branch, #733. @seanshpark thank you! that's fantastic news! A patch is recently landed wich adds the support.
gharchive/issue
2015-08-07T08:45:19
2025-04-01T04:55:36.474746
{ "authors": [ "JerrySievert", "lemmaa", "maxogden", "nkolban", "seanshpark", "usefulthink", "zherczeg" ], "repo": "Samsung/jerryscript", "url": "https://github.com/Samsung/jerryscript/issues/546", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2576939651
Windows Error When Downloading or Launching This Launcher Used this launcher quite a bit throughout my playthroughs but recently Windows is making it real annoying due to it declaring it as a virus. Anyway to remove this as a false positive? I know you can disable App screening in the action center but it will always turn it back on. Any update or solution would be appreciated! you can add the folder as an exception but it's better if i find a way to recompile it without malware detection or report it as a false positive so give me some time @thevelcrosasquatch re-uploaded the latest package, see if the problem is still there @thevelcrosasquatch re-uploaded the latest package, see if the problem is still there It downloads and the zip file is not instantly deleted. but Windows still flags it as a virus and doesn't allow it. I appreciate anything you do and am also trying to report it as a false positive i dont have it report on my end and virustotal doesn't mention microsoft, so i can't report it myself reported it as a false positive to malwarebytes and microsoft idc about the rest thanks 1.4 has false positive issues across many engines https://www.virustotal.com/gui/file/2e670db6722e4f9b1fa984b72f9386d20207d3843af70f5b1a13c6affb6ed49a/summary
gharchive/issue
2024-10-09T20:56:10
2025-04-01T04:55:36.497308
{ "authors": [ "SandeMC", "mer2329", "thevelcrosasquatch" ], "repo": "SandeMC/Kingdom-Hearts-Launchers", "url": "https://github.com/SandeMC/Kingdom-Hearts-Launchers/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1474684084
Small nitpick with the divide examples https://github.com/SandroMaglione/fpdart/blob/2c8cdd73f7d774a18d806743357a77f7569726d6/example/src/option/overview.dart#L4-L10 This is kind of a red herring but I still wanted to address it. Floating point division by zero is not really breaking. void main() { final allPassed = [ double.nan != double.nan, !double.infinity.isNaN, !double.negativeInfinity.isNaN, (0 / 0).isNaN, (1 / 0) == double.infinity, (-1 / 0) == double.negativeInfinity, ].every((x) => x); print('All True: $allPassed'); // Output: All True: true } A better example that definitely breaks: int divide(int a, int b) => a ~/ b; divide(10, 0); // Unhandled exception: IntegerDivisionByZeroException Hi @jcdang You are right, thanks for pointing this out 👍. Would you be open to send a PR to fix this?
gharchive/issue
2022-12-04T10:27:41
2025-04-01T04:55:36.506785
{ "authors": [ "SandroMaglione", "jcdang" ], "repo": "SandroMaglione/fpdart", "url": "https://github.com/SandroMaglione/fpdart/issues/71", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
784513397
Suggesting Music🎵 based on Facial Emotion predicted This is my solution to issue #2 . Whenever an emotion is predicted, the user can choose from 5 music options related to that emotion. I have made the suggested changes. Do let me know if it needs any other changes I have made the suggested changes. Do let me know if it needs any other changes @all-contributors please add @anushavc for Code @all-contributors please add @anushavc for Code
gharchive/pull-request
2021-01-12T19:06:36
2025-04-01T04:55:36.509247
{ "authors": [ "SanjayMarreddi", "anushavc" ], "repo": "SanjayMarreddi/Facial-Expression-Recognition-Classifier-Model", "url": "https://github.com/SanjayMarreddi/Facial-Expression-Recognition-Classifier-Model/pull/23", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1571182515
Usage of cancelAuctions() unclear Hi, If possible, could you explain a little bit how we should call the "cancelAuctions()" function? Can it be called directly after "CommodityBuyerAndSeller.sell" or should it be used in some other way? Also is there a way to always cancel auctions that have been undercut? Thanks in advance! Hi, in the current state it can be called manually to cancel auctions with: /run CommodityBuyerAndSeller.cancelAuctions(). Hi Sanjo - Also curious how this works. When you run the .cancelAuctions() command manually, what is the logic behind what it should cancel? Is it looking for undercut auctions, or some other flag? Based on the metric for how many auctions seem to be sold per day (from TradeSkillMaster), it cancels everything where it seems that the auction runs out before it seems sold. With this implementation, even when there are some auctions for a lower price in the auction house, if enough of the commodity is sold per day, the auction is kept in the auction house. The implementation can be found here: https://github.com/SanjoSolutions/CommodityBuyerAndSeller/blob/3b558e0df1d566821270093f3f29b69f17091915/CommodityBuyerAndSeller/CommodityBuyerAndSeller.lua#L162-L179. Ah I see, interesting! I can see how that would be useful for certain items. I was hoping there might be a Cancel function that can cycle through existing auctions and check for undercuts and cancel those auctions. When posting Region-Wide Commodities quickly, it's helpful to be able to cancel auctions immediately after running out of stock. Maybe even define just specific items to cancel on the AH via the API. The same way we define items to post via the CommodityBuyerAndSellerData.lua file. Thank you for explaining, however I do agree with fatmagic that it would be nice to have a function that does not consider the auctions sold per day before canceling. A function that only checks for undercuts.
gharchive/issue
2023-02-05T00:28:50
2025-04-01T04:55:36.513382
{ "authors": [ "SanjoSolutions", "fatmagic", "steenert" ], "repo": "SanjoSolutions/CommodityBuyerAndSeller", "url": "https://github.com/SanjoSolutions/CommodityBuyerAndSeller/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1674431608
🛑 Mail is down In 096133b, Mail (https://mail.selgaraje.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Mail is back up in 4da0386.
gharchive/issue
2023-04-19T08:28:20
2025-04-01T04:55:36.549154
{ "authors": [ "Sasillo" ], "repo": "Sasillo/upptime", "url": "https://github.com/Sasillo/upptime/issues/108", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
783553136
Custom Chart Component: Hexagonal Map of U.S. States & Territories Rather than displaying data for U.S. states as a geographic (choropleth) map, display them as an abstracted map of gridded / tessellated shapes such as squares, circles, or hexagons. This approach has a few added benefits over the geographic map counterpart: Small states and U.S. territories are easier to find and interact with (e.g. hover / click with a mouse) Helps solve the common problem with choropleth maps of viewers incorrectly associating geographic area with a value Are better than choropleth maps with racial/ethnic representation reference article One downside is that these maps may confuse users who are not familiar with them, though they are becoming much more common in reporting (e.g. election maps) and are used in quite a few health equity related websites and dashboards. Having well written titles, descriptions, and labeling each shape with its corresponding state abbreviation can also help with this. Currently I have not seen an example of this type of map being created with Vega/VegaLite, so it seems warranted to implement it as a custom component. A couple ways of accomplishing this could be: Use inline SVG augmented with data-id attributes for binding data (example / explanation) Create a data structure that contains a record describing each shape (e.g. the coordinates necessary for constructing an SVG path "d" attribute, in the case of hexagons) and use that data structure with an HTML/SVG templating system (in this case JSX) to draw the shapes. The benefit of approach #1 is that if the shapes needed design modifications, the SVG could be edited using software such as Figma or Sketch. The benefit of approach #2 is that the map graphic could be scaled to any arbitrary size using a linear scale. (Approach #1 could be made responsive by using the viewBox attribute on the SVG element, but this also auto-scales any text / labels). Either approach will require a method for binding data to the shapes, or in other words programmatically determining what fill color is used for each shape and whether or not to draw shapes representing U.S. territories as not all health data includes information on territories. It might be worth prototyping this component first, or looking for other ready made solutions in the wild. didn't move forward with this design
gharchive/issue
2020-11-16T18:35:36
2025-04-01T04:55:36.554835
{ "authors": [ "clhenrick", "kkatzen" ], "repo": "SatcherInstitute/health-equity-tracker", "url": "https://github.com/SatcherInstitute/health-equity-tracker/issues/233", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1789115710
Creating a duplicate default testnet wallet and removing main net default wallet when I look into transaction on mempool @yshwanth can u please add more details on what the bug is and how you encountered it ? When I am on Transaction page, on click of Tx I'd or address ,it goes mempool space. After clicking back from mempool , the above bug arises Nice catch @yshwanth I was able to replicate it: start a fresh install of bull wallet goto testnet mode receive some funds open tx details click txid for mempool web view go back return to main net mode WALLET WILL BE GONE! Nice catch @yshwanth I was able to replicate it: * start a fresh install of bull wallet * goto testnet mode * receive some funds * open tx details * click txid for mempool web view * go back * return to main net mode * Lands on create wallet page * EDIT: Wallet is not removed. It's still there on restart. So much be just a router issue. This seems to have gotten fixed over the past few releases. Just tested this flow and not able to replicate it anymore.
gharchive/issue
2023-07-05T09:07:27
2025-04-01T04:55:36.592701
{ "authors": [ "i5hi", "mocodesmo", "yshwanth" ], "repo": "SatoshiPortal/bullbitcoin-mobile", "url": "https://github.com/SatoshiPortal/bullbitcoin-mobile/issues/23", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2290920987
Like button : we can add like button that displays the number of likes for the song Describe the feature @Satyam1923 can you please assign this task for me under GSSOC24 Add ScreenShots ![image](https://github.com/Satyam1923/Spring-Music-Player/assets/105491821/bc924948-3f7b-4e1e-8b37-cd32a85c17ee Record [X] I agree to follow this project's Code of Conduct [X] I have checked the existing issues [X] I'm a GSSOC'24 contributor [X] I want to work on this issue Hey @Varnika2526, I'd like to work on adding a like button feature that displays the number of likes for the song. Could you please assign this task to me with the GSSOC tag? Thanks! @sahithi000 do this task sooner. @saurabhsingh720 sure will do this @saurabhsingh720 is this all or something else missing @sahithi000 just type the name of song and search it, then u will get interface @saurabhsingh720 This is all i am able to render - is there something missing @sahithi000 Go for it and add this feature
gharchive/issue
2024-05-11T13:46:15
2025-04-01T04:55:36.598082
{ "authors": [ "Priyanka-Ray-Choudhury", "sahithi000", "saurabhsingh720" ], "repo": "Satyam1923/Spring-Music-Player", "url": "https://github.com/Satyam1923/Spring-Music-Player/issues/50", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
925907583
I2C GPS + RTC are not compatible Check why there is a problem: same address? power problem? pull up resistor? The only problem is when RTC is used (ON) and GPS is OFF but connected. Simple solution: leave GPS ON during LOG even if not used (RTC is used instead)
gharchive/issue
2021-06-21T07:18:14
2025-04-01T04:55:36.602794
{ "authors": [ "Saultes45" ], "repo": "Saultes45/ESP32-Seismometer-Datalogger", "url": "https://github.com/Saultes45/ESP32-Seismometer-Datalogger/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2296465873
[FEATURE] Adding a Chatbot Description I suggest adding a chatbot that answers common questions related to the CodeIN website quickly as well as links to certain sections. We can replace the feedback form button with the chatbot as we already have a seperate feedback section. I can implement this chatbot as a GSSOC'24 contributor. Please assign me this task. Screenshots No response @sanskritilabroo You can work on this issue! Thanks for choosing our community project. 🙌🏻 ✅ For any queries, Please reach out to me here: https://www.linkedin.com/in/sauravmukherjee44/
gharchive/issue
2024-05-14T21:47:42
2025-04-01T04:55:36.605074
{ "authors": [ "SauravMukherjee44", "sanskritilabroo" ], "repo": "SauravMukherjee44/CodeIN-Community-Website", "url": "https://github.com/SauravMukherjee44/CodeIN-Community-Website/issues/835", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2328240056
Navbar style changed and fixed Navbar style changed and fixed @Sayak-Bhunia please review the PR issue #183 solved
gharchive/pull-request
2024-05-31T16:56:19
2025-04-01T04:55:36.606378
{ "authors": [ "RamakrushnaBiswal" ], "repo": "Sayak-Bhunia/mystory", "url": "https://github.com/Sayak-Bhunia/mystory/pull/201", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
174446571
Add parallel orderbook This PR provides a new package implementing order books whose methods are capable of operating in parallel. Coverage decreased (-3.05%) to 92.899% when pulling 0a186c57bcc15326393a33090f2b45f3928d2ff7 on add-parallel-orderbook into 63b877a2b51b02b931431c874f7c7f895249177d on develop. Coverage decreased (-3.05%) to 92.899% when pulling 0e5e3b7e97ea14f3a0a701a64140e52696e71ded on add-parallel-orderbook into 63b877a2b51b02b931431c874f7c7f895249177d on develop. Coverage increased (+1.2%) to 97.159% when pulling 0fbd285db349ccdefc44bcb4703d5559f45be886 on add-parallel-orderbook into 63b877a2b51b02b931431c874f7c7f895249177d on develop.
gharchive/pull-request
2016-09-01T06:30:50
2025-04-01T04:55:36.616581
{ "authors": [ "coveralls", "davidrpugh" ], "repo": "ScalABM/markets-sandbox", "url": "https://github.com/ScalABM/markets-sandbox/pull/169", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2270929760
[MODEL EVALUATION REQUEST] syvai/llama3-da-base Model ID syvai/llama3-da-base Model type Decoder model (e.g., GPT) Model languages [X] Danish [ ] Swedish [ ] Norwegian (Bokmål or Nynorsk) [ ] Icelandic [ ] Faroese [ ] German [ ] Dutch [ ] English Merged model Not a merged model {"dataset": "angry-tweets", "task": "sentiment-classification", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"mcc": 0.48578669475498515, "macro_f1": 0.5756074875014533}, {"mcc": 0.5451341385080095, "macro_f1": 0.6440269667992441}, {"mcc": 0.5227556106816867, "macro_f1": 0.6198220175920836}, {"mcc": 0.4549742914924587, "macro_f1": 0.5671032255914509}, {"mcc": 0.5786792580939469, "macro_f1": 0.7152859477124184}, {"mcc": 0.49080920574546677, "macro_f1": 0.5924390761222383}, {"mcc": 0.4668025294430722, "macro_f1": 0.597952923055018}, {"mcc": 0.48730996713632285, "macro_f1": 0.6003524161940004}, {"mcc": 0.5252555882558146, "macro_f1": 0.6619089576396068}, {"mcc": 0.4519940521524307, "macro_f1": 0.5662222690500022}]}, "total": {"test_mcc": 50.095013362641936, "test_mcc_se": 2.551819498716857, "test_macro_f1": 61.407212872575165, "test_macro_f1_se": 2.947555653496753}}, "num_model_parameters": -1, "max_sequence_length": 8192, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": true, "scandeval_version": "12.9.0"} {"dataset": "dansk", "task": "named-entity-recognition", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"micro_f1_no_misc": 0.5454545454545455, "micro_f1": 0.40713536201469047}, {"micro_f1_no_misc": 0.48306332842415317, "micro_f1": 0.40217391304347827}, {"micro_f1_no_misc": 0.44741873804971316, "micro_f1": 0.34292866082603257}, {"micro_f1_no_misc": 0.5071315372424723, "micro_f1": 0.41745283018867924}, {"micro_f1_no_misc": 0.5543766578249337, "micro_f1": 0.4809322033898305}, {"micro_f1_no_misc": 0.39278557114228463, "micro_f1": 0.32552404438964244}, {"micro_f1_no_misc": 0.4952380952380952, "micro_f1": 0.3904555314533622}, {"micro_f1_no_misc": 0.5054945054945056, "micro_f1": 0.4390832328106152}, {"micro_f1_no_misc": 0.625, "micro_f1": 0.5067873303167422}, {"micro_f1_no_misc": 0.5052264808362369, "micro_f1": 0.40208333333333335}]}, "total": {"test_micro_f1_no_misc": 50.611894597069394, "test_micro_f1_no_misc_se": 3.862473048440808, "test_micro_f1": 41.14556441766406, "test_micro_f1_se": 3.4216346088864853}}, "num_model_parameters": -1, "max_sequence_length": 8317, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": true, "scandeval_version": "12.9.0"} {"dataset": "scala-da", "task": "linguistic-acceptability", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"mcc": 0.3467728773357996, "macro_f1": 0.606567985116881}, {"mcc": 0.30523384783367985, "macro_f1": 0.5734203903965177}, {"mcc": 0.39958010799281674, "macro_f1": 0.6805883898082479}, {"mcc": 0.21991635455480876, "macro_f1": 0.5894339622641509}, {"mcc": 0.19537991267766006, "macro_f1": 0.5733333333333333}, {"mcc": 0.21771619517396076, "macro_f1": 0.6032388663967612}, {"mcc": 0.18358731423084815, "macro_f1": 0.5536145638367647}, {"mcc": 0.29677977453837695, "macro_f1": 0.6399103485586213}, {"mcc": 0.23492706697300833, "macro_f1": 0.5305213518470002}, {"mcc": 0.3322003278622785, "macro_f1": 0.6652668174196625}]}, "total": {"test_mcc": 27.32093779173238, "test_mcc_se": 4.52250751255337, "test_macro_f1": 60.15896008977941, "test_macro_f1_se": 2.9856376529477107}}, "num_model_parameters": -1, "max_sequence_length": 8192, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": true, "scandeval_version": "12.9.0"} {"dataset": "scandiqa-da", "task": "question-answering", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"em": 55.357142857142854, "f1": 62.96957671957673}, {"em": 60.0, "f1": 64.12896825396825}, {"em": 62.04819277108434, "f1": 66.48594377510042}, {"em": 57.73809523809524, "f1": 63.8128306878307}, {"em": 59.354838709677416, "f1": 65.99436763952895}, {"em": 58.23529411764706, "f1": 64.29161138139467}, {"em": 67.48466257668711, "f1": 71.47823546596553}, {"em": 60.24844720496895, "f1": 65.47964113181504}, {"em": 65.18987341772151, "f1": 67.73558368495078}, {"em": 57.57575757575758, "f1": 63.355940355940355}]}, "total": {"test_em": 60.32323044687821, "test_em_se": 2.2843013779381365, "test_f1": 65.57326990960715, "test_f1_se": 1.5881087791350552}}, "num_model_parameters": -1, "max_sequence_length": 8221, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": true, "scandeval_version": "12.9.0"} {"dataset": "nordjylland-news", "task": "summarization", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"bertscore": 0.6666320929070935, "rouge_l": 0.2058250782315233}, {"bertscore": 0.6737563784699887, "rouge_l": 0.22166790899050892}, {"bertscore": 0.6400531205581501, "rouge_l": 0.19714933578314667}, {"bertscore": 0.6744423654163256, "rouge_l": 0.21594586589408826}, {"bertscore": 0.6439992199884728, "rouge_l": 0.19159034738509856}, {"bertscore": 0.6679514087736607, "rouge_l": 0.21379192973044353}, {"bertscore": 0.6552678904263303, "rouge_l": 0.1846247075747077}, {"bertscore": 0.6591290490468964, "rouge_l": 0.1805290442556633}, {"bertscore": 0.6659495243802667, "rouge_l": 0.2009899162103309}, {"bertscore": 0.6703548389486969, "rouge_l": 0.20564027484019184}]}, "total": {"test_bertscore": 66.17535888915882, "test_bertscore_se": 0.7431799750954451, "test_rouge_l": 20.17754408895703, "test_rouge_l_se": 0.8375811790095886}}, "num_model_parameters": -1, "max_sequence_length": 8445, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": true, "scandeval_version": "12.9.0"} {"dataset": "danske-talemaader", "task": "knowledge", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"mcc": 0.6030800344400068, "accuracy": 0.69921875}, {"mcc": 0.553221350485432, "accuracy": 0.66796875}, {"mcc": 0.5575239513432603, "accuracy": 0.66796875}, {"mcc": 0.5193061200548681, "accuracy": 0.63671875}, {"mcc": 0.6052081533379606, "accuracy": 0.70703125}, {"mcc": 0.5718578040164519, "accuracy": 0.67578125}, {"mcc": 0.5165724875512112, "accuracy": 0.63671875}, {"mcc": 0.5491412540113148, "accuracy": 0.66796875}, {"mcc": 0.5813671897234177, "accuracy": 0.6875}, {"mcc": 0.5889524748933258, "accuracy": 0.6875}]}, "total": {"test_mcc": 56.46230819857247, "test_mcc_se": 1.9417613719502544, "test_accuracy": 67.34375, "test_accuracy_se": 1.4580416374941654}}, "num_model_parameters": -1, "max_sequence_length": 8192, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": true, "scandeval_version": "12.9.0"} {"dataset": "danish-citizen-tests", "task": "knowledge", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"mcc": 0.4613738338317101, "accuracy": 0.6484375}, {"mcc": 0.6102337871808171, "accuracy": 0.7421875}, {"mcc": 0.5380750385837817, "accuracy": 0.6875}, {"mcc": 0.6157677332388264, "accuracy": 0.734375}, {"mcc": 0.6844294910429646, "accuracy": 0.7890625}, {"mcc": 0.6094010421946812, "accuracy": 0.734375}, {"mcc": 0.5425073579510877, "accuracy": 0.6953125}, {"mcc": 0.5676531741521174, "accuracy": 0.703125}, {"mcc": 0.6020677180639061, "accuracy": 0.7265625}, {"mcc": 0.4982675721980892, "accuracy": 0.6484375}]}, "total": {"test_mcc": 57.29776748437981, "test_mcc_se": 4.028763349833513, "test_accuracy": 71.09375, "test_accuracy_se": 2.7008711300451025}}, "num_model_parameters": -1, "max_sequence_length": 8192, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": true, "scandeval_version": "12.9.0"} {"dataset": "hellaswag-da", "task": "common-sense-reasoning", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"accuracy": 0.359375, "mcc": 0.20300685930771878}, {"accuracy": 0.37890625, "mcc": 0.17731662243422025}, {"accuracy": 0.39453125, "mcc": 0.21835098490320076}, {"accuracy": 0.32421875, "mcc": 0.13681729983361274}, {"accuracy": 0.44921875, "mcc": 0.2812738332990418}, {"accuracy": 0.4140625, "mcc": 0.2613022689367176}, {"accuracy": 0.41015625, "mcc": 0.24066807507021618}, {"accuracy": 0.35546875, "mcc": 0.17447302547057214}, {"accuracy": 0.4375, "mcc": 0.29307972433588125}, {"accuracy": 0.3671875, "mcc": 0.18955891606529846}]}, "total": {"test_accuracy": 38.90625, "test_accuracy_se": 2.432391147144787, "test_mcc": 21.7584760965648, "test_mcc_se": 3.1470006095223524}}, "num_model_parameters": -1, "max_sequence_length": 8192, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": true, "scandeval_version": "12.9.0"} {"dataset": "speed", "task": "speed", "dataset_languages": ["ab", "aa", "af", "sq", "am", "ar", "an", "hy", "as", "av", "ae", "ay", "az", "bm", "ba", "eu", "be", "bn", "bi", "bs", "br", "bg", "my", "ca", "ch", "ce", "ny", "zh", "cu", "cv", "kw", "co", "cr", "hr", "cs", "da", "dv", "nl", "dz", "en", "eo", "et", "ee", "fo", "fj", "fi", "fr", "fy", "ff", "gd", "gl", "lg", "ka", "de", "el", "kl", "gn", "gu", "ht", "ha", "he", "hz", "hi", "ho", "hu", "is", "io", "ig", "id", "ia", "ie", "iu", "ik", "ga", "it", "ja", "kn", "kr", "ks", "kk", "km", "ki", "rw", "ky", "kv", "kg", "ko", "kj", "ku", "lo", "la", "lv", "li", "ln", "lt", "lu", "lb", "mk", "mg", "ms", "ml", "mt", "gv", "mi", "mr", "mh", "mn", "na", "nv", "nd", "nr", "ng", "ne", "no", "nb", "nn", "ii", "oc", "oj", "or", "om", "os", "pi", "ps", "fa", "pl", "pt", "pa", "qu", "ro", "rm", "rn", "ru", "se", "sm", "sg", "sa", "sc", "sr", "sn", "sd", "si", "sk", "sl", "so", "st", "es", "su", "sw", "ss", "sv", "tl", "ty", "tg", "ta", "tt", "te", "th", "bo", "ti", "to", "ts", "tn", "tr", "tk", "tw", "ug", "uk", "ur", "uz", "ve", "vi", "vo", "wa", "cy", "wo", "xh", "yi", "yo", "za", "zu"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"test_speed": 616.64, "test_speed_short": 81.8}, {"test_speed": 1221.48, "test_speed_short": 144.0}, {"test_speed": 1819.84, "test_speed_short": 271.32}, {"test_speed": 2392.46, "test_speed_short": 332.64}, {"test_speed": 2918.52, "test_speed_short": 396.0}, {"test_speed": 3513.78, "test_speed_short": 520.74}, {"test_speed": 4057.64, "test_speed_short": 580.16}, {"test_speed": 4590.3, "test_speed_short": 641.24}, {"test_speed": 5140.64, "test_speed_short": 706.5}, {"test_speed": 5654.099999999999, "test_speed_short": 776.16}]}, "total": {"test_speed": 3192.54, "test_speed_se": 1048.9928626307697, "test_speed_short": 445.05600000000004, "test_speed_short_se": 147.29295704505446}}, "num_model_parameters": -1, "max_sequence_length": 8190, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": true, "scandeval_version": "12.9.0"} @saattrupdan sure. Seems like a lot of the models on the leaderboard use the validation dataset though. What are the rules? @saattrupdan sure. Seems like a lot of the models on the leaderboard use the validation dataset though. What are the rules? Yeah, I understand the confusion. Here are the current rules, along with their reasoning: OpenAI models are mostly evaluated on validation splits, to avoid spending too much money on their API. Large models (which currently means >60B) are evaluated on validation splits, as it simply takes too long to evaluate them on the full test split. These would ideally be replaced with test performance, but we wanted the 70B models on the leaderboard asap. Merged models are only allowed to be evaluated on validation splits, to avoid overfitting the test splits. @saattrupdan here is the full test set. {"dataset": "angry-tweets", "task": "sentiment-classification", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"mcc": 0.4696411710308399, "macro_f1": 0.5689088437559591}, {"mcc": 0.4783072941320707, "macro_f1": 0.6064846392432599}, {"mcc": 0.5048132999312483, "macro_f1": 0.6262060450130525}, {"mcc": 0.48082449595985216, "macro_f1": 0.5827575804086761}, {"mcc": 0.5039576194682494, "macro_f1": 0.6559966993874401}, {"mcc": 0.47692483982088185, "macro_f1": 0.5648128093595589}, {"mcc": 0.4907810646537454, "macro_f1": 0.6245595547024162}, {"mcc": 0.5103644688156992, "macro_f1": 0.6109627105261105}, {"mcc": 0.505733348275026, "macro_f1": 0.6491378479346875}, {"mcc": 0.4393512405941747, "macro_f1": 0.5627873001235221}]}, "total": {"test_mcc": 48.60698842681788, "test_mcc_se": 1.3538823811908938, "test_macro_f1": 60.52614030454683, "test_macro_f1_se": 2.128340943556976}}, "num_model_parameters": -1, "max_sequence_length": 8192, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": false, "scandeval_version": "12.9.0"} {"dataset": "dansk", "task": "named-entity-recognition", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"micro_f1_no_misc": 0.5048104131295981, "micro_f1": 0.32262014080195894}, {"micro_f1_no_misc": 0.42061855670103093, "micro_f1": 0.3161073825503356}, {"micro_f1_no_misc": 0.4327798607391537, "micro_f1": 0.2846465275615073}, {"micro_f1_no_misc": 0.5171281088690756, "micro_f1": 0.35988200589970504}, {"micro_f1_no_misc": 0.5409683426443203, "micro_f1": 0.40038192234245706}, {"micro_f1_no_misc": 0.4943820224719101, "micro_f1": 0.3696468820435762}, {"micro_f1_no_misc": 0.47228691197019096, "micro_f1": 0.29665618748213773}, {"micro_f1_no_misc": 0.5407594936708862, "micro_f1": 0.4412225705329153}, {"micro_f1_no_misc": 0.5801133436373004, "micro_f1": 0.38125597704813513}, {"micro_f1_no_misc": 0.49290593799264315, "micro_f1": 0.31412977562158884}]}, "total": {"test_micro_f1_no_misc": 49.967529918261086, "test_micro_f1_no_misc_se": 3.046468213709984, "test_micro_f1": 34.86549371884317, "test_micro_f1_se": 3.10619097319635}}, "num_model_parameters": -1, "max_sequence_length": 8317, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": false, "scandeval_version": "12.9.0"} {"dataset": "scala-da", "task": "linguistic-acceptability", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"mcc": 0.2909047529431938, "macro_f1": 0.6117338354016496}, {"mcc": 0.2230574954642969, "macro_f1": 0.4953854270049298}, {"mcc": 0.3190845541826493, "macro_f1": 0.6432908509096258}, {"mcc": 0.29427572171645977, "macro_f1": 0.6070426023045632}, {"mcc": 0.18649944446476618, "macro_f1": 0.5727118923948437}, {"mcc": 0.3098060222524053, "macro_f1": 0.6505678904420525}, {"mcc": 0.31740902887177114, "macro_f1": 0.5926223009719411}, {"mcc": 0.3220416633819579, "macro_f1": 0.6458542138775158}, {"mcc": 0.23488725485413886, "macro_f1": 0.517865135179449}, {"mcc": 0.3182183982344284, "macro_f1": 0.6497240761551745}]}, "total": {"test_mcc": 28.161843363660676, "test_mcc_se": 3.0195073014356284, "test_macro_f1": 59.86798224641745, "test_macro_f1_se": 3.4364072979850233}}, "num_model_parameters": -1, "max_sequence_length": 8192, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": false, "scandeval_version": "12.9.0"} {"dataset": "scandiqa-da", "task": "question-answering", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"em": 58.79163439194423, "f1": 65.63016988218365}, {"em": 58.91472868217054, "f1": 64.70418756396865}, {"em": 62.364760432766616, "f1": 66.91442006993417}, {"em": 61.6822429906542, "f1": 67.55571718916958}, {"em": 58.3011583011583, "f1": 64.95205914201928}, {"em": 57.05474171164225, "f1": 63.85024540015692}, {"em": 57.32725892179195, "f1": 64.05410779632373}, {"em": 59.50349107835532, "f1": 65.26705872961514}, {"em": 62.509803921568626, "f1": 66.78999459432893}, {"em": 59.006211180124225, "f1": 65.32129657216186}]}, "total": {"test_em": 59.54560316121763, "test_em_se": 1.2263301129220445, "test_f1": 65.50392569398619, "test_f1_se": 0.7658081470552555}}, "num_model_parameters": -1, "max_sequence_length": 8221, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": false, "scandeval_version": "12.9.0"} {"dataset": "nordjylland-news", "task": "summarization", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"bertscore": 0.6720210982457502, "rouge_l": 0.21981335760765502}, {"bertscore": 0.6782984648452839, "rouge_l": 0.23576159574932398}, {"bertscore": 0.6426302872423548, "rouge_l": 0.20095338498339652}, {"bertscore": 0.6763368048268603, "rouge_l": 0.230150609632282}, {"bertscore": 0.6483787225588458, "rouge_l": 0.20681575666146607}, {"bertscore": 0.6698031227570027, "rouge_l": 0.21904983152090168}, {"bertscore": 0.6483039667800767, "rouge_l": 0.1785807992755065}, {"bertscore": 0.6498676358751254, "rouge_l": 0.180089508970512}, {"bertscore": 0.6600310464855283, "rouge_l": 0.19310263853776272}, {"bertscore": 0.6702250030211871, "rouge_l": 0.21830647392677194}]}, "total": {"test_bertscore": 66.15896152638015, "test_bertscore_se": 0.8258502677445179, "test_rouge_l": 20.826239568655787, "test_rouge_l_se": 1.2302837868787073}}, "num_model_parameters": -1, "max_sequence_length": 8445, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": false, "scandeval_version": "12.9.0"} {"dataset": "danske-talemaader", "task": "knowledge", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"mcc": 0.5517720051679987, "accuracy": 0.65625}, {"mcc": 0.5495761161654226, "accuracy": 0.6611328125}, {"mcc": 0.5092245230176828, "accuracy": 0.625}, {"mcc": 0.48581108828897157, "accuracy": 0.6025390625}, {"mcc": 0.5913112532680613, "accuracy": 0.681640625}, {"mcc": 0.5863301539155569, "accuracy": 0.6826171875}, {"mcc": 0.5521940579086276, "accuracy": 0.6552734375}, {"mcc": 0.5101407479709369, "accuracy": 0.625}, {"mcc": 0.5615994442484297, "accuracy": 0.6640625}, {"mcc": 0.557205620582808, "accuracy": 0.6669921875}]}, "total": {"test_mcc": 54.55165010534495, "test_mcc_se": 2.106095737366316, "test_accuracy": 65.205078125, "test_accuracy_se": 1.627153276087195}}, "num_model_parameters": -1, "max_sequence_length": 8192, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": false, "scandeval_version": "12.9.0"} {"dataset": "danish-citizen-tests", "task": "knowledge", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"mcc": 0.6468501709056725, "accuracy": 0.755859375}, {"mcc": 0.6045907638585358, "accuracy": 0.732421875}, {"mcc": 0.7037107952506362, "accuracy": 0.80078125}, {"mcc": 0.6118635158111443, "accuracy": 0.736328125}, {"mcc": 0.5749724096059811, "accuracy": 0.71484375}, {"mcc": 0.6507803034201805, "accuracy": 0.755859375}, {"mcc": 0.6274317264510327, "accuracy": 0.74609375}, {"mcc": 0.5973916219955222, "accuracy": 0.724609375}, {"mcc": 0.5838446789453279, "accuracy": 0.71484375}, {"mcc": 0.617418634994055, "accuracy": 0.73828125}]}, "total": {"test_mcc": 62.18854621238088, "test_mcc_se": 2.342150079479917, "test_accuracy": 74.19921875, "test_accuracy_se": 1.5695312499999998}}, "num_model_parameters": -1, "max_sequence_length": 8192, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": false, "scandeval_version": "12.9.0"} {"dataset": "hellaswag-da", "task": "common-sense-reasoning", "dataset_languages": ["da"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"accuracy": 0.3193359375, "mcc": 0.13163249732552004}, {"accuracy": 0.41162109375, "mcc": 0.2306322150220266}, {"accuracy": 0.43408203125, "mcc": 0.2582696052362616}, {"accuracy": 0.34716796875, "mcc": 0.17752581042088006}, {"accuracy": 0.40625, "mcc": 0.23309539033513088}, {"accuracy": 0.37548828125, "mcc": 0.1944480497433213}, {"accuracy": 0.37744140625, "mcc": 0.18311200439947584}, {"accuracy": 0.3740234375, "mcc": 0.2064809101108482}, {"accuracy": 0.38720703125, "mcc": 0.21531993481195724}, {"accuracy": 0.39453125, "mcc": 0.2059679515332374}]}, "total": {"test_accuracy": 38.271484375, "test_accuracy_se": 2.0267984825902023, "test_mcc": 20.36484368938659, "test_mcc_se": 2.1770244718658343}}, "num_model_parameters": -1, "max_sequence_length": 8192, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": false, "scandeval_version": "12.9.0"} {"dataset": "speed", "task": "speed", "dataset_languages": ["ab", "aa", "af", "sq", "am", "ar", "an", "hy", "as", "av", "ae", "ay", "az", "bm", "ba", "eu", "be", "bn", "bi", "bs", "br", "bg", "my", "ca", "ch", "ce", "ny", "zh", "cu", "cv", "kw", "co", "cr", "hr", "cs", "da", "dv", "nl", "dz", "en", "eo", "et", "ee", "fo", "fj", "fi", "fr", "fy", "ff", "gd", "gl", "lg", "ka", "de", "el", "kl", "gn", "gu", "ht", "ha", "he", "hz", "hi", "ho", "hu", "is", "io", "ig", "id", "ia", "ie", "iu", "ik", "ga", "it", "ja", "kn", "kr", "ks", "kk", "km", "ki", "rw", "ky", "kv", "kg", "ko", "kj", "ku", "lo", "la", "lv", "li", "ln", "lt", "lu", "lb", "mk", "mg", "ms", "ml", "mt", "gv", "mi", "mr", "mh", "mn", "na", "nv", "nd", "nr", "ng", "ne", "no", "nb", "nn", "ii", "oc", "oj", "or", "om", "os", "pi", "ps", "fa", "pl", "pt", "pa", "qu", "ro", "rm", "rn", "ru", "se", "sm", "sg", "sa", "sc", "sr", "sn", "sd", "si", "sk", "sl", "so", "st", "es", "su", "sw", "ss", "sv", "tl", "ty", "tg", "ta", "tt", "te", "th", "bo", "ti", "to", "ts", "tn", "tr", "tk", "tw", "ug", "uk", "ur", "uz", "ve", "vi", "vo", "wa", "cy", "wo", "xh", "yi", "yo", "za", "zu"], "model": "syvai/llama3-da-base", "results": {"raw": {"test": [{"test_speed": 621.5600000000001, "test_speed_short": 83.10000000000001}, {"test_speed": 1215.0, "test_speed_short": 141.29999999999998}, {"test_speed": 1790.8000000000002, "test_speed_short": 270.98}, {"test_speed": 2373.14, "test_speed_short": 326.76}, {"test_speed": 2922.54, "test_speed_short": 396.5}, {"test_speed": 3533.06, "test_speed_short": 527.34}, {"test_speed": 4057.64, "test_speed_short": 590.52}, {"test_speed": 4596.72, "test_speed_short": 654.36}, {"test_speed": 5133.42, "test_speed_short": 707.4}, {"test_speed": 5670.14, "test_speed_short": 757.5400000000001}]}, "total": {"test_speed": 3191.4019999999996, "test_speed_se": 1052.8881977460358, "test_speed_short": 445.58000000000004, "test_speed_short_se": 147.2742039592934}}, "num_model_parameters": -1, "max_sequence_length": 8190, "vocabulary_size": 128256, "generative": true, "few_shot": true, "validation_split": false, "scandeval_version": "12.9.0"} Thanks! It's live now 🙂
gharchive/issue
2024-04-30T09:07:28
2025-04-01T04:55:36.659329
{ "authors": [ "mhenrichsen", "saattrupdan" ], "repo": "ScandEval/ScandEval", "url": "https://github.com/ScandEval/ScandEval/issues/424", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2023886954
Enable Flash Attention 2 and rename TEXT_TO_TEXT task to SUMMARIZATION This PR: Adds a --use-flash-attention flag, which for instance allows benchmarking Mistral-based models. If the model has "mistral" in its name (case-insensitive) then this is automatically set. Renames the TEXT_TO_TEXT task to SUMMARIZATION, while keeping 'text-to-text' as supertask. This makes it slightly more intuitive, e.g., scandeval --dataset-task summarization -m gpt2. If we add an abstractive question-answering task in the future it would also use the 'text-to-text' supertask, making it use the same evaluation script, but also allowing different metrics and few-shot prompts, for instance. This PR replaces the PR #68 @KennethEnevoldsen I've also added you as a maintainer now, given all that you've helped with these PRs - is that okay with you, or would you prefer to stay informal? Thanks @saattrupdan, that is perfectly fine with me
gharchive/pull-request
2023-12-04T13:40:30
2025-04-01T04:55:36.663421
{ "authors": [ "KennethEnevoldsen", "saattrupdan" ], "repo": "ScandEval/ScandEval", "url": "https://github.com/ScandEval/ScandEval/pull/80", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
505732486
BarcodePicker dissapears right after showing without any errors Hi, I've tryied sample code and can't get it work. I've changed only type of barcodes I want to scan. Scandit.License.setAppKey('my license'); var settings = new Scandit.ScanSettings(); settings.setSymbologyEnabled(Scandit.Barcode.Symbology.CODE128, true); // Instantiate the barcode picker by using the settings defined above. var picker = new Scandit.BarcodePicker(settings); const success = function(session) { alert('Scanned ' + session.newlyRecognizedCodes[0].symbology + ' code: ' + session.newlyRecognizedCodes[0].data); // If you are using continuous scanning you might want to stop here. Please note that // you will have to use session.stopScanning()/session.pauseScanning() instead of the // corresponding method on the picker. This will avoid a race condition and immediately stop // the scanning process after the success callback has finished executing. session.stopScanning(); }; const manual = function(content) { console.log('Manual: ' + content); }; const failure = function(error) { console.log('Failed: ' + error); }; picker.show(success, manual, failure); Picker is opened, but is dissapeared in fraction of second. I ran the app in the following environment: com.mirasense.scanditsdk.plugin: 5.12.1 cordova 6.5 iOS 13.1.2 yes, unfortunately this is an issue when you use Xcode 11 and deploy to iOS 13, we actually fixed it yesterday. It will be released soon as part of 5.13. This is the MR that fixed it: https://github.com/Scandit/barcodescanner-sdk-cordova/pull/245 Hi! @lucatorella, when are you going to release it? We're expecting the stable release of 5.13 later this month, while you will be able to already access the beta release this week. Thank you! Any news about beta? 5.13.0BETA2 is already available from the Scandit dashboard. The source code is here GitHub in the 5.13 branch. Nice, will try it. Thanks! It works. Thanks! This issue can be closed, I believe. I'm glad the fix worked for you. Thanks for getting back.
gharchive/issue
2019-10-11T09:09:21
2025-04-01T04:55:36.669214
{ "authors": [ "abondarenko", "lucatorella", "petradonka" ], "repo": "Scandit/barcodescanner-sdk-cordova", "url": "https://github.com/Scandit/barcodescanner-sdk-cordova/issues/246", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
788768830
Webpanel polls too often. Also trello does not respect settings.HttpWait The webpanel polls too often which may cause the HttpService to get overloaded by betting over the 500 requests per minute. The delay should be 3 seconds and not wait(). wait() is too fast. Instead it should be wait(3) or wait(4) Make the trello polling seperate from the webpanel polling so it can respect settings.HttpWait Also because some games have many games servers and they all poll you could make it so only 1 server is polling and the others lisen to changes with MessagingService and coordinate the stuff trough a datastore. Then change the server which polls regularly for redundancy. And make it so that the trello and WebPanel check for changes and if they see no changes they will not fire the events and update any data. The webpanel uses long polling. As for trello make it respect it yourself, idk what to tell you for that The webpanel uses long polling. As for trello make it respect it yourself, idk what to tell you for that Oh. But do you think it would be a good idea ot have just one server to be polling and then send changed data trough MessagingService? Oh. But do you think it would be a good idea ot have just one server to be polling and then send changed data trough MessagingService? Oh. But do you think it would be a good idea ot have just one server to be polling and then send changed data trough MessagingService? Seems like that would be unnecessary. Assuming the panel is not being actively used by the game's creator, it is only sending one request per minute. Oh. But do you think it would be a good idea ot have just one server to be polling and then send changed data trough MessagingService? Seems like that would be unnecessary. Assuming the panel is not being actively used by the game's creator, it is only sending one request per minute. oki will make new issue about trello not respecting settings and will close this 1. oki will make new issue about trello not respecting settings and will close this 1. remade into #243 remade into #243
gharchive/issue
2021-01-19T07:09:47
2025-04-01T04:55:36.677229
{ "authors": [ "Cald-fan", "ccuser44", "joritochip" ], "repo": "Sceleratis/Adonis", "url": "https://github.com/Sceleratis/Adonis/issues/240", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1850377710
Fix 'Bad state: No element' - Empty Path Metrics I ran into the problem of 'computeMetrics' returning an empty list, getting the error 'Bad state: No element' because the code uses 'first' and 'last' getters, therefore I was getting a grey screen in a production build. Quick fix, to validate if the list is not empty, it probably requires some analysis as to why 'computeMetrics' is not returning any values. Thanks for the PR! It would be nice to know why the metrics are empty, but it's not good with a grey screen either :) Just a few nits, looks good otherwise Hi @Schwusch, thanks for taking a look, comments have been addressed and pushed. Perfect! I'll push a release as soon as possible Pushed to Pub: https://pub.dev/packages/widget_arrows/versions/0.6.0
gharchive/pull-request
2023-08-14T19:07:08
2025-04-01T04:55:36.712632
{ "authors": [ "HSCOGT", "Schwusch" ], "repo": "Schwusch/widget_arrows", "url": "https://github.com/Schwusch/widget_arrows/pull/18", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
983280918
add non-ascii character warning Temporary warning for #43 I would put the check and the warning in the constructor of the ODE object as one can pass non-ascii characters there directly
gharchive/pull-request
2021-08-30T23:08:12
2025-04-01T04:55:36.754544
{ "authors": [ "iliailmer", "pogudingleb" ], "repo": "SciML/StructuralIdentifiability.jl", "url": "https://github.com/SciML/StructuralIdentifiability.jl/pull/45", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
194313347
Add Matrix::vectorToDiag static factory Add a static factory like this one: /** * @param int[]|float[] $vec */ public static function vectorToDiag(array $vec) : Matrix example: /** * $m == [ * [4, 0, 0, 0], * [0, 3, 0, 0], * [0, 0, 2, 0], * [0, 0, 0, 1], * ]; */ $m = IntMatrix::vectorToDiag([4, 3, 2, 1]); done
gharchive/issue
2016-12-08T11:34:40
2025-04-01T04:55:36.755983
{ "authors": [ "castarco" ], "repo": "SciPHPy/php-sds-polyfill", "url": "https://github.com/SciPHPy/php-sds-polyfill/issues/31", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1186432975
MeshCoord metadata should include the "mesh" property We initially chose to leave the 'mesh' property out of the MeshCoordMetadata, largely because at that time we were wary of implementing equivalence on Mesh objects. As noted here IIRC this was largely due to concerns about the efficiency of calculating mesh equality (i.e. potentially comparing large arrays), but we since essentially changed our minds about this, and implemented full equality testing on meshes anyway So, in all probability, it would make more sense now to reverse this, and make 'mesh' a participant in MeshCooord.metadata Present effects With same 'location' and 'axis' but different 'mesh', we can find that meshcoord1 != meshcoord2, while at the same time meshcoord1.metadata == meshcoord2.metadata. This is not a logical flaw (can be true for coords with different data), but we would expect mesh to participate + it does not. Arguably, metadata comparison should work as an 'identity equivalence' for data components, this allows e.g. co = cube1.coord('longitude') ; co2 = cube2.coord(co) to work even when the 2 coords in question have different data, shapes or dtypes. This aspect is basically not working for meshcoords, at this point. The code here would not be reachable, or would not be needed, if this worked as it does for the other coordinate types -- instead, the resolve operation "ought to" be able to decide whether 2 meshcoords can be treated as equivalent or not. It might also make sense to permit some aspects of lenient mesh comparison within this -- e.g. if ~equivalent meshes from different files had component with different var_names -- which is not possible at present. At present, cubes can be merged between meshes that compare equal, but it requires an exact match in all details (e.g. including var_names) Similar considerations could also apply to merge+concatenate. Effects of change It's not clear, if we simply change this, what knock-on problems could be introduced. Possibly, merging 'equal' meshes, loaded from different files, would stop working (it currently works). Up to now, no other metadata component has been a complex object -- the most complex is attribute dictionaries, which themselves require special consideration in comparison/combination. After some thought TL;DR: I now think we shouldn't do this after all. I think metadata is generally to be viewed as analagous to "phenomenon" As on cubes or coordinates, for instance .. it distinguishes one thing from another, where directly combining/comparing them would be a categorical mistake, like a units error (i.e. like adding apples to elephants) it characterises the "meaning" of the numbers stored the object, e.g. the numbers are "a longitude" or "a height in metres". so in that way, it's a bit like an extended units concept So for a cube, the metadata is those characteristics which give it a distinct identity as a "type of value" -- but not the values themselves (.data), their type (.dtype) or structure (.shape) I think with that in mind, adding a mesh to the MeshCoordMetadata would be like adding the coords or coord-system to a CubeMetadata : It would mean that changing these makes an entirely different phenomenon, which can't be compared with original. So, at present, this is mostly used in 3 places: in comparison (==): But here, it is just a fast check, as non-metadata properties are generally also compared anyway in arithmetic : but this doesn't really apply to MeshCoords in selecting objects ... read on ... At present, component access methods like Cube.coords use metadata as an identity, so you can ... co = cube1.coord(name) meta = co.metadata assert cube2.coord(co).metadata == cube2.coord(meta).metadata == meta That is, you can get the equivalent coord in a different cube by using the coord (or equivalently, in fact, its metadata) as a reference. This works even when the 2 cubes have different dimensions or coordinate values, because those aspects are outside the metadata. At present, we can do the same with MeshCoords.. If we adde @stephenworsley please "review", i.e. see what you think of the above ideas. Do you agree that we should close this ? @pp-mo This makes sense, I agree that the status of the mesh property should stay as is given that: metadata is a "lightweight" description and meshes are large. mesh already exists as a property of MeshCoords. meshes are more similar to objects which are not considered suitable for including in other metadata (e.g. coords, data) than they are to objects included in metadata (e.g. names, attributes). adding meshes to metadata would break some existing ehaviour. Ok, thanks @stephenworsley . I think we are done with this. Anyone disagreeing can re-open !
gharchive/issue
2022-03-30T13:05:42
2025-04-01T04:55:36.782253
{ "authors": [ "pp-mo", "stephenworsley" ], "repo": "SciTools/iris", "url": "https://github.com/SciTools/iris/issues/4673", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1877620284
whatsnew for #5331 🚀 Pull Request Description Follows #5331. Consult Iris pull request check list Renders like this https://scitools-iris--5458.org.readthedocs.build/en/5458/whatsnew/index.html#bugs-fixed
gharchive/pull-request
2023-09-01T15:22:59
2025-04-01T04:55:36.784939
{ "authors": [ "rcomer" ], "repo": "SciTools/iris", "url": "https://github.com/SciTools/iris/pull/5458", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1911800915
[Bug] Description 打开托盘无效 Reproduction Steps 设置页面打开托盘,在系统托盘区无软件图标 Expected Behavior No response Screenshots Starward Version 0.10.0 Windows Version 22621.2283 Log fail: Starward.Services.SystemTrayService[0] Initialize system tray System.InvalidOperationException: TryCreate failed. at H.NotifyIcon.Core.TrayIcon.Create() in /_/src/libs/H.NotifyIcon/Core/TrayIcon.cs:line 324 at H.NotifyIcon.TaskbarIcon.ForceCreate(Boolean ) in /_/src/libs/H.NotifyIcon.Shared/TaskbarIcon.cs:line 134 at Starward.Services.SystemTrayService.Initialize(GameBiz gameBiz) in D:\a\Starward\Starward\src\Starward\Servi Additional Context No response Closed for no title or content.
gharchive/issue
2023-09-25T15:44:35
2025-04-01T04:55:36.788012
{ "authors": [ "Endless2333", "Scighost" ], "repo": "Scighost/Starward", "url": "https://github.com/Scighost/Starward/issues/403", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1679899772
yumi-exfat: Update to version 1.0.1.6a and fix checkver yumi-exfat: couldn't match 'YUMI-exFAT-([\d.]+).exe' in https://www.pendrivelinux.com/yumi-multiboot-usb-creator/ [X] I have read the Contributing Guide. /verify
gharchive/pull-request
2023-04-23T07:21:38
2025-04-01T04:55:36.844729
{ "authors": [ "StarsbySea" ], "repo": "ScoopInstaller/Extras", "url": "https://github.com/ScoopInstaller/Extras/pull/11108", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
914600107
How to render the Signals and Axes at different quality ScottPlot Version: 4.1.16 Operating System: W10 Application Type: WinForms Question: Is there a way to render the Signals and Axes at different quality? For instance, I would like to render the Signals at Low quality always (for speed purposes), while the Axes at High quality. Basically, to be able to configure the plot something like this: formsPlot1.Configuration.QualitySignals = ScottPlot.Control.QualityMode.Low; formsPlot1.Configuration.QualityOther = ScottPlot.Control.QualityMode.High; I am not sure if it is possible, because I can imagine that anti-aliasing is applied on the whole image, but I decided to still ask maybe there is a posibility for it. Hi Scott, I just came acrros one of your videos :) very impressive what you did. In the video (at least to me) it seems like the Axes ticks have High quality, while the Data is displayed at Low quality (no anti-aliasing). This is exactly what I meant in this topic. Please correct me if I am wrong in observing the video. This is a minor topic. I will close.
gharchive/issue
2021-06-08T08:42:47
2025-04-01T04:55:36.888529
{ "authors": [ "EmanuelFeru" ], "repo": "ScottPlot/ScottPlot", "url": "https://github.com/ScottPlot/ScottPlot/issues/1114", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1221632234
Manual tick spacing alongside automatic Feature Suggestion Feature goes with PR #1814 Feature description: I'd like to be able to manually label certain major ticks, but keep the automatic ticks as well. Code example: // a code sample may improve communication var plt = new ScottPlot.Plot(); double[] myCustomTicks = { Math.PI, 2 * Math.PI }; string[] myCustomLabels = { "π", "2π" }; plt.XAxis.ManualTickPositions(myCustomTicks, myCustomLabels, union: true); Example implementation from #1814: @swharden I decided to remove the closest tick to each custom tick so that the labels didn't overlap. Some users may not want this, though. Should I add a flag for users who want to keep all labels? Or maybe an enum? Thanks for this @Xerxes004! Following-up, I modified your #1814 so that this feature is an addition to automatic ticks (rather than a third type of ticks). Users can now enable automatic ticks like this: plt.XAxis.AutomaticTickPositions(); and if they want to add additional manual ticks which will always be displayed they can now do this: double[] positions = { Math.PI, 2 * Math.PI }; string[] labels = { "π", "2π" }; plt.XAxis.AutomaticTickPositions(positions, labels);
gharchive/issue
2022-04-29T22:29:42
2025-04-01T04:55:36.892317
{ "authors": [ "Xerxes004", "swharden" ], "repo": "ScottPlot/ScottPlot", "url": "https://github.com/ScottPlot/ScottPlot/issues/1815", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1838462372
SP5: consider cutting dependency on System.Drawing.Primitives it doesn't seem to be used much anyway Ha! It's not used once 🤦
gharchive/issue
2023-08-06T23:36:12
2025-04-01T04:55:36.893519
{ "authors": [ "swharden" ], "repo": "ScottPlot/ScottPlot", "url": "https://github.com/ScottPlot/ScottPlot/issues/2844", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1647591208
Convert WpfPlot to CustomControl Purpose: Use Control as base class for WpfPlot instead of UserControl to enable customization. (cf #2509) add property to show / hide legend add property to define legend position add property to select traces palette add property to select plot style remove dead code It is now possible to select WpfPlot palette and style directly from XAML This control replace the previous without any change on existing XAML or C# code. NOTE I could not change the ScottPlot5 version, because I am not able to load the solution (cf #2525) Hi @KroMignon, thanks for this PR! I won't be able to review it thoroughly until tonight, but I'm happy to see you figured out the PR process and the tests are passing and I look forward to merging this soon! 🚀 I have one bit of feedback about a feature that was added: It is now possible to select WpfPlot palette and style directly from XAML I'm extremely hesitant to allow plot customizations from XAML. There are potentially thousands of little customizations that could be added (e.g., axis tick mark length), and if we allow some it will open the door for others to request more. Also, if we add XAML customizations to WPF then I would think we need to add the same to WinUI, Avalonia, Eto, Blazor, etc., and the maintenance burden grows as the number of supported controls grows. Although I think what you did here with properties and XAML is elegant and would be an ideal solution if ScottPlot were a WPF-only control, for the sake of maintainability I would feel better not customizing plots using styling, but instead interacting with the Plot object as demonstrated by the cookbook. However, I custom XAML customizations are the type of thing that will be enabled by the new inherit-and-extend functionality that the core of this PR now allows 😉 Thanks again for your work on this, and I'm looking forward to hearing your thoughts! Scott Good morning @swharden I can now open the ScottPlot 5 solution and I have started to look at the code. I have first to understand how the WPF control has been design before I implement my modification. SkiaSharp is new to me. I will look at this more closely this week-end, I've got some ideas how it could be more WPF / MVVM "friendly" but it needs some tests. Is it okay if I do some modification on ScottPlot.Control.Interaction, I am not sure I have to, but maybe 🤔 Hi @KroMignon, thanks again for your work on this! To try to keep this PR clean and isolated to a single purpose I reverted some of your changes related to styling enhancements. This makes it easier to do a line-by-line evaluation of https://github.com/ScottPlot/ScottPlot/pull/2526/files I don't have a lot of experience with WPF. Can you advise us about the purpose of Generic.xaml and whether it is required? It seems like we're pretty close to having a XAML-free control which makes me really happy. Is this functionality required and/or could it be performed programmatically? Thanks for your input! I think we're getting pretty close to being ready to merge this in 🚀 Hi @KroMignon, thanks again for your work on this! To try to keep this PR clean and isolated to a single purpose I reverted some of your changes related to styling enhancements. This makes it easier to do a line-by-line evaluation of https://github.com/ScottPlot/ScottPlot/pull/2526/files That's fine for me 👍 I don't have a lot of experience with WPF. Can you advise us about the purpose of Generic.xaml and whether it is required? It seems like we're pretty close to having a XAML-free control which makes me really happy. Is this functionality required and/or could it be performed programmatically? A WPF Control requires at least a default style, which will be loaded by OnApplyTemplate(). In the style, we will define the content of the control. If there is no default style, the component will be empty/transparent. For this control, the style contains a picture (to show the graph) and a label (to show the error messages). In fact, it is the same structure as the previous UserControl. The name of those items (picture and label) is important to be able to extract them from the style, with Template.FindName(). The difference with a UserControl, is that now it is possible to customize the control, which is not possible with a UserControl (cf. https://www.wpftutorial.net/CustomVsUserControl.html) Thanks for your input! I think we're getting pretty close to being ready to merge this in 🚀 Your welcome 😉 EDIT: After this is merged we can revisit styling and palette improvements in a separate issue/PR Thanks for the clarification @KroMignon! I read some more information about custom controls and WPF templates and it makes a lot more sense now - thanks for pointing me in the right direction. I plan to release a new package today on NuGet, but I'm still hesitant to merge this in. I found one bug that I'll try to fix now, and depending on how easy it is to correct and/or how many other bugs I find, I may push this back until I have more time to review and test it thoroughly. Here's the bug I found this morning. I suspect it relates to events not being handled as expected... <Window x:Class="WpfApp.MainWindow" xmlns="http://schemas.microsoft.com/winfx/2006/xaml/presentation" xmlns:x="http://schemas.microsoft.com/winfx/2006/xaml" xmlns:d="http://schemas.microsoft.com/expression/blend/2008" xmlns:mc="http://schemas.openxmlformats.org/markup-compatibility/2006" xmlns:local="clr-namespace:WpfApp" mc:Ignorable="d" xmlns:ScottPlot="clr-namespace:ScottPlot;assembly=ScottPlot.WPF" Title="ScottPlot Sandbox - WPF" WindowStartupLocation="CenterScreen" Height="450" Width="800"> <Grid> <Grid.RowDefinitions> <RowDefinition /> <RowDefinition /> </Grid.RowDefinitions> <ScottPlot:WpfPlot Name="WpfPlot1" Grid.Row="0"/> <ScottPlot:WpfPlot Name="WpfPlot2" Grid.Row="1"/> </Grid> </Window> using System.Windows; namespace WpfApp { public partial class MainWindow : Window { public MainWindow() { InitializeComponent(); WpfPlot1.Plot.AddSignal(ScottPlot.DataGen.Sin(51, 2)); WpfPlot2.Plot.AddSignal(ScottPlot.DataGen.Cos(51, 2)); WpfPlot1.Refresh(); WpfPlot2.Refresh(); WpfPlot1.Configuration.AddLinkedControl(WpfPlot2); // update plot 2 when plot 1 changes WpfPlot2.Configuration.AddLinkedControl(WpfPlot1); // update plot 1 when plot 2 changes } } } main branch this PR Maybe this is a niche case, because this works... so I'm guessing it's some type of difference in window size evaluation inside the initializer? The size of the control is evaluated to determine how large to make the plot area. using System.Windows; namespace WpfApp { public partial class MainWindow : Window { public MainWindow() { InitializeComponent(); Loaded += (s, e) => { WpfPlot1.Plot.AddSignal(ScottPlot.DataGen.Sin(51, 2)); WpfPlot2.Plot.AddSignal(ScottPlot.DataGen.Cos(51, 2)); WpfPlot1.Refresh(); WpfPlot2.Refresh(); WpfPlot1.Configuration.AddLinkedControl(WpfPlot2); WpfPlot2.Configuration.AddLinkedControl(WpfPlot1); }; } } } However, it's these type of unexpected changes in behavior that concern me considering how many existing users have code in the wild that depends on the functionality of WpfPlot to remain stable 😅 Looking at the original WpfPlot control, OnSizeChanged is not invoked by the main window or the image, but the grid... I wonder if this difference is what is causing the discrepancy https://github.com/ScottPlot/ScottPlot/blob/065383a3b6182e35edbe39c2606528bd147bb013/src/ScottPlot4/ScottPlot.WPF/WpfPlot.xaml#L22 I'm probably going to have to put this down for a while because I don't want it to consume a disproportionate amount of time considering the other issues I'm trying to clear today 😅 Let me know if you have any thoughts or suggestions for how to track this down, otherwise I'll pick it up again in a day or two! Thanks again for your work on this Looking at the original WpfPlot control, OnSizeChanged is not invoked by the main window or the image, but the grid... I wonder if this difference is what is causing the discrepancy https://github.com/ScottPlot/ScottPlot/blob/065383a3b6182e35edbe39c2606528bd147bb013/src/ScottPlot4/ScottPlot.WPF/WpfPlot.xaml#L22 I am not a WPF specialist, so maybe I am wrong. You cannot use the OnSizeChanged from the Image itself, because it depends on the load image. I have used the size of the control itself, because the Grid should follow the size of the control. It looks like I missed something 😞. I'm probably going to have to put this down for a while because I don't want it to consume a disproportionate amount of time considering the other issues I'm trying to clear today 😅 Let me know if you have any thoughts or suggestions for how to track this down, otherwise I'll pick it up again in a day or two! Thanks again for your work on this As I am really new to PR, could you please tell me how I can import the changes you have done into my clone (I create a branch to test the changes)? Or what would be the best way to continue with your changes? I would look at this more closely tomorrow. Perhaps you should way before merge this change to the main branch. It looks like I missed something No worries! There are tons of little corner cases and it's hard to catch everything. The things that I frequently miss are when I make a change that breaks support for display scaling, because that's really difficult to test 😅 As I am really new to PR, could you please tell me how I can import the changes you have done into my clone (I create a branch to test the changes)? I recommend using the GitHub Desktop app. It's really easy to use (command line git and git inside Visual Studio can both be pretty confusing) https://desktop.github.com/ Using the GitHub desktop you can open this repository, switch to the branch for this PR, and click "fetch" or "pull" to bring the latest code onto your computer. I recognize dealing with multi-author branches and pull requests is pretty confusing at first, but once you figure it out once it gets a lot easier! 🚀 Let me know if you have any thoughts or suggestions for how to track this down, otherwise I'll pick it up again in a day or two! Thanks again for your work on this I think I found what's going wrong after reading Microsoft documentation. The SizeChanged event is not the best choice, it should be replaced with LayoutUpdated, because every time LayoutUpdated is emitted ActualHeight and ActualWidth are valid. This is not always the case with SizeChanged event. The SizeChanged event ... should be replaced with LayoutUpdated I gave this a quick try, but the control crashes maybe because of some type of infinite loop 😅 Hello @swharden, I made some changes to try to fix the issue with WPF Sandbox app. Now it looks good for me 🚀 😄 Hope it is okay for you: little change in Backend to avoid creating a new Bitmap when resizing is done with same size update WpfPlot to override Control function instead of using events WpfPlot: always create Backend instance with 1x1size to ensure a Bitmap will be created and then resized. This solves the issue. Fantastic work @KroMignon! That was a complicated behavior to figure out. I'm excited to try this out and merge it in! 🚀 This works great - merging it in now! I made one small change so the scroll wheel respects direction: https://github.com/ScottPlot/ScottPlot/pull/2526/commits/8d3c065fe247ccb6c11bf2fa402bcd944a579ada I learned a lot about WPF by studying the topics discussed above - thanks again for the initial suggestion and creating the PR! Very good news this morning 🎉, I am excited to use new ScottPlotV4 release ! I am very happy I could help, and will starts soon a new PR for ScottPlotV5, this should be done faster... I hope 🙈
gharchive/pull-request
2023-03-30T13:08:49
2025-04-01T04:55:36.920631
{ "authors": [ "KroMignon", "swharden" ], "repo": "ScottPlot/ScottPlot", "url": "https://github.com/ScottPlot/ScottPlot/pull/2526", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
119735985
FEATURE: Optimizer creates repairing policies in the TOSCA accepted by the Deployer This PR makes the Optimizer module create repairing policies in the syntax given in Section 7.3 of this document https://docs.google.com/document/d/1Z9RdaWn54IomgI-jCiw-tdutsCdDoHfk67PNUASOFmY/edit# . There is also a bit of refactoring of previous repeated code. In order to have a finished way to create these policies, for the next PRs it is still required to specify which modules can scale. Following the last directives, I will create the documentation of the module description, usage and run (I do not add it now due to lack of time as somebody may want to see the code for discussion in the next meeting). Thanks @perezp. Could you add some test to check the policies generation? I'm writing tests for the generated policies. Thanks @kiuby88 Current coverage is 32.14% Branch #170 has no coverage reports uploaded yet. No diff could be generated. No reports for master found. Powered by Codecov. Updated on successful CI builds. @kiuby88 I addressed your comments and I solved a bug which made travis fail. There are now a lot of commits, do you want me to squash now or after you check the comments and changes? +1 Thanks for the review @kiuby88 . Merging....
gharchive/pull-request
2015-12-01T15:07:04
2025-04-01T04:55:36.945438
{ "authors": [ "codecov-io", "kiuby88", "perezp" ], "repo": "SeaCloudsEU/SeaCloudsPlatform", "url": "https://github.com/SeaCloudsEU/SeaCloudsPlatform/pull/170", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1277773455
Does sea-orm-cli generate MigrationName by current datetime, not Fixed Date code link:https://github.com/SeaQL/sea-orm/blob/fff738a706d8097bb2b92d7a1fad3b67c2cde653/sea-orm-cli/src/commands.rs#L232 Now: Suggestion: Hey @baoyachi, sorry for the delay. Good suggestions! Actually, it's very similar to this feature: creating new migration file with filename of current date time https://github.com/SeaQL/sea-orm/pull/656 This line, to be exact: https://github.com/SeaQL/sea-orm/pull/656/files#diff-0ad0c0489cc7359f48740c645608cad9aaf0318787c7d1b7fc0910828f2429e4R234 Btw... with this PR landed. We no longer has to specify name of the migration explicitly https://github.com/SeaQL/sea-orm/pull/736 So, the template now becomes: https://github.com/SeaQL/sea-orm/blob/beef8211d82567858a9ea403d3cb7e54176fa92a/sea-orm-cli/template/migration/src/m20220101_000001_create_table.rs#L3-L4 I'll open this for anyone who feel interested to contribute :) I want try. Please go ahead! Thanks in advance! Feel free to ping me here or at Discord if you need help. Just open a draft PR and ask for comments / review :) Hey @baoyachi, I'll assign this to you
gharchive/issue
2022-06-21T03:20:26
2025-04-01T04:55:36.950865
{ "authors": [ "baoyachi", "billy1624" ], "repo": "SeaQL/sea-orm", "url": "https://github.com/SeaQL/sea-orm/issues/806", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1038515322
EOS-25262 Merging Main branch to Kubernetes branch for hare fixes and verification Adding the important commits present in main branch to kubernetes branch Rebased with latest kubernetes branch Testing Testing done with Motr's branch - https://github.com/Seagate/cortx-motr/tree/kubernetes_main in 3 Node deployment in containers custom image - http://eos-jenkins.colo.seagate.com/job/GitHub-custom-ci-builds/job/centos-7.9/job/cortx-all-image-custom-ci/1033/parameters/ (build 1033) motr branch - https://github.com/Seagate/cortx-motr/tree/kubernetes_main Deployed docker image [root@ssc-vm-g2-rhev4-1756 ~]# docker images ghcr.io/seagate/cortx-all REPOSITORY TAG IMAGE ID CREATED SIZE ghcr.io/seagate/cortx-all 2.0.0-1033-custom-ci 778a9ea53044 4 hours ago 1.91GB kubectl get pods [root@ssc-vm-g2-rhev4-1756 ~]# kubectl get pods NAME READY STATUS RESTARTS AGE consul-6v5t2 1/1 Running 0 4h5m consul-bnzzj 1/1 Running 0 4h5m consul-gkqlm 1/1 Running 0 4h5m consul-server-0 1/1 Running 0 4h5m consul-server-1 1/1 Running 0 4h5m consul-server-2 1/1 Running 0 4h5m cortx-control-pod-7c64974dd6-94p4p 4/4 Running 34 3h57m cortx-data-pod-ssc-vm-g2-rhev4-1734-65887b8fc9-rghzp 9/9 Running 0 3h57m cortx-data-pod-ssc-vm-g2-rhev4-1756-65f6f9fddc-zrpm7 9/9 Running 0 3h57m cortx-data-pod-ssc-vm-g2-rhev4-1913-66df965968-rtcj4 9/9 Running 0 3h57m gluster-ssc-vm-g2-rhev4-1756-7c99c7fc4d-47xz2 1/1 Running 0 4h3m kafka-0 1/1 Running 0 4h4m kafka-1 1/1 Running 0 4h4m kafka-2 1/1 Running 0 4h4m openldap-0 1/1 Running 0 4h5m openldap-1 1/1 Running 0 4h5m openldap-2 1/1 Running 0 4h5m zookeeper-0 1/1 Running 0 4h4m zookeeper-1 1/1 Running 0 4h4m zookeeper-2 1/1 Running 0 4h4m All the services are up and running. S3 services are stable. [root@cortx-data-headless-svc-ssc-vm-g2-rhev4-1734 /]# hctl status Data pool: # fid name 0x6f00000000000001:0x75 'storage-set-1__sns' Profile: # fid name: pool(s) 0x7000000000000001:0xba 'Profile_the_pool': 'storage-set-1__sns' 'storage-set-1__dix' None Services: cortx-data-headless-svc-ssc-vm-g2-rhev4-1756 (RC) [started] hax 0x7200000000000001:0x2d inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1756@2001 [started] ioservice 0x7200000000000001:0x30 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1756@3001 [started] ioservice 0x7200000000000001:0x3b inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1756@3002 [started] confd 0x7200000000000001:0x46 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1756@3003 [started] s3server 0x7200000000000001:0x49 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1756@4001 [started] s3server 0x7200000000000001:0x4c inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1756@4002 cortx-data-headless-svc-ssc-vm-g2-rhev4-1913 [started] hax 0x7200000000000001:0x7 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1913@2001 [started] ioservice 0x7200000000000001:0xa inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1913@3001 [started] ioservice 0x7200000000000001:0x15 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1913@3002 [started] confd 0x7200000000000001:0x20 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1913@3003 [started] s3server 0x7200000000000001:0x23 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1913@4001 [started] s3server 0x7200000000000001:0x26 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1913@4002 cortx-data-headless-svc-ssc-vm-g2-rhev4-1734 [started] hax 0x7200000000000001:0x53 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1734@2001 [started] ioservice 0x7200000000000001:0x56 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1734@3001 [started] ioservice 0x7200000000000001:0x61 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1734@3002 [started] confd 0x7200000000000001:0x6c inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1734@3003 [started] s3server 0x7200000000000001:0x6f inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1734@4001 [started] s3server 0x7200000000000001:0x72 inet:tcp:cortx-data-headless-svc-ssc-vm-g2-rhev4-1734@4002 [root@cortx-data-headless-svc-ssc-vm-g2-rhev4-1734 /]# Performed S3 IOs successfully Created 2 files - 10MB 100MB Uploaded to the bucket and downloaded and verified. [root@ssc-vm-g2-rhev4-1756 ~]# aws s3 ls [root@ssc-vm-g2-rhev4-1756 ~]# aws s3 mb s3://mybucket make_bucket: mybucket [root@ssc-vm-g2-rhev4-1756 ~]# dd if=/dev/zero of=file10mb bs=1M count=10 10+0 records in 10+0 records out 10485760 bytes (10 MB) copied, 0.038858 s, 270 MB/s [root@ssc-vm-g2-rhev4-1756 ~]# aws s3 cp file10mb s3://mybucket/file10MB upload: ./file10mb to s3://mybucket/file10MB [root@ssc-vm-g2-rhev4-1756 ~]# aws s3 ls s3://mybucket/ 2021-10-29 05:23:05 10485760 file10MB [root@ssc-vm-g2-rhev4-1756 ~]# aws s3 cp s3://mybucket/file10MB file10mbDn download: s3://mybucket/file10MB to ./file10mbDn [root@ssc-vm-g2-rhev4-1756 ~]# diff file10mb file10mbDn [root@ssc-vm-g2-rhev4-1756 ~]# dd if=/dev/zero of=file100mb bs=10M count=10 10+0 records in 10+0 records out 104857600 bytes (105 MB) copied, 0.238556 s, 440 MB/s [root@ssc-vm-g2-rhev4-1756 ~]# ls cortx-all-1009.tar cortx-all-1033.tar cortx-k8s cortx-re file100mb file10mb file10mbDn get_helm.sh [root@ssc-vm-g2-rhev4-1756 ~]# aws s3 cp file100mb s3://mybucket/file100MB upload: ./file100mb to s3://mybucket/file100MB View rendered README.md View rendered Testing_Verification_History.md View rendered ha-simulator/README.md @mssawant please review and lets see if we can merge
gharchive/pull-request
2021-10-28T13:24:57
2025-04-01T04:55:36.960094
{ "authors": [ "Shreya-18", "vaibhavparatwar" ], "repo": "Seagate/cortx-hare", "url": "https://github.com/Seagate/cortx-hare/pull/1855", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1498910624
🛑 IPFS Gateway 1/2 is down In 0cc608d, IPFS Gateway 1/2 (https://ipfs.seanstaffiery.com) was down: HTTP code: 0 Response time: 0 ms Resolved: IPFS Gateway 1/2 is back up in 383cef4.
gharchive/issue
2022-12-15T18:38:09
2025-04-01T04:55:37.006787
{ "authors": [ "SeanStaffiery" ], "repo": "SeanStaffiery/status.seanstaffiery.com", "url": "https://github.com/SeanStaffiery/status.seanstaffiery.com/issues/1021", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2291072886
Belindas-closet-nextjs_6_369_enable-auto-assign-issues Resolves #369 Once merged, this should auto-assign developers to the issues they create for this repo. Looks good to me. Thanks for the work on this issue. Looks good to me. Thanks, Taylor This will be a great feature enhancement added to the project. Looks great Great job! This feature is very convenient! I found an interesting link too: https://github.com/andrewlock/auto-assign-issues
gharchive/pull-request
2024-05-11T21:35:31
2025-04-01T04:55:37.021848
{ "authors": [ "Diego-Cano", "MaddasaHatter", "MuhammadNSC", "kaimanasse15", "taylorpapke" ], "repo": "SeattleColleges/belindas-closet-nextjs", "url": "https://github.com/SeattleColleges/belindas-closet-nextjs/pull/370", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2535733670
undefined field 'tags' Thanks for the info. tags was missing in the type definition. I added them to the latest release.
gharchive/issue
2024-09-19T09:00:59
2025-04-01T04:55:37.023507
{ "authors": [ "Sebaestschjin", "sulfur17" ], "repo": "Sebaestschjin/typed-tabletop-simulator", "url": "https://github.com/Sebaestschjin/typed-tabletop-simulator/issues/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
191589899
Refactor suggestions about @Input and implementing a custom decorator The code is currently using inputs array of strings on component metadata to tell Angular that properties with these names are to be considered inputs. This way of listing such properties is currently considered bad practice according to official Angular styleguide. It makes sense: maintaining is difficult because it's scattered around, and let's not even begin on typos and refactoring. So my first suggestion is moving away from inputs and using @Input() instead. In a similar fashion, we're maintaining an array of strings _mapOptionsAttributes in order to iterate over them more easily. This can be solved with a custom decorator; here's a proposed simple implementation. function MapOptionAttribute() { return function (target: SebmGoogleMap, key: string) { if (!target.constructor._mapOptionsAttributes) { target.constructor._mapOptionsAttributes = []; } target.constructor._mapOptionsAttributes.push(key); } } And example usage: class SebmGoogleMap { @Input() @MapOptionAttribute() public longitude: number; // We basically need this just for typing information in here. // The actual value is assigned through the decorator. private static _mapOptionsAttributes: string[]; } The accompanying JS bin in action is also available. How does it seem? Any suggestions? @SebastianM, if you like the idea, I could work out a PR. It was mentioned in https://github.com/SebastianM/angular2-google-maps/pull/863#issuecomment-274323914 that a reason for not doing this is being unable to generate docs. There doesn't seem to be a proper document generation tool for Angular (yet). However, ng-bootstrap has a custom script for generating docs which is probably worth checking out.
gharchive/issue
2016-11-24T20:53:59
2025-04-01T04:55:37.030164
{ "authors": [ "lazarljubenovic" ], "repo": "SebastianM/angular2-google-maps", "url": "https://github.com/SebastianM/angular2-google-maps/issues/767", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
100288246
Weirdness with sampler Dear Sebastian, I seem to be having a difficulty with the Bingham sampler-- it always produces the same sample (which is a degenerate sample). I am going to include some example code and then the output. I compiled with "-lbingham -lm" and nothing else interesting. First, the code: #include <stdio.h> #include <stdlib.h> #include <string.h> #include <math.h> #include "bingham.h" #include "bingham/util.h" #include "bingham/bingham_constants.h" #include "bingham/hypersphere.h" int main(){ // create n 4-D unit vectors int n = 10,i,j; bingham_t B_true; /* Make a non-uniform Bingham distribution */ double Z[3] = {1, 2, 5}; double V[3][4] = {{1,0,0,0}, {0,0,1,0}, {0,0,0,1}}; double *Vp[3] = {&V[0][0], &V[1][0], &V[2][0]}; bingham_new(&B_true, 4, Vp, Z); /* Print out some details */ printf("B_true:\n Z:"); for (i=0;i<3;i++){ printf(" %.3f",B_true.Z[i]); } printf("\n"); /* Draw some samples from that distribution */ double **X = new_matrix2(n,4); bingham_sample(X, &B_true, n); /* Print the first few samples */ for (i=0;i<3;i++){ printf(" Sample %d: ",i); for (j=0;j<4;j++){ printf(" %.3f",X[i][j]); } printf("\n"); } /* cleanup */ free_matrix2(X); bingham_free(&B_true); return(0); } Next, the output: B_true: Z: 1.000 2.000 5.000 ********* seed = 1439290056 d_off = nan, d_diag = nan Sample 0: 0.000 1.000 0.000 0.000 Sample 1: 0.000 1.000 0.000 0.000 Sample 2: 0.000 1.000 0.000 0.000 The concentration parameters Z should be in the range [-900,0]. 0 represents a uniform distribution, towards -900 the distribution gets more concentrated and peaked. Can you try correct Z parameters and report back? On August 11, 2015 1:02:15 PM CEST, wfbradley notifications@github.com wrote: Dear Sebastian, I seem to be having a difficulty with the Bingham sampler-- it always produces the same sample (which is a degenerate sample). I am going to include some sample code and then the output. I compiled with "-lbingham -lm" and nothing else interesting. First, the code: #include <stdio.h> #include <stdlib.h> #include <string.h> #include <math.h> #include "bingham.h" #include "bingham/util.h" #include "bingham/bingham_constants.h" #include "bingham/hypersphere.h" int main(){ // create n 4-D unit vectors int n = 10,i,j; bingham_t B_true; /* Make a non-uniform Bingham distribution */ double Z[3] = {1, 2, 5}; double V[3][4] = {{1,0,0,0}, {0,0,1,0}, {0,0,0,1}}; double *Vp[3] = {&V[0][0], &V[1][0], &V[2][0]}; bingham_new(&B_true, 4, Vp, Z); /* Print out some details */ printf("B_true:\n Z:"); for (i=0;i<3;i++){ printf(" %.3f",B_true.Z[i]); } printf("\n"); /* Draw some samples from that distribution */ double **X = new_matrix2(n,4); bingham_sample(X, &B_true, n); /* Print the first few samples */ for (i=0;i<3;i++){ printf(" Sample %d: ",i); for (j=0;j<4;j++){ printf(" %.3f",X[i][j]); } printf("\n"); } /* cleanup */ free_matrix2(X); bingham_free(&B_true); return(0); } Next, the output: B_true: Z: 1.000 2.000 5.000 ********* seed = 1439290056 d_off = nan, d_diag = nan Sample 0: 0.000 1.000 0.000 0.000 Sample 1: 0.000 1.000 0.000 0.000 Sample 2: 0.000 1.000 0.000 0.000 Reply to this email directly or view it on GitHub: https://github.com/SebastianRiedel/bingham/issues/7 Dear Sebastian, Thank you, that fixed it! For future reference, how does the code define a Bingham distribution (i.e. "V" and "Z")? Based on the parameterizations I am more familiar with, I was assuming that Z[i]>=0 and V[i][] were unit length, both of which appear to be false. Is there some document I should be reading to figure this stuff out? The code is originally Jared Glover's. He describes the convention and code in his Bingham related publications. You can easily find them in Google scholar. Fyi, the V vectors are unit vectors. On August 11, 2015 4:43:49 PM CEST, wfbradley notifications@github.com wrote: Dear Sebastian, Thank you, that fixed it! For future reference, how does the code define a Bingham distribution (i.e. "V" and "Z")? Based on the parameterizations I am more familiar with, I was assuming that Z[i]>=0 and V[i][] were unit length, both of which appear to be false. Is there some document I should be reading to figure this stuff out? Reply to this email directly or view it on GitHub: https://github.com/SebastianRiedel/bingham/issues/7#issuecomment-129912448 Ah, of course. For other readers, a definition can be found on page two of Tracking the Spin on a Ping Pong Ball with the Quaternion Bingham Filter. I'd copy the definition, but I don't know how to write the math with Git's markdown format. And on more careful examination, the V vectors are unit length, as you mentioned; I'd made a mistake interpreting the precision.
gharchive/issue
2015-08-11T11:02:13
2025-04-01T04:55:37.040518
{ "authors": [ "SebastianRiedel", "wfbradley" ], "repo": "SebastianRiedel/bingham", "url": "https://github.com/SebastianRiedel/bingham/issues/7", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
1585891530
Installing module causes database error Trying to install this on our server causes: AppApi: Error creating db-tables: SQLSTATE[42000]: Syntax error or access violation: 1064 You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version for the right syntax to use near 'JSON NOT NULL, token_secretvarchar(100) NOT NULL,expires_in int(11' at line 10 Rest of the error is missing. Trying to create new Application causes: The application could not be saved: SQLSTATE[42S02]: Base table or view not found: 1146... So something is breaking during installation. Just posting this here incase there's something unorthodox going on with the modules installation. First time we've encountered this and we've done dozens of Processwire sites with many modules before. Thank you @Hurmeli ! I noticed problems with the installation routine a few days ago when I was working on another module. Somehow the code that is defined in the module in the ___install function doesn't seem to be executed anymore. As soon as I add my own ___install implementation, the installation throws an error. In the AppApi module, all database tables are created in this function. I don't know yet exactly where this error comes from and I am debugging it. Right now I'm assuming a bug in the ProcessWire core, but I need to test further. Please let me know if you find out more in the meantime! Hey @Hurmeli, I think your issue could be solved. I have found out that some versions of MariaDB seem to have problems with JSON for the database column type. It's recommended to use LONGTEXT instead, so I changed the db-initialization logic for it. Please report back if the issue is still apparent!
gharchive/issue
2023-02-15T13:47:51
2025-04-01T04:55:37.045171
{ "authors": [ "Hurmeli", "Sebiworld" ], "repo": "Sebiworld/AppApi", "url": "https://github.com/Sebiworld/AppApi/issues/38", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2245880344
Remove JwksSupplier from the public API JwksSupplier interface has 2 implementations: JwksSupplierJwksUri JwksSupplierEmbeddedJwks The first impl works with a SoftwareStatement that has a jwks_uri defined, the second impl works with a SoftwareStatement that has an embedded jwks defined. The issue with the current impl is, because these classes are public, you can call the wrong supplier type for the SoftwareStatement. These classes are currently only used by RegistrationRequestJwtSignatureValidationService to validate the signature of the registration request, as there is no way to plug behaviour into this service then it makes sense to move the supplier classes in here and make them private. Additional cleanup task, rename the suppliers to make them follow the naming convention of using a prefix: JwksUriJwksSupplier EmbeddedJwksSupplier Alternatively, we could allow the JwksSupplier behaviour to be plugged into the RegistrationRequestJwtSignatureValidationService to allow customers to extend it if necessary.
gharchive/issue
2024-04-16T12:07:53
2025-04-01T04:55:37.066115
{ "authors": [ "dbadham-fr" ], "repo": "SecureApiGateway/SecureApiGateway", "url": "https://github.com/SecureApiGateway/SecureApiGateway/issues/1375", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
895367281
Partnership > Marketing user can create "Partnership" even user have only "Global R" access Roles: Marketing Describe the bug Partnership > Marketing user can create Partnership even user have only Global R access Steps to Reproduce Login as Admin user Create a project and go to the project detail page Login as a Marketing user Open created project Click on the partnership button Add partnership Expected Result The partnership should not be created Actual Result Marketing user can create Partnership even user have only Global R access See video - https://user-images.githubusercontent.com/5212585/118809669-8a686700-b8c8-11eb-8a47-57531f9e3a06.mp4 FYI: @ParkeBrown @michaelmarshall cord-field# 1d6e8100 @ParkeBrown @CarsonF @parashara1 Tested with latest develop d49b012 and it's working fine now. Attaching a video for reference. Closing the issue. Thanks! See video : https://user-images.githubusercontent.com/5212585/132639169-1d15c800-3606-4634-890f-e5eb9c268f11.mp4
gharchive/issue
2021-05-19T12:12:11
2025-04-01T04:55:37.082486
{ "authors": [ "iPiyushPatel" ], "repo": "SeedCompany/cord-field", "url": "https://github.com/SeedCompany/cord-field/issues/850", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1249681192
[Bugfix] Allow PSBTOverviewScreen to truncate "self-transfer" Fixes #215 Truncating "self-transfer" gives the UI the horizontal room it needs to properly render. LGTM Code reviewed and tested on local device without issue LGTM Code reviewed and tested on local device without issue
gharchive/pull-request
2022-05-26T14:55:49
2025-04-01T04:55:37.084551
{ "authors": [ "kdmukai", "newtonick" ], "repo": "SeedSigner/seedsigner", "url": "https://github.com/SeedSigner/seedsigner/pull/217", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
160309744
executable damaged Hi Andrei &Elika, when I double click the executable I get this message: "idslabel" is damaged and can't be opened. You should move it to the Trash. I am on OS 10.11.5. It does work when I launch it from command line. Thank you for your work on this! well that's not good... I'll look into this
gharchive/issue
2016-06-14T23:45:41
2025-04-01T04:55:37.086474
{ "authors": [ "AnneSWarlaumont", "andreiamatuni" ], "repo": "SeedlingsBabylab/idslabel", "url": "https://github.com/SeedlingsBabylab/idslabel/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
205490986
Is it possible to read I-Bus with some modifications? Sorry if this is a stupid question, I am on a crash course trying to learn all this stuff, but Is it possible to read I-Bus with this same shield and or some modifications to the code? Hello, I think I-Bus is not supported by this shield. Thanks. Loovee
gharchive/issue
2017-02-06T05:26:49
2025-04-01T04:55:37.087932
{ "authors": [ "DarkN3ss61", "loovee" ], "repo": "Seeed-Studio/CAN_BUS_Shield", "url": "https://github.com/Seeed-Studio/CAN_BUS_Shield/issues/42", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1605402792
pt estimate for T5s in parseT5 function is deprecated The pT for T5s in the output NTuple are computed as follows: float pt = quintupletsInGPU.innerRadius[quintupletIndex] * kRinv1GeVf; ... ana.tx->pushbackToBranch<float>("t5_pt", pt); However, the pT estimate in parseT5 is just the average of the T3 pT's (which is somewhat questionable per #214): // Compute pt estimates from inner and outer triplets const float ptAv_in = abs(dr_in * k2Rinv1GeVf / sin((betaIn_in + betaOut_in) / 2.)); const float ptAv_out = abs(dr_out * k2Rinv1GeVf / sin((betaIn_out + betaOut_out) / 2.)); // T5 pt is average of the two pt estimates const float pt = (ptAv_in + ptAv_out) / 2.; It's my understanding that the pT as computed for the T5s in the output NTuple are "more correct," and that both the pTs from the output NTuple and parseT5 are used and therefore giving inconsistent results. I am working on a PR (will include link here when available) that will add some additional branches for the GNN NTuple, so I can make the fix in that PR as well if desired. @jkguiang Is this still an issue? Did you deal with it in any of your PRs,as mentioned in the description? Closed by #388
gharchive/issue
2023-03-01T17:47:14
2025-04-01T04:55:37.091116
{ "authors": [ "VourMa", "jkguiang" ], "repo": "SegmentLinking/TrackLooper", "url": "https://github.com/SegmentLinking/TrackLooper/issues/249", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2189074760
The previous step2 file is incompatible with CMSSW_14_1_X The CI was running into segfaults with CMSSW_14_1_0_pre0 after I switched to using the full file with 100 events, but it worked briefly while it was still using a newer file that only had 10 events. I was able to reproduce the issue on the UCSD and Cornell machines, and after some testing I verified that the issue was the old step2 file. I'm generating a new step2 file with 100 events using CMSSW_14_1_0_pre0 and we'll have to deploy it to the CI, and the UCSD and Cornell machines. @ariostas what's the cmsDriver command used for testing? where is the reference step2 file coming from? is it one from evourlio/LST/samples/RelValTTbarForCMSSW1300pre4/CMSSW_13_0_0_pre4/RelValTTbar_14TeV/GEN-SIM-DIGI-RAW/130X_mcRun4_realistic_v2_2026D88PU200-v1/ ? @slava77 it's this one: evourlio/LST/step2_21034.1_100Events.root @slava77 it's this one: evourlio/LST/step2_21034.1_100Events.root the provenance shows that this was in CMSSW_13_0_0_pre4. I'm aware of some L1-related data format changes; but then we are not reading L1 objects, not obvious if it's still possibly related. anyways, it sounds like it's not a part of a large production. It should be more straightforward to reproduce this file. I suggest to use HLT:@fake in step2; perhaps even try without L1TrackTrigger,L1: the output file will be a bit smaller and will also not have products that we are not using. @slava77 @VourMa Could someone who has write permissions to /data2/segmentlinking on cgpu-1 create a CMSSW_14_1_0_pre0 directory and make a symlink to /ceph/cms/store/user/anriosta/LST/samples/CMSSW_14_1_0_pre0/step2_21034.1_100Events.root? I can do it. Done!
gharchive/issue
2024-03-15T16:47:08
2025-04-01T04:55:37.096039
{ "authors": [ "VourMa", "ariostas", "slava77" ], "repo": "SegmentLinking/TrackLooper", "url": "https://github.com/SegmentLinking/TrackLooper/issues/378", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2471478116
🛑 Strapi is down In db5a66c, Strapi (https://strapi.sekai.best) was down: HTTP code: 525 Response time: 1385 ms Resolved: Strapi is back up in 4350a84 after 13 minutes.
gharchive/issue
2024-08-17T11:02:10
2025-04-01T04:55:37.108244
{ "authors": [ "dnaroma" ], "repo": "Sekai-World/uptime-monitor", "url": "https://github.com/Sekai-World/uptime-monitor/issues/459", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1696844278
Rewrite the plugin to run in an exclusive system The motivation for this was that it lets us get rid of all the Bevy reflection machinery and the requirement to register every trigger (which makes 'trigger combinators' much harder to use). The downside is that the triggers and transitions now run in an exclusive system, which could be worse for performance. We can ameliorate this using the ComputeTaskPool to spread running the machine across threads. I also added some functionality tests. I kept the Reflect bound on MachineState, even though it's probably not necessary, in case it winds up being useful for better debugging information or something later. (Suggestion: don't review the PR one commit at a time, it's a bit messy; just look at the final diff.) Breaking changes TriggerPlugin no longer exists and is not necessary; StateMachinePlugin is all that's needed. State builders now take their input by value instead of as a reference Triggers can now only take ReadOnlyStateParams The Trigger set no longer exists, since we check triggers and run transitions in the same system Thanks!
gharchive/pull-request
2023-05-05T00:05:25
2025-04-01T04:55:37.111453
{ "authors": [ "Seldom-SE", "deifactor" ], "repo": "Seldom-SE/seldom_state", "url": "https://github.com/Seldom-SE/seldom_state/pull/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
85162183
Right sidenav opening from left side of page. Just getting started with semantic-ui. Seem to have run into a problem with the right sidebar opening from the left side of the page. Only happens while there is a visible left sidebar. Do I need to put each sidebar in a different container div? http://jsfiddle.net/rduvhn8u/30/ click inside the Need Help input box to load right sidebar. Prelim debug: If I take out the left sidebar the code runs and the right sidebar works great. If you call sidebar with an init object it will reinitialize and destroy the previous instance. Try syntax like this http://jsfiddle.net/bzx2x929/ Thanks. Worked great!
gharchive/issue
2015-06-04T16:04:14
2025-04-01T04:55:37.250473
{ "authors": [ "debutmyname", "jlukic" ], "repo": "Semantic-Org/Semantic-UI", "url": "https://github.com/Semantic-Org/Semantic-UI/issues/2362", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
176204458
[Menu] Make a Menu containing another Menu stackable Currently a menu containing another menu does not stack properly. For example in http://jsfiddle.net/05d2qsgy/2/ you can see that the menu button "help" in the second menu is missing when the menu is stacked. My solution is to add the following css code: @media only screen and (max-width: 767px) { .ui.stackable.menu .right.menu, .ui.stackable.menu .right.item { display: block; } } Now you can see that the menu button "help" in the second menu is stacked properly http://jsfiddle.net/05d2qsgy/1/. See #3604.
gharchive/issue
2016-09-10T21:39:21
2025-04-01T04:55:37.253294
{ "authors": [ "Banandrew", "gujiman" ], "repo": "Semantic-Org/Semantic-UI", "url": "https://github.com/Semantic-Org/Semantic-UI/issues/4545", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
203454775
Is the gap between .ui.fixed.menu and .ui.grid enough? An grid after an fixed top menu receives a top padding of 2.75rem. Is it working for you? In my case, the grid is still rendered over the top menu. Is that "by design"? This issue is not valid.
gharchive/issue
2017-01-26T18:45:45
2025-04-01T04:55:37.254201
{ "authors": [ "imaia" ], "repo": "Semantic-Org/Semantic-UI", "url": "https://github.com/Semantic-Org/Semantic-UI/issues/4949", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
214938126
[Dropdown] Change direction of Dropdown menu I'm using Semantic-UI and Angular for developing a web app. Following to Semantic-UI documentation about dropdown, I created the following button: <div class="ui dropdown button"> <div class="text">Menu</div> <i class="dropdown icon"></i> <div class="menu"> <a class="item" (click)="onClickSuItem(m)" *ngFor="let m of menu"> <i class="icon {{m.icon}}"></i>{{m.name}}</a> </div> </div> I would like to change the direction of the menu, and set it on the left, instead of right. Do you know how can I reach that? Could be a good idea to have a direction/orientation option? Here is a Plunker. Did you get a chance to read : Menu Direction in Dropdown ? Yes, I read it, but it wasn't what I'm looking for. I mean direction of the dropdown menu, not direction for a sub-menu of the dropdown menu. Guess, what you're asking is similar to this issue: #3716 But still, that's again dropdown, then (left)/(right) side menu.. I see, thank you anyway. So that's a feature request that has not yet been accomplished. Am I right? Hi @smartm0use, try to use <div class="ui left pointing dropdown button">, it’s the only thing that can work out of the box. Closing the issue as a duplicate of #3716, I see no difference between the two. @Banandrew As I said before that solution wasn't I looking for. It should be like Floating appearance, but I need one button instead of two. I am looking for bottom left direction dropdown
gharchive/issue
2017-03-17T08:35:23
2025-04-01T04:55:37.259575
{ "authors": [ "Banandrew", "Utsho-Sadhak-Joy", "smartm0use", "vinayakkulkarni" ], "repo": "Semantic-Org/Semantic-UI", "url": "https://github.com/Semantic-Org/Semantic-UI/issues/5179", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1053563463
Добавить больше рассчетных примеров в лекцию про кубит Рядом с примерами на NumPy прямо матрички расписать еще для наглядности. Добрый день. Могу взяться за этот issue. И вопрос - для измерений ("Еще пара слов об измерениях") так же нужны формулы? Спасибо Леонид @flatslm Я тогда лучше объясню, откуда этот issue взялся. Был фидбек о том, что все очень сложно: У numpy расчета есть большой недостаток. Он что-то посчитал, кажется понятно что, а потом надо посчитать в первом задании скалярное произведение, я его считаю. А ответ не получается я сейчас перечитываю статью про кубиты. Все начинается просто и понятно, а потом фигак фигак Правило Борна. И приходит понимание, что чет я не умею считать ожидаемое значение оператора. Общая проблема вызвана наверное слабой математической подготовкой и быстрым усложнением. Кажется, что если бы я в ручную посчитал побольше операций над векторами и матрицами, то лучше бы понял. Этим и занимаюсь по вечерам. Потом расскажу, помогло ли ) Отсюда и родилась идея о том, что помимо операций на NumPy надо рядом расписать все то же в виде матриц, явно показать как все получается и т.д. Именно примеры, прямо с цифрами. Наверное в измерения тоже надо, так как людям может быть трудно понять, как работает правило Борна, как считаются мат-ожидания результата измерения какого-то оператора в каком-то состоянии и т.д. @flatslm Я добавил тебя в коллабораторы репозитория. Просто у нас ломается деплой стэйджинг версии и все проверки, если Pull-Request идет из форка. Поэтому просьба делать сразу внутри этого репозитория, просто создав ветку.
gharchive/issue
2021-11-15T11:38:07
2025-04-01T04:55:37.279230
{ "authors": [ "SemyonSinchenko", "flatslm" ], "repo": "SemyonSinchenko/qmlcourse", "url": "https://github.com/SemyonSinchenko/qmlcourse/issues/277", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1047585457
[WIP] Lecture/simon algorithm Тут удалилась лекция Саймона, возвращаю. Скоро разрулю Не, чет тут мешанина
gharchive/pull-request
2021-11-08T15:25:41
2025-04-01T04:55:37.281288
{ "authors": [ "Yorko" ], "repo": "SemyonSinchenko/qmlcourse", "url": "https://github.com/SemyonSinchenko/qmlcourse/pull/265", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1561016833
change hover/focus border on header nav surface-4 border by default text-color on hover/focus add !important to rad/kanj/vocab nav colors I think this looks pretty good. I left the check for min-width: 1024 px to style the border around the magnifying glass. Probably unnecessary, but wanted to tread lightly. I've not done a lot of testing of this theme at different screen resolutions in general. I only added selectors for :hover and :focus. I think that's right, but let me know if I should add :active as well. After looking a bit more into it, the bevel values are both set to 4px. So it's either an issue with rendering on my browser or the negative margin of the span element is not taken into consideration, when calculating the bevels. There are 3 potential solutions I can think of: Leave it be, it's a minor issue, that might not be worth fixing. Make the border-radius of the whole button larger. Make the border-radius of the span smaller. Im gonna trust your judgement here. Do what you consider to be the best solution and if that happens to be leaving it as is, feel free to merge into main. Yup, that's how it works with the default WK styling. The span with the count doesn't get a visible border (on hover/focus or normally). I think it's due simply to the span having a higher z-index. They've got a weird negative margin thing going on, too. Let me poke a bit and see if it's easy to fix. @Sepitus-exe How does this look? Normal: Hovering over lessons: I just gave a z-index of -1 to the span. Changing the z-index introduces a whole seperate, more sever issue. The span is no loger visible on narrow resolutions. It would probably be safer to remove the border-radius and the negative margins from it. The span is no longer visible on narrow resolutions. How did you create that screenshot? At resolutions below 1024px wide, the shortcuts and levels/radicals/kanji/vocab/help nav items disappear for me. The header stops being sticky, too. I'm not able to reproduce your screenshot. I just scrolled down a bit so the L/R buttons would appear and used the F12 console to shrink the width of the screen. The header has its position set as fixed, so it shouldn't disappear on its own. F12 console to shrink the width of the screen You mean the inspector in the browser? (I'm on a mac.) I'm unable to reproduce on either Brave or Firefox. What browser are you using? The entire mini-nav thing for reviews and lessons shouldn't appear at screen resolutions below 1024px wide, I think. That seems to be an independent issue, regardless. I'm tempted to merge with the z-index fix as the span "belongs" to the link as it were. It makes sense for the the border-outline to cover everything. it seems to be an issue only on non-dash pages. Aha! Let me look into that (was only looking at the dashboard). I'm struggling to figure out why the span disappears at widths below 1024px. Very weird. I'm tempted to just add a media query so the z-index change only affects resolutions above 1024px wide, and revert to the original behavior for narrow screens. If I can't figure it out, that's what I will merge. Aha. It's something to do with using display: flex; at lower resolutions. This is getting too painful to futz with. I'm just going to use a media query. Okay, we can revisit this at some point if desired. Current behavior is to have a border surrounding the entire nav link at higher screen resolutions. At lower screen resolutions it reverts to the top "leave it be" behavior shown in this comment
gharchive/pull-request
2023-01-28T21:14:54
2025-04-01T04:55:37.310292
{ "authors": [ "Sepitus-exe", "wrex" ], "repo": "Sepitus-exe/WKElementaryDark", "url": "https://github.com/Sepitus-exe/WKElementaryDark/pull/60", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1687872656
amazon_cognito_identity_dart_2 new edition conflict I use ^3.3.0 version of amazon_cognito_identity_dart_2 I don't want to downgrade\ 0.1.25 I guess it is the best module in Dart modules. bad: SimpleS3 minio If you have the time to fork this library and test that it works with the latest version of amazon_cognito_identity_dart_2, a PR would be incredibly welcome! FYI flutter pub add aws_s3_upload command's result Because aws_s3_upload <1.2.0 depends on amazon_cognito_identity_dart_2 ^0.1.25+1 and aws_s3_upload >=1.2.0 depends on amazon_cognito_identity_dart_2 ^1.0.3, every version of aws_s3_upload requires amazon_cognito_identity_dart_2 ^0.1.25+1 or ^1.0.3. So, because MKBContainerFortune depends on both amazon_cognito_identity_dart_2 ^3.3.0 and aws_s3_upload any, version solving failed. pub finished with exit code 65 dependency_overrides: amazon_cognito_identity_dart_2: ^3.3.0 in pubspec.yaml
gharchive/issue
2023-04-28T04:19:26
2025-04-01T04:55:37.319375
{ "authors": [ "Serdnad", "mkbsugita" ], "repo": "Serdnad/aws_s3_upload", "url": "https://github.com/Serdnad/aws_s3_upload/issues/16", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1800121158
Support for Inline arrays in C# 12 preview? https://devblogs.microsoft.com/dotnet/new-csharp-12-preview-features/#inline-arrays Is anyone working on this? I may be willing to do so, but I'm hesitant because I don't have much experience with source generators. This is planned (along with all other new polyfills for C# 12), just waiting for the language to be stable first 🙂 The design of several of the new attribute types has changed over time, so I want them to stabilize before adding polyfills. Hey Sergio, I love your work! Following on from the C#12 theme, with .NET 8 now released I can’t wait to start using it! Are you able to share the plans for C#12 support? Is there anything the community can do to help? Thanks :) Hmm, today I was curious how we can do things by ourselves. So I decided to play around with things myself in order to see how far I can get, and what happened was actually quite interesting. After setting my csproj TFM to net481 I got several errors, but most obviously that the InlineArray attribute was missing. So I just decided to copy the official InlineArrayAttribute.cs .NET source file into my project. Which solved that part of the problems. Another error shown about missing System.Runtime.InteropServices.MemoryMarshal.CreateSpan was shown by IntelliSense (but somehow not by the compiler output itself) at that point on my code where I assigned values to my internal array for all the indices. I was eventually able to remove it by making my own public static partial class MemoryMarshal implementation with a CreateSpan method and modify that a bit so that it could compile on .NET Framework as well. However, after these actions, the following more serious compiler error still remained: error CS9171: Target runtime doesn't support inline array types. To me, that suggests that the current situation is that inline arrays are actually impossible on .NET Framework at all, or at least that Roslyn compiler itself does not allow you to do this at least. So in that case I guess PolySharp alone will not be able to solve it I guess... Oh, I forgot this one. Actually completed by #81. "inline arrays are actually impossible on .NET Framework at all" That is correct, it requires runtime support. This is why it's part of the "runtime supported" attributes which are dummy and don't actually do anything (like eg. [UnmanagedCallersOnly]), and are opt-in. They're only meant to make multi-targeted code cleaner.
gharchive/issue
2023-07-12T03:58:53
2025-04-01T04:55:37.364316
{ "authors": [ "Sergio0694", "ds5678", "eduarddejong", "optiks", "sdcb" ], "repo": "Sergio0694/PolySharp", "url": "https://github.com/Sergio0694/PolySharp/issues/78", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
294540250
Didn't run on Ubuntu Traceback (most recent call last): File "Training Data Collection/caputre_training_date.py", line 1, in from sneakysnek.recorder import Recorder File "/home/dao/.local/lib/python2.7/site-packages/sneakysnek/recorder.py", line 26 raise RecorderError(f"Unsupported platform '{sys.platform}'") ^ SyntaxError: invalid syntax Could it be as Ubuntu returns 'linux2' for sys.platform while expected is "linux2" ? The invalid syntax is because you aren't running on Python 3.6+ The platform for Ubuntu is covered: https://github.com/SerpentAI/sneakysnek/blob/master/sneakysnek/recorder.py#L16 Thanks!
gharchive/issue
2018-02-05T20:59:37
2025-04-01T04:55:37.387148
{ "authors": [ "nbrochu", "suchitsha" ], "repo": "SerpentAI/sneakysnek", "url": "https://github.com/SerpentAI/sneakysnek/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
345944520
Application failing on Safari 6.2.8 System – OS X 10.8.5 Processor 2.66 GHz Intel Core 2 Duo Memory – 8GB 1067 MHz DDR3 Various tests to test eligibility (eg Income approx $38000 1 dependent) and front page worked correctly by showing eligibility. Clicked on apply but failed to get to next screen with just white screen. Shut down browser and tried again. Still failed. @Br3nda @ssibbehh I dont really have a way to test this. My Safari is 9.2.1 and it works fine for this version. Only other option I can think of is by purchasing browserstack but its not free @gthurstonnz can you bring your safari device to the lab? I'll see what I can do. I'll check in with Bron and Jess. When do you need it by? On Sat, Aug 4, 2018 at 4:30 PM Brenda Wallace notifications@github.com wrote: @gthurstonnz https://github.com/gthurstonnz can you bring your safari device to the lab? — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/ServiceInnovationLab/pancake-frontend/issues/504#issuecomment-410422667, or mute the thread https://github.com/notifications/unsubscribe-auth/An_Eq_HdjtnrygNKHAFVvlG-fML6ywm6ks5uNSNtgaJpZM4VnNj- .
gharchive/issue
2018-07-30T21:35:53
2025-04-01T04:55:37.392312
{ "authors": [ "Br3nda", "dlouise64", "gthurstonnz" ], "repo": "ServiceInnovationLab/pancake-frontend", "url": "https://github.com/ServiceInnovationLab/pancake-frontend/issues/504", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2025285021
Fix chrome ts sorry potato pc people Fixes problems that rose after typescript migration in chrome branch hahaha ofc ofc. Wonderful!
gharchive/pull-request
2023-12-05T04:54:33
2025-04-01T04:55:37.406364
{ "authors": [ "Crazypersonalph", "SethBurkart123" ], "repo": "SethBurkart123/EvenBetterSEQTA", "url": "https://github.com/SethBurkart123/EvenBetterSEQTA/pull/73", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
391488804
Make boiling oil engineers pour automatically. It is hard to use boiling oil pot engineers as a human player, you have to manually pour the oil in a certain direction which makes them a poor option for defence compared with fire throwers, as it takes a lot of time to manually make several oil pot throwers use their oil, when there may be other things you need to do with castle defense such as recruiting new troops. It would be good if you could have the option to put oil pot engineers on aggressive mode and they would automatically pour the oil if units were in range, then they would automatically return and get oil and go back to their position. Actually, you can do that! On aggressive mode, they will pour their oil as soon as 1 enemy unit is in range, and on defensive mode they will pour their oil as soon as 3 enemy units are in range. Right after they emptied their oil pot, they will run back to the oil smelter, get their pot filled up again and then return to their previous position on some wall/tower/gate. Oops, had the original pre-Warchest version of Stronghold Crusader where the oil engineers didn't appear to autopour and haven't used oil engineers since, didn't realise this was fixed now officially!
gharchive/issue
2018-12-16T17:40:07
2025-04-01T04:55:37.419661
{ "authors": [ "Heroesflorian", "PitchNeeded" ], "repo": "Sh0wdown/UnofficialCrusaderPatch", "url": "https://github.com/Sh0wdown/UnofficialCrusaderPatch/issues/132", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
505471576
Some video for firefly about bugs and glitches to fix, but also some useful hints(tricks) for overall playing. I'm already talking about some SHC tricks/glitches/bugs. I'm going to showcase them on Crusader(game) soon too. Let me know what you think about it, maybe underneath the video in the comments section(so it'll be forever visible, even for future for firefly?)? ;) Help push up the topic, so Firefly might finally take an interest in fixing/removing/altering those, once Warlords will come out. Cheers! ~Rakso Youtube video I like how everybody is trying to use "issue" section as promotion stand I like how everybody is trying to use "issue" section as promotion stand i think that shows 2 things: we are in a desperate need for fixing bugs of the game, the video showcases alot of bugs, which apperently are in present in Stronghold Crusader. I know that your mod fixed some issues, that were not adressed by sh0wdown, which are mostly multiplayer related, while sh0wdowns mod fixes so many issues with the AI (which i am really exited and happy about!). We are in a desperate need to unite Crusader friends on a commen plattform. There is no official plattform from firefly that works. So there are many plattforms popping up, each claming to be the one, but not really all the people know of all of these and they are scattered around the world. Even the chinese community must have their own plattform for the game. This is how the Firefly Forum looks like. http://archive.fireflyworlds.com/index.php?option=com_fireboard&Itemid=271&func=showcat&catid=17 If i try to enter the general section of Stronghold Cruisader Original, i am promptet with 502 bad gateway. It seems like not a hole lot of care is put into this. But at least they have a working discord now. And maybe, just maybe, they will notice their huge fanbase about the original 2D titles some day and adress some issues. Keep up the great work guys and spread the word about the game! Yeah, I have to agree with Krarilotus. There is no central point for the Stronghold and Stronghold Crusader community, and this bug-tracker is at the moment the most active thing. And as long as Sh0wdown and his mods accept 'issues'/posts like this, it is fine I guess. (But obviously, not the perfect solution, but I guess there will be no alternative in the future).
gharchive/issue
2019-10-10T19:41:17
2025-04-01T04:55:37.424759
{ "authors": [ "J-T-de", "JuGGerNaunT", "Krarilotus", "Rakso69" ], "repo": "Sh0wdown/UnofficialCrusaderPatch", "url": "https://github.com/Sh0wdown/UnofficialCrusaderPatch/issues/463", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1862492477
Error 405 when sending email Description I've been experiencing an issue when attempting to send an email using this library. Specifically, I'm receiving a HTTP 405 error. I suspect this issue may be related to a change in mailchannels. This was working fine for a couple months and stop working today. Additional Information From preliminary research, MailChannels, has recently published an article detailing various security measures against domain name spoofing, called "Domain Lockdown". This might be a potential reason for the issues we're facing. Linked below are their help posts: Sending Email from Cloudflare Workers using MailChannels Send API Secure your domain name against spoofing with Domain Lockdown™ Note that participation in Domain Lockdown is now mandatory for Cloudflare Workers users. I've tried to configure the Domain Lockdown they write about with no luck. Does anybody now how to set this up? Thank you. Facing the same issue here Sometimes Vercel returns a 405 when a response isn't sent, like when an error is thrown. For me it's failing when calling Email.send with the error Error sending email.
gharchive/issue
2023-08-23T03:16:09
2025-04-01T04:55:37.428963
{ "authors": [ "abhishekmg", "francofantini", "treyhuffine" ], "repo": "Sh4yy/vercel-email", "url": "https://github.com/Sh4yy/vercel-email/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
419122289
0MC3 Use the preview tab to see the example given below. Edit the relevant information and preview the changes before submitting. You may delete these instructions. Key Value ID 0MC3 Source Vaaran Bhai Gurdas Ji Page 28 Line gur isKI DIrju Drmu; iprm ipAwlw Ajru jrxw [ Correction THIS ≠ THAT PROOF (EDITION) Image: Correction: gur isKI ≠ gurisKI Proof: Vaaran Bhai Gurdaas Ji | November 2011 | Shamsher Singh Ashok | SGPC, Sri Amritsar Image:
gharchive/issue
2019-03-09T22:27:09
2025-04-01T04:55:37.436876
{ "authors": [ "UmeetSingh", "bhajneet" ], "repo": "ShabadOS/database", "url": "https://github.com/ShabadOS/database/issues/609", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1683026915
useMutation without id Hi, I have question about mutations. I have this backend controller: React-query-swagger generates this: And now, because I have ID in url, it is needed while initializing useUsersPUTMutation. So I cannot use useMutation for multiple users. Or sometimes I dont know ID on initialization. So I'm doing workaround and using Client.usersPUT(id, body) directly when I know ID. But I'm wondering if it is good practice to using usersPUT directly, and if I'm not loosing any of features that useUsersPUTMutation adds. Thank you in advance for your time :) Personally I'm not using mutations that often. I mean, I almost never send GET requests directly (always work with useQuery), but for POST/PUT/DELETE requests, in over 60% of the cases I send them via client methods, and not via mutations. So for me it feels ok :)
gharchive/issue
2023-04-25T11:57:10
2025-04-01T04:55:37.461462
{ "authors": [ "Shaddix", "panda7789" ], "repo": "Shaddix/react-query-swagger", "url": "https://github.com/Shaddix/react-query-swagger/issues/22", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
456468589
Problem passing multiple parameters to "RegisterMethod" registered method Currently i'm implementing a library in Go which needs to get multiple parameters on a method call: https://github.com/chiguireitor/godot-gobtcsuite/blob/154bc8fec32c87dc9a8c8f7dba0a67c3850f5605/src/gobtcsuite.go#L280 That call is getting the first parameter right, but the rest of the Variants in the args array come with weird values. For example, calling with 3 ints i do: fmt.Sprintf("%d %d %d", args[0].GetType(), args[1].GetType(), args[2].GetType()) and get 2 81748992 81748992 as result Edit: It should be noted that the expected result should be 2 2 2 seems like the args array is only getting initialized on the first value, i've been trying to debug this but can't find where's my method getting called to change it on godot's or godot-go's code. I'm seeing the same issue here - any update?
gharchive/issue
2019-06-14T23:06:05
2025-04-01T04:55:37.464318
{ "authors": [ "1800alex", "chiguireitor" ], "repo": "ShadowApex/godot-go", "url": "https://github.com/ShadowApex/godot-go/issues/48", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1054880134
How to turn on and turn off botplay with lua script? What is your question? How to turn on and turn off botplay with lua script? setProperty('botPlay',false) thanks
gharchive/issue
2021-11-16T13:26:01
2025-04-01T04:55:37.468030
{ "authors": [ "AziziPGM", "NS-Unfair", "zwl1619" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/1070", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1096415870
arrows and notesplashes changes mid song Describe your problem here. something like in endless Are you modding a build from source or with Lua? Lua What is your build target? Windows x64 Did you edit anything in this build? If so, mention or summarize your changes. No response This is an event that sets the note skins and splashes. function onCreate() for i = 0, getProperty('unspawnNotes.length')-1 do if getPropertyFromGroup('unspawnNotes', i, 'noteType') == '' then setPropertyFromGroup('unspawnNotes', i, 'texture', 'your note here', 'your note here'); setPropertyFromGroup('unspawnNotes', i, 'noteSplashTexture', 'your splash here'); end end end If you want to send more than one: function onUpdate(elapsed) if curStep >= (your number here) then setPropertyFromGroup('unspawnNotes', i, 'texture', 'your note here', 'your note here'); setPropertyFromGroup('unspawnNotes', i, 'noteSplashTexture', 'your splash here'); (repeat as many times as you want)
gharchive/issue
2022-01-07T15:24:23
2025-04-01T04:55:37.471989
{ "authors": [ "Itz-Miles", "MrSomethingoridk" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/3271", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1196462073
Black screen or crash on song Describe your bug here. So uh, I just loaded the game by lime test windows, and when I select a song, it crashes. But black screen is on the character editor. Any help please? I worked on it and it looks like I lost most of the work on it. Command Prompt/Terminal logs (if existing) No response Are you modding a build from source or with Lua? Source What is your build target? Windows Did you edit anything in this build? If so, mention or summarize your changes. I remember adding a new credit section, no errors to it tho. also i have a good laptop/pc so yea
gharchive/issue
2022-04-07T19:09:42
2025-04-01T04:55:37.474763
{ "authors": [ "indeedbadundertaleplayer" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/8004", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1217865653
in the character editor there should be an option to put a "beat" on certain frames What feature do you want to get added on the base engine? in the character editor there should be an option to put a "beat" on certain frames as to make it on that frame a beat gets hit To test your sight, and reliability, please select the option of what should NOT be requested. Stage Editor, 6K+ support, and winning icons. What exactly is a "beat"? so you know when a character puts there head down on there idle I'm saying there should be a feature where you can pick the frame or frames it happens on that means changing how beats are calculated for a feature with little use that means changing how beats are calculated for a feature with little use thats not exactly what it would do for example for the people in front player and opponent there idle animation plays once every other beat what I'm saying is that you could put the frame on wich the animations starts each beat or other beat like how GF head hits every beat it wouldn't change how it was calculated because it doesn't change anything besides an animation I barely got that but isn't that what indices are for character.danceEveryNumBeats? You could probably do this using a LUA event. Either using characterDance('dad/boyfriend/gf'), or characterPlayAnim('dad/boyfriend/gf', 'idleanimationname', true) in a LUA event, and then placing the event wherever you want a beat in the chart editor. I haven't tested it yet, so I'm not sure if it would work. One main problem I could think of is that the normal idle animation would play in the middle of the new beat, but I'm not sure.
gharchive/issue
2022-04-27T20:46:44
2025-04-01T04:55:37.478820
{ "authors": [ "DefNotDylan", "Slimquickhunt", "frantastic24", "tylerandari12" ], "repo": "ShadowMario/FNF-PsychEngine", "url": "https://github.com/ShadowMario/FNF-PsychEngine/issues/8365", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2760454761
Tier augments still not applied in some vanilla methods which spawn mobs The augments from the current world tier are not applied when mobs are spawned by the following vanilla methods: Generating chunks Generating structures containing mobs Breeding Throwing eggs Building iron or snow golems In all five cases, apotheosis:tier_augments_applied is 1b in the entity's NBT data, but no tier augments are applied. However, the same mobs spawned naturally, from a spawner, or from the /summon command do get the tier augments applied. Mod list: Apotheosis-1.21.1-8.0.1 ApothicAttributes-1.21.1-2.5.0 ApothicEnchanting-1.21.1-1.2.5 ApothicSpawners-1.21.1-1.1.1 cloth-config-15.0.140-neoforge emi_enchanting-0.1.2+1.21+neoforge emi-1.1.18+1.21.1+neoforge NoChatReports-NEOFORGE-1.21.1-v2.9.1 Patchouli-1.21-87-NEOFORGE Placebo-1.21.1-9.6.1 Looks like the check is ignoring creative players when considering options for applying the tier augments, and it still marks the entity when the augments cannot be applied due to no available players. https://github.com/Shadows-of-Fire/Apotheosis/commit/d48d65cc6fb932eb05f9b7c2fba27b8755692f80
gharchive/issue
2024-12-27T06:17:59
2025-04-01T04:55:37.484743
{ "authors": [ "James103", "Shadows-of-Fire" ], "repo": "Shadows-of-Fire/Apotheosis", "url": "https://github.com/Shadows-of-Fire/Apotheosis/issues/1420", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
675650261
NullPointerException MC Version: 1.15.2 Forge Version: 31.2.30 Mod Version: 3.2.4 Placebo Ver: 3.1.0 I get a crash due to a Null Pointer Exception during the Mod loading process before MC starts Crash Log crash-2020-08-08_23.57.59-client.txt Latest.log latest.log That's a nonsense crash caused by another mod erroring earlier in the loading process. The debug.log has the actual information, but its not likely an apoth error
gharchive/issue
2020-08-09T07:04:21
2025-04-01T04:55:37.487645
{ "authors": [ "Shadows-of-Fire", "ThatBenderGuy" ], "repo": "Shadows-of-Fire/Apotheosis", "url": "https://github.com/Shadows-of-Fire/Apotheosis/issues/198", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1062655553
incompatible with curios Placebo-1.16.5-4.6.0 is incompatible with the curios api https://www.curseforge.com/minecraft/mc-mods/curios This is not a valid report by any means, apoth uses curios alongside placebo 4.6, and you have't provided any actual data how come you think that apotheosis uses curios? curios is not listed in the relations section of apotheosis. ok, sorry, let me rephrase: placebo 4.6.0 does not run.
gharchive/issue
2021-11-24T16:49:14
2025-04-01T04:55:37.489657
{ "authors": [ "PhiCecHey", "Shadows-of-Fire" ], "repo": "Shadows-of-Fire/Placebo", "url": "https://github.com/Shadows-of-Fire/Placebo/issues/28", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
369933776
i got "File Not Found" i got "File Not Found", any hints? Document Details ⚠ Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking. ID: aac83aaa-403c-bbb5-1381-9e1e39913e41 Version Independent ID: 6cb358f2-72b2-aa3d-6ac4-48352f199eda Content: Add a web part to a page in a SharePoint-hosted SharePoint Add-in Content Source: docs/sp-add-ins/add-a-web-part-to-a-page-in-a-sharepoint-hosted-sharepoint-add-in.md Product: sharepoint GitHub Login: @spdevdocs Microsoft Alias: spdevdocs Thanks @peterremote1980 , could you please provide more information about your issue? such as what did you do and what is the problem. Closing due to no response.
gharchive/issue
2018-10-14T19:13:03
2025-04-01T04:55:37.533478
{ "authors": [ "KingKangMSFT", "VesaJuvonen", "peterremote1980" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/issues/2763", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
421532167
Can't use loadLegacyFabricCss in SPFX 1.8 Category [ ] Question [ ] Typo [x] Bug [ ] Additional article idea Expected or Desired Behavior According to SPFX 1.8 documentation, it is possible to add "loadLegacyFabricCss": true to web part manifest file in order to force old fabric CSS to be loaded in the page. Observed Behavior After adding the mentioned property ("loadLegacyFabricCss": true) to web part manifest file, execution of commands gulp serve or gulp bundle failed and after webpacking the following error message is displayed: [15:29:01] Error - [write-manifests] Manifest validation error (./src/webparts/helloWorld/HelloWorldWebPart.manifest.json): (#/) Data does not match any schemas from 'oneOf' (#/) Missing required property: description (#/) Missing required property: extensionType (#/componentType) No enum match for: WebPart (#/) Additional properties not allowed: loadLegacyFabricCss (#/) Missing required property: items (#/) Missing required property: rootComponentId Steps to Reproduce Create new spfx 1.8 project containing web part. Add "loadLegacyFabricCss": true to web part manifest file. Here is Example of such manifest file: Run command gulp serve or gulp bundle and you will get the mentioned error: Thank you to @srideshpande for finding a temporary solution. The final one will require an update to our public packages. Go to the below location node_modules\@microsoft\sp-module-interfaces\lib\manifestSchemas\jsonSchemas\client-side-web-part-manifest.schema.json Add the following line to the file before the 'canUpdateConfiguration' entry "loadLegacyFabricCss": { "$ref": "any-value.schema.json" } save, and run gulp or gulp build or gulp bundle or gulp --ship. @patmill Is this fixed yet? I am facing the same issue when I used it yesterday. @araLT Can you reopen the bug as it is not fixed yet and the temporary solution would not work in build pipelines? For now, I am just including the fabric css file directly using require('../../../../node_modules/office-ui-fabric-core/dist/css/fabric.min.css'); till this is fixed. @patmill is this fixed yet? I am also facing same issue but with another property 'supportedHosts:["SharePointWebPart","TeamsTab"]' When I am passing this values in manifiest file of project I am getting error. Fixed with the v1.8.1 release - see release notes at - https://github.com/SharePoint/sp-dev-docs/wiki/SharePoint-Framework-v1.8.1-release-notes
gharchive/issue
2019-03-15T13:58:58
2025-04-01T04:55:37.542746
{ "authors": [ "VesaJuvonen", "abhishek-raj", "araLT", "patmill", "salahsaleh" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/issues/3612", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
564604937
Filtering in Office Fabric UI React Hi Team, I am using SPFX with React and using the Office fabric UI for design. As of now I am able to add drop down icon near to the column header in Detailslist by using ColumnActionsMode.hasDropdown. I want to show the drop down menu by clicking on the drop down icon as like below image @sang1205 Can you please provide more information? Not clear if this is a question about SPFx (doesn't appear so) or Office UI Fabric React (appears so). If you're having an issue, please add details around your context. If it's a question on how to do something, respectfully you need to provide a lot more "here's what I've tried" instead of a generic "someone help me do this." @sang1205 I was able to implement something like this by passing an IContextualMenu props object into the menuProps of a CommandBarItem in a CommandBar. Though we don't use DetailsList, we made our own list component instead, but I'm sure it can be done with a column header like that. So it can be done, it's just a matter of putting the right props in the right place, once I figured that out I think I finally understand Fabric, after two years of using it... Perhaps look at these props for DetailsList onColumnHeaderClick onColumnHeaderContextMenu IDetailsList documentation @sang1205 Can you please provide more information? I have updated my question . Please check and let me know if you can help This question isn't so much about SharePoint dev / SPFx, it's more specific on how to use the Office UI Fabric React library. I suggest you please post your question to the OUIFR projects as they have more context and can likely provide a solution. Feel free to cross-link this issue. https://github.com/OfficeDev/office-ui-fabric-react
gharchive/issue
2020-02-13T11:03:51
2025-04-01T04:55:37.548577
{ "authors": [ "KaushalKhamar", "andrewconnell", "ng-lboyle", "sang1205" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/issues/5341", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
639231083
Custom Sharepoint Permissions using Microsoft Graph Category [x] Question [ ] Typo [x] Additional article idea Question Is there a way to modify custom Sharepoint Online permissions using Microsft Graph API. This article explains the entire process but uses the old api. Is there a way to do the above steps but using Graph API v1.0 Environment details (development & target environment) Your Developer Environment: MacOS Target Environment: SharePoint Online Framework: Node.js v12 Browser(s): FireFox Tooling: VS Code No, the MS Graph only supports CRUD operations on sites, lists, & list items along with following sites: https://docs.microsoft.com/en-us/graph/api/resources/sharepoint?view=graph-rest-1.0
gharchive/issue
2020-06-15T22:48:56
2025-04-01T04:55:37.553190
{ "authors": [ "andrewconnell", "gurleensethi" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/issues/5888", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1453321807
Viva Connections Desktop App Issue: Custom API calls are failing What type of issue is this? other What SharePoint development model, framework, SDK or API is this about? 💥 SharePoint Framework Target SharePoint environment SharePoint Online What browser(s) / client(s) have you tested [ ] 💥 Internet Explorer [ ] 💥 Microsoft Edge [ ] 💥 Google Chrome [ ] 💥 FireFox [ ] 💥 Safari [ ] mobile (iOS/iPadOS) [ ] mobile (Android) [ ] not applicable [ ] other (enter in the "Additional environment details" area below) Additional environment details Desktop App for Microsoft Teams Issue description Following authentication calls in MS Teams Desktop client are failing while calling custom AD authenticated API calls. https://tenant.sharepoint.com/_api/Microsoft.SharePoint.Internal.ClientSideComponent.Token.AcquireOBOToken?resource=resource id&clientId=SharePointOnlineClientExtensibilityWebApplicationPrincipalID Error messages differ from time to time. Following have been observed: AADSTS500131: Assertion audience does not match the Client app presenting the assertion. The audience in the assertion was 'https://tenant.sharepoint.com/' and the expected audience is 'api://tenantid/microsoft.spfx3rdparty.com' or one of the Application Uris of this application with App ID (SharePoint Online Client Extensibility Web Application Principal). The downstream client must request a token for the expected audience (the application that made the OBO request) and this application should use that token as the assertion.\ 2.One or more errors occurred. 3.Authentication Method is not allowed. Please refer below screenshots for the same. @Srpatil212 - are you using Sync to Teams to add the SPFx app to the Teams? or do you have a custom manifest? If the latter - could you please share the manifest? Thanks! @AJIXuMuK Since this is Viva Connections app of Microsoft teams , we have just added the app in Teams. Our home site for sharepoint tenant has custom spfx webparts and in them we have api calls. Viva connections app is opening our sharepoint home site and the custom api calls in spfx webparts failing as mentioned above. @Srpatil212 could you please send the request id and request date? You can find those in the request headers. I can share the request id and date with you separately. Please let me know your microsoft communication id. @Srpatil212 it's aterentiev@microsoft.com Marking as feedback needed as haven't heard anything back. We also made some changes that can potentially fix this issue. These changes should be rolled out next week.
gharchive/issue
2022-11-17T13:10:45
2025-04-01T04:55:37.563130
{ "authors": [ "AJIXuMuK", "Srpatil212" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/issues/8591", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
1396694269
Update toc.yml Category [X] Content fix/update What's in this Pull Request? Adding the "What's new in the Migration API" to the TOC. The article is already published and live. Learn Build status updates of commit 3db8d28: :white_check_mark: Validation status: passed File Status Preview URL Details docs/toc.yml :white_check_mark:Succeeded View For more details, please refer to the build report. Note: Broken links written as relative paths are included in the above build report. For broken links written as absolute paths or external URLs, see the broken link report. For any questions, please:Try searching the learn.microsoft.com contributor guidesPost your question in the Learn support channel
gharchive/pull-request
2022-10-04T18:40:28
2025-04-01T04:55:37.569290
{ "authors": [ "JoanneHendrickson", "VesaJuvonen" ], "repo": "SharePoint/sp-dev-docs", "url": "https://github.com/SharePoint/sp-dev-docs/pull/8481", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }